Skip to content

Commit

Permalink
modified README to include the blogpost link
Browse files Browse the repository at this point in the history
  • Loading branch information
shashank committed Sep 24, 2024
1 parent bfb1216 commit 2a5991f
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion README.md
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,7 @@
This repository contains the implementation of two Transformer variants (Performer and Longformer) along with the base Transformer. These are more memory efficient variations proposed after the Transformer particularly when we have to deal with longer sequences. We have also compared their performance on a classification task. Specifically, we have focused on sentiment classification using only the encoder portion of the Transformer model, since the decoder is not needed for this task. The base transformer model is derived from the Annotated Transformer tutorial, and the two variants—Performer and Longformer—are implemented by modifying just the self-attention mechanism. The focus is on enhancing performance (both in terms of time and memory), and comparing the efficacy of these models on a sentiment classification dataset.


For a deeper theoretical understanding, please refer to the blog post titled [Benchmarking Efficient Transformer Variants](https://www.infocusp.com/blogs/benchmarking-efficient-transformers-variants/) which covers these two vaiants along with the base Transformer.
For a deeper theoretical understanding, please refer to the blog post titled [Benchmarking Efficient Transformer Variants](https://www.infocusp.com/blogs/benchmarking-efficient-transformers-variants/) which covers these two variants along with the base Transformer.

## Dataset

Expand Down

0 comments on commit 2a5991f

Please sign in to comment.