File tree Expand file tree Collapse file tree 3 files changed +5
-0
lines changed Expand file tree Collapse file tree 3 files changed +5
-0
lines changed Original file line number Diff line number Diff line change 22
22
* [Transformer building blocks](transformers/models.html)
23
23
* [Transformer XL](transformers/xl/index.html)
24
24
* [Relative multi-headed attention](transformers/xl/relative_mha.html)
25
+ * [Rotary Positional Embeddings](transformers/rope/index.html)
25
26
* [Compressive Transformer](transformers/compressive/index.html)
26
27
* [GPT Architecture](transformers/gpt/index.html)
27
28
* [GLU Variants](transformers/glu_variants/simple.html)
Original file line number Diff line number Diff line change 22
22
This implements Transformer XL model using
23
23
[relative multi-head attention](xl/relative_mha.html)
24
24
25
+ ## [Rotary Positional Embeddings](rope/index.html)
26
+ This implements Rotary Positional Embeddings (RoPE)
27
+
25
28
## [Compressive Transformer](compressive/index.html)
26
29
27
30
This is an implementation of compressive transformer
Original file line number Diff line number Diff line change @@ -24,6 +24,7 @@ implementations almost weekly.
24
24
* [ Transformer building blocks] ( https://nn.labml.ai/transformers/models.html )
25
25
* [ Transformer XL] ( https://nn.labml.ai/transformers/xl/index.html )
26
26
* [ Relative multi-headed attention] ( https://nn.labml.ai/transformers/xl/relative_mha.html )
27
+ * [ Rotary Positional Embeddings] ( https://nn.labml.ai/transformers/rope/index.html )
27
28
* [ Compressive Transformer] ( https://nn.labml.ai/transformers/compressive/index.html )
28
29
* [ GPT Architecture] ( https://nn.labml.ai/transformers/gpt/index.html )
29
30
* [ GLU Variants] ( https://nn.labml.ai/transformers/glu_variants/simple.html )
You can’t perform that action at this time.
0 commit comments