Skip to content

Commit 3a54a20

Browse files
committed
links
1 parent 95bc354 commit 3a54a20

File tree

5 files changed

+15
-5
lines changed

5 files changed

+15
-5
lines changed

docs/index.html

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -90,7 +90,8 @@ <h4>✨ <a href="transformers/index.html">Transformers</a></h4>
9090
<li><a href="transformers/mlp_mixer/index.html">MLP-Mixer: An all-MLP Architecture for Vision</a> </li>
9191
<li><a href="transformers/gmlp/index.html">Pay Attention to MLPs (gMLP)</a> </li>
9292
<li><a href="transformers/vit/index.html">Vision Transformer (ViT)</a> </li>
93-
<li><a href="transformers/primer_ez/index.html">Primer EZ</a></li></ul>
93+
<li><a href="transformers/primer_ez/index.html">Primer EZ</a> </li>
94+
<li><a href="transformers/hourglass/index.html">Hourglass</a></li></ul>
9495
<h4><a href="recurrent_highway_networks/index.html">Recurrent Highway Networks</a></h4>
9596
<h4><a href="lstm/index.html">LSTM</a></h4>
9697
<h4><a href="hypernetworks/hyper_lstm.html">HyperNetworks - HyperLSTM</a></h4>

docs/transformers/index.html

Lines changed: 6 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -104,13 +104,15 @@ <h2><a href="vit/index.html">Vision Transformer (ViT)</a></h2>
104104
<p>This is an implementation of the paper <a href="https://papers.labml.ai/paper/2010.11929">An Image Is Worth 16x16 Words: Transformers For Image Recognition At Scale</a>.</p>
105105
<h2><a href="primer_ez/index.html">Primer EZ</a></h2>
106106
<p>This is an implementation of the paper <a href="https://papers.labml.ai/paper/2109.08668">Primer: Searching for Efficient Transformers for Language Modeling</a>.</p>
107+
<h2><a href="hour_glass/index.html">Hourglass</a></h2>
108+
<p>This is an implementation of the paper <a href="https://papers.labml.ai/paper/2110.13711">Hierarchical Transformers Are More Efficient Language Models</a></p>
107109

108110
</div>
109111
<div class='code'>
110-
<div class="highlight"><pre><span class="lineno">98</span><span></span><span class="kn">from</span> <span class="nn">.configs</span> <span class="kn">import</span> <span class="n">TransformerConfigs</span>
111-
<span class="lineno">99</span><span class="kn">from</span> <span class="nn">.models</span> <span class="kn">import</span> <span class="n">TransformerLayer</span><span class="p">,</span> <span class="n">Encoder</span><span class="p">,</span> <span class="n">Decoder</span><span class="p">,</span> <span class="n">Generator</span><span class="p">,</span> <span class="n">EncoderDecoder</span>
112-
<span class="lineno">100</span><span class="kn">from</span> <span class="nn">.mha</span> <span class="kn">import</span> <span class="n">MultiHeadAttention</span>
113-
<span class="lineno">101</span><span class="kn">from</span> <span class="nn">labml_nn.transformers.xl.relative_mha</span> <span class="kn">import</span> <span class="n">RelativeMultiHeadAttention</span></pre></div>
112+
<div class="highlight"><pre><span class="lineno">103</span><span></span><span class="kn">from</span> <span class="nn">.configs</span> <span class="kn">import</span> <span class="n">TransformerConfigs</span>
113+
<span class="lineno">104</span><span class="kn">from</span> <span class="nn">.models</span> <span class="kn">import</span> <span class="n">TransformerLayer</span><span class="p">,</span> <span class="n">Encoder</span><span class="p">,</span> <span class="n">Decoder</span><span class="p">,</span> <span class="n">Generator</span><span class="p">,</span> <span class="n">EncoderDecoder</span>
114+
<span class="lineno">105</span><span class="kn">from</span> <span class="nn">.mha</span> <span class="kn">import</span> <span class="n">MultiHeadAttention</span>
115+
<span class="lineno">106</span><span class="kn">from</span> <span class="nn">labml_nn.transformers.xl.relative_mha</span> <span class="kn">import</span> <span class="n">RelativeMultiHeadAttention</span></pre></div>
114116
</div>
115117
</div>
116118
<div class='footer'>

labml_nn/__init__.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -36,6 +36,7 @@
3636
* [Pay Attention to MLPs (gMLP)](transformers/gmlp/index.html)
3737
* [Vision Transformer (ViT)](transformers/vit/index.html)
3838
* [Primer EZ](transformers/primer_ez/index.html)
39+
* [Hourglass](transformers/hourglass/index.html)
3940
4041
#### ✨ [Recurrent Highway Networks](recurrent_highway_networks/index.html)
4142

labml_nn/transformers/__init__.py

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -93,6 +93,11 @@
9393
9494
This is an implementation of the paper
9595
[Primer: Searching for Efficient Transformers for Language Modeling](https://papers.labml.ai/paper/2109.08668).
96+
97+
## [Hourglass](hour_glass/index.html)
98+
99+
This is an implementation of the paper
100+
[Hierarchical Transformers Are More Efficient Language Models](https://papers.labml.ai/paper/2110.13711)
96101
"""
97102

98103
from .configs import TransformerConfigs

readme.md

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -38,6 +38,7 @@ implementations almost weekly.
3838
* [Pay Attention to MLPs (gMLP)](https://nn.labml.ai/transformers/gmlp/index.html)
3939
* [Vision Transformer (ViT)](https://nn.labml.ai/transformers/vit/index.html)
4040
* [Primer EZ](https://nn.labml.ai/transformers/primer_ez/index.html)
41+
* [Hourglass](https://nn.labml.ai/transformers/hourglass/index.html)
4142

4243
#### [Recurrent Highway Networks](https://nn.labml.ai/recurrent_highway_networks/index.html)
4344

0 commit comments

Comments
 (0)