File tree 1 file changed +32
-0
lines changed
1 file changed +32
-0
lines changed Original file line number Diff line number Diff line change
1
+ Attention Is All You Need
2
+ https://arxiv.org/abs/1706.03762
3
+
4
+ BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
5
+ https://arxiv.org/abs/1810.04805v2
6
+
7
+ Improving Language Understanding by Generative Pre-Training (GPT)
8
+ https://s3-us-west-2.amazonaws.com/openai-assets/research-covers/language-unsupervised/language_understanding_paper.pdf
9
+
10
+ Improving Language Understanding with Unsupervised Learning
11
+ https://openai.com/blog/language-unsupervised/
12
+
13
+ Language Models are Unsupervised Multitask Learners (GPT-2)
14
+ https://d4mucfpksywv.cloudfront.net/better-language-models/language_models_are_unsupervised_multitask_learners.pdf
15
+
16
+ Better Language Models and Their Implications
17
+ https://openai.com/blog/better-language-models/
18
+
19
+ Language Models are Few-Shot Learners (GPT-3)
20
+ https://arxiv.org/abs/2005.14165
21
+
22
+ List of Hugging Face Pipelines for NLP
23
+ https://lazyprogrammer.me/list-of-hugging-face-pipelines-for-nlp/
24
+
25
+ BitFit: Simple Parameter-efficient Fine-tuning for Transformer-based Masked Language-models
26
+ https://arxiv.org/abs/2106.10199
27
+
28
+ Translation Datasets
29
+ https://opus.nlpl.eu/KDE4.php
30
+
31
+ Layer Normalization
32
+ https://arxiv.org/abs/1607.06450
You can’t perform that action at this time.
0 commit comments