This is a collection/reading-list of awesome Natural Language Processing papers sorted by date.
-
Unsupervised Machine Translation Using Monolingual Corpora Only, Lample et al.
Paper
-
On the Dimensionality of Word Embeddings, Yin et al.
Paper
-
An efficient framework for learning sentence representations, Logeswaran et al.
Paper
-
Refining Pretrained Word Embeddings Using Layer-wise Relevance Propagation, Akira Utsumi
Paper
-
Domain Adapted Word Embeddings for Improved Sentiment Classification, Sarma et al.
Paper
-
In-domain Context-aware Token Embeddings Improve Biomedical Named Entity Recognition, Sheikhshab et al.
Paper
-
Generalizing Word Embeddings using Bag of Subwords, Zhao et al.
Paper
-
What's in Your Embedding, And How It Predicts Task Performance, Rogers et al.
Paper
-
On Learning Better Word Embeddings from Chinese Clinical Records: Study on Combining In-Domain and Out-Domain Data Wang et al.
Paper
-
Predicting and interpreting embeddings for out of vocabulary words in downstream tasks, Garneau et al.
Paper
-
Addressing Low-Resource Scenarios with Character-aware Embeddings, Papay et al.
Paper
-
Domain Adaptation for Disease Phrase Matching with Adversarial Networks, Liu et al.
Paper
-
Investigating Effective Parameters for Fine-tuning of Word Embeddings Using Only a Small Corpus, Komiya et al.
Paper
-
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding, Devlin et al.
Paper
-
Adapting Word Embeddings from Multiple Domains to Symptom Recognition from Psychiatric Notes, Zhang et al.
Paper
-
Evaluation of sentence embeddings in downstream and linguistic probing tasks, Perone et al.
Paper
-
Universal Sentence Encoder, Cer et al.
Paper
-
Deep Contextualized Word Representations, Peters et al.
Paper
-
Learned in Translation: Contextualized Word Vectors, McCann et al.
Paper
-
Concatenated p-mean Word Embeddings as Universal Cross-Lingual Sentence Representations, Rücklé et al.
paper
-
A Compressed Sensing View of Unsupervised Text Embeddings, Bag-Of-n-Grams, and LSTMs, Arora et al.
Paper
-
Attention Is All You Need, Vaswani et al.
Paper
-
Skip-Gram – Zipf + Uniform = Vector Additivity, Gittens et al.
Paper
-
A Simple but Tough-to-beat Baseline for Sentence Embeddings, Arora et al.
Paper
-
Fast and Accurate Entity Recognition with Iterated Dilated Convolutions, Strubell et al.
Paper
-
Advances in Pre-Training Distributed Word Representations, Mikolov et al.
Paper
-
Replicability Analysis for Natural Language Processing: Testing Significance with Multiple Datasets, Dror et al.
Paper
-
Towards Universal Paraphrastic Sentence Embeddings, Wieting et al.
Paper
-
Bag of Tricks for Efficient Text Classification, Joulin et al.
Paper
-
Enriching Word Vectors with Subword Information, Bojanowski et al.
Paper
-
Assessing the Corpus Size vs. Similarity Trade-off for Word Embeddings in Clinical NLP, Kirk Roberts
Paper
-
How to Train Good Word Embeddings for Biomedical NLP, Chiu et al.
Paper
-
Log-Linear Models, MEMMs, and CRFs, Michael Collins
Paper
-
Counter-fitting Word Vectors to Linguistic Constraints, Mrkšić et al.
Paper
-
Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation, Wu et al.
Paper
-
Semi-supervised Sequence Learning, Dai et al.
Paper
-
Evaluating distributed word representations for capturing semantics of biomedical concepts, Th et al.
Paper
-
GloVe: Global Vectors for Word Representation, Pennington et al.
Paper
-
Linguistic Regularities in Sparse and Explicit Word Representations, Levy and Goldberg.
Paper
-
Neural Word Embedding as Implicit Matrix Factorization, Levy and Goldberg.
Paper
-
word2vec Explained: deriving Mikolov et al.'s negative-sampling word-embedding method, Goldberg and Levy.
Paper
-
What’s in a p-value in NLP?, Søgaard et al.
Paper
-
How transferable are features in deep neural networks?, Yosinski et al.
Paper
-
Improving lexical embeddings with semantic knowledge, Yu et al.
Paper
-
Retrofitting word vectors to semantic lexicons, Faruqui et al.
Paper
-
Efficient Estimation of Word Representations in Vector Space, Mikolov et al.
Paper
-
Linguistic Regularities in Continuous Space Word Representations, Mikolov et al.
Paper
-
Distributed Representations of Words and Phrases and their Compositionality, Mikolov et al.
Paper
- An Empirical Investigation of Statistical Significance in NLP, Berg-Kirkpatrick et al.
Paper
- Word representations: A simple and general method for semi-supervised learning, Turian et al.
Paper
- A Unified Architecture for Natural Language Processing: Deep Neural Networks with Multitask Learning, Collobert and Weston.
Paper
- Domain adaptation with structural correspondence learning, Blitzer et al.
Paper
- A Neural Probabilistic Language Model, Bengio et al.
Paper
- Distributed Representations, Hinton et al.
Paper