Skip to content

helboukkouri/awesome-nlp-papers

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 

Repository files navigation

Awesome NLP Papers

This is a collection/reading-list of awesome Natural Language Processing papers sorted by date.

2018

  • Unsupervised Machine Translation Using Monolingual Corpora Only, Lample et al. Paper

  • On the Dimensionality of Word Embeddings, Yin et al. Paper

  • An efficient framework for learning sentence representations, Logeswaran et al. Paper

  • Refining Pretrained Word Embeddings Using Layer-wise Relevance Propagation, Akira Utsumi Paper

  • Domain Adapted Word Embeddings for Improved Sentiment Classification, Sarma et al. Paper

  • In-domain Context-aware Token Embeddings Improve Biomedical Named Entity Recognition, Sheikhshab et al. Paper

  • Generalizing Word Embeddings using Bag of Subwords, Zhao et al. Paper

  • What's in Your Embedding, And How It Predicts Task Performance, Rogers et al. Paper

  • On Learning Better Word Embeddings from Chinese Clinical Records: Study on Combining In-Domain and Out-Domain Data Wang et al. Paper

  • Predicting and interpreting embeddings for out of vocabulary words in downstream tasks, Garneau et al. Paper

  • Addressing Low-Resource Scenarios with Character-aware Embeddings, Papay et al. Paper

  • Domain Adaptation for Disease Phrase Matching with Adversarial Networks, Liu et al. Paper

  • Investigating Effective Parameters for Fine-tuning of Word Embeddings Using Only a Small Corpus, Komiya et al. Paper

  • BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding, Devlin et al. Paper

  • Adapting Word Embeddings from Multiple Domains to Symptom Recognition from Psychiatric Notes, Zhang et al. Paper

  • Evaluation of sentence embeddings in downstream and linguistic probing tasks, Perone et al. Paper

  • Universal Sentence Encoder, Cer et al. Paper

  • Deep Contextualized Word Representations, Peters et al. Paper

  • Learned in Translation: Contextualized Word Vectors, McCann et al. Paper

  • Concatenated p-mean Word Embeddings as Universal Cross-Lingual Sentence Representations, Rücklé et al. paper

  • A Compressed Sensing View of Unsupervised Text Embeddings, Bag-Of-n-Grams, and LSTMs, Arora et al. Paper

2017:

  • Attention Is All You Need, Vaswani et al. Paper

  • Skip-Gram – Zipf + Uniform = Vector Additivity, Gittens et al. Paper

  • A Simple but Tough-to-beat Baseline for Sentence Embeddings, Arora et al. Paper

  • Fast and Accurate Entity Recognition with Iterated Dilated Convolutions, Strubell et al. Paper

  • Advances in Pre-Training Distributed Word Representations, Mikolov et al. Paper

  • Replicability Analysis for Natural Language Processing: Testing Significance with Multiple Datasets, Dror et al. Paper

2016:

  • Towards Universal Paraphrastic Sentence Embeddings, Wieting et al. Paper

  • Bag of Tricks for Efficient Text Classification, Joulin et al. Paper

  • Enriching Word Vectors with Subword Information, Bojanowski et al. Paper

  • Assessing the Corpus Size vs. Similarity Trade-off for Word Embeddings in Clinical NLP, Kirk Roberts Paper

  • How to Train Good Word Embeddings for Biomedical NLP, Chiu et al. Paper

  • Log-Linear Models, MEMMs, and CRFs, Michael Collins Paper

  • Counter-fitting Word Vectors to Linguistic Constraints, Mrkšić et al. Paper

  • Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation, Wu et al. Paper

2015:

  • Semi-supervised Sequence Learning, Dai et al. Paper

  • Evaluating distributed word representations for capturing semantics of biomedical concepts, Th et al. Paper

2014:

  • GloVe: Global Vectors for Word Representation, Pennington et al. Paper

  • Linguistic Regularities in Sparse and Explicit Word Representations, Levy and Goldberg. Paper

  • Neural Word Embedding as Implicit Matrix Factorization, Levy and Goldberg. Paper

  • word2vec Explained: deriving Mikolov et al.'s negative-sampling word-embedding method, Goldberg and Levy. Paper

  • What’s in a p-value in NLP?, Søgaard et al. Paper

  • How transferable are features in deep neural networks?, Yosinski et al. Paper

  • Improving lexical embeddings with semantic knowledge, Yu et al. Paper

  • Retrofitting word vectors to semantic lexicons, Faruqui et al. Paper

2013:

  • Efficient Estimation of Word Representations in Vector Space, Mikolov et al. Paper

  • Linguistic Regularities in Continuous Space Word Representations, Mikolov et al. Paper

  • Distributed Representations of Words and Phrases and their Compositionality, Mikolov et al. Paper

2012:

  • An Empirical Investigation of Statistical Significance in NLP, Berg-Kirkpatrick et al. Paper

2010:

  • Word representations: A simple and general method for semi-supervised learning, Turian et al. Paper

2008:

  • A Unified Architecture for Natural Language Processing: Deep Neural Networks with Multitask Learning, Collobert and Weston. Paper

2006:

  • Domain adaptation with structural correspondence learning, Blitzer et al. Paper

2003:

  • A Neural Probabilistic Language Model, Bengio et al. Paper

1986:

  • Distributed Representations, Hinton et al. Paper

About

A collection/reading-list of awesome Natural Language Processing papers sorted by date.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published