Skip to content

Generative language model in Theano/Lasagne using noise contrastive estimation (NCE)

License

Notifications You must be signed in to change notification settings

cjratcliff/nce-lm

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

nce-lm

Generative language model in Theano/Lasagne using noise contrastive estimation (NCE)

Trains a word-level language model on a text file using an LSTM or GRU. Once the model is trained a piece of text is generated.

Using NCE greatly improves efficiency for word-level language models where the large vocabulary size makes computing softmax inefficient.

NCE can only be used during the training of the model as it describes a loss function, not a way of outputting predictions. A full softmax is used for generating samples.

Adapted from the character-level model by Eben Olson here.

About

Generative language model in Theano/Lasagne using noise contrastive estimation (NCE)

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages