Skip to content
master
Switch branches/tags
Code

Latest commit

 

Git stats

Files

Permalink
Failed to load latest commit information.
Type
Name
Latest commit message
Commit time
 
 

NLP on IMDB Dataset for sentimental analysis

Blogs

Dataset

Large Movie Review Dataset This is a dataset for binary sentiment classification containing substantially more data than previous benchmark datasets. We provide a set of 25,000 highly polar movie reviews for training, and 25,000 for testing. There is additional unlabeled data for use as well. Raw text and already processed bag of words formats are provided. See the README file contained in the release for more details. :http://ai.stanford.edu/~amaas/data/sentiment/

Research Papers

Universal Language Model Fine-tuning for Text Classification: https://arxiv.org/abs/1801.06146

Pretrained model that we used for Transfer Learning

WikiText-2(SalesForce) by Stephen Merity

The WikiText language modeling dataset is a collection of over 100 million tokens extracted from the set of verified Good and Featured articles on Wikipedia. The dataset is available under the Creative Commons Attribution-ShareAlike License.

https://blog.einstein.ai/the-wikitext-long-term-dependency-language-modeling-dataset/

Implementation of RNN

Machine Info

Trained on: AWS Instance p2.xlarge 16 GIGS vRAM 64 GB RAM

About

No description, website, or topics provided.

Resources

Releases

No releases published

Packages

No packages published