Skip to content

bgavran/LSTM

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

20 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

LSTM

Playing around with various LSTM architectures and figuring out TensorFlow. Using the IMDb review data to do that.

Achieved ~85% accuracy.

Visualization

The network yields different types of activations on different neurons. The visualizations of the most interesting neurons are shown below. Each image represents the activations of one single neuron in the network. Each row represents one input example (one review). Each column represents the activations of that neuron at time t, for all the shown input examples.

Neuron shown below fires positively when detects words such as worst, bad, awful and positively when it detects words such as classic, good, finest etc. It also puts them on a linear scale where some words are more positive/negative than others. alt text

This neuron learned to count zeroes: alt text

For comparison, here is a neuron that is not very interesting, which models some uninterpretable features. alt text

The whole data structure (activations of all the neurons for all the examples for all the time steps) could be seen as a 3d tensor from which we're extracting various lower dimensional slices (in this case, abstractions over two variables, i.e., images).

Inspired by Karpathy's amazing article.

About

Playing around with various LSTM architectures and figuring out TensorFlow

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages