A library for Recurrent Neural Networks built on top of Theano. It allows quick prototyping of complicated structures of Recurrent Neural Networks. Checkout some of the existing architectures (and research papers) implemented using NeuralModels https://github.com/asheshjain399/RNNexp/
python setup.py develop
In order to check NeuralModels is correctly installed, try the character-rnn example.
neuralmodels: Python module containing definition of layers, optimization methods, and few models.
NeuralModels comes with some pre-implemented models in the models directory
is the structural-RNN (S-RNN) code for doing deep learning on spatio-temporal graphs. The paper is present here http://www.cs.stanford.edu/people/ashesh/srnn See the repository https://github.com/asheshjain399/RNNexp/ for examples to use S-RNN
This is a simple RNN implementation.
In order to implement strucutures of RNN see examples such as models/SharedRNN.py
Many standard layers are pre-implemented in NeuralModels:
implements the standard LSTM with gradient clipping (tune-able parameters)
Use this to create a stack of LSTM with skip-input and output connections. If skip-input and output connections are not desired, then simply create an array of layers/LSTM.py
implements the Recurrent Neural Network
this is the softmax layer, or commonly the output layer of the architecture.
this layer generates the one-hot vector representation, this is commonly the input layer of the architecture when dealing with finite size vocabulary (eg. in NLP)
this is the input layer of the architecture when we have precomputed feature vectors. The shape of input is
T x N x D
T is the number of time-steps.
N is the number of sequences.
D is the dimension of each feature vector.
Creating a new deep architecture
Say if you need to create a new architecture with 3-layers of LSTM, and the input into the architecture are precomputed feature vectors, and the output is softmax scores. We first import the relevant python modules:
import theano from theano import tensor as T from neuralmodels.layers import * from neuralmodels.models import * from neuralmodels.updates import Adagrad from neuralmodels.costs import softmax_loss
Next we define our architecture:
layers = [TemporalInputFeatures(inputDim),LSTM(size=512),LSTM(size=512),LSTM(size=512),softmax(size=numClasses)]
Now we will create the model and build its computation graph:
trY = T.lmatrix() initial_step_size = 1e-3 model = RNN(layers,softmax_loss,trY,initial_step_size,Adagrad())
In order to train the architecture see the
fitModel() function in