Skip to content

evolvingstuff/SimpleLSTM

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

This is a recurrent neural network architecture inspired by Long Short Term Memory, but with a much simpler architecture.

I will be adding more documentation to describe this architecture shortly. For now, a brief summary is that it has only a single gate (similar to the Forget Gate in LSTM) and an input squashing function. Rather than acting to reset the state of the cell, the gate modulates a weighted average of the current input with the state of the cell at the previous time step. So if the gate is fully active, the cell will ignore its current input and fully retain its state. If the gate is inactive then the cell will lose all traces of its previous state and shift to match the current input.

So far on a few tests I've done, it appears to perform fairly well, although it tends to need more cell blocks than LSTM to do the same thing.

About

A recurrent neural network heavily inspired by Long Short Term Memory, but simpler.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages