A recurrent neural network heavily inspired by Long Short Term Memory, but simpler.
License
evolvingstuff/SimpleLSTM
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
This is a recurrent neural network architecture inspired by Long Short Term Memory, but with a much simpler architecture. I will be adding more documentation to describe this architecture shortly. For now, a brief summary is that it has only a single gate (similar to the Forget Gate in LSTM) and an input squashing function. Rather than acting to reset the state of the cell, the gate modulates a weighted average of the current input with the state of the cell at the previous time step. So if the gate is fully active, the cell will ignore its current input and fully retain its state. If the gate is inactive then the cell will lose all traces of its previous state and shift to match the current input. So far on a few tests I've done, it appears to perform fairly well, although it tends to need more cell blocks than LSTM to do the same thing.
About
A recurrent neural network heavily inspired by Long Short Term Memory, but simpler.
Resources
License
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published