Skip to content

coxlab/tsnet

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

TSNet

This repository provides code and scripts for reproducing the experimental results of Tensor Switching Networks by Chuan-Yung Tsai, Andrew Saxe, and David Cox.

TSNet is a novel neural network algorithm, which generalizes the Rectified Linear Unit (ReLU) nonlinearity to tensor-valued hidden units, and avoids the vanishing gradient problem by construction. Our experimental results show that the TSNet is not only more expressive, but also consistently learns faster than standard ReLU networks.

Requirements

Keras, Kerosene, Blessings, and libsvm-compact.

Usage

Use bash nips06/run.sh if you wish to run all of our experiments. Otherwise, use python tsnet_cli.py to run single models. For example, you can use the following commands to compare a simple single-hidden-layer ReLU network (first line) and its TS counterpart using the inverted backpropagation learning (second line):

python tsnet_cli.py -d mnist -n conv:0/16 relu:0 flat:0 sfmx:0/10 -e 10 -lrnparam 1e-3 1e-3 0.9 -v 1
python tsnet_cli.py -d mnist -n conv:2/16 relu:2 flat:0 sfmx:0/10 -e 10 -lrnparam 1e-3 1e-3 0.9 -v 1

By default, the Numpy backend supporting all learning algorithms is used, but you can also switch to the simpler and faster Keras backend by using -k. Please refer to cmp_mlp.py and cmp_cnn.py for more examples of how to define networks.

Releases

No releases published

Packages

No packages published