Experiments with Group Equivariant Convolutional Networks
Switch branches/tags
Nothing to show
Clone or download
Fetching latest commit…
Cannot retrieve the latest commit at this time.

README.md

G-CNN Experiments

Code for reproducing the experiments reported in: T.S. Cohen, M. Welling, Group Equivariant Convolutional Networks. Proceedings of the International Conference on Machine Learning (ICML), 2016

p4_anim

A rotating feature map on the rotation-translation group p4. See section 4 of the paper.

Results

Comparison to other methods

There are two common regimes for comparing classifiers on CIFAR10:

  • No data augmentation (CIFAR10)
  • Modest augmentation with small translations and horizontal flips (CIFAR10+)
Network CIFAR10 CIFAR10+
Maxout [8] 11.68 9.38
DropConnect [9] 9.32
NiN [10] 10.41 8.81
DSN [13] 9.69 7.97
All-CNN-C [11] 9.07 7.25
Highway nets [15] 7.6
ELU [12] 6.55
Generalized Pooling [14] 7.62 6.05
ResNet1001 [18] 4.62
ResNet101 [17] 6.43
Wide ResNet 28 [16] 4.17
p4m-Resnet26 (ours) 5.74 4.19

Comparison of G-CNNs

We compare the following group convolutions:

  • Z2 is the 2D translation group (used in a standard CNN)
  • p4 is the group of translations and rotations by multiples of 90 degrees
  • p4m is the group of translations, reflections and rotations by multiples of 90 degrees

We compare the following architectures:

  • All-CNN-C is described in [11].
  • ResNet44 is a 44 layer residual network [18] with 32, 64, 128 filters per stage. For the p4m version, we use 11, 23, 45 filters per stage.
  • ResNet26 is a 26 layer residual network [18] with 71, 142, 284 filters per stage. For the p4m version, we use 25, 50, 100 filters per stage.
Network G CIFAR10 CIFAR10+ Param.
All-CNN-C Z2 9.44 8.86 1.37M
p4 8.84 7.67 1.37M
p4m 7.59 7.04 1.22M
ResNet44 Z2 9.45 5.61 2.64M
p4m 6.46 4.94 2.62M
ResNet26 Z2 8.95 5.27 7.24M
p4m 5.74 4.19 7.17M

We see a very consistent behaviour:

  • p4-CNNs outperform Z2-CNNs
  • p4m-CNNs outperform p4-CNNs
  • Data augmentation is beneficial for all CNNs.

Installation

Install scientific python stack and progressbar

$ pip install ipython numpy scipy matplotlib progressbar2 skimage

Install chainer with CUDNN and HDF5: installation instructions

Install GrouPy

Add the gconv_experiments folder to your PYTHONPATH.

Download data

CIFAR10

$ cd [datadir]
$ wget http://www.cs.toronto.edu/~kriz/cifar-10-python.tar.gz
$ tar zxvf cifar-10-python.tar.gz
$ rm cifar-10-python.tar.gz

MNIST-rot

$ cd [datadir]
$ wget http://www.iro.umontreal.ca/~lisa/icml2007data/mnist_rotation_new.zip
$ unzip mnist_rotation_new.zip 
$ rm mnist_rotation_new.zip
$ ipython /path/to/gconv_experiments/gconv_experiments/MNIST_ROT/mnist_rot.py -- --datadir=./

Train a G-CNN

MNIST-rot

To run the MNIST-rot experiments:

$ ipython MNIST_ROT/experiment.py -- --trainfn=[datadir]/train_all.npz --valfn=[datadir]/test.npz

You can also call train.py directly to train a single model.

CIFAR10

The first time you run an experiment, the code will preprocess the dataset and leave a preprocessed copy in [datadir].

$ cd gconv_experiments
$ ipython CIFAR10/train.py -- --datadir=[datadir] --resultdir=[resultdir] --modelfn=CIFAR10/models/P4AllCNNC.py

For other options, see train.py.

References

Related work

  1. Kivinen, Jyri J. and Williams, Christopher K I. Transformation equivariant Boltzmann machines. In 21st International Conference on Artificial Neural Networks, 2011.
  2. Sifre, Laurent and Mallat, Stephane. Rotation, Scaling and Deformation Invariant Scattering for Texture Discrimination. IEEE conference on Computer Vision and Pattern Recognition (CVPR), 2013.
  3. Gens, R. and Domingos, P. Deep Symmetry Networks. In Advances in Neural Information Processing Systems (NIPS), 2014.
  4. Jaderberg, M., Simonyan, K., Zisserman, A., and Kavukcuoglu, K. Spatial Transformer Networks. In Advances in Neural Information Processing Systems 28 (NIPS 2015), 2015.
  5. Oyallon, E. and Mallat, S. Deep Roto-Translation Scattering for Object Classification. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2015.
  6. Zhang, C., Voinea, S., Evangelopoulos, G., Rosasco, L., and Poggio, T. Discriminative template learning in group-convolutional networks for invariant speech representations. InterSpeech, 2015.
  7. Dieleman, S., De Fauw, J., and Kavukcuoglu, K. Exploiting Cyclic Symmetry in Convolutional Neural Networks. In International Conference on Machine Learning (ICML), 2016.

Papers reporting results on CIFAR10

  1. Goodfellow, I. J., Warde-Farley, D., Mirza, M., Courville, A., and Bengio, Y. Maxout Networks. In Proceedings of the 30th International Conference on Machine Learning (ICML), pp. 1319–1327, 2013.
  2. Wan, L., Zeiler, M., Zhang, S., LeCun, Y., and Fergus, R. Regularization of neural networks using dropconnect. International Conference on Machine Learning (ICML), 2013.
  3. Lin, M., Chen, Q., and Yan, S. Network In Network. International Conference on Learning Representations (ICLR), 2014.
  4. Springenberg, J.T., Dosovitskiy, A., Brox, T., and Riedmiller, M. Striving for Simplicity: The All Convolutional Net. Proceedings of the International Conference on Learning Representations (ICLR), 2015.
  5. Clevert, D., Unterthiner, T., and Hochreiter, S. Fast and Accurate Deep Network Learning by Exponential Linear Units (ELUs). arXiv:1511.07289v3, 2015.
  6. Lee, C., Xie, S., Gallagher, P.W., Zhang, Z., and Tu, Z. Deeply-Supervised Nets. In Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics (AISTATS), 2015.
  7. Lee, C., Gallagher, P. W., and Tu, Z. Generalizing Pooling Functions in Convolutional Neural Networks: Mixed, Gated, and Tree. ArXiv:1509.08985, 2015.
  8. Srivastava, Rupesh Kumar, Greff, Klaus, and Schmidhuber, Jurgen. Training Very Deep Networks. Advances in Neural Information Processing Systems (NIPS), 2015.
  9. Zagoruyko, S. and Komodakis, N. Wide Residual Networks. arXiv:1605.07146, 2016.
  10. He, K., Zhang, X., Ren, S., and Sun, J. Deep Residual Learning for Image Recognition. arXiv:1512.03385, 2015.
  11. He, Kaiming, Zhang, Xiangyu, Ren, Shaoqing, and Sun, Jian. Identity Mappings in Deep Residual Networks. arXiv:1603.05027, 2016.