Skip to content
ImageNet pre-trained models with batch normalization for the Caffe framework
Python
Branch: master
Clone or download
Latest commit a0ade4d Nov 27, 2017
Permalink
Type Name Latest commit message Commit time
Failed to load latest commit information.
AlexNet_cvgj Added Baidu and FSU Jena download links and sha256 hash Dec 8, 2016
ResNet_preact fix typo in table of README.md Sep 8, 2017
VGG19_cvgj Added Baidu and FSU Jena download links and sha256 hash Dec 8, 2016
.gitignore Added gitignore Dec 7, 2016
LICENSE Fixed typo Nov 26, 2017
README.md Update README.md Sep 19, 2017

README.md

CNN Models by CVGJ

Intro

This repository contains convolutional neural network (CNN) models trained on ImageNet by Marcel Simon at the Computer Vision Group Jena (CVGJ) using the Caffe framework as published in the accompanying technical report. Each model is in a separate subfolder and contains everything needed to reproduce the results. This repository focuses currently contains the batch-normalization-variants of AlexNet and VGG19 as well as the training code for Residual Networks (Resnet).

How to use

No mean subtraction is required for the pre-trained models! We have a batch-normalization layer which basically does the same.

The pre-trained models can be obtained by the download link written in model_download_link.txt.

If you want to train on your own dataset, simply execute caffe train --solver train.solver --gpu 0 2> train.log to start the training and write the output to the log file train.log. Please note, that we pre-scaled all images of the dataset such that the smaller side has a length of 256 pixels while keeping the aspect ratio. You can use convert input.jpg -resize 256x256^ output.jpg to convert the images using the command line.

To evaluate the final model, execute caffe train --solver test.solver --gpu 0 2> test.log.

Accuracy on ImageNet

Single-crop error rates on the validation set of the ILSVRC 2012--16 classification task.

Model Top-1 error (vs. original) Top-5 error (vs. original)
AlexNet_cvgj 39.9% (vs. 42.6%) 18.1% (vs. 19.6%)
VGG19_cvgj 26.9% (vs. 28.7%) 8.8% (vs. 9.9%)
ResNet10_cvgj 36.1% 14.8%
ResNet50_cvgj 24.6% (vs. 24.7%) 7.6% (vs. 7.8%)

Citation

Please cite the following technical report if our models helped your research:

@article{simon2016cnnmodels,
  Author = {Simon, Marcel and Rodner, Erik and Denzler, Joachim},
  Journal = {arXiv preprint arXiv:1612.01452},
  Title = {ImageNet pre-trained models with batch normalization},
  Year = {2016}
}

The report also contains an overview and analysis of the models shown here.

Appendix: convergence plots

AlexNet_cvgj

Convergence plot of AlexNet with batch normalization

VGG19_cvgj

Convergence plot of AlexNet with batch normalization

ResNet10_cvgj

Convergence plot of AlexNet with batch normalization

Further details

Please see the accompanying technical report for further information about the models and the training procedure:

@article{simon2016cnnmodels,
  Author = {Simon, Marcel and Rodner, Erik and Denzler, Joachim},
  Journal = {arXiv preprint arXiv:1612.01452},
  Title = {ImageNet pre-trained models with batch normalization},
  Year = {2016}
}

License and support

The models are released under BSD 2-clause license allowing both academic and commercial use. I would appreciate if you give credit to this work by citing our paper in academic works and referencing to this Github repository in commercial works. If you need any support, please open an issue or contact Marcel Simon.

You can’t perform that action at this time.