This is a fast implementation of integration Chainer with Intel® Math Kernel Library for Deep Neural Networks (Intel® MKL-DNN). It accelerates Chainer on CPU, esp. Intel® Xeon® and Intel® Xeon Phi™ processors. Current optimized layers (operations) includes convolution (2D), local response normalization, ReLU, linear (inner product), pooling, concat, sum and gradient accumulation. Validated topologies includes Alexnet, Overfeat, VGGA, VGG-16, VGG-19 and GoogLeNet-v1 with performance gain from 50-250X on Xeon and Xeon Phi.
Chainer is tested on Ubuntu 14.04 and CentOS 7. We recommend them to use Chainer, though it may run on other systems as well.
Minimum requirements:
- Python 2.7.6+, 3.4.3+, 3.5.1+, 3.6.0+
- NumPy 1.9, 1.10, 1.11, 1.12
- Six 1.9
Requirements for some features:
- Intel MKL-DNN support
- mkl-dnn 0.7
- g++ 4.8.4+
- swig 3.0
- glog 0.3.3
- gflags 2.0
- python-setuptools 3.3
- boost 1.58
- CUDA support
- CUDA 6.5, 7.0, 7.5, 8.0
- filelock
- g++ 4.8.4+
- cuDNN support
- cuDNN v2, v3, v4, v5, v5.1
- Caffe model support
- Protocol Buffers (pip install protobuf)
- protobuf>=3.0.0 is required for Py3
- Protocol Buffers (pip install protobuf)
- Image dataset support
- Pillow
- HDF5 serialization support
- h5py 2.5.0
- Testing utilities
- Mock
- Nose
If you use old setuptools
, upgrade it:
pip install -U setuptools
Then, install Chainer via PyPI:
pip install chainer
You can also install Chainer from the source code:
python setup.py install
To enable MKL-DNN, first you have to install MKL-DNN library.
git clone https://github.com/01org/mkl-dnn.git
cd scripts && ./prepare_mkl.sh && cd -
mkdir -p build && cd build && cmake .. && make -j
sudo make install
python setup.py build
python setup.py install
If you want to enable CUDA, first you have to install CUDA and set the environment variable PATH
and LD_LIBRARY_PATH
for CUDA executables and libraries.
For example, if you are using Ubuntu and CUDA is installed by the official distribution, then CUDA is installed at /usr/local/cuda
.
In this case, you have to add the following lines to .bashrc
or .zshrc
(choose which you are using):
export PATH=/usr/local/cuda/bin:$PATH
export LD_LIBRARY_PATH=/usr/local/cuda/lib64:$LD_LIBRARY_PATH
Chainer had chainer-cuda-deps
module to enable CUDA in previous version.
Recent version (>=1.3) does not require this module.
So you do not have to install chainer-cuda-deps
.
If you want to enable cuDNN, add a directory containing cudnn.h
to CFLAGS
, and add a directory containing libcudnn.so
to LDFLAGS
and LD_LIBRARY_PATH
:
export CFLAGS=-I/path/to/cudnn/include
export LDFLAGS=-L/path/to/cudnn/lib
export LD_LIBRARY_PATH=/path/to/cudnn/lib:$LD_LIBRARY_PATH
Do not forget to restart your terminal session (or source
it) to enable these changes.
And then, reinstall Chainer.
If you want to use Image dataset (chainer/datasets/ImageDataset
), please install Pillow manually.
Supported image format depends on your environment.
pip install pillow
If you want to use HDF5 serialization, please install h5py manually. h5py requires libhdf5. Anaconda distribution includes this package. If you are using another Python distribution, use either of the following commands to install libhdf5 depending on your Linux environment:
apt-get install libhdf5-dev
yum install hdf5-devel
And then, install h5py via PyPI. You may need to install Cython for h5py.
pip install cython
pip install h5py
Set environment variable LD_LIBRARY_PATH
for MKL-DNN library before run, most likely it will be /usr/local/lib:
export LD_LIBRARY_PATH=/usr/local/lib:$LD_LIBRARY_PATH
The rest of the steps is the same as before. To run convnet-benchmarks on IA, please check out convnet-benchmarks repo:
cd chainer
./train_imagenet.py -a alexnet -B 128 -g -1
Note: if an error of "AttributeError: 'module' object has no attribute 'cupy'" is reported, please refer to the following PR for the fix: Timer fix for IA
We provide the official Docker image. Use nvidia-docker command to run Chainer image with GPU. You can login to the environment with bash, and run the Python interpreter.
$ nvidia-docker run -it chainer/chainer /bin/bash
Tokui, S., Oono, K., Hido, S. and Clayton, J., Chainer: a Next-Generation Open Source Framework for Deep Learning, Proceedings of Workshop on Machine Learning Systems(LearningSys) in The Twenty-ninth Annual Conference on Neural Information Processing Systems (NIPS), (2015) URL, BibTex
- Official site: http://chainer.org/
- Official document: http://docs.chainer.org/
- github: https://github.com/pfnet/chainer
- Forum: https://groups.google.com/forum/#!forum/chainer
- Forum (Japanese): https://groups.google.com/forum/#!forum/chainer-jp
- Twitter: https://twitter.com/ChainerOfficial
- Twitter (Japanese): https://twitter.com/chainerjp
- External examples: https://github.com/pfnet/chainer/wiki/External-examples
- Research projects using Chainer: https://github.com/pfnet/chainer/wiki/Research-projects-using-Chainer
MIT License (see LICENSE
file).