Skip to content
master
Go to file
Code

Latest commit

 

Git stats

Files

Permalink
Failed to load latest commit information.
Type
Name
Latest commit message
Commit time
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

README.md

N2N: Network to Network Compression using Policy Gradient Reinforcement Learning (ICLR 2018)

This is the code to run the model compression algorithm described in the paper. It currently supports trained models in pytorch. If you would like to use it with a model in another deep learning framework, it would have to be converted to pytorch first. Link to ICLR paper

Dependencies

There are some dependencies for running this

  1. python >= 2.7
  2. pytorch >= 0.2
  3. torchvision >= 0.19

How to run

  1. Clone this repository using
git clone https://github.com/anubhavashok/N2N.git
  1. Download teacher models from the links below

  2. Layer removal and Layer shrinkage instructions are described below Additional detailed instructions can be found in the help menu in run.py

Removal

Here is an example command to train the layer removal policy on the cifar10 dataset using the resnet-18 model

python run.py removal cifar10 teacherModels/resnet18_cifar10.net --cuda True 

Shrinkage

NOTE: To run shrinkage, specify both teacher model and reduced model from stage1

python run.py shrinkage cifar10 teacherModels/resnet18_cifar10.net --model Stage1_cifar10/reduced_model1.net --cuda True 

Downloading models

All models can be downloaded at this link

Pre-trained teacher models

The teacher models are to be specified to run.py to train.

Pre-trained student models

The pre-trained student models are given to show the performance of the models described in the paper. They can be tested using test_model.py Test using

python test_model.py studentModels/resnet18_cifar10.net cifar10

Pre-trained policies

The pre-trained polcies are specified to run the transfer learning experiments

Experiments folder

The experiments folder contains various variants of layer removal and shrinkage that were tried for the actual paper. These were mainly experiments which require substantial modifications to the main code or were used on earlier iterations of the project. They have to be moved to the main folder before being run. The following describes each experiment

  1. ar_run_layer_clean.py - Layer removal using the Autoregressive controller
  2. ar_run_param_clean.py - Layer shrinkage for Non-ResNet convolutional models
  3. bd_run_layer_clean.py - Layer removal for Non-ResNet convolutional models using the bidirectional controller
  4. ed_run_layer_general.py - Layer removal for Non-ResNet convolutional models using the encoder-decoder controller
  5. resnet_actor_critic_layer.py - Layer removal using the Actor-Critic controller
  6. resnet_ar_run_layer_clean.py - Layer removal for ResNet models using the Autoregressive controller

Citing

Please use the following bibtex to cite the paper:

@inproceedings{
ashok2018nn,
title={N2N learning: Network to Network Compression via Policy Gradient Reinforcement Learning},
author={Anubhav Ashok and Nicholas Rhinehart and Fares Beainy and Kris M. Kitani},
booktitle={International Conference on Learning Representations},
year={2018},
url={https://openreview.net/pdf?id=B1hcZZ-AW},
}

About

No description, website, or topics provided.

Resources

License

Releases

No releases published

Packages

No packages published

Languages

You can’t perform that action at this time.