Here is a detailed schedule, session by session:
-
Introduction / Refresher on Deep Learning
- General Intro - Why do we need to optimize deep learning ? Introduction of the MicroNet Challenge.
- Course - Deep Learning and Transfer Learning.
- Practical session - introduction to PyTorch, transfer learning.
- Short project - exploring hyper parameters on a fixed architecture
-
Quantification
- Short evaluation on Deep Learning Essentials
- Student's presentation of short project - exploring hyper parameters on a fixed architecture
- Course - Quantifying Deep neural networks
- Practical session - quantification on a small convolutional network
- Long project 1 - MicroNet Challenge
-
Pruning
- Short evaluation on Quantification
- Course - Pruning Deep neural networks
- Practical session - pruning on a small convolutional network.
- Long project 2 - MicroNet Challenge
-
Factorization
- Short evaluation on Pruning
- Student's presentation on current work on MicroNet
- Course - Factorizing Deep neural networks
- Practical session - factorizing a small convolutional network
- Long Project 3 - MicroNet Challenge
-
Factorisation - Part 2 - Operators and Architectures
- Course - Factorization Pt2, alternative operators and efficient architectures
- Long Project 5 - MicroNet Challenge
-
Distillation
- Short evaluation on Factorization Pt1 and Pt2 and previous courses
- Course - Distillation of knowledge and features between neural networks
- Long Project 4 - MicroNet Challenge
-
Embedded Software and Hardware for Deep Learning
- Short evaluation on Distillation
- Course - Embedded Software and Hardware for Deep Learning
- Long Project 6 - MicroNet Challenge
-
Final Session
- Short evaluation on embedded software and hardware for Deep Learning
- Long Project 7 - MicroNet Challenge
- Student's presentation - Final results on MicroNet
There are short written evaluations during the first 10 minutes of each session starting from session 2. Don't be late!
For the final session, we ask you to prepare a 20 minutes presentation, that will be followed by 10 Minutes of question.
You'll find in the micronet-ressources folder, presentations from the winners of the 2019, and rules for the 2020 challenge.
List of references IMT Atlantique and AI
Amazon Book - Dive into Deep learning
Tutorial presentation on Efficient Deep Learning from NeurIPS'19
Here are some academic papers discussing learning rate strategies :
- Cyclic learning rates
- Demystifying Learning Rate Policies for High Accuracy Training of Deep Neural Networks
- A Closer Look at Deep Learning Heuristics: Learning rate restarts, Warmup and Distillation
Main strategies are readily available in pytorch.
Start page to access the full python API of pytorch, to check all existing functions.
A useful tutorial on Saving and Loading models.
Popular methods :
Other ressources :
A list of papers and code for data augmentation
IMGAUG and Colab Notebook showing how to use IMGAUG with pytorch
A popular python package in Kaggle competitions : Albumentations
Pruning Filters for Efficient ConvNets
AutoML for Model Compression (AMC)
Pruning Channel with Attention Statistics (PCAS)
BitPruning: Learning Bitlengths for Aggressive and Accurate Quantization
Distilling the knowledge in a neural network
Fitnets: Hints for thin deep nets
LIT: Learned Intermediate Representation Training for Model Compression
A Comprehensive Overhaul of Feature Distillation
And the bit goes down: Revisiting the quantization of neural networks
See references section of Tutorial presentation on Efficient Deep Learning from NeurIPS'19.