Skip to content

phquang/Contextual-Transformation-Network

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

42 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

CONTEXTUALTRANSFORMATIONNETWORKS FORONLINECONTINUALLEARNING

This project contains the implementation of the following ICLR 2021 paper:

Title: Contextual Transformation Networks for Online Continual Learning (ICLR 2021). [openreview], [pdf].

Authors: Quang Pham, Chenghao Liu, Doyen Sahoo, and Steven Hoi

CTN proposes a novel network design with a controller that can efficiently extract task-specific features from a base network. Both the base network and the controller have access to their own memory units and are joinly trained via a bilevel optimization strategy.

CTN

Requirements

  • Pytorch 1.5.0
  • CUDA 10.2

All experiments in this work was run on a single K80 GPU with 12Gb memory.

Benchmarks

1. Prepare data

Follow the instructions in the data/ folders to prepare the benchmarks.

2. Run experiments

To replicate our results on the Permuted MNIST, Split CIFAR100, CORE50, and Split miniImagenet, run

chmod 777 scripts/run.sh
./scripts/run.sh

The results will be put in the resuts/ folders.

3. Parameter Setting

The provided script scripts/run.sh includes the best hyper-parameter cross-validated from the cross-validation tasks. The following is the list of parameters you can experiment with

Parameter Description Values
data_path path where the data sets are saved e.g. data/
data_file name of the data file e.g. mnist_permutations.pt
use randomly use a subset of data. When use < 1, use% of the original data, when use > 1, select use samples from the data e.g. 0.5 (select 50% of data), 1000 (select 1000 data samples)
n_memories number of data stored per task e.g. 50
memory_strength value of the regularizer's coefficient e.g. 100
temperature temperature of the softmax in knowledge distillation e.g. 5
lr (inner) learning rate e.g. 0.1
beta (outer) learning rate e.g. 0.3
inner_steps number of SGD udpates per samples e.g. 2
n_meta number of outer updates per samples e.g. 2
n_val percentage of the total memory used for the semantic memory (in CTN) e.g. 0.2
replay_batch_size number of data in the memory used per experience replay step e.g. 64

Cite

If you found CTN useful for your research, please consider citing.

@inproceedings{pham2020contextual,
  title={Contextual transformation networks for online continual learning},
  author={Pham, Quang and Liu, Chenghao and Sahoo, Doyen and Steven, HOI},
  booktitle={International Conference on Learning Representations},
  year={2021}
}

Acknowledgement

This project structure is based on the GEM repository with additional methods, metrics and implementation improvements. For the CORe50 benchmark, we modify the data loader from MIR.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages