Skip to content

uzh-rpg/rpg_event_representation_learning

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

27 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Event Representation Learning

Event Representation Learning

This repository contains learning code that implements event representation learning as described in Gehrig et al. ICCV'19. The paper can be found here

If you use this code in an academic context, please cite the following work:

Daniel Gehrig, Antonio Loquercio, Konstantinos G. Derpanis, Davide Scaramuzza, "End-to-End Learning of Representations for Asynchronous Event-Based Data", The International Conference on Computer Vision (ICCV), 2019

@InProceedings{Gehrig_2019_ICCV,
  author = {Daniel Gehrig and Antonio Loquercio and Konstantinos G. Derpanis and Davide Scaramuzza},
  title = {End-to-End Learning of Representations for Asynchronous Event-Based Data},
  booktitle = {Int. Conf. Comput. Vis. (ICCV)},
  month = {October},
  year = {2019}
}

Requirements

  • Python 3.7
  • virtualenv
  • cuda 10

Dependencies

Create a virtual environment with python3.7 and activate it

virtualenv venv -p /usr/local/bin/python3.7
source venv/bin/activate

Install all dependencies by calling

pip install -r requirements.txt

Training

Before training, download the N-Caltech101 dataset and unzip it

wget http://rpg.ifi.uzh.ch/datasets/gehrig_et_al_iccv19/N-Caltech101.zip 
unzip N-Caltech101.zip

Then start training by calling

python main.py --validation_dataset N-Caltech101/validation/ --training_dataset N-Caltech101/training/ --log_dir log/temp --device cuda:0

Here, validation_dataset and training_dataset should point to the folders where the training and validation set are stored. log_dir controls logging and device controls on which device you want to train. Checkpoints and models with lowest validation loss will be saved in the root folder of log_dir.

The N-Cars dataset can be downloaded here.

Additional parameters

  • --num_worker how many threads to use to load data
  • --pin_memory wether to pin memory or not
  • --num_epochs number of epochs to train
  • --save_every_n_epochs save a checkpoint every n epochs.
  • --batch_size batch size for training

Visualization

Training can be visualized by calling tensorboard

tensorboard --logdir log/temp

Training and validation losses as well as classification accuracies are plotted. In addition, the learnt representations are visualized. The training and validation curves should look something like this:
alt_text

Testing

Once trained, the models can be tested by calling the following script:

python testing.py --test N-Caltech101/testing/ --device cuda:0

Which will print the test score after iteration through the whole dataset.

About

Repo for learning event representations

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages