Skip to content
Code for the paper "High Speed and High Dynamic Range Video with an Event Camera" (arXiv, 2019).
Branch: master
Clone or download
Fetching latest commit…
Cannot retrieve the latest commit at this time.
Type Name Latest commit message Commit time
Failed to load latest commit information.
base Initial commit Jun 14, 2019
model Initial commit Jun 14, 2019
options Initial commit Jun 14, 2019
scripts Initial commit Jun 14, 2019
utils Initial commit Jun 14, 2019
LICENSE License Jun 18, 2019 Add link to video + URL of arXiv paper. Jul 2, 2019 Initial commit Jun 14, 2019 Initial commit Jun 14, 2019

High Speed and High Dynamic Range Video with an Event Camera

High Speed and High Dynamic Range Video with an Event Camera

This is the code for the paper High Speed and High Dynamic Range Video with an Event Camera by Henri Rebecq, Rene Ranftl, Vladlen Koltun and Davide Scaramuzza:

You can find a pdf of the paper here. If you use any of this code, please cite the following publications:

  author        = {Henri Rebecq and Ren{\'{e}} Ranftl and Vladlen Koltun and Davide Scaramuzza},
  title         = {High Speed and High Dynamic Range Video with an Event Camera},
  journal       = {ar{X}iv e-prints},
  url           = {}
  year          = 2019
  author        = {Henri Rebecq and Ren{\'{e}} Ranftl and Vladlen Koltun and Davide Scaramuzza},
  title         = {Events-to-Video: Bringing Modern Computer Vision to Event Cameras},
  journal       = {{IEEE} Conf. Comput. Vis. Pattern Recog. (CVPR)},
  year          = 2019



Install with Anaconda

The installation requires Anaconda3. You can create a new Anaconda environment with the required dependencies as follows (make sure to adapt the CUDA toolkit version according to your setup):

conda create -n E2VID
conda activate E2VID
conda install pytorch torchvision cudatoolkit=10.0 -c pytorch
conda install pandas
conda install -c conda-forge opencv


  • Download the pretrained model:
mkdir -p pretrained
wget "" -O pretrained/E2VID.pth.tar
  • Download an example file with event data:
mkdir -p data
wget "" -O data/

Before running the reconstruction, make sure the conda environment is sourced:

conda activate E2VID
  • Run reconstruction:
python \
  -c pretrained/E2VID.pth.tar \
  -i data/ \
  --auto_hdr \
  --display \


Below is a description of the most important parameters:

Main parameters

  • --window_size / -N (default: None) Number of events per tensor. This is the parameter that has the most influence of the image reconstruction quality. If set to None, this number will be automatically computed based on the sensor size, as N = width * height * num_events_per_pixel (see description of that parameter below).
  • --auto_hdr (default: True) If True, the output image values (in the range [0,1]) will be rescaled using (robust) min/max normalization. See also the parameters --auto_hdr_min_percentile, --auto_hdr_max_percentile, and --auto_hdr_moving_average_size.
  • --safety_margin (default: 5): zero-pad the input event tensors with zeros to avoid boundary effects (typical range: [0,5]). A small value reduces computation time but may introduce boundary effects.
  • --hot_pixels_file (default: None): Path to a file specifying the locations of hot pixels (such a file can be obtained with this tool for example). These pixels will be ignored (i.e. zeroed out in the event tensors).
  • --color (default: False): if True, will perform color reconstruction as described in the paper. Only use this with a color event camera such as the Color DAVIS346. When --color is set, the --auto_hdr flag will be automatically set. Do not forget to set width=346 and height=260 for the ColorDAVIS346.

Output parameters

  • --output_folder: path of the output folder. If not set, the image reconstructions will not be saved to disk.
  • --dataset_name: name of the output folder directory (default: 'reconstruction').

Display parameters

  • --display (default: False): display the video reconstruction in real-time in an OpenCV window.
  • --show_events (default: False): show the input events side-by-side with the reconstruction. If --output_folder is set, the previews will also be saved to disk in /path/to/output/folder/events.

Additional parameters

  • --num_events_per_pixel (default: 0.35): Parameter used to automatically estimate the window size based on the sensor size. The value of 0.35 was chosen to correspond to ~ 15,000 events on a 240x180 sensor such as the DAVIS240C.
  • --no-normalize (default: False): Disable event tensor normalization: this will improve speed a bit, but might degrade the image quality a bit.
  • --no-recurrent (default: False): Disable the recurrent connection (i.e. do not maintain a state). For experimenting only, the results will be flickering a lot.

Example datasets

We provide a list of example (publicly available) event datasets to get started with E2VID.

Working with ROS

Because PyTorch recommends Python 3 and ROS is only compatible with Python2, it is not straightforward to have the PyTorch reconstruction code and ROS code running in the same environment. To make things easy, the reconstruction code we provide has no dependency on ROS, and simply read events from a text file or ZIP file. We provide convenience functions to convert ROS bags (a popular format for event datasets) into event text files. In addition, we also provide scripts to convert a folder containing image reconstructions back to a rosbag (or to append image reconstructions to an existing rosbag).

Note: it is not necessary to have a sourced conda environment to run the following scripts. However, ROS needs to be installed and sourced.

rosbag -> events.txt

To extract the events from a rosbag to a zip file containing the event data:

python scripts/ /path/to/rosbag.bag \
  --output_folder=/path/to/output/folder \

image reconstruction folder -> rosbag

python scripts/ \
  --rosbag_folder /path/to/rosbag/folder \
  --datasets dynamic_6dof \
  --image_folder /path/to/image/folder \
  --output_folder /path/to/output_folder \
  --image_topic /dvs/image_reconstructed

Append image_reconstruction_folder to an existing rosbag

cd scripts
python \
  --rosbag_folder /path/to/rosbag/folder \
  --datasets dynamic_6dof \
  --image_folder /path/to/image/folder \
  --output_folder /path/to/output_folder\
  --image_topic /dvs/image_reconstructed

Generating a video reconstruction (with a fixed framerate)

It can be convenient to convert an image folder to a video with a fixed framerate (for example for use in a video editing tool). You can proceed as follows:

export FRAMERATE=30
python -i /path/to/input_folder -o /tmp/resampled -r $FRAMERATE
ffmpeg -framerate $FRAMERATE -i /tmp/resampled/frame_%010d.png video_"$FRAMERATE"Hz.mp4


This code borrows from the following open source projects, whom we would like to thank:

You can’t perform that action at this time.