Joint learning of object and action detectors
Branch: master
Clone or download
Fetching latest commit…
Cannot retrieve the latest commit at this time.
Permalink
Type Name Latest commit message Commit time
Failed to load latest commit information.
detections/-4KMQsUaSAQ
src
utils
LICENSE
README.md
vic_A2D_eval.m
vic_A2D_eval_zeroshot.m
vic_eval_segmentation.m

README.md

Joint learning of object and action detectors

By Vicky Kalogeiton, Philippe Weinzaepfel, Vittorio Ferrari, Cordelia Schmid

Introduction

In our ICCV17 paper we jointly detect objects performing various avtions in videos. Here, we provide the evaluation code used in our experiments.

Citing Joint learning of object and action detectors

If you find our evaluation code useful in your research, please cite:

@inproceedings{kalogeiton17biccv,
  TITLE = {Joint learning of object and action detectors},
  AUTHOR = {Kalogeiton, Vicky and Weinzaepfel, Philippe and Ferrari, Vittorio and Schmid, Cordelia},
  YEAR = {2017},
  BOOKTITLE = {ICCV},
}

Evalutation of all architectures

You can download our detection results (multitask, hierarchical and cartesian):

curl http://pascal.inrialpes.fr/data2/joint-objects-actions/JointLearningDetections.tar.gz | tar xz 

To run the mAP evaluation function for the multitask, hierarchical and cartesian cases (Table 4 in our paper), run:

vic_A2D_eval(learning_case) # learning_case: 1, 2 or 3 for multitask, hierarchical and cartesian, respectively

Zero shot learning

To run the mAP evaluation function for the zero shot learning (Table 5 in our paper), run:

vic_A2D_eval_zeroshot 

Object-action semantic segmentation

You can download our semantic segmentation images:

curl http://pascal.inrialpes.fr/data2/joint-objects-actions/JointLearningSegmentationResults.tar.gz | tar xz 

To run the semantic segmentation evaluation function (Table 6 in our paper), run:

vic_eval_segmentation