Attend Refine Repeat: Active Box Proposal Generation via In-Out Localization
Matlab Cuda C++
Switch branches/tags
Nothing to show
Clone or download
Fetching latest commit…
Cannot retrieve the latest commit at this time.
Permalink
Failed to load latest commit information.
code
examples
external
sup_data
.gitignore
AttractioNet_build.m
DATASETS.md
README.md
generate_multi_nms_setting_attend.m
startup.m

README.md

Attend Refine Repeat: Active Box Proposal Generation via In-Out Localization

Introduction

The AttractioNet code implements the following arxiv paper:
Title: "Attend Refine Repeat: Active Box Proposal Generation via In-Out Localization"
Authors: Spyros Gidaris, Nikos Komodakis
Institution: Universite Paris Est, Ecole des Ponts ParisTech
Code: https://github.com/gidariss/AttractioNet
Arxiv Link: https://arxiv.org/abs/1606.04446

Abstract:
The problem of computing category agnostic bounding box proposals is utilized as a core component in many computer vision tasks and thus has lately attracted a lot of attention. In this work we propose a new approach to tackle this problem that is based on an active strategy for generating box proposals that starts from a set of seed boxes, which are uniformly distributed on the image, and then progressively moves its attention on the promising image areas where it is more likely to discover well localized bounding box proposals. We call our approach AttractioNet and a core component of it is a CNN-based category agnostic object location refinement module that is capable of yielding accurate and robust bounding box predictions regardless of the object category. We extensively evaluate our AttractioNet approach on several image datasets (i.e. COCO, PASCAL, ImageNet detection and NYU-Depth V2 datasets) reporting on all of them state-of-the-art results that surpass the previous work in the field by a significant margin and also providing strong empirical evidence that our approach is capable to generalize to unseen categories. Furthermore, we evaluate our AttractioNet proposals in the context of the object detection task using a VGG16-Net based detector and the achieved detection performance on COCO manages to significantly surpass all other VGG16-Net based detectors while even being competitive with a heavily tuned ResNet-101 based detector.

Sample experimental results

Average recall results of AttractioNet box proposal generation approach:

Test set AR@10 AR@100 AR@1000 AR@Small AR@Medium AR@Large time/img
COCO 2014 val. 0.326 0.532 0.660 0.317 0.621 0.771 1.63secs
VOC 2007 test 0.547 0.740 0.848 0.575 0.666 0.788 1.63secs
ImageNet detection task val. 0.412 0.618 0.748 - - - 1.63secs
NYU-Depth V2 0.159 0.389 0.579 0.205 0.419 0.498 1.63secs

Average precision results on COCO test-dev2015 set of a VGG16-Net based object detector with AttractioNet proposals:

Method AP@0.50 AP@0.75 AP@0.50:0.95 AP@Small AP@Medium AP@Large
AttractioNet based detector 0.537 0.363 0.341 0.175 0.365 0.469

The AttractioNet model used for the generation of the above results was trained on the COCO 2014 train set.

Citing AttractioNet

If you find AttractioNet useful in your research, please consider citing our BMVC paper:

@inproceedings{gidaris2016attendbmvc,  
  title={Attend Refine Repeat: Active Box Proposal Generation via In-Out Localization},  
  author={Gidaris, Spyros and Komodakis, Nikos},   
  booktitle={BMVC},  
  year={2016}  
}

or the significantly extended arxiv version of the former:

@article{gidaris2016attend,
  title={Attend Refine Repeat: Active Box Proposal Generation via In-Out Localization},
  author={Gidaris, Spyros and Komodakis, Nikos},
  journal={arXiv preprint arXiv:1606.04446},
  year={2016}
}

License

This code is released under the MIT License (refer to the LICENSE file for details).

Contents

  1. Requirements
  2. Installation
  3. Downloading pre-computed bounding box proposals
  4. Preparing and using the COCO and PASCAL datasets

Requirements

Hardware. In order to use AttractioNet for extracting bounding box proposals from an image you will require a GPU with at least 4 Gbytes of memory

Software.

  1. A modified, in order to support AttractioNet, version of Caffe [link] (you must use the AttractioNet branch).
  2. The cuDNN(-v5) library during Caffe installation.
  3. Matlab (tested with R2014b).

Installation

  1. Download and install this modified version of Caffe, initially developed to supprot LocNet and now extended to support AttractioNet (use the AttractioNet branch for the later). Clone Caffe in your local machine:
     
     # $caffe_AttractioNet: directory where Caffe will be cloned 
     git clone https://github.com/gidariss/caffe_LocNet.git $caffe_AttractioNet  
     # switch on the AttractioNet branch  
     git checkout AttractioNet                              

Then follow the Caffe and Matcaffe installation instructions here. Note that you have to install Caffe with the cuDNN(-v5) library.
2. Clone the AttractioNet code in your local machine:

 # $AttractioNet: directory where AttractioNet will be cloned    
 git clone https://github.com/gidariss/AttractioNet $AttractioNet  
 ```   
From now on, the directory where *AttractioNet* is cloned will be called `$AttractioNet`.  
3. Create a symbolic link of [Caffe](https://github.com/gidariss/caffe_LocNet/tree/AttractioNet) installatation directory at `$AttractioNet/external/caffe_AttractioNet`:  
```Shell

 # $AttractioNet: directory where AttractioNet is cloned   
 # $caffe_AttractioNet: directory where caffe is cloned and installed    
 ln -sf $caffe_AttractioNet $AttractioNet/external/caffe_AttractioNet   
 ```      
4.  Open Matlab from the `$AttractioNet/` directory and run the `AttractioNet_build.m` script:  
```Shell

 $ cd $AttractioNet  
 $ matlab   
 
 # matlab command line enviroment
 >> AttractioNet_build   
 ``` 
 Do not worry about the warning messages. They also appear on my machine. 
5. Download the ***AttractioNet pre-trained model*** from [Google drive](https://drive.google.com/file/d/0BwxkAdGoNzNTV2N3RjN5dXNpWVE/view?usp=sharing) or [Mega](https://mega.nz/#!r90F3KJS!TM-slQgOF4OFWnVJrgZVr0S8aSn-f_s_IVFigbnMWrI). Note that the provided model is actually the fast version of AttractioNet model that is described on section 3.1.3 of the technical report. *It is trained on the COCO 2014 train set*. After downloading, unzip and untar the .tar.gz archive file with the AttractioNet model files on the directory `$AttractioNet/models-exps/AttractioNet_Model` by running:   
```Shell

tar xvfz AttractioNet_Model.tar.gz -C $AttractioNet/models-exps/    

Demo

After having complete the installation, you will be able to use AttractioNet for extracting bounding box proposals from any image. For a demo see the demo_AttractioNet.m script. Note that you will require a GPU with at least 4 Gbytes of memory in order to run the demo.

Downloading pre-computed bounding box proposals

We provide pre-computed bounding box proposals using the same AttractioNet model we also provide here (trained on the COCO 2014 train set) for the following datasets:

PASCAL VOC:

MS-COCO:

Each package contains the AttractioNet box proposals of the corresponding data set, stored using a separate box proposal file per image. Specifically, the box proposals of each image are stored in Matlab files (.mat files) using the same filenames as those that the images have. Each Matlab file contains a single data field, called boxes, that is a K x 5 single precision array with the box proposals of the corresponding image (where K is the number of box proposals). Each row of the boxes array contains a single bounding box proposal represented by the 5 values [x0,y0,x1,y1,obj], where (x0,y0) and (x1,y1) are its top-left and bottom-right coordinates correspondingly (specifically, 1-based pixel coordinates) and obj is its objectness score. Note that the box proposals are already sorted w.r.t. their objectness score in decreasing order.

Preparing and using the COCO and PASCAL datasets

In case you need to set up and use the COCO and/or PASCAL datasets (e.g. generating or evaluating AttractioNet proposals) then follow the instructions on the DATASET.md file.