Skip to content

HaozhiQi/RPIN

master
Switch branches/tags
Code

Latest commit

 

Git stats

Files

Permalink
Failed to load latest commit information.
Type
Name
Latest commit message
Commit time
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Region Proposal Interaction Network

PWC PWC

This repository is an official PyTorch implementation of the ICLR paper:

Learning Long-term Visual Dynamics with Region Proposal Interaction Networks
Haozhi Qi, Xiaolong Wang, Deepak Pathak, Yi Ma, Jitendra Malik
International Conference on Learning Representations (ICLR), 2021
[Website], [Paper]

Ground-Truth Our Prediction Ground-Truth Our Prediction
image image image image

Introduction

In this project, we aim to leverage the ideas from success stories in visual recognition tasks to build object representations that can capture inter-object and object-environment interactions over a long range. To this end, we propose Region Proposal Interaction Networks (RPIN), which reason about each object's trajectory in a latent region-proposal feature space. Our approach outperforms prior methods by a significant margin both in terms of prediction quality and their ability to plan for downstream tasks, and also generalize well to novel environments.

Method

image

Our model takes N video frames as inputs and predicts the object locations for the future T timesteps, as illustrated above. We first extract the image feature representation using a ConvNet for each frame, and then apply RoI pooling to obtain object-centric visual features. These object feature representations are forwarded to the interaction modules to perform interaction reasoning and predict future object locations. The whole pipeline is trained end-to-end by minimizing the loss between predicted and the ground-truth object locations. Since the parameters of each interaction module is shared so we can apply this process recurrently over time to an arbitrary T during testing.

Using this codebase

Installation

This codebase is developed and tested with python 3.6, PyTorch 1.4, and cuda 10.1. But any version newer than that should work.

Here we gave an example of installing RPIN using the conda virtual environment:

conda create -y -n rpin python=3.6
conda activate rpin
# install pytorch according to https://pytorch.org/
conda install -y pytorch==1.4 torchvision cudatoolkit=10.1 -c pytorch
pip install yacs hickle tqdm matplotlib
# OpenCV changes their way of reading image and has different results
# We don't use later version for consistency
pip install opencv-python==3.4.2.17 
# to evaluate phyre planning, you also need to also do
pip install phyre
# (optional) install gdown for downloading
pip install gdown

Then

git clone https://github.com/HaozhiQi/RPIN
cd RPIN

Data Preparation & Training & Evaluation

This codebase supports all four datasets we used in our paper. We provide the instructions for each dataset separately. Please see PHYRE, SS, RealB, and SimB for detailed instructions.

For results and models we provided in this codebase, see the Model Zoo.

Citing RPIN

If you find RPIN or this codebase helpful in your research, please consider citing:

@InProceedings{qi2021learning,
  author={Qi, Haozhi and Wang, Xiaolong and Pathak, Deepak and Ma, Yi and Malik, Jitendra},
  title={Learning Long-term Visual Dynamics with Region Proposal Interaction Networks},
  booktitle={ICLR},
  year={2021}
}

Log of main updates:

2022-04-28 (Fix Download Script)

Use gdown from pip to download dataset adn models.

2021-03-31 (ICLR camera ready)

Update the PHYRE reasoning results. Importantly, we remove the 2stage method for PHYRE reasoning, which significantly simplify the pipeline. See our camera ready version for detail.

2021-01-27 (ICLR version)

Update the PHYRE reasoning results. Not it contains the full evaluation result for the benchmark.

2020-08-05 (Init)

Initial Release of this repo.

About

Learning Long-term Visual Dynamics with Region Proposal Interaction Networks (ICLR 2021)

Topics

Resources

Stars

Watchers

Forks