Skip to content
This's an implementation of deepmind Visual Interaction Networks paper using pytorch
Python
Branch: master
Clone or download
Latest commit 27ae3ff Feb 2, 2018
Permalink
Type Name Latest commit message Commit time
Failed to load latest commit information.
.idea readme Feb 2, 2018
figures readme Feb 2, 2018
.gitignore finish Feb 1, 2018
LICENSE Initial commit Oct 13, 2017
README.md
config.py finish Oct 21, 2017
logger.py finish Feb 1, 2018
model.py finish Feb 1, 2018
train.py finish Feb 1, 2018
utils.py finish Feb 1, 2018
vin.py finish Feb 1, 2018
vin_dataset.py

README.md

Visual-Interaction-Networks

An implementation of Deepmind visual interaction networks in Pytorch.

Introduction

For the purpose of understanding the challenge of relational reasoning. they publised VIN that involves predicting the future in a physical scene. From just a glance, humans can infer not only what objects are where, but also what will happen to them over the upcoming seconds, minutes and even longer in some cases. For example, if you kick a football against a wall, your brain predicts what will happen when the ball hits the wall and how their movements will be affected afterwards (the ball will ricochet at a speed proportional to the kick and - in most cases - the wall will remain where it is).

Architecture

Data

I used I@jaesik817 physics engine to generate the data.

Just run the physics_engine.py

Usage

Main Dependencies

Python 3.5
pytorch 0.3
numpy 1.13.1

RUN

  • Edit configration file to meet your need.
  • Run vin.py

References

You can’t perform that action at this time.