Skip to content
Code for SplitNet paper
Python Shell
Branch: master
Clone or download
Fetching latest commit…
Cannot retrieve the latest commit at this time.
Permalink
Type Name Latest commit message Commit time
Failed to load latest commit information.
assets
configs
dataset_dump removing references Aug 16, 2019
networks
reinforcement_learning small bug fix and dg_util update Aug 2, 2019
supervised_learning removing references Aug 16, 2019
utils
.gitignore
CODE_OF_CONDUCT.md initial commit of splitnet May 15, 2019
CONTRIBUTING.md
LICENSE initial commit of splitnet May 15, 2019
README.md
arguments.py
base_habitat_rl_runner.py
download_dataset.sh removing references Aug 16, 2019
download_weights.sh removing references Aug 16, 2019
environment.yml
eval_splitnet.py
eval_splitnet.sh
flee_test.py
generate_videos.sh mit -> cc May 18, 2019
requirements.txt
train_splitnet.py small bug fix and dg_util update Aug 2, 2019

README.md

SplitNet: Sim2Sim and Task2Task Transfer for Embodied Visual Navigation

Requirements

  1. Python 3.6+ (to be compatible with Habitat, though SplitNet is Python2.7 compatible).
  2. Recommended: anaconda/miniconda
  3. Slightly less recommended: virtualenv pip3 install virtualenv or sudo apt-get install python-virtualenv
  4. A good attitude.

First Time Setup

Conda

  1. Create your virtual environment: conda create -n habitat-env python=3.6
  2. Run this code (replace the env name if necessary)
git clone git@github.com:facebookresearch/splitnet.git
cd splitnet
conda deactivate
conda env update -n habitat-env -f environment.yml
conda activate habitat-env
  1. Clone and install habitat-sim from the github repo. You must be in the same virtualenv as before. Alternatively, run pip install git+https://github.com/facebookresearch/habitat-sim.git@v0.1.2 to install it locally. Note that the pip method does not support headless compilation. SplitNet's codebase is known to be compatible with https://github.com/facebookresearch/habitat-sim/releases/tag/v0.1.2
  2. Clone and install habitat-api from the github repo. You must be in the same virtualenv as before. Alternatively, run pip install git+https://github.com/facebookresearch/habitat-api.git@7015813aefae99233864c4ffcf7e52e9097392a4 to install it locally. SplitNet's codebase is known to be compatible with https://github.com/facebookresearch/habitat-api/commit/7015813aefae99233864c4ffcf7e52e9097392a4

Virtualenv

  1. Create your virtual environment: virtualenv --python=python3.6 habitat-env && source habitat-env/bin/activate
  2. Run this code (replace the env name if necessary)
git clone git@github.com:facebookresearch/splitnet.git
cd splitnet
source habitat-env/bin/activate
pip install -r requirements.txt
  1. Clone and install habitat-sim from the github repo. You must be in the same virtualenv as before. Alternatively, run pip install git+https://github.com/facebookresearch/habitat-sim.git@v0.1.2 to install it locally. Note that the pip method does not support headless compilation. SplitNet's codebase is known to be compatible with https://github.com/facebookresearch/habitat-sim/releases/tag/v0.1.2
  2. Clone and install habitat-api from the github repo. You must be in the same virtualenv as before. Alternatively, run pip install git+https://github.com/facebookresearch/habitat-api.git@7015813aefae99233864c4ffcf7e52e9097392a4 to install it locally. SplitNet's codebase is known to be compatible with https://github.com/facebookresearch/habitat-api/commit/7015813aefae99233864c4ffcf7e52e9097392a4

Data

We use the data sources linked from the public habitat-api repository. You will need to individually download MP3D, and Gibson from their sources. habitat-sim and habitat-api share the links to the files. We additionally use the Point-Nav datasets from habitat-api, but we also provide a script for generating new datasets.

To use the same data from the dataset:

  1. Create a symlink to where you downloaded the directory containing the scene_datasets asset files for each of the datasets. Call this folder data.
ln -s /path/to/habitat/data data
  1. Verify that this is set up in the expected way by running ls data/scene_datasets. You should see: gibson mp3d based on which you have downloaded.
  2. Download and extract the premade training datasets by running sh download_dataset.sh.
  3. Copy/Move the downloaded datasets into the data folder.
mv downloaded_data/* data

Pretrained Weights

Edit download_weights.sh WEIGHT_DIRECTORY to be a non-existing directory where the files should be downloaded. Download the pretrained weights for all experiments by running sh download_weights.sh. The download contains a README on how the files are organized.

Training

What's the matter, you don't trust our results? Good. You shouldn't. We provide example training scripts to recreate our experiments so you don't have to trust us.

First you should look at all the options of arguments in the arguments.py file. Many of them are useful for tweaking parameters or adjusting how the learning is performed. All baselines and experiments from the paper can be recreated using various arguments, but shell scripts with these exact parameters are not provided for brevity. They can be added upon request.

Downloading the training/val datasets

See Data

Retraining the Point-Nav model

  1. You may wish to modify some of the variables in the shell script file before running such as the log location, dataset, gpus to use etc. It should be fairly obvious how these work.
  2. Pretrain the visual encoder/decoder: sh supervised_learning/train_supervised_visual_only.sh
  3. Train the visual encoder and the visual, motion, and policy decoders on pointnav using Behavioral Cloning: sh supervised_learning/train_splitnet_supervised.sh
  4. Finetune the policy, freezing the visual encoder and the visual and motion decoders on pointnav using Behavioral Cloning: sh supervised_learning/finetune_splitnet_policy_supervised.sh
  5. Finetune the policy, freezing the visual encoder and the visual and motion decoders on pointnav using PPO: sh reinforcement_learning/train_splitnet_rl.sh
  6. reinforcement_learning/train_splitnet_rl_debug.sh is provided as a convenience file to show how to test code without generating lots of extra log files or updating the weights. This is useful for developing your own code. Similar arguments also work for supervised learning.
  7. You can retrain the ImageNet encoder with sh supervised_learning/train_imagenet.sh if you really want to. But then you also have to have all the ImageNet files and that's just annoying.

Retraining Sim2Sim

  1. Copy your pretrained weights from the initial simulator to a new folder so you don't overwrite them.
  2. Modify the parameters in reinforcement_learning/train_splitnet_rl_sim2sim.sh if need be and run it.

Retraining Task2Task

  1. Copy your pretrained weights from the initial simulator to a new folder so you don't overwrite them.
  2. Modify the parameters in reinforcement_learning/train_splitnet_rl_task2task.sh if need be and run it.

Evaluation

Evaluation can be performed during training using the --eval-interavl flag, but you may also wish to evaluate an individual file on its own. eval_splitnet.sh makes this possible.

  1. Edit the DATASET, TASK, and LOG_LOCATION in eval_splitnet.sh and any other variables you wish.
  2. By default, the code restores the most recently modified weights file in the checkpoints folder. If this is not the one you want to evaluate, you will have to edit base_habitat_rl_runner.py restore function to point to the proper file.
  3. Run sh eval_splitnet.sh

Generating Debug/Output Videos

This can be useful for visually examining the learned policies and debugging mistakes that the agent may perform. The instructions are much the same as the Evaluation section, so read that first.

  1. Modify parameters in generate_videos.sh and run it.
  2. Play back the generated videos.
  3. ???
  4. Profit.

License

SplitNet is licensed under Creative Commons-Non Commercial 4.0. See the LICENSE file for details.

Citation

@inproceedings{gordon2019splitnet, 
    title={SplitNet: Sim2Sim and Task2Task Transfer for Embodied Visual Navigation}, 
    author={Gordon, Daniel and Kadian, Abhishek and Parikh, Devi and Hoffman, Judy and Batra, Dhruv},
    booktitle = {The IEEE International Conference on Computer Vision (ICCV)},
    month = {October},
    year={2019},
}
You can’t perform that action at this time.