Skip to content

PlusLabNLP/follower_aware_speaker

Repository files navigation

FOAM: A Follower-aware Speaker Model for Vision-and-Language Navigation

Requirements

pip install -r requirements.txt

Install Matterport3D simulators:

git submodule update --init --recursive 
sudo apt-get install libjsoncpp-dev libepoxy-dev libglm-dev libosmesa6 libosmesa6-dev libglew-dev libopencv-dev
sudo apt-get install libopencv-dev
mkdir build && cd build
cmake -DEGL_RENDERING=ON ..
# Replace the above line with following if it doesn't work:
#   cmake -DOSMESA_RENDERING=ON ..
make -j8

Data

bash ./tasks/R2R/data/download.sh

Image Features

Training

bash run/agent_clip_vit16.bash 0 # 0 is the id of GPU
bash run/speaker_clip_vit16.bash 0
bash run/foam_envdrop_clip_vit16.bash 0

Pretrained Checkpoint

Citation

@inproceedings{dou2022foam,
  title={FOAM: A Follower-aware Speaker Model for Vision-and-Language Navigation},
  author={Dou, Zi-Yi and Peng, Nanyun},
  booktitle={Conference of the North American Chapter of the Association for Computational Linguistics (NAACL)},
  year={2022},
}

Acknowledgement

The code is based on EnvDrop and CLIP-ViL-VLN. We thank Hao Tan for the help with preprocessing.

About

No description, website, or topics provided.

Resources

License

MIT, MIT licenses found

Licenses found

MIT
LICENSE
MIT
LICENSE_R2REnvDrop

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published