Models of APS
Since the code is too messy (several settings for different architectures), we currently provide the trained models for further experiments. The architectures of R2R and Speaker-Follower are the same as original, so you can apply the code and load the model directly. We have verified that the performance is the same as our report in the paper.
Please apply code in R2R and Speaker-Follower to load the model
@inproceedings{fu2020aps,
author = {Tsu-Jui Fu and Xin Eric Wang and Matthew Peterson and Scott Grafton and Miguel Eckstein and William Yang Wang},
title = {{Counterfactual Vision-and-Language Navigation via Adversarial Path Sampler}},
booktitle = {European Conference on Computer Vision (ECCV)},
year = {2020}
}