Skip to content

The repository of ECCV 2020 paper `Active Visual Information Gathering for Vision-Language Navigation`

License

Notifications You must be signed in to change notification settings

HanqingWangAI/Active_VLN

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

20 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Active VLN

This repository is the implementation of our ECCV 2020 paper:

Active Visual Information Gathering for Vision-Language Navigation

Hanqing Wang, Wenguan Wang, Tianmin Shu, Wei Liang, Jianbing Shen.


Introduction

This work draws inspiration from human navigation behavior and endows an agent with an active information gathering ability for a more intelligent vision-language navigation policy.

To achieve this, we develop an active exploration module, which learns to 1) decide when the exploration is necessary, 2) identify which part of the surroundings is worth exploring, and 3) gather useful knowledge from the environment to support more robust navigation.

Please refer to our paper for the detailed formulations.

Results

Here are some results on R2R dataset reported in our paper.

Single Run Setting

Set SR↑ NE↓ TL↓ OR↑ SPL↑
Validation Seen 0.70 3.20 19.7 0.80 0.52
Validation Unseen 0.58 4.36 20.6 0.70 0.40
Test Unseen 0.60 4.33 21.6 0.71 0.41

Pre-explore Setting

Set SR↑ NE↓ TL↓ OR↑ SPL↑
Test Unseen 0.70 3.30 9.85 0.77 0.68

Beam-Search Setting

Set SR↑ TL↓ SPL↑
Test Unseen 0.71 176.2 0.05

Please refer to our paper for the comparsions with previous arts.

Environment Installation

  1. Install Jupyter Install jupyter using the following scripts. pip install jupyter

  2. Install R2R environment via Jupyter Our code is built basing on R2R-EnvDrop, please install the R2R environment for the python interpreter used in Jupyter following the installation instructions.

Quick Start

Inference:

  1. Download the checkpoint of the agent to directory snap/agent/state_dict/best_val_unseen. The checkpoint is available on Google Drive.
  2. Start a Jupyter service and run the jupyter notebook evaluation.ipynb.

TODO

  • Release the checkpoint.
  • Add training code.

Citation

Please cite this paper in your publications if it helps your research:

@inproceedings{wang2020active,
    title={Active Visual Information Gathering for Vision-Language Navigation},
    author={Wang, Hanqing and Wang, Wenguan and Shu, Tianmin and Liang, Wei and Shen, Jianbing},
    booktitle=ECCV,
    year={2020}
}

License

Active VLN is freely available for non-commercial use, and may be redistributed under these conditions. Please see the license for further details. For commercial license, please contact the authors.

Contact Information

  • hanqingwang[at]bit[dot]edu[dot]cn, Hanqing Wang
  • wenguanwang[dot]ai[at]gmail[dot]com, Wenguan Wang

About

The repository of ECCV 2020 paper `Active Visual Information Gathering for Vision-Language Navigation`

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published