Skip to content

Skylark0924/awesome-target-driven-navigation

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

15 Commits
 
 
 
 

Repository files navigation

#! https://zhuanlan.zhihu.com/p/338404758

Awesome Target-driven Navigation 资源汇总

Alias Target Sensor SLAM Framework Dataset SR SPL Note
t-SNE Image RGB No ResNet + A3C AI2THOR - -
SPTM Image RGB No DL + Graph game Doom 1 -
NTS panoramic image RGB/RGB Yes Graph Construction + Global Policy +
Local Policy
Gibson 0.55/0.63 0.38/0.43
PTVN panoramic image RGBD No autoencoder + policy model +
goal checking
MP3D 0.8125 0.6614
CMP Label RGB/D Yes Encoder-decoder+value iteration S3DIS + MP3D 0.638 0.228
Scene Priors Label RGB No GCN+LSTM+A3C AI2THOR 0.649 0.304
VR Label RGBD No CNN+LSTM Active Vision Dataset (AVD) 0.54 -
SAVN Label RGB No ResNet18+GloVe+
LSTM+MAML
AI2THOR 0.287 0.139
BRM Label (Room) RGBD No CNN+Graph+LSTM House3D 0.231 0.045
SF Label RGBD No Situational Fusion MP3D 0.44 -
ORG Label RGB No GCN+LSTM+IL+TPN AI2THOR 0.607 0.385
MJOLNIR Label RGB No GCN+LSTM+A3C AI2THOR 0.50 0.209
Attention 3D Image RGB No LSTM A3C+KG+Attention AI2THOR 0.419 0.072
YouTube Label RGB No Double DQN YouTube House Tours Dataset 0.79 0.53
SemExp Label RGBD No Mask RCNN+Voxel Representation+RL MP3D 0.360 0.144

$^\dagger$ For methods designed for single room navigation, I use the result about a subset of targets whose optimal trajectory length is greater than 5 ($L\ge 5$).

ImageGoal Navigation

  1. Target-driven Visual Navigation in Indoor Scenes using Deep Reinforcement Learning (t-SNE)

    Yuke Zhu, Roozbeh Mottaghi, Eric Kolve, Joseph J. Lim, Abhinav Gupta, Li Fei-Fei, Ali Farhadi
    ICRA, 2017. [Paper] [Website]

  2. Semi-Parametric Topological Memory for Navigation (SPTM)

    Nikolay Savinov*, Alexey Dosovitskiy*, Vladlen Koltun
    ICLR, 2018. [Paper] [Code] [Website]

  3. Neural Topological SLAM for Visual Navigation (NTS)

    Devendra Singh Chaplot, Ruslan Salakhutdinov, Abhinav Gupta, Saurabh Gupta
    CVPR, 2020. [Paper] [Website]

  4. Learning Your Way Without Map or Compass: Panoramic Target Driven Visual Navigation (PTVN)

    David Watkins-Valls, Jingxi Xu, Nicholas Waytowich, Peter Allen
    [Paper]

LabelGoal Navigation

  1. Cognitive Mapping and Planning for Visual Navigation (CMP)

    Saurabh Gupta, Varun Tolani, James Davidson, Sergey Levine, Rahul Sukthankar, Jitendra Malik
    CVPR, 2017. [Paper]

    Summary

    Two key ideas:

    • a unified joint architecture for mapping and planning, such that the mapping is driven by the needs of the task;
    • a spatial memory with the ability to plan given an incomplete set of observations about the world.

    CMP constructs a top-down belief map of the world and applies a differentiable neural net planner to produce the next action at each time step.

    Network architecture

    Architecture of the mapper

    Architecture of the hierarchical planner

  2. Visual Semantic Navigation using Scene Priors (Scene Priors)

    Wei Yang, Xiaolong Wang, Ali Farhadi, Abhinav Gupta, Roozbeh Mottaghi
    ICLR, 2019. [Paper]

    Summary

    They address navigation to novel objects or navigating in unseen scenes using scene priors, like human does.

    Scene Priors

    Architecture

    GCN for relation graph embedding

  3. Visual Representations for Semantic Target Driven Navigation (VR)

    Arsalan Mousavian, Alexander Toshev, Marek Fiser, Jana Kosecka, Ayzaan Wahid, James Davidson
    ICRA, 2019. [Paper] [Code]

    Summary

This paper focuses on finding a good visual representation.

  1. Learning to Learn How to Learn: Self-Adaptive Visual Navigation using Meta-Learning (SAVN)

    Mitchell Wortsman, Kiana Ehsani, Mohammad Rastegari, Ali Farhadi, Roozbeh Mottaghi
    CVPR, 2019. [Paper] [Code] [Website]

    Summary

    This paper uses Meta-reinforcement learning to construct an interaction loss for self-adaptive visual navigation.

  2. Bayesian Relational Memory for Semantic Visual Navigation (BRM)

    Yi Wu, Yuxin Wu, Aviv Tamar, Stuart Russell, Georgia Gkioxari, Yuandong Tian
    ICCV, 2019. [Paper] [Code]

    Summary

    Construct a probabilistic relation graph to learn the relationship or a topological memory of house layout.

  3. Situational Fusion of Visual Representation for Visual Navigation (SF)

    William B. Shen, Danfei Xu, Yuke Zhu, Leonidas J. Guibas, Li Fei-Fei, Silvio Savarese
    ICCV, 2019. [Paper]

    Summary

    This paper aims at fusing multiple visual representations, such as Semantic Segment, Depth Perception, Object Class, Room Layout and Scene Class.

    They develop an action-level representation fusion scheme, which predicts an action candidate from each representation and adaptively consolidate these action candidates into the final action.

  4. Learning Object Relation Graph and Tentative Policy for Visual Navigation (ORG)

    Heming Du, Xin Yu, Liang Zheng
    ECCV, 2020. [Paper]

    Summary

    Aiming to learn informative visual representation and robust navigation policy, this paper proposes three complementary techniques, object relation graph (ORG), trial-driven imitation learning (IL), and a memory-augmented tentative policy network (TPN).

    • ORG improves visual representation learning by integrating object relationships;
    • Both Trial-driven IL and TPN underlie robust navigation policy, in- structing the agent to escape from deadlock states, such as looping or being stuck;
    • IL is used in training, TPN for testing.

  5. Target driven visual navigation exploiting object relationships (MJOLNIR)

    Yiding Qiu, Anwesan Pal, Henrik I. Christensen
    arxiv, 2020. [Paper]

    Summary

    They present Memory-utilized Joint hierarchical Object Learning for Navigation in Indoor Rooms (MJOLNIR)1, a target-driven visual navigation algorithm, which considers the inherent relationship between target objects, along with the more salient parent objects occurring in its surrounding.

    MJOLNIR architecture

    The novel CGN architecture

  6. Improving Target-driven Visual Navigation with Attention on 3D Spatial Relationships (Attention 3D)

    Yunlian Lv, Ning Xie, Yimin Shi, Zijiao Wang, and Heng Tao Shen
    arxiv, 2020. [Paper]

    Summary

    To address the generalization and automatic obstacle avoidance issues, we incorporate two designs into classic DRL framework: attention on 3D knowledge graph (KG) and target skill extension (TSE) module.

    • visual features and 3D spatial representations to learn navigation policy;
    • TSE module is used to generate sub-targets which allow agent to learn from failures.

    Framework

    3D spatial representations

  7. Semantic Visual Navigation by Watching YouTube Videos (YouTube)

    Matthew Chang, Arjun Gupta, Saurabh Gupta
    arXiv, 2020. [Paper] [Website]

    Summary

    This paper learns and leverages such semantic cues for navigating to objects of interest in novel environments, by simply watching YouTube videos. They believe that these priors can improve efficiency for navigation in novel environments.

  8. Object Goal Navigation using Goal-Oriented Semantic Exploration (SemExp)

    Devendra Singh Chaplot, Dhiraj Gandhi, Abhinav Gupta*, Ruslan Salakhutdinov
    arXiv, 2020. [Paper] [Website]

    Win CVPR2020 Habitat ObjectNav Challenge

    Summary

    They propose a modular system called, ‘Goal- Oriented Semantic Exploration (SemExp)’ which builds an episodic semantic map and uses it to explore the environment efficiently based on the goal object category.

    • It builds top-down metric maps, which adds extra channels to encode semantic categories explicitly;
    • Instead of using a coverage maximizing goal-agnostic exploration policy based only on obstacle maps, we train a goal-oriented semantic exploration policy which learns semantic priors for efficient navigation.

    Framework

    Semantic Mapping

  9. ObjectNav Revisited: On Evaluation of Embodied Agents Navigating to Objects

    Dhruv Batra, Aaron Gokaslan, Aniruddha Kembhavi, Oleksandr Maksymets, Roozbeh Mottaghi, Manolis Savva, Alexander Toshev, Erik Wijmans
    arXiv, 2020. [Paper]

    Summary

    This paper is not a research paper. They summarize the ObjectNav task and introduce popular datasets (Matterport3D, AI2-THOR) and Challenges (Habitat 2020 Challenge Habitat, RoboTHOR 2020 Challenge RoboTHOR).


知乎没法显示Summary部分的折叠和表格内部跳转,还不支持目录,格式真的太丑了,再次请求支持完整版的 Markdown + HTML!! @知乎小管家

欢迎移步 Github 观看正常版,记得 Fork & Star 哦!

Awesome-Target-driven-Navigation/Skylark0924

About

Paper and summaries about state-of-the-art robot Target-driven Navigation task

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published