Visual Navigation

107 papers with code • 6 benchmarks • 16 datasets

Visual Navigation is the problem of navigating an agent, e.g. a mobile robot, in an environment using camera input only. The agent is given a target image (an image it will see from the target position), and its goal is to move from its current position to the target by applying a sequence of actions, based on the camera observations only.

Source: Vision-based Navigation Using Deep Reinforcement Learning

Libraries

Use these libraries to find Visual Navigation models and implementations

What do navigation agents learn about their environment?

allenai/isee CVPR 2022

We use iSEE to probe the dynamic representations produced by these agents for the presence of information about the agent as well as the environment.

17
17 Jun 2022

SoundSpaces 2.0: A Simulation Platform for Visual-Acoustic Learning

facebookresearch/sound-spaces 16 Jun 2022

We introduce SoundSpaces 2. 0, a platform for on-the-fly geometry-based audio rendering for 3D environments.

313
16 Jun 2022

Zero-shot object goal visual navigation

pioneer-innovation/zero-shot-object-navigation 15 Jun 2022

Object goal visual navigation is a challenging task that aims to guide a robot to find the target object based on its visual observation, and the target is limited to the classes pre-defined in the training stage.

24
15 Jun 2022

Contrastive Learning for Image Registration in Visual Teach and Repeat Navigation

Zdeeno/Siamese-network-image-alignment MDPI Sensors 2022

Visual teach and repeat navigation (VT&R) is popular in robotics thanks to its simplicity and versatility.

7
13 Apr 2022

A Visual Navigation Perspective for Category-Level Object Pose Estimation

wrld/visual_navigation_pose_estimation 25 Mar 2022

In this paper, we take a deeper look at the inference of analysis-by-synthesis from the perspective of visual navigation, and investigate what is a good navigation policy for this specific task.

19
25 Mar 2022

Benchmarking Visual Localization for Autonomous Navigation

lasuomela/carla_vloc_benchmark 24 Mar 2022

The experimental part of the paper studies the effects of four such variables by evaluating state-of-the-art visual localization methods as part of the motion planning module of an autonomous navigation stack.

21
24 Mar 2022

HOP: History-and-Order Aware Pre-training for Vision-and-Language Navigation

yanyuanqiao/hop-vln CVPR 2022

Pre-training has been adopted in a few of recent works for Vision-and-Language Navigation (VLN).

26
22 Mar 2022

Think Global, Act Local: Dual-scale Graph Transformer for Vision-and-Language Navigation

cshizhe/vln-duet CVPR 2022

To balance the complexity of large action space reasoning and fine-grained language grounding, we dynamically combine a fine-scale encoding over local observations and a coarse-scale encoding on a global map via graph transformers.

92
23 Feb 2022

Sound Adversarial Audio-Visual Navigation

yyf17/saavn ICLR 2022

In this work, we design an acoustically complex environment in which, besides the target sound, there exists a sound attacker playing a zero-sum game with the agent.

15
22 Feb 2022

RARA: Zero-shot Sim2Real Visual Navigation with Following Foreground Cues

kkelchte/fgbg 8 Jan 2022

In this work, we tackle this gap for the specific case of camera-based navigation, formulating it as following a visual cue in the foreground with arbitrary backgrounds.

2
08 Jan 2022