Visual Navigation

101 papers with code • 6 benchmarks • 16 datasets

Visual Navigation is the problem of navigating an agent, e.g. a mobile robot, in an environment using camera input only. The agent is given a target image (an image it will see from the target position), and its goal is to move from its current position to the target by applying a sequence of actions, based on the camera observations only.

Source: Vision-based Navigation Using Deep Reinforcement Learning

Libraries

Use these libraries to find Visual Navigation models and implementations

Learning Navigational Visual Representations with Semantic Map Supervision

yiconghong/ego2map-navit ICCV 2023

Being able to perceive the semantics and the spatial structure of the environment is essential for visual navigation of a household robot.

24
23 Jul 2023

Online Self-Supervised Thermal Water Segmentation for Aerial Vehicles

connorlee77/uav-thermal-water-segmentation 18 Jul 2023

We present a new method to adapt an RGB-trained water segmentation network to target-domain aerial thermal imagery using online self-supervision by leveraging texture and motion cues as supervisory signals.

11
18 Jul 2023

The Drunkard's Odometry: Estimating Camera Motion in Deforming Scenes

UZ-SLAMLab/DrunkardsOdometry 29 Jun 2023

Estimating camera motion in deformable scenes poses a complex and open research challenge.

45
29 Jun 2023

HabiCrowd: A High Performance Simulator for Crowd-Aware Visual Navigation

habicrowd/HabiCrowd 20 Jun 2023

Visual navigation, a foundational aspect of Embodied AI (E-AI), has been significantly studied in the past few years.

17
20 Jun 2023

Sonicverse: A Multisensory Simulation Platform for Embodied Household Agents that See and Hear

stanfordvl/sonicverse 1 Jun 2023

We introduce Sonicverse, a multisensory simulation platform with integrated audio-visual simulation for training household agents that can both see and hear.

16
01 Jun 2023

NavGPT: Explicit Reasoning in Vision-and-Language Navigation with Large Language Models

gengzezhou/navgpt 26 May 2023

Trained with an unprecedented scale of data, large language models (LLMs) like ChatGPT and GPT-4 exhibit the emergence of significant reasoning abilities from model scaling.

73
26 May 2023

POPGym: Benchmarking Partially Observable Reinforcement Learning

proroklab/popgym 3 Mar 2023

Real world applications of Reinforcement Learning (RL) are often partially observable, thus requiring memory.

134
03 Mar 2023

Learning by Asking for Embodied Visual Navigation and Task Completion

plan-lab/elba 9 Feb 2023

The research community has shown increasing interest in designing intelligent embodied agents that can assist humans in accomplishing tasks.

0
09 Feb 2023

Offline Reinforcement Learning for Visual Navigation

arjunbhorkar/ReViND 16 Dec 2022

Reinforcement learning can enable robots to navigate to distant goals while optimizing user-specified reward functions, including preferences for following lanes, staying on paved paths, or avoiding freshly mowed grass.

17
16 Dec 2022

BEVBert: Multimodal Map Pre-training for Language-guided Navigation

marsaki/vln-bevbert ICCV 2023

Concretely, we build a local metric map to explicitly aggregate incomplete observations and remove duplicates, while modeling navigation dependency in a global topological map.

156
08 Dec 2022