GoToNet: Fast Monocular Scene Exposure and Exploration

06/13/2022
by   Tom Avrech, et al.
4

Autonomous scene exposure and exploration, especially in localization or communication-denied areas, useful for finding targets in unknown scenes, remains a challenging problem in computer navigation. In this work, we present a novel method for real-time environment exploration, whose only requirements are a visually similar dataset for pre-training, enough lighting in the scene, and an on-board forward-looking RGB camera for environmental sensing. As opposed to existing methods, our method requires only one look (image) to make a good tactical decision, and therefore works at a non-growing, constant time. Two direction predictions, characterized by pixels dubbed the Goto and Lookat pixels, comprise the core of our method. These pixels encode the recommended flight instructions in the following way: the Goto pixel defines the direction in which the agent should move by one distance unit, and the Lookat pixel defines the direction in which the camera should be pointing at in the next step. These flying-instruction pixels are optimized to expose the largest amount of currently unexplored areas. Our method presents a novel deep learning-based navigation approach that is able to solve this problem and demonstrate its ability in an even more complicated setup, i.e., when computational power is limited. In addition, we propose a way to generate a navigation-oriented dataset, enabling efficient training of our method using RGB and depth images. Tests conducted in a simulator evaluating both the sparse pixels' coordinations inferring process, and 2D and 3D test flights aimed to unveil areas and decrease distances to targets achieve promising results. Comparison against a state-of-the-art algorithm shows our method is able to overperform it, that while measuring the new voxels per camera pose, minimum distance to target, percentage of surface voxels seen, and compute time metrics.

READ FULL TEXT

page 2

page 11

page 14

page 15

page 16

page 17

page 22

research
03/31/2021

SOON: Scenario Oriented Object Navigation with Graph-based Exploration

The ability to navigate like a human towards a language-guided target fr...
research
04/11/2022

Bimodal Camera Pose Prediction for Endoscopy

Deducing the 3D structure of endoscopic scenes from images remains extre...
research
09/26/2016

From Monocular SLAM to Autonomous Drone Exploration

Micro aerial vehicles (MAVs) are strongly limited in their payload and p...
research
06/24/2021

Planetary UAV localization based on Multi-modal Registration with Pre-existing Digital Terrain Model

The autonomous real-time optical navigation of planetary UAV is of the k...
research
04/12/2022

Sonified distance in sensory substitution does not always improve localization: comparison with a 2D and 3D handheld device

Early visual to auditory substitution devices encode 2D monocular images...
research
10/01/2021

Vision-Only Robot Navigation in a Neural Radiance World

Neural Radiance Fields (NeRFs) have recently emerged as a powerful parad...
research
01/12/2021

Binary TTC: A Temporal Geofence for Autonomous Navigation

Time-to-contact (TTC), the time for an object to collide with the observ...

Please sign up or login with your details

Forgot password? Click here to reset