On the Importance of Accurate Geometry Data for Dense 3D Vision Tasks

03/26/2023
by   HyunJun Jung, et al.
9

Learning-based methods to solve dense 3D vision problems typically train on 3D sensor data. The respectively used principle of measuring distances provides advantages and drawbacks. These are typically not compared nor discussed in the literature due to a lack of multi-modal datasets. Texture-less regions are problematic for structure from motion and stereo, reflective material poses issues for active sensing, and distances for translucent objects are intricate to measure with existing hardware. Training on inaccurate or corrupt data induces model bias and hampers generalisation capabilities. These effects remain unnoticed if the sensor measurement is considered as ground truth during the evaluation. This paper investigates the effect of sensor errors for the dense 3D vision tasks of depth estimation and reconstruction. We rigorously show the significant impact of sensor characteristics on the learned predictions and notice generalisation issues arising from various technologies in everyday household environments. For evaluation, we introduce a carefully designed dataset[dataset available at https://github.com/Junggy/HAMMER-dataset] comprising measurements from commodity sensors, namely D-ToF, I-ToF, passive/active stereo, and monocular RGB+P. Our study quantifies the considerable sensor noise impact and paves the way to improved dense vision estimates and targeted data fusion.

READ FULL TEXT

page 15

page 17

page 21

page 27

page 29

page 30

page 31

page 32

research
05/09/2022

Is my Depth Ground-Truth Good Enough? HAMMER – Highly Accurate Multi-Modal Dataset for DEnse 3D Scene Regression

Depth estimation is a core task in 3D computer vision. Recent methods in...
research
07/20/2023

Probabilistic Multimodal Depth Estimation Based on Camera-LiDAR Sensor Fusion

Multi-modal depth estimation is one of the key challenges for endowing a...
research
03/20/2022

Depth Estimation by Combining Binocular Stereo and Monocular Structured-Light

It is well known that the passive stereo system cannot adapt well to wea...
research
12/30/2022

X-MAS: Extremely Large-Scale Multi-Modal Sensor Dataset for Outdoor Surveillance in Real Environments

In robotics and computer vision communities, extensive studies have been...
research
01/26/2022

On the Issues of TrueDepth Sensor Data for Computer Vision Tasks Across Different iPad Generations

In 2017 Apple introduced the TrueDepth sensor with the iPhone X release....
research
08/18/2020

DeepLiDARFlow: A Deep Learning Architecture For Scene Flow Estimation Using Monocular Camera and Sparse LiDAR

Scene flow is the dense 3D reconstruction of motion and geometry of a sc...
research
06/27/2012

Efficient Selection of Disambiguating Actions for Stereo Vision

In many domains that involve the use of sensors, such as robotics or sen...

Please sign up or login with your details

Forgot password? Click here to reset