Monitored Distillation for Positive Congruent Depth Completion

03/30/2022
by   Tian Yu Liu, et al.
1

We propose a method to infer a dense depth map from a single image, its calibration, and the associated sparse point cloud. In order to leverage existing models that produce putative depth maps (teacher models), we propose an adaptive knowledge distillation approach that yields a positive congruent training process, where a student model avoids learning the error modes of the teachers. We consider the scenario of a blind ensemble where we do not have access to ground truth for model selection nor training. The crux of our method, termed Monitored Distillation, lies in a validation criterion that allows us to learn from teachers by choosing predictions that best minimize the photometric reprojection error for a given image. The result of which is a distilled depth map and a confidence map, or "monitor", for how well a prediction from a particular teacher fits the observed image. The monitor adaptively weights the distilled depth where, if all of the teachers exhibit high residuals, the standard unsupervised image reconstruction loss takes over as the supervisory signal. On indoor scenes (VOID), we outperform blind ensembling baselines by 13.3 model size reduction while maintaining comparable performance to the best supervised method. For outdoors (KITTI), we tie for 5th overall on the benchmark despite not using ground truth.

READ FULL TEXT

page 5

page 6

page 10

page 22

page 24

page 25

page 26

page 27

research
03/28/2023

Enhancing Depth Completion with Multi-View Monitored Distillation

This paper presents a novel method for depth completion, which leverages...
research
12/30/2021

Confidence-Aware Multi-Teacher Knowledge Distillation

Knowledge distillation is initially introduced to utilize additional sup...
research
01/28/2019

Dense Depth Posterior (DDP) from Single Image and Sparse Range

We present a deep learning system to infer the posterior distribution of...
research
06/25/2023

Feature Adversarial Distillation for Point Cloud Classification

Due to the point cloud's irregular and unordered geometry structure, con...
research
03/11/2019

Refine and Distill: Exploiting Cycle-Inconsistency and Knowledge Distillation for Unsupervised Monocular Depth Estimation

Nowadays, the majority of state of the art monocular depth estimation te...
research
06/25/2020

Discontinuous and Smooth Depth Completion with Binary Anisotropic Diffusion Tensor

We propose an unsupervised real-time dense depth completion from a spars...

Please sign up or login with your details

Forgot password? Click here to reset