How to Train a CAT: Learning Canonical Appearance Transformations for Direct Visual Localization Under Illumination Change

09/09/2017
by   Lee Clement, et al.
0

Direct visual localization has recently enjoyed a resurgence in popularity with the increasing availability of cheap mobile computing power. The competitive accuracy and robustness of these algorithms compared to state-of-the-art feature-based methods, as well as their natural ability to yield dense maps, makes them an appealing choice for a variety of mobile robotics applications. However, direct methods remain brittle in the face of appearance change due to their underlying assumption of photometric consistency, which is commonly violated in practice. In this paper, we propose to mitigate this problem by training deep convolutional encoder-decoder models to transform images of a scene such that they correspond to a previously-seen canonical appearance. We validate our method in multiple environments and illumination conditions using high-fidelity synthetic RGB-D datasets, and integrate the trained models into a direct visual localization pipeline, yielding improvements in visual odometry (VO) accuracy through time-varying illumination conditions, as well as improved metric relocalization performance under illumination change, where conventional methods normally fail. We further provide a preliminary investigation of transfer learning from synthetic to real environments in a localization context. An open-source implementation of our method using PyTorch is available at https://github.com/utiasSTARS/cat-net.

READ FULL TEXT

page 1

page 4

page 5

page 6

research
08/10/2020

IF-Net: An Illumination-invariant Feature Network

Feature descriptor matching is a critical step is many computer vision a...
research
04/01/2019

Learning Matchable Colorspace Transformations for Long-term Metric Visual Localization

Long-term metric localization is an essential capability of autonomous m...
research
04/07/2022

Deep Visual Geo-localization Benchmark

In this paper, we propose a new open-source benchmarking framework for V...
research
03/26/2018

REST: Real-to-Synthetic Transform for Illumination Invariant Camera Localization

Accurate camera localization is an essential part of tracking systems. H...
research
03/15/2022

DSOL: A Fast Direct Sparse Odometry Scheme

In this paper, we describe Direct Sparse Odometry Lite (DSOL), an improv...
research
04/14/2023

DCFace: Synthetic Face Generation with Dual Condition Diffusion Model

Generating synthetic datasets for training face recognition models is ch...
research
11/06/2020

Illumination Normalization by Partially Impossible Encoder-Decoder Cost Function

Images recorded during the lifetime of computer vision based systems und...

Please sign up or login with your details

Forgot password? Click here to reset