Real-Time Facial Segmentation and Performance Capture from RGB Input

04/10/2016
by   Shunsuke Saito, et al.
0

We introduce the concept of unconstrained real-time 3D facial performance capture through explicit semantic segmentation in the RGB input. To ensure robustness, cutting edge supervised learning approaches rely on large training datasets of face images captured in the wild. While impressive tracking quality has been demonstrated for faces that are largely visible, any occlusion due to hair, accessories, or hand-to-face gestures would result in significant visual artifacts and loss of tracking accuracy. The modeling of occlusions has been mostly avoided due to its immense space of appearance variability. To address this curse of high dimensionality, we perform tracking in unconstrained images assuming non-face regions can be fully masked out. Along with recent breakthroughs in deep learning, we demonstrate that pixel-level facial segmentation is possible in real-time by repurposing convolutional neural networks designed originally for general semantic segmentation. We develop an efficient architecture based on a two-stream deconvolution network with complementary characteristics, and introduce carefully designed training samples and data augmentation strategies for improved segmentation accuracy and robustness. We adopt a state-of-the-art regression-based facial tracking framework with segmented face images as training, and demonstrate accurate and uninterrupted facial performance capture in the presence of extreme occlusion and even side views. Furthermore, the resulting segmentation can be directly used to composite partial 3D face models on the input images and enable seamless facial manipulation tasks, such as virtual make-up or face replacement.

READ FULL TEXT

page 5

page 10

page 12

page 13

research
12/15/2021

Segmentation-Reconstruction-Guided Facial Image De-occlusion

Occlusions are very common in face images in the wild, leading to the de...
research
08/16/2018

Self-supervised CNN for Unconstrained 3D Facial Performance Capture from an RGB-D Camera

We present a novel method for real-time 3D facial performance capture wi...
research
08/16/2018

Self-supervised CNN for Unconstrained 3D Facial Performance Capture from a Single RGB-D Camera

We present a novel method for real-time 3D facial performance capture wi...
research
09/21/2022

Self-adversarial Multi-scale Contrastive Learning for Semantic Segmentation of Thermal Facial Images

Reliable segmentation of thermal facial images in unconstrained settings...
research
07/05/2022

Latents2Segments: Disentangling the Latent Space of Generative Models for Semantic Segmentation of Face Images

With the advent of an increasing number of Augmented and Virtual Reality...
research
06/21/2021

Normalized Avatar Synthesis Using StyleGAN and Perceptual Refinement

We introduce a highly robust GAN-based framework for digitizing a normal...
research
04/05/2019

Prediction-Tracking-Segmentation

We introduce a prediction driven method for visual tracking and segmenta...

Please sign up or login with your details

Forgot password? Click here to reset