Improving Self-supervised Learning with Automated Unsupervised Outlier Arbitration

12/15/2021
by   Yu Wang, et al.
0

Our work reveals a structured shortcoming of the existing mainstream self-supervised learning methods. Whereas self-supervised learning frameworks usually take the prevailing perfect instance level invariance hypothesis for granted, we carefully investigate the pitfalls behind. Particularly, we argue that the existing augmentation pipeline for generating multiple positive views naturally introduces out-of-distribution (OOD) samples that undermine the learning of the downstream tasks. Generating diverse positive augmentations on the input does not always pay off in benefiting downstream tasks. To overcome this inherent deficiency, we introduce a lightweight latent variable model UOTA, targeting the view sampling issue for self-supervised learning. UOTA adaptively searches for the most important sampling region to produce views, and provides viable choice for outlier-robust self-supervised learning approaches. Our method directly generalizes to many mainstream self-supervised learning approaches, regardless of the loss's nature contrastive or not. We empirically show UOTA's advantage over the state-of-the-art self-supervised paradigms with evident margin, which well justifies the existence of the OOD sample issue embedded in the existing approaches. Especially, we theoretically prove that the merits of the proposal boil down to guaranteed estimator variance and bias reduction. Code is available: at https://github.com/ssl-codelab/uota.

READ FULL TEXT
research
01/27/2023

Cross-Architectural Positive Pairs improve the effectiveness of Self-Supervised Learning

Existing self-supervised techniques have extreme computational requireme...
research
07/22/2022

Adaptive Soft Contrastive Learning

Self-supervised learning has recently achieved great success in represen...
research
07/18/2022

ExAgt: Expert-guided Augmentation for Representation Learning of Traffic Scenarios

Representation learning in recent years has been addressed with self-sup...
research
08/19/2021

Concurrent Discrimination and Alignment for Self-Supervised Feature Learning

Existing self-supervised learning methods learn representation by means ...
research
02/18/2022

Masked prediction tasks: a parameter identifiability view

The vast majority of work in self-supervised learning, both theoretical ...
research
10/31/2022

DUEL: Adaptive Duplicate Elimination on Working Memory for Self-Supervised Learning

In Self-Supervised Learning (SSL), it is known that frequent occurrences...
research
05/25/2023

Sample and Predict Your Latent: Modality-free Sequential Disentanglement via Contrastive Estimation

Unsupervised disentanglement is a long-standing challenge in representat...

Please sign up or login with your details

Forgot password? Click here to reset