Divide and Contrast: Self-supervised Learning from Uncurated Data

05/17/2021
by   Yonglong Tian, et al.
0

Self-supervised learning holds promise in leveraging large amounts of unlabeled data, however much of its progress has thus far been limited to highly curated pre-training data such as ImageNet. We explore the effects of contrastive learning from larger, less-curated image datasets such as YFCC, and find there is indeed a large difference in the resulting representation quality. We hypothesize that this curation gap is due to a shift in the distribution of image classes – which is more diverse and heavy-tailed – resulting in less relevant negative samples to learn from. We test this hypothesis with a new approach, Divide and Contrast (DnC), which alternates between contrastive learning and clustering-based hard negative mining. When pretrained on less curated datasets, DnC greatly improves the performance of self-supervised learning on downstream tasks, while remaining competitive with the current state-of-the-art on curated datasets.

READ FULL TEXT
research
06/22/2021

Unsupervised Object-Level Representation Learning from Scene Images

Contrastive self-supervised learning has largely narrowed the gap to sup...
research
01/27/2023

Leveraging the Third Dimension in Contrastive Learning

Self-Supervised Learning (SSL) methods operate on unlabeled data to lear...
research
02/18/2023

Data-Efficient Contrastive Self-supervised Learning: Easy Examples Contribute the Most

Self-supervised learning (SSL) learns high-quality representations from ...
research
03/03/2023

Towards Democratizing Joint-Embedding Self-Supervised Learning

Joint Embedding Self-Supervised Learning (JE-SSL) has seen rapid develop...
research
07/30/2023

Self-Supervised Learning of Gait-Based Biomarkers

Markerless motion capture (MMC) is revolutionizing gait analysis in clin...
research
05/17/2023

Sharpness Shift-Aware Self-Supervised Learning

Self-supervised learning aims to extract meaningful features from unlabe...
research
06/10/2022

Is Self-Supervised Learning More Robust Than Supervised Learning?

Self-supervised contrastive learning is a powerful tool to learn visual ...

Please sign up or login with your details

Forgot password? Click here to reset