Object-wise Masked Autoencoders for Fast Pre-training

05/28/2022
by   Jiantao Wu, et al.
0

Self-supervised pre-training for images without labels has recently achieved promising performance in image classification. The success of transformer-based methods, ViT and MAE, draws the community's attention to the design of backbone architecture and self-supervised task. In this work, we show that current masked image encoding models learn the underlying relationship between all objects in the whole scene, instead of a single object representation. Therefore, those methods bring a lot of compute time for self-supervised pre-training. To solve this issue, we introduce a novel object selection and division strategy to drop non-object patches for learning object-wise representations by selective reconstruction with interested region masks. We refer to this method ObjMAE. Extensive experiments on four commonly-used datasets demonstrate the effectiveness of our model in reducing the compute cost by 72 investigate the inter-object and intra-object relationship and find that the latter is crucial for self-supervised pre-training.

READ FULL TEXT

page 1

page 2

page 3

page 4

page 6

page 7

page 11

research
04/20/2022

Self-supervised Learning for Sonar Image Classification

Self-supervised learning has proved to be a powerful approach to learn i...
research
04/12/2023

Hard Patches Mining for Masked Image Modeling

Masked image modeling (MIM) has attracted much research attention due to...
research
08/14/2023

PatchContrast: Self-Supervised Pre-training for 3D Object Detection

Accurately detecting objects in the environment is a key challenge for a...
research
05/08/2023

Self-supervised Pre-training with Masked Shape Prediction for 3D Scene Understanding

Masked signal modeling has greatly advanced self-supervised pre-training...
research
05/28/2022

SupMAE: Supervised Masked Autoencoders Are Efficient Vision Learners

Self-supervised Masked Autoencoders (MAE) are emerging as a new pre-trai...
research
10/12/2020

Improving Self-supervised Pre-training via a Fully-Explored Masked Language Model

Masked Language Model (MLM) framework has been widely adopted for self-s...
research
11/23/2022

Self-Supervised Learning based on Heat Equation

This paper presents a new perspective of self-supervised learning based ...

Please sign up or login with your details

Forgot password? Click here to reset