BigDatasetGAN: Synthesizing ImageNet with Pixel-wise Annotations

by   Daiqing Li, et al.

Annotating images with pixel-wise labels is a time-consuming and costly process. Recently, DatasetGAN showcased a promising alternative - to synthesize a large labeled dataset via a generative adversarial network (GAN) by exploiting a small set of manually labeled, GAN-generated images. Here, we scale DatasetGAN to ImageNet scale of class diversity. We take image samples from the class-conditional generative model BigGAN trained on ImageNet, and manually annotate 5 images per class, for all 1k classes. By training an effective feature segmentation architecture on top of BigGAN, we turn BigGAN into a labeled dataset generator. We further show that VQGAN can similarly serve as a dataset generator, leveraging the already annotated data. We create a new ImageNet benchmark by labeling an additional set of 8k real images and evaluate segmentation performance in a variety of settings. Through an extensive ablation study we show big gains in leveraging a large generated dataset to train different supervised and self-supervised backbone models on pixel-wise tasks. Furthermore, we demonstrate that using our synthesized datasets for pre-training leads to improvements over standard ImageNet pre-training on several downstream datasets, such as PASCAL-VOC, MS-COCO, Cityscapes and chest X-ray, as well as tasks (detection, segmentation). Our benchmark will be made public and maintain a leaderboard for this challenging task. Project Page:


page 7

page 12

page 13

page 14

page 15

page 16

page 17

page 18


Mix-and-Match Tuning for Self-Supervised Semantic Segmentation

Deep convolutional networks for semantic image segmentation typically re...

Spurious Features Everywhere – Large-Scale Detection of Harmful Spurious Features in ImageNet

Benchmark performance of deep learning classifiers alone is not a reliab...

Additive Class Distinction Maps using Branched-GANs

We present a new model, training procedure and architecture to create pr...

DreamTeacher: Pretraining Image Backbones with Deep Generative Models

In this work, we introduce a self-supervised feature representation lear...

PourIt!: Weakly-supervised Liquid Perception from a Single Image for Visual Closed-Loop Robotic Pouring

Liquid perception is critical for robotic pouring tasks. It usually requ...

Self-Supervised One-Shot Learning for Automatic Segmentation of StyleGAN Images

We propose in this paper a framework for automatic one-shot segmentation...

Overcoming Small Minirhizotron Datasets Using Transfer Learning

Minirhizotron technology is widely used for studying the development of ...

Please sign up or login with your details

Forgot password? Click here to reset