generating annotated high-fidelity images containing multiple coherent objects

06/22/2020
by   B. G. Cardenas, et al.
0

Recent developments related to generative models have made it possible to generate diverse high-fidelity images. In particular, layout-to-image generation models have gained significant attention due to their capability to generate realistic complex images containing distinct objects. These models are generally conditioned on either semantic layouts or textual descriptions. However, unlike natural images, providing auxiliary information can be extremely hard in domains such as biomedical imaging and remote sensing. In this work, we propose a multi-object generation framework that can synthesize images with multiple objects without explicitly requiring their contextual information during the generation process. Based on a vector-quantized variational autoencoder (VQ-VAE) backbone, our model learns to preserve spatial coherency within an image as well as semantic coherency between the objects and the background through two powerful autoregressive priors: PixelSNAIL and LayoutPixelSNAIL. While the PixelSNAIL learns the distribution of the latent encodings of the VQ-VAE, the LayoutPixelSNAIL is used to specifically learn the semantic distribution of the objects. An implicit advantage of our approach is that the generated samples are accompanied by object-level annotations. We demonstrate how coherency and fidelity are preserved with our method through experiments on the Multi-MNIST and CLEVR datasets; thereby outperforming state-of-the-art multi-object generative methods. The efficacy of our approach is demonstrated through application on medical imaging datasets, where we show that augmenting the training set with generated samples using our approach improves the performance of existing models.

READ FULL TEXT

page 2

page 6

page 8

page 9

page 15

page 16

page 17

page 18

research
06/02/2019

Generating Diverse High-Fidelity Images with VQ-VAE-2

We explore the use of Vector Quantized Variational AutoEncoder (VQ-VAE) ...
research
12/03/2021

Global Context with Discrete Diffusion in Vector Quantised Modelling for Image Generation

The integration of Vector Quantised Variational AutoEncoder (VQ-VAE) wit...
research
09/19/2022

MoVQ: Modulating Quantized Vectors for High-Fidelity Image Generation

Although two-stage Vector Quantized (VQ) generative models allow for syn...
research
11/08/2019

Collapse Resistant Deep Convolutional GAN for Multi-Object Image Generation

This work introduces a novel system for the generation of images that co...
research
06/27/2020

Generate High Fidelity Images With Generative Variational Autoencoder

In this work, we address the problem of blurred images which are often g...
research
12/04/2018

Generating High Fidelity Images with Subscale Pixel Networks and Multidimensional Upscaling

The unconditional generation of high fidelity images is a longstanding b...
research
04/20/2021

VideoGPT: Video Generation using VQ-VAE and Transformers

We present VideoGPT: a conceptually simple architecture for scaling like...

Please sign up or login with your details

Forgot password? Click here to reset