Closed-Loop Transcription via Convolutional Sparse Coding

02/18/2023
by   Xili Dai, et al.
0

Autoencoding has achieved great empirical success as a framework for learning generative models for natural images. Autoencoders often use generic deep networks as the encoder or decoder, which are difficult to interpret, and the learned representations lack clear structure. In this work, we make the explicit assumption that the image distribution is generated from a multi-stage sparse deconvolution. The corresponding inverse map, which we use as an encoder, is a multi-stage convolution sparse coding (CSC), with each stage obtained from unrolling an optimization algorithm for solving the corresponding (convexified) sparse coding program. To avoid computational difficulties in minimizing distributional distance between the real and generated images, we utilize the recent closed-loop transcription (CTRL) framework that optimizes the rate reduction of the learned sparse representations. Conceptually, our method has high-level connections to score-matching methods such as diffusion models. Empirically, our framework demonstrates competitive performance on large-scale datasets, such as ImageNet-1K, compared to existing autoencoding and generative methods under fair conditions. Even with simpler networks and fewer computational resources, our method demonstrates high visual quality in regenerated images. More surprisingly, the learned autoencoder performs well on unseen datasets. Our method enjoys several side benefits, including more structured and interpretable representations, more stable convergence, and scalability to large datasets. Our method is arguably the first to demonstrate that a concatenation of multiple convolution sparse coding/decoding layers leads to an interpretable and effective autoencoder for modeling the distribution of large-scale natural image datasets.

READ FULL TEXT

page 9

page 10

page 11

page 17

page 18

page 19

research
06/01/2023

StableRep: Synthetic Images from Text-to-Image Models Make Strong Visual Representation Learners

We investigate the potential of learning visual representations using sy...
research
10/24/2022

Revisiting Sparse Convolutional Model for Visual Recognition

Despite strong empirical performance for image classification, deep neur...
research
08/31/2019

Stochastic Convolutional Sparse Coding

State-of-the-art methods for Convolutional Sparse Coding usually employ ...
research
11/12/2021

Closed-Loop Data Transcription to an LDR via Minimaxing Rate Reduction

This work proposes a new computational framework for learning an explici...
research
11/22/2022

Convolutional Neural Generative Coding: Scaling Predictive Coding to Natural Images

In this work, we develop convolutional neural generative coding (Conv-NG...
research
02/13/2021

On the convergence of group-sparse autoencoders

Recent approaches in the theoretical analysis of model-based deep learni...
research
06/10/2014

Optimization Methods for Convolutional Sparse Coding

Sparse and convolutional constraints form a natural prior for many optim...

Please sign up or login with your details

Forgot password? Click here to reset