On Energy-Based Models with Overparametrized Shallow Neural Networks

04/15/2021
by   Carles Domingo Enrich, et al.
15

Energy-based models (EBMs) are a simple yet powerful framework for generative modeling. They are based on a trainable energy function which defines an associated Gibbs measure, and they can be trained and sampled from via well-established statistical tools, such as MCMC. Neural networks may be used as energy function approximators, providing both a rich class of expressive models as well as a flexible device to incorporate data structure. In this work we focus on shallow neural networks. Building from the incipient theory of overparametrized neural networks, we show that models trained in the so-called "active" regime provide a statistical advantage over their associated "lazy" or kernel regime, leading to improved adaptivity to hidden low-dimensional structure in the data distribution, as already observed in supervised learning. Our study covers both maximum likelihood and Stein Discrepancy estimators, and we validate our theoretical results with numerical experiments on synthetic data.

READ FULL TEXT
research
07/11/2021

Dual Training of Energy-Based Models with Overparametrized Shallow Neural Networks

Energy-based models (EBMs) are generative models that are usually traine...
research
02/02/2023

Over-parameterised Shallow Neural Networks with Asymmetrical Node Scaling: Global Convergence Guarantees and Feature Learning

We consider the optimisation of large and shallow neural networks via gr...
research
07/12/2023

Energy Discrepancies: A Score-Independent Loss for Energy-Based Models

Energy-based models are a simple yet powerful class of probabilistic mod...
research
02/03/2022

Generative Flow Networks for Discrete Probabilistic Modeling

We present energy-based generative flow networks (EB-GFN), a novel proba...
research
07/14/2023

Training Discrete Energy-Based Models with Energy Discrepancy

Training energy-based models (EBMs) on discrete spaces is challenging be...
research
06/29/2023

Sampling weights of deep neural networks

We introduce a probability distribution, combined with an efficient samp...
research
10/08/2020

No MCMC for me: Amortized sampling for fast and stable training of energy-based models

Energy-Based Models (EBMs) present a flexible and appealing way to repre...

Please sign up or login with your details

Forgot password? Click here to reset