Exact information propagation through fully-connected feed forward neural networks

06/17/2018
by   Rebekka Burkholz, et al.
0

Neural network ensembles at initialisation give rise to the trainability and training speed of neural networks and thus support parameter choices at initialisation. These insights rely so far on mean field approximations that assume infinite layer width and study average squared signals. Thus, information about the full output distribution gets lost. Therefore, we derive the output distribution exactly (without mean field assumptions), for fully-connected networks with Gaussian weights and biases. The layer-wise transition of the signal distribution is guided by a linear integral operator, whose kernel has a closed form solution in case of rectified linear units for nonlinear activations. This enables us to analyze some of its spectral properties, for instance, the shape of the stationary distribution for different parameter choices and the dynamics of signal propagation.

READ FULL TEXT
research
10/17/2021

A Riemannian Mean Field Formulation for Two-layer Neural Networks with Batch Normalization

The training dynamics of two-layer neural networks with batch normalizat...
research
07/14/2020

Global Convergence of Second-order Dynamics in Two-layer Neural Networks

Recent results have shown that for two-layer fully connected neural netw...
research
07/12/2023

Quantitative CLTs in Deep Neural Networks

We study the distribution of a fully connected neural network with rando...
research
01/11/2018

Which Neural Net Architectures Give Rise To Exploding and Vanishing Gradients?

We give a rigorous analysis of the statistical behavior of gradients in ...
research
02/01/2019

Signal propagation in continuous approximations of binary neural networks

The training of stochastic neural network models with binary (±1) weight...
research
06/09/2023

Deterministic equivalent of the Conjugate Kernel matrix associated to Artificial Neural Networks

We study the Conjugate Kernel associated to a multi-layer linear-width f...
research
05/15/2023

Introduction to dynamical mean-field theory of generic random neural networks

Dynamical mean-field theory is a powerful physics tool used to analyze t...

Please sign up or login with your details

Forgot password? Click here to reset