Collective evolution of weights in wide neural networks

10/09/2018
by   Dmitry Yarotsky, et al.
0

We derive a nonlinear integro-differential transport equation describing collective evolution of weights under gradient descent in large-width neural-network-like models. We characterize stationary points of the evolution and analyze several scenarios where the transport equation can be solved approximately. We test our general method in the special case of linear free-knot splines, and find good agreement between theory and experiment in observations of global optima, stability of stationary points, and convergence rates.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/12/2019

Convective transport in nanofluids: the stationary problem

We analyze the existence of solutions to the stationary problem from a m...
research
11/13/2019

Asymptotics of Reinforcement Learning with Neural Networks

We prove that a single-layer neural network trained with the Q-learning ...
research
02/14/2022

Continuous-time stochastic gradient descent for optimizing over the stationary distribution of stochastic differential equations

We develop a new continuous-time stochastic gradient descent method for ...
research
01/05/2019

Analysis of a Two-Layer Neural Network via Displacement Convexity

Fitting a function by using linear combinations of a large number N of `...
research
10/12/2021

On Convergence of Training Loss Without Reaching Stationary Points

It is a well-known fact that nonconvex optimization is computationally i...
research
07/25/2021

Estimation of Stationary Optimal Transport Plans

We study optimal transport problems in which finite-valued quantities of...
research
01/01/2023

Yuille-Poggio's Flow and Global Minimizer of polynomials through convexification by Heat Evolution

In this paper, we investigate the possibility of the backward-differenti...

Please sign up or login with your details

Forgot password? Click here to reset