No bad local minima: Data independent training error guarantees for multilayer neural networks

05/26/2016
by   Daniel Soudry, et al.
0

We use smoothed analysis techniques to provide guarantees on the training loss of Multilayer Neural Networks (MNNs) at differentiable local minima. Specifically, we examine MNNs with piecewise linear activation functions, quadratic loss and a single output, under mild over-parametrization. We prove that for a MNN with one hidden layer, the training error is zero at every differentiable local minimum, for almost every dataset and dropout-like noise realization. We then extend these results to the case of more than one hidden layer. Our theoretical guarantees assume essentially nothing on the training data, and are verified numerically. These results suggest why the highly non-convex loss of such MNNs can be easily optimized using local updates (e.g., stochastic gradient descent), as observed empirically.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/19/2017

Exponentially vanishing sub-optimal local minima in multilayer neural networks

Background: Statistical mechanics results (Dauphin et al. (2014); Chorom...
research
12/20/2019

Landscape Connectivity and Dropout Stability of SGD Solutions for Over-parameterized Neural Networks

The optimization of multilayer neural networks typically leads to a solu...
research
01/22/2013

Piecewise Linear Multilayer Perceptrons and Dropout

We propose a new type of hidden layer for a multilayer perceptron, and d...
research
07/17/2016

Piecewise convexity of artificial neural networks

Although artificial neural networks have shown great promise in applicat...
research
05/31/2022

Feature Learning in L_2-regularized DNNs: Attraction/Repulsion and Sparsity

We study the loss surface of DNNs with L_2 regularization. We show that ...
research
11/09/2016

Diverse Neural Network Learns True Target Functions

Neural networks are a powerful class of functions that can be trained wi...
research
02/23/2022

On the Omnipresence of Spurious Local Minima in Certain Neural Network Training Problems

We study the loss landscape of training problems for deep artificial neu...

Please sign up or login with your details

Forgot password? Click here to reset