No Spurious Local Minima: on the Optimization Landscapes of Wide and Deep Neural Networks

10/02/2020
by   Johannes Lederer, et al.
0

Empirical studies suggest that wide neural networks are comparably easy to optimize, but mathematical support for this observation is scarce. In this paper, we analyze the optimization landscapes of deep learning with wide networks. We prove especially that constraint and unconstraint empirical-risk minimization over such networks has no spurious local minima. Hence, our theories substantiate the common belief that increasing network widths not only improves the expressiveness of deep-learning pipelines but also facilitates their optimizations.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/16/2018

Non-attracting Regions of Local Minima in Deep and Wide Neural Networks

Understanding the loss surface of neural networks is essential for the d...
research
10/01/2019

Truth or Backpropaganda? An Empirical Investigation of Deep Learning Theory

We empirically evaluate common assumptions about neural networks that ar...
research
02/25/2021

Spurious Local Minima Are Common for Deep Neural Networks with Piecewise Linear Activations

In this paper, it is shown theoretically that spurious local minima are ...
research
12/13/2017

Regularization and Optimization strategies in Deep Convolutional Neural Network

Convolution Neural Networks, known as ConvNets exceptionally perform wel...
research
05/09/2022

Statistical Guarantees for Approximate Stationary Points of Simple Neural Networks

Since statistical guarantees for neural networks are usually restricted ...
research
03/28/2017

Theory II: Landscape of the Empirical Risk in Deep Learning

Previous theoretical work on deep learning and neural network optimizati...
research
02/11/2020

Unique Properties of Wide Minima in Deep Networks

It is well known that (stochastic) gradient descent has an implicit bias...

Please sign up or login with your details

Forgot password? Click here to reset