A survey of deep learning optimizers-first and second order methods

11/28/2022
by   Rohan V Kashyap, et al.
0

Deep Learning optimization involves minimizing a high-dimensional loss function in the weight space which is often perceived as difficult due to its inherent difficulties such as saddle points, local minima, ill-conditioning of the Hessian and limited compute resources. In this paper, we provide a comprehensive review of 12 standard optimization methods successfully used in deep learning research and a theoretical assessment of the difficulties in numerical optimization from the optimization literature.

READ FULL TEXT

page 1

page 3

page 4

research
04/08/2021

Exact Stochastic Second Order Deep Learning

Optimization in Deep Learning is mainly dominated by first-order methods...
research
02/19/2023

Optimization Methods in Deep Learning: A Comprehensive Overview

In recent years, deep learning has achieved remarkable success in variou...
research
05/30/2015

Saddle-free Hessian-free Optimization

Nonconvex optimization problems such as the ones in training deep neural...
research
05/02/2013

Deep Learning of Representations: Looking Forward

Deep learning research aims at discovering learning algorithms that disc...
research
04/04/2020

Optimization methods for achieving high diffraction efficiency with perfect electric conducting gratings

This work presents the implementation, analysis, and convergence study o...
research
04/16/2022

Analytical Benchmark Problems for Multifidelity Optimization Methods

The paper presents a collection of analytical benchmark problems specifi...

Please sign up or login with your details

Forgot password? Click here to reset