Gradient descent revisited via an adaptive online learning rate

01/27/2018
by   Mathieu Ravaut, et al.
0

Any gradient descent optimization requires to choose a learning rate. With deeper and deeper models, tuning that learning rate can easily become tedious and does not necessarily lead to an ideal convergence. We propose a variation of the gradient descent algorithm in the which the learning rate is not fixed. Instead, we learn the learning rate itself, either by another gradient descent (first-order method), or by Newton's method (second-order). This way, gradient descent for any machine learning algorithm can be optimized.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/27/2020

Faster Biological Gradient Descent Learning

Back-propagation is a popular machine learning algorithm that uses gradi...
research
10/19/2022

Differentiable Self-Adaptive Learning Rate

Learning rate adaptation is a popular topic in machine learning. Gradien...
research
09/29/2019

Gradient Descent: The Ultimate Optimizer

Working with any gradient-based machine learning algorithm involves the ...
research
04/12/2021

Meta-Regularization: An Approach to Adaptive Choice of the Learning Rate in Gradient Descent

We propose Meta-Regularization, a novel approach for the adaptive choice...
research
10/09/2020

Reparametrizing gradient descent

In this work, we propose an optimization algorithm which we call norm-ad...
research
10/15/2020

Neograd: gradient descent with an adaptive learning rate

Since its inception by Cauchy in 1847, the gradient descent algorithm ha...
research
01/19/2020

Dual Stochastic Natural Gradient Descent

Although theoretically appealing, Stochastic Natural Gradient Descent (S...

Please sign up or login with your details

Forgot password? Click here to reset