Cyclical Learning Rates for Training Neural Networks

06/03/2015
by   Leslie N. Smith, et al.
0

It is known that the learning rate is the most important hyper-parameter to tune for training deep neural networks. This paper describes a new method for setting the learning rate, named cyclical learning rates, which practically eliminates the need to experimentally find the best values and schedule for the global learning rates. Instead of monotonically decreasing the learning rate, this method lets the learning rate cyclically vary between reasonable boundary values. Training with cyclical learning rates instead of fixed values achieves improved classification accuracy without a need to tune and often in fewer iterations. This paper also describes a simple way to estimate "reasonable bounds" -- linearly increasing the learning rate of the network for a few epochs. In addition, cyclical learning rates are demonstrated on the CIFAR-10 and CIFAR-100 datasets with ResNets, Stochastic Depth networks, and DenseNets, and the ImageNet dataset with the AlexNet and GoogLeNet architectures. These are practical tools for everyone who trains neural networks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/13/2020

k-decay: A New Method For Learning Rate Schedule

It is well known that the learning rate is the most important hyper-para...
research
02/20/2019

A novel adaptive learning rate scheduler for deep neural networks

Optimizing deep neural networks is largely thought to be an empirical pr...
research
03/23/2021

Evolving Learning Rate Optimizers for Deep Neural Networks

Artificial Neural Networks (ANNs) became popular due to their successful...
research
08/21/2023

We Don't Need No Adam, All We Need Is EVE: On The Variance of Dual Learning Rate And Beyond

In the rapidly advancing field of deep learning, optimising deep neural ...
research
03/22/2020

Tune smarter not harder: A principled approach to tuning learning rates for shallow nets

Effective hyper-parameter tuning is essential to guarantee the performan...
research
03/01/2021

Statistically Significant Stopping of Neural Network Training

The general approach taken when training deep learning classifiers is to...
research
05/27/2021

Training With Data Dependent Dynamic Learning Rates

Recently many first and second order variants of SGD have been proposed ...

Please sign up or login with your details

Forgot password? Click here to reset