SING: A Plug-and-Play DNN Learning Technique

05/25/2023
by   Adrien Courtois, et al.
0

We propose SING (StabIlized and Normalized Gradient), a plug-and-play technique that improves the stability and generalization of the Adam(W) optimizer. SING is straightforward to implement and has minimal computational overhead, requiring only a layer-wise standardization of the gradients fed to Adam(W) without introducing additional hyper-parameters. We support the effectiveness and practicality of the proposed approach by showing improved results on a wide range of architectures, problems (such as image classification, depth estimation, and natural language processing), and in combination with other optimizers. We provide a theoretical analysis of the convergence of the method, and we show that by virtue of the standardization, SING can escape local minima narrower than a threshold that is inversely proportional to the network's depth.

READ FULL TEXT
research
12/10/2017

Gradient Normalization & Depth Based Decay For Deep Learning

In this paper we introduce a novel method of gradient normalization and ...
research
12/24/2019

An Analysis of the Expressiveness of Deep Neural Network Architectures Based on Their Lipschitz Constants

Deep neural networks (DNNs) have emerged as a popular mathematical tool ...
research
09/20/2017

Estimated Depth Map Helps Image Classification

We consider image classification with estimated depth. This problem fall...
research
02/01/2023

Weight Prediction Boosts the Convergence of AdamW

In this paper, we introduce weight prediction into the AdamW optimizer t...
research
03/16/2020

Active Depth Estimation: Stability Analysis and its Applications

Recovering the 3D structure of the surrounding environment is an essenti...
research
09/18/2014

Fingerprint Classification Based on Depth Neural Network

Fingerprint classification is an effective technique for reducing the ca...

Please sign up or login with your details

Forgot password? Click here to reset