Generalization Comparison of Deep Neural Networks via Output Sensitivity

07/30/2020
by   Mahsa Forouzesh, et al.
0

Although recent works have brought some insights into the performance improvement of techniques used in state-of-the-art deep-learning models, more work is needed to understand their generalization properties. We shed light on this matter by linking the loss function to the output's sensitivity to its input. We find a rather strong empirical relation between the output sensitivity and the variance in the bias-variance decomposition of the loss function, which hints on using sensitivity as a metric for comparing the generalization performance of networks, without requiring labeled data. We find that sensitivity is decreased by applying popular methods which improve the generalization performance of the model, such as (1) using a deep network rather than a wide one, (2) adding convolutional layers to baseline classifiers instead of adding fully-connected layers, (3) using batch normalization, dropout and max-pooling, and (4) applying parameter initialization techniques.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/04/2015

Max-Pooling Dropout for Regularization of Convolutional Neural Networks

Recently, dropout has seen increasing use in deep learning. For deep con...
research
12/01/2015

Towards Dropout Training for Convolutional Neural Networks

Recently, dropout has seen increasing use in deep learning. For deep con...
research
08/01/2022

Improving the Trainability of Deep Neural Networks through Layerwise Batch-Entropy Regularization

Training deep neural networks is a very demanding task, especially chall...
research
10/31/2018

Understanding Deep Neural Networks through Input Uncertainties

Techniques for understanding the functioning of complex machine learning...
research
08/11/2019

Enhanced Seismic Imaging with Predictive Neural Networks for Geophysics

We propose a predictive neural network architecture that can be utilized...
research
08/24/2022

On a Built-in Conflict between Deep Learning and Systematic Generalization

In this paper, we hypothesize that internal function sharing is one of t...

Please sign up or login with your details

Forgot password? Click here to reset