Measures of Information Reflect Memorization Patterns

10/17/2022
by   Rachit Bansal, et al.
1

Neural networks are known to exploit spurious artifacts (or shortcuts) that co-occur with a target label, exhibiting heuristic memorization. On the other hand, networks have been shown to memorize training examples, resulting in example-level memorization. These kinds of memorization impede generalization of networks beyond their training distributions. Detecting such memorization could be challenging, often requiring researchers to curate tailored test sets. In this work, we hypothesize – and subsequently show – that the diversity in the activation patterns of different neurons is reflective of model generalization and memorization. We quantify the diversity in the neural activations through information-theoretic measures and find support for our hypothesis on experiments spanning several natural language and vision tasks. Importantly, we discover that information organization points to the two forms of memorization, even for neural activations computed on unlabeled in-distribution examples. Lastly, we demonstrate the utility of our findings for the problem of model selection. The associated code and other resources for this work are available at https://linktr.ee/InformationMeasures .

READ FULL TEXT
research
04/09/2022

Neural networks embrace learned diversity

Diversity conveys advantages in nature, yet homogeneous neurons typicall...
research
08/03/2022

Improving Meta-Learning Generalization with Activation-Based Early-Stopping

Meta-Learning algorithms for few-shot learning aim to train neural netwo...
research
10/03/2017

Training Feedforward Neural Networks with Standard Logistic Activations is Feasible

Training feedforward neural networks with standard logistic activations ...
research
06/28/2023

On information captured by neural networks: connections with memorization and generalization

Despite the popularity and success of deep learning, there is limited un...
research
06/10/2021

Within-layer Diversity Reduces Generalization Gap

Neural networks are composed of multiple layers arranged in a hierarchic...
research
05/26/2022

Emergent organization of receptive fields in networks of excitatory and inhibitory neurons

Local patterns of excitation and inhibition that can generate neural wav...
research
06/20/2023

GIO: Gradient Information Optimization for Training Dataset Selection

It is often advantageous to train models on a subset of the available tr...

Please sign up or login with your details

Forgot password? Click here to reset