Some like it tough: Improving model generalization via progressively increasing the training difficulty

10/25/2021
by   Hannes Fassold, et al.
0

In this work, we propose to progressively increase the training difficulty during learning a neural network model via a novel strategy which we call mini-batch trimming. This strategy makes sure that the optimizer puts its focus in the later training stages on the more difficult samples, which we identify as the ones with the highest loss in the current mini-batch. The strategy is very easy to integrate into an existing training pipeline and does not necessitate a change of the network model. Experiments on several image classification problems show that mini-batch trimming is able to increase the generalization ability (measured via final test error) of the trained model.

READ FULL TEXT

page 1

page 2

research
07/14/2021

Disparity Between Batches as a Signal for Early Stopping

We propose a metric for evaluating the generalization ability of deep ne...
research
04/07/2023

Can we learn better with hard samples?

In deep learning, mini-batch training is commonly used to optimize netwo...
research
06/20/2019

Submodular Batch Selection for Training Deep Neural Networks

Mini-batch gradient descent based methods are the de facto algorithms fo...
research
08/08/2022

GRIT-VLP: Grouped Mini-batch Sampling for Efficient Vision and Language Pre-training

Most of the currently existing vision and language pre-training (VLP) me...
research
11/20/2017

A Novel Convolutional Neural Network for Image Steganalysis with Shared Normalization

Deep learning based image steganalysis has attracted increasing attentio...
research
06/19/2017

An Empirical Study of Mini-Batch Creation Strategies for Neural Machine Translation

Training of neural machine translation (NMT) models usually uses mini-ba...
research
09/29/2022

Batch Normalization Explained

A critically important, ubiquitous, and yet poorly understood ingredient...

Please sign up or login with your details

Forgot password? Click here to reset