SLACK: Stable Learning of Augmentations with Cold-start and KL regularization

06/16/2023
by   Juliette Marrie, et al.
0

Data augmentation is known to improve the generalization capabilities of neural networks, provided that the set of transformations is chosen with care, a selection often performed manually. Automatic data augmentation aims at automating this process. However, most recent approaches still rely on some prior information; they start from a small pool of manually-selected default transformations that are either used to pretrain the network or forced to be part of the policy learned by the automatic data augmentation algorithm. In this paper, we propose to directly learn the augmentation policy without leveraging such prior knowledge. The resulting bilevel optimization problem becomes more challenging due to the larger search space and the inherent instability of bilevel optimization algorithms. To mitigate these issues (i) we follow a successive cold-start strategy with a Kullback-Leibler regularization, and (ii) we parameterize magnitudes as continuous distributions. Our approach leads to competitive results on standard benchmarks despite a more challenging setting, and generalizes beyond natural images.

READ FULL TEXT

page 2

page 8

page 14

page 15

page 16

page 17

research
09/29/2022

Automatic Data Augmentation via Invariance-Constrained Learning

Underlying data structures, such as symmetries or invariances to transfo...
research
10/20/2021

Improving Model Generalization by Agreement of Learned Representations from Data Augmentation

Data augmentation reduces the generalization error by forcing a model to...
research
03/16/2018

A Kernel Theory of Modern Data Augmentation

Data augmentation, a technique in which a training set is expanded with ...
research
10/19/2022

G-Augment: Searching for the Meta-Structure of Data Augmentation Policies for ASR

Data augmentation is a ubiquitous technique used to provide robustness t...
research
05/03/2019

Information asymmetry in KL-regularized RL

Many real world tasks exhibit rich structure that is repeated across dif...
research
12/17/2020

Joint Search of Data Augmentation Policies and Network Architectures

The common pipeline of training deep neural networks consists of several...
research
05/17/2020

FuCiTNet: Improving the generalization of deep learning networks by the fusion of learned class-inherent transformations

It is widely known that very small datasets produce overfitting in Deep ...

Please sign up or login with your details

Forgot password? Click here to reset