Going Further: Flatness at the Rescue of Early Stopping for Adversarial Example Transferability

04/05/2023
by   Martin Gubri, et al.
0

Transferability is the property of adversarial examples to be misclassified by other models than the surrogate model for which they were crafted. Previous research has shown that transferability is substantially increased when the training of the surrogate model has been early stopped. A common hypothesis to explain this is that the later training epochs are when models learn the non-robust features that adversarial attacks exploit. Hence, an early stopped model is more robust (hence, a better surrogate) than fully trained models. We demonstrate that the reasons why early stopping improves transferability lie in the side effects it has on the learning dynamics of the model. We first show that early stopping benefits transferability even on models learning from data with non-robust features. We then establish links between transferability and the exploration of the loss landscape in the parameter space, on which early stopping has an inherent effect. More precisely, we observe that transferability peaks when the learning rate decays, which is also the time at which the sharpness of the loss significantly drops. This leads us to propose RFN, a new approach for transferability that minimizes loss sharpness during training in order to maximize transferability. We show that by searching for large flat neighborhoods, RFN always improves over early stopping (by up to 47 points of transferability rate) and is competitive to (if not better than) strong state-of-the-art baselines.

READ FULL TEXT

page 17

page 29

research
07/26/2022

LGV: Boosting Adversarial Example Transferability from Large Geometric Vicinity

We propose transferability from Large Geometric Vicinity (LGV), a new te...
research
01/27/2021

An Early Stopping Bayesian Data Assimilation Approach for Mixed-Logit Estimation

The mixed-logit model is a flexible tool in transportation choice analys...
research
08/21/2023

Enhancing Adversarial Attacks: The Similar Target Method

Deep neural networks are vulnerable to adversarial examples, posing a th...
research
10/16/2022

Non-Transferability in Communication Channels and Tarski's Truth Theorem

This article aims to study transferability issues in communication chann...
research
08/27/2021

Disrupting Adversarial Transferability in Deep Neural Networks

Adversarial attack transferability is a well-recognized phenomenon in de...
research
03/21/2018

Mislearning from Censored Data: Gambler's Fallacy in a Search Problem

In the context of a sequential search problem, I explore large-generatio...
research
09/26/2019

Towards Understanding the Transferability of Deep Representations

Deep neural networks trained on a wide range of datasets demonstrate imp...

Please sign up or login with your details

Forgot password? Click here to reset