Toward Few-step Adversarial Training from a Frequency Perspective

10/13/2020
by   Hans Shih-Han Wang, et al.
23

We investigate adversarial-sample generation methods from a frequency domain perspective and extend standard l_∞ Projected Gradient Descent (PGD) to the frequency domain. The resulting method, which we call Spectral Projected Gradient Descent (SPGD), has better success rate compared to PGD during early steps of the method. Adversarially training models using SPGD achieves greater adversarial accuracy compared to PGD when holding the number of attack steps constant. The use of SPGD can, therefore, reduce the overhead of adversarial training when utilizing adversarial generation with a smaller number of steps. However, we also prove that SPGD is equivalent to a variant of the PGD ordinarily used for the l_∞ threat model. This PGD variant omits the sign function which is ordinarily applied to the gradient. SPGD can, therefore, be performed without explicitly transforming into the frequency domain. Finally, we visualize the perturbations SPGD generates and find they use both high and low-frequency components, which suggests that removing either high-frequency components or low-frequency components is not an effective defense.

READ FULL TEXT

page 5

page 8

page 9

research
02/05/2021

Robust Single-step Adversarial Training with Regularizer

High cost of training time caused by multi-step adversarial example gene...
research
02/28/2019

On the Effectiveness of Low Frequency Perturbations

Carefully crafted, often imperceptible, adversarial perturbations have b...
research
01/12/2023

Phase-shifted Adversarial Training

Adversarial training has been considered an imperative component for saf...
research
11/26/2018

EnResNet: ResNet Ensemble via the Feynman-Kac Formalism

We propose a simple yet powerful ResNet ensemble algorithm which consist...
research
02/10/2022

Domain Adversarial Training: A Game Perspective

The dominant line of work in domain adaptation has focused on learning i...
research
02/06/2021

Understanding the Interaction of Adversarial Training with Noisy Labels

Noisy labels (NL) and adversarial examples both undermine trained models...
research
03/01/2021

Learning Frequency Domain Approximation for Binary Neural Networks

Binary neural networks (BNNs) represent original full-precision weights ...

Please sign up or login with your details

Forgot password? Click here to reset