Adversarial vulnerability for any classifier

02/23/2018
by   Alhussein Fawzi, et al.
0

Despite achieving impressive and often superhuman performance on multiple benchmarks, state-of-the-art deep networks remain highly vulnerable to perturbations: adding small, imperceptible, adversarial perturbations can lead to very high error rates. Provided the data distribution is defined using a generative model mapping latent vectors to datapoints in the distribution, we prove that no classifier can be robust to adversarial perturbations when the latent space is sufficiently large and the generative model sufficiently smooth. Under the same conditions, we prove the existence of adversarial perturbations that transfer well across different models with small risk. We conclude the paper with experiments validating the theoretical bounds.

READ FULL TEXT
research
12/19/2019

Adversarial Perturbations on the Perceptual Ball

We present a simple regularisation of Adversarial Perturbations based up...
research
01/09/2018

Adversarial Spheres

State of the art computer vision models have been shown to be vulnerable...
research
10/29/2018

Adversarial Risk and Robustness: General Definitions and Implications for the Uniform Distribution

We study adversarial perturbations when the instances are uniformly dist...
research
10/29/2020

Robustifying Binary Classification to Adversarial Perturbation

Despite the enormous success of machine learning models in various appli...
research
03/24/2019

Variational Inference with Latent Space Quantization for Adversarial Resilience

Despite their tremendous success in modelling high-dimensional data mani...
research
08/03/2018

Ask, Acquire, and Attack: Data-free UAP Generation using Class Impressions

Deep learning models are susceptible to input specific noise, called adv...
research
06/20/2021

Generative Model Adversarial Training for Deep Compressed Sensing

Deep compressed sensing assumes the data has sparse representation in a ...

Please sign up or login with your details

Forgot password? Click here to reset