Generating Adversarial Attacks in the Latent Space

04/10/2023
by   Nitish Shukla, et al.
0

Adversarial attacks in the input (pixel) space typically incorporate noise margins such as L_1 or L_∞-norm to produce imperceptibly perturbed data that confound deep learning networks. Such noise margins confine the magnitude of permissible noise. In this work, we propose injecting adversarial perturbations in the latent (feature) space using a generative adversarial network, removing the need for margin-based priors. Experiments on MNIST, CIFAR10, Fashion-MNIST, CIFAR100 and Stanford Dogs datasets support the effectiveness of the proposed method in generating adversarial attacks in the latent space while ensuring a high degree of visual realism with respect to pixel-based adversarial attack methods.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset