Improved robustness to adversarial examples using Lipschitz regularization of the loss

10/01/2018
by   Chris Finlay, et al.
0

Adversarial training is an effective method for improving robustness to adversarial attacks. We show that adversarial training using the Fast Signed Gradient Method can be interpreted as a form of regularization. We implemented a more effective form of adversarial training, which in turn can be interpreted as regularization of the loss in the 2-norm, ∇_x ℓ(x)_2. We obtained further improvements to adversarial robustness, as well as provable robustness guarantees, by augmenting adversarial training with Lipschitz regularization.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset