Second-Order Adversarial Attack and Certifiable Robustness

09/10/2018
by   Bai Li, et al.
0

We propose a powerful second-order attack method that outperforms existing attack methods on reducing the accuracy of state-of-the-art defense models based on adversarial training. The effectiveness of our attack method motivates an investigation of provable robustness of a defense model. To this end, we introduce a framework that allows one to obtain a certifiable lower bound on the prediction accuracy against adversarial examples. We conduct experiments to show the effectiveness of our attack method. At the same time, our defense models obtain higher accuracies compared to previous works under our proposed attack.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset