Adversarial Risk and Robustness: General Definitions and Implications for the Uniform Distribution

10/29/2018
by   Dimitrios I. Diochnos, et al.
0

We study adversarial perturbations when the instances are uniformly distributed over {0,1}^n. We study both "inherent" bounds that apply to any problem and any classifier for such a problem as well as bounds that apply to specific problems and specific hypothesis classes. As the current literature contains multiple definitions of adversarial risk and robustness, we start by giving a taxonomy for these definitions based on their goals, we identify one of them as the one guaranteeing misclassification by pushing the instances to the error region. We then study some classic algorithms for learning monotone conjunctions and compare their adversarial risk and robustness under different definitions by attacking the hypotheses using instances drawn from the uniform distribution. We observe that sometimes these definitions lead to significantly different bounds. Thus, this study advocates for the use of the error-region definition, even though other definitions, in other contexts, may coincide with the error-region definition. Using the error-region definition of adversarial perturbations, we then study inherent bounds on risk and robustness of any classifier for any classification problem whose instances are uniformly distributed over {0,1}^n. Using the isoperimetric inequality for the Boolean hypercube, we show that for initial error 0.01, there always exists an adversarial perturbation that changes O(√(n)) bits of the instances to increase the risk to 0.5, making classifier's decisions meaningless. Furthermore, by also using the central limit theorem we show that when n→∞, at most c ·√(n) bits of perturbations, for a universal constant c< 1.17, suffice for increasing the risk to 0.5, and the same c ·√(n) bits of perturbations on average suffice to increase the risk to 1, hence bounding the robustness by c ·√(n).

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/23/2018

Adversarial vulnerability for any classifier

Despite achieving impressive and often superhuman performance on multipl...
research
09/18/2023

Evaluating Adversarial Robustness with Expected Viable Performance

We introduce a metric for evaluating the robustness of a classifier, wit...
research
06/03/2019

Adversarial Risk Bounds for Neural Networks through Sparsity based Compression

Neural networks have been shown to be vulnerable against minor adversari...
research
04/06/2021

Taming Adversarial Robustness via Abstaining

In this work, we consider a binary classification problem and cast it in...
research
04/20/2023

Certified Adversarial Robustness Within Multiple Perturbation Bounds

Randomized smoothing (RS) is a well known certified defense against adve...
research
01/08/2018

Why informatics and general science need a conjoint basic definition of information

First the basic definition of information as a selection from a set of p...
research
06/23/2018

Defending Malware Classification Networks Against Adversarial Perturbations with Non-Negative Weight Restrictions

There is a growing body of literature showing that deep neural networks ...

Please sign up or login with your details

Forgot password? Click here to reset