Adversarially Robust Learning with Tolerance
Abstract
We initiate the study of tolerant adversarial PAC-learning with respect to metric perturbation sets. In adversarial PAC-learning, an adversary is allowed to replace a test point with an arbitrary point in a closed ball of radius centered at . In the tolerant version, the error of the learner is compared with the best achievable error with respect to a slightly larger perturbation radius . This simple tweak helps us bridge the gap between theory and practice and obtain the first PAC-type guarantees for algorithmic techniques that are popular in practice.
Our first result concerns the widely-used “perturb-and-smooth” approach for adversarial learning. For perturbation sets with doubling dimension , we show that a variant of these approaches PAC-learns any hypothesis class with VC-dimension in the -tolerant adversarial setting with samples. This is in contrast to the traditional (non-tolerant) setting in which, as we show, the perturb-and-smooth approach can provably fail.
Our second result shows that one can PAC-learn the same class using samples even in the agnostic setting. This result is based on a novel compression-based algorithm, and achieves a linear dependence on the doubling dimension as well as the VC-dimension. This is in contrast to the non-tolerant setting where there is no known sample complexity upper bound that depends polynomially on the VC-dimension.
1 Introduction
Several empirical studies (Szegedy et al., 2014; Goodfellow et al., 2018) have demonstrated that models trained to have a low accuracy on a data set often have the undesirable property that a small perturbation to an input instance can change the label outputted by the model. For most domains this does not align with human perception and thus indicates that the learned models are not representing the ground truth despite obtaining good accuracy on test sets.
The theory of PAC-learning characterizes the conditions under which learning is possible. For binary classification, the following conditions are sufficient: a) unseen data should arrive from the same distribution as training data, and b) the class of models should have a low capacity (as measured, for example, by its VC-dimension). If these conditions are met, an Empirical Risk Minimizer (ERM) that simply optimizes model parameters to maximize accuracy on the training set learns successfully.
Recent work has studied test-time adversarial perturbations under the PAC-learning framework. If an adversary is allowed to perturb data during test time then the conditions above do not hold, and we cannot hope for the model to learn to be robust just by running ERM. Thus, the goal here is to bias the learning process towards finding models where label-changing perturbations are rare. This is achieved by defining a loss function that combines both classification error and the probability of seeing label-changing perturbations, and learning models that minimize this loss on unseen data. It has been shown that even though (robust) ERM can fail in this setting, PAC-learning is still possible as long as we know during training the kinds of perturbations we want to guard against at test time (Montasser et al., 2019). This result holds for all perturbation sets. However, the learning algorithm is significantly more complex than robust ERM and requires a large number of samples (with the best known sample complexity bounds potentially being exponential in the VC-dimension).
We study a tolerant version of the adversarially robust learning framework and restrict the perturbations to balls in a general metric space with finite doubling dimension. We show this slight shift in the learning objective yields significantly improved sample complexity bounds through a simpler learning paradigm than what was previously known. In fact, we show that a version of the common “perturb-and-smooth” paradigm successfully PAC-learns any class of bounded VC-dimension in this setting.
Learning in general metric spaces. What kinds of perturbations should a learning algorithm guard against? Any transformation of the input that we believe should not change its label could be a viable perturbation for the adversary to use. The early works in this area considered perturbations contained within a small -ball of the input. More recent work has considered other transformations such as a small rotation, or translation of an input image (Engstrom et al., 2019; Fawzi and Frossard, 2015; Kanbak et al., 2018; Xiao et al., 2018), or even adding small amounts of fog or snow (Kang et al., 2019). It has also been argued that small perturbations in some feature space should be allowed as opposed to the input space (Inkawhich et al., 2019; Sabour et al., 2016; Xu et al., 2020; Song et al., 2018; Hosseini and Poovendran, 2018). This motivates the study of more general perturbations.
We consider a setting where the input comes from a domain that is equipped with a distance metric and allows perturbations to be within a small metric ball around the input. Earlier work on general perturbation sets (for example, (Montasser et al., 2019)) considered arbitrary perturbations. In this setting one does not quantify the magnitude of a perturbation and thus cannot talk about small versus large perturbations. Modeling perturbations using a metric space enables us to do that while also keeping the setup general enough to be able to encode a large variety of perturbation sets by choosing appropriate distance functions.
Learning with tolerance. In practice, we often believe that small perturbations of the input should not change its label but we do not know precisely what small means. However, in the PAC-learning framework for adversarially robust classification, we are required to define a precise perturbation set and learn a model that has error arbitrarily close to the smallest error that can be achieved with respect to that perturbation set. In other words, we aim to be arbitrarily close to a target that was picked somewhat arbitrarily to begin with. Due to the uncertainty about the correct perturbation size, it is more meaningful to allow for a wider range of error values. To achieve this, we introduce the concept of tolerance. In the tolerant setting, in addition to specifying a perturbation size , we introduce a tolerance parameter that encodes our uncertainty about the size of allowed perturbations. Then, for any given , we aim to learn a model whose error with respect to perturbations of size is at most more than the smallest error achievable with respect to perturbations of size .
2 Our results
In this paper we formalize and initiate the study of the problem of adversarially robust learning in the tolerant setting for general metric spaces and provide two algorithms for the task. Both of our algorithms rely on: 1) modifying the training data by randomly sampling points from the perturbation sets around each data point, and 2) smoothing the output of the model by taking a majority over the labels returned by the model for nearby points.
Our first algorithm starts by modifying the training set by randomly perturbing each training point using a certain distribution (see Section 5 for details). It then trains a (non-robust) PAC-learner (such as ERM) on the perturbed training set to find a hypothesis . Finally, it outputs a smooth version of . The smoothing step replaces at each point with the a majority label outputted by on the points around . We show that for metric spaces of a fixed doubling dimension, this algorithm successfully learns in the tolerant setting assuming tolerant realizability.
Theorem 2.1 (Informal version of Theorem 5.2).
Let be a metric space with doubling dimension and a hypothesis class. Assuming tolerant realizability, can be learned tolerantly in the adversarially robust setting using samples, where encodes the amount of allowed tolerance, and is the desired accuracy.
An interesting feature of the above result is the linear dependence of the sample complexity with respect to . This is in contrast to the best known upper bound for non-tolerant adversarial setting (Montasser et al., 2019) which depends on the dual VC-dimension of the hypothesis class and in general is exponential in . Moreover, this is the first PAC-type guarantee for the general perturb-and-smooth paradigm, indicating that the tolerant adversarial learning is the “right” learning model for studying these approaches. While the above method enjoys simplicity and can be computationally efficient, one downside is that its sample complexity grows exponentially with the doubling dimension. For instance, such algorithm cannot be used on high-dimensional data in the Euclidean space. Another limitation is that the guarantee holds only in the (robustly) realizable setting.
In the second main part of our submission (Section 6) we show that, surprisingly, these limitations can be overcome by incorporating ideas from the tolerant framework and perturb-and-smooth algorithms into a novel compression scheme for robust learning. The resulting algorithm improves the dependence on the doubling dimension, and works in the general agnostic setting.
Theorem 2.2 (Informal version of Corollary 6.5).
Let be a metric space with doubling dimension and a hypothesis class. Then can be learned tolerantly in the adversarially robust setting using samples, where hides logarithmic factors, encodes the amount of allowed tolerance, and is the desired accuracy.
This algorithm exploits the connection between sample compression and adversarially robust learning Montasser et al. (2019). However, unlike Montasser et al. (2019), our new compression scheme sidesteps the dependence on the dual VC-dimension (refer to the discussion at the end of Section 6 for more details). As a result, we get an exponential improvement over the best known (nontolerant) sample complexity in terms of dependence on VC-dimension.
3 Related work
PAC-learning for adversarially robust classification has been studied extensively in recent years (Cullina et al., 2018; Awasthi et al., 2019; Montasser et al., 2019; Feige et al., 2015; Attias et al., 2019; Montasser et al., 2020a; Ashtiani et al., 2020). These works provide learning algorithms that guarantee low generalization error in the presence of adversarial perturbations in various settings. The most general result is due to Montasser et al. (2019), and is proved for general hypothesis classes and perturbation sets. All of the above results assume that the learner knows the kinds of perturbations allowed for the adversary. Some more recent papers have considered scenarios where the learner does not even need to know that. Goldwasser et al. (2020) allow the adversary to perturb test data in unrestricted ways and are still able to provide learning guarantees. The catch is that it only works in the transductive setting and only if the learner is allowed to abstain from making a prediction on some test points. Montasser et al. (2021a) consider the case where the learner needs to infer the set of allowed perturbations by observing the actions of the adversary.
Tolerance was introduced by Ashtiani et al. (2020) in the context of certification. They provide examples where certification is not possible unless we allow some tolerance. Montasser et al. (2021b) study transductive adversarial learning and provide a “tolerant” guarantee. Note that unlike our work, the main focus of that paper is on the transductive setting. Moreover, they do not specifically study tolerance with respect to metric perturbation sets. Without a metric, it is not meaningful to expand perturbation sets by a factor (as we do in the our definition of tolerance). Instead, they expand their perturbation sets by applying two perturbations in succession, which is akin to setting . In contrast, our results hold in the more common inductive setting, and capture a more realistic setting where can be any (small) real number larger than zero.
Subsequent to our work, Bhattacharjee et al. (2022) study adversarially robust learning with tolerance for “regular” VC-classes and show that a simple modification of robust ERM achieves a sample complexity polynomial in both VC-dimension and doubling dimension. In a similar vein, Raman et al. (2022) identify a more general property of hypothesis classes for which robust ERM is sufficient for adversarially robust learning with tolerance.
Like many recent adversarially robust learning algorithms (Feige et al., 2015; Attias et al., 2019), our first algorithm relies on calls to a non-robust PAC-learner. Montasser et al. (2020b) formalize the question of reducing adversarially robust learning to non-robust learning and study finite perturbation sets of size . They show a reduction that makes calls to the non-robust learner and also prove a lower bound of . It will be interesting to see if our algorithms can be used to obtain better bounds for the tolerant setting. Our first algorithm makes one call to the non-robust PAC-learner at training time, but needs to perform potentially expensive smoothing for making actual predictions (see Theorem 5.2).
A related line of work studies smallest achievable robust loss for various distributions and hypothesis classes. For example, Bubeck and Sellke (2021) show that hypothesis classes with low robust loss must be overparametrized. Yang et al. (2020b) explore real-world datasets and provide evidence that they are separable and therefore there must exist locally Lipschitz hypotheses with low robust loss. Note that the existence of such hypotheses does not immediately imply that PAC-learning is possible.
The techniques of randomly perturbing the training data and smoothing the output classifier has been extensively used in practice and has shown good empirical success. Augmenting the training data with some randomly perturbed samples was used for handwriting recognition as early as by Yaeger et al. (1996). More recently, “stability training” was introduced by Zheng et al. (2016) for state of the art image classifiers where training data is augmented with Gaussian perturbations. Empirical evidence was provided that the technique improved the accuracy against naturally occurring perturbations. Augmentations with non-Gaussian perturbations of a large variety were considered by Hendrycks et al. (2019).
Smoothing the output classifier using random samples around the test point is a popular technique for producing certifiably robust classifiers. A certification, in this context, is a guarantee that given a test point , all points within a certain radius of receive the same label as . Several papers have provided theoretical analyses to show that smoothing produces certifiably robust classifiers (Cao and Gong, 2017; Cohen et al., 2019; Lecuyer et al., 2019; Li et al., 2019; Liu et al., 2018; Salman et al., 2019; Levine and Feizi, 2020), whereas others have identified cases where smoothing does not work Yang et al. (2020a); Blum et al. (2020).
However, to the best of our knowledge, a PAC-type guarantee has not been shown for any algorithm that employs training data perturbations or output classifier smoothing, and our paper provides the first such analysis.
4 Notations and setup
We denote by the input domain and by the binary label space. We assume that is equipped with a metric . A hypothesis is a function that assigns a label to each point in the domain. A hypothesis class is a set of such hypotheses. For a sample , we use the notation to denote the collection of domain points occurring in . The binary (also called 0-1) loss of on data point is defined by
where is the indicator function. Let by a probability distribution over . Then the expected binary loss of with respect to is defined by
Similarly, the empirical binary loss of on sample is defined as . We also define the approximation error of with respect to as .
A learner is a function that takes in a finite sequence of labeled instances and outputs a hypothesis . The following definition abstracts the notion of PAC-learning Vapnik and Chervonenkis (1971); Valiant (1984).
Definition 4.1 (PAC-learner).
Let be a set of distributions over and a hypothesis class. We say PAC-learns with samples if the following holds: for every distribution over , and every , if is an i.i.d. sample of size at least from , then with probability at least (over the randomness of ) we have
is called an agnostic learner if is the set of all distributions on , and a realizable learner if .
The smallest function for which there exists a learner that satisfies the above definition with is referred to as the (realizable or agnostic) sample complexity of learning .
The existence of sample-efficient PAC-learners for VC classes is a standard result Vapnik and Chervonenkis (1971). We state the results formally in Appendix A.
4.1 Tolerant adversarial PAC-learning
Let be a function that maps each point in the domain to the set of its “admissible” perturbations. We call this function the perturbation type. The adversarial loss of with respect to on is defined by
The expected adversarial loss with respect to is defined by . The empirical adversarial loss of on sample is defined by . Finally, the adversarial approximation error of with respect to and is defined by .
The following definition generalizes the setting of PAC adversarial learning to what we call the tolerant setting, where we consider two perturbation types and . We say is contained in and and write it as if for all .
Definition 4.2 (Tolerant Adversarial PAC-learner).
Let be a set of distributions over , a hypothesis class, and two perturbation types. We say tolerantly PAC-learns with samples if the following holds: for every distribution and every , if is an i.i.d. sample of size at least from , then with probability at least (over the randomness of ) we have
We say is a tolerant PAC-learner in the agnostic setting if is the set of all distributions over , and in the tolerantly realizable setting if .
In the above context, we refer to as the actual perturbation type and to as the reference perturbation type. The case where for all corresponds to the usual adversarial learning scenario (with no tolerance).
4.2 Tolerant adversarial PAC-learning in metric spaces
If is equipped with a metric , then can be naturally defined by a ball of radius around , i.e., . To simplify the notation, we sometimes use instead of to denote the adversarial loss with respect to .
In the tolerant setting, we consider the perturbation sets and , where is called the tolerance parameter. Note that . We now define PAC-learning with respect to the metric space.
Definition 4.3 (Tolerant Adversarial Learning in metric spaces).
Let be a metric space, a hypothesis class, and a set of distributions of . We say is tolerantly PAC-learnable with samples when for every there exist a PAC-learner for that uses samples.
Remark 4.4.
In this definition the learner receives and as input but its sample complexity does not depend on (but can depend on ). Also, as in Definition 4.2, the tolerantly realizable setting corresponds to while in the agnostic setting is the set of all distributions over .
The doubling dimension and the doubling measure of the metric space will play important roles in our analysis. We refer the reader to Appendix B for their definitions.
We will use the following lemma in our analysis, whose proof can be found in Appendix B:
Lemma 4.5.
For any family of complete, doubling metric spaces, there exist constants such that for any metric space with doubling dimension , there exists a measure such that if a ball of radius is completely contained inside a ball of radius (with potentially a different center) for any , then . Furthermore, if we have a constant such that we know that then the bound can be simplified to , where depends on and .
Later, we will set where is the tolerance parameter. Since we are mostly interested in small values of , suppose we decide on some loose upper bound . This corresponds to saying that there exists some such that .
It is worth noting that in the special case of Euclidean metric spaces, we can set both and to be 1. In the rest of the paper, we will assume we have a loose upper bound and use the simpler bound from Lemma B.5 extensively.
Given a metric space and a measure defined over it, for any subset for which is non-zero and finite, induces a probability measure over as follows. For any set in the -algebra over , we define . With a slight abuse of notation, we write to mean whenever we know from the context.
Our learners rely on being able to sample from . Thus we define the following oracle, which can be implemented efficiently for spaces.
Definition 4.6 (Sampling Oracle).
Given a metric space equipped with a doubling measure , a sampling oracle is an algorithm that when queried with a such that is finite, returns a sample drawn from . We will use the notation for queries to this oracle.
5 The perturb-and-smooth approach for tolerant adversarial learning
In this section we focus on tolerant adversarial PAC-learning in metric spaces (Definition 4.3), and show that VC classes are tolerantly PAC-learnable in the tolerantly realizable setting. Interestingly, we prove this result using an approach that resembles the “perturb-and-smooth” paradigm which is used in practice (for example by Cohen et al. (2019)). The overall idea is to “perturb” each training point , train a classifier on the “perturbed” points, and “smooth out” the final hypothesis using a certain majority rule.
We employ three perturbation types: and play the role of the actual and the reference perturbation type respectively. Additionally, we consider a perturbation type , which is used for smoothing. We assume and . For this section, we will use metric balls for the three types. Specifically, if consists of balls of radius for some , then will consists of balls of radius and will consist of balls of radius .
Definition 5.1 (Smoothed classifier).
For a hypothesis , and perturbation type , we let denote the classifier resulting from replacing the label with the average label over , that is
For metric perturbation types, where is a ball of some radius , we also use the notation and when the type is clear from context, we may omit the subscript altogether and simply write for the smoothed classifier.
The tolerant perturb-and-smooth algorithm
We propose the following learning algorithm, TPaS, for tolerant learning in metric spaces. Let the perturbation radius be for the actual type , and let be the training sample. For each , the learner samples a point (using the sampling oracle) from the expanded reference perturbation set . Let . TPaS then invokes a (standard, non-robust) PAC-learner for the hypothesis class on the perturbed data . We let denote the output of this PAC-learner. Finally, TPaS outputs the -smoothed version of for . That is, is simply the majority label in a ball of radius around with respect to the distribution defined by , see also Definition 5.1. We will prove below that this has a small -adversarial loss. Algorithm 1 below summarizes our learning procedure.
The following is the main result of this section.
Theorem 5.2.
Let be an any metric space with doubling dimension and doubling measure . Let be a sampling oracle for . Let be a hypothesis class and a set of distributions over . Assume PAC-learns with samples in the realizable setting. Then there exists a learner , namely TPaS, that
-
•
Tolerantly PAC-learns in the tolerantly realizable setting with sample complexity bounded by , where is the tolerance parameter and is the doubling dimension.
-
•
Makes only one query to
-
•
Makes queries to sampling oracle
The proof of this theorem uses the following key technical lemma (its proof can be found in Appendix C):
Lemma 5.3.
Let be a perturbation radius, a tolerance parameter, and a classifier. For and , we define
Then implies that for all .
Proof of Theorem 5.2.
Consider some and to be given (we will pick a suitable value of later), and assume the PAC-learner was invoked on the perturbed sample of size at least . According to definition 4.1, this implies that with probability , the output has (binary) loss at most with respect to the data-generating distribution. Note that the relevant distribution here is the two-stage process of the original data generating distribution and the perturbation sampling according to . Since is -robustly realizable, the two-stage process yields a realizable distribution with respect to the standard -loss. Thus, we have
With Lemma 5.3, this becomes . For , Markov’s inequality then yields :
(1) |
Thus setting and plugging in the result of the Lemma 5.3 to equation (1), we get
Since implies that using the definition of the smoothed classifier we get
which implies . Thus, for the robust learning problem, if we are given a desired accuracy and we want , we can pick . Putting it all together, we get sample complexity where , and . Therefore, . ∎
Computational complexity of the learner. Assuming we have access to and an efficient algorithm for non-robust PAC-learning in the realizable setting, we can compute efficiently. Therefore, the learning can be done efficiently in this case. However, at the prediction time, we need to compute on new test points which requires us to compute an expectation. We can instead estimate the expectations using random samples from the sampling oracle. For a single test point , if the number of samples we draw is then with probability at least we get the same result as that of the optimal . Using more samples we can boost this probability to guarantee a similar output to that of on a larger set of test points.
The traditional non-tolerant framework does not justify the use of perturb-and-smooth-type approaches. The introduction of the tolerance in the adversarial learning framework is crucial for being able to prove guarantees for perturb-and-smooth-type algorithms. To see why, consider a simple case where the domain is the real line, the perturbation set is an open Euclidean ball of radius 1, and the hypothesis class is the set of all thresholds. Assume that the underlying distribution is supported only on two points: . This distribution is robustly realizable, but the threshold should be set exactly to to get a small error. However, the perturb-and-smooth method will fail because the only way the PAC-learner sets the threshold to is if it receives a (perturbed) sample exactly at , whose probability is 0.
6 Improved tolerant learning guarantees through sample compression
The perturb-and-smooth approach discussed in the previous section offers a general method for tolerant robust learning. However, one shortcoming of this approach is the exponential dependence of its sample complexity with respect to the doubling dimension of the metric space. Furthermore, the tolerant robust guarantee relied on the data generating distribution being tolerantly realizable. In this section, we propose another approach that addresses both of these issues. The idea is to adopt the perturb-and-smooth approach within a sample compression argument. We introduce the notion of a -tolerant sample compression scheme and present a learning bound based on such a compression scheme, starting with the realizable case. We then show that this implies learnability in the agnostic case as well. Remarkably, this tolerant compression based analysis will yield bounds on the sample complexity that avoid the exponential dependence on the doubling dimension.
For a compact representation, we will use the general notation and for the three perturbation types (actual, reference and smoothing type) in this section and will assume that they satisfy the Property 1 below for some parameter . Lemma 5.3 implies that, in the metric setting, for any radius and tolerance parameter the perturbation types , , and have this property for .
Property 1.
For a fixed , we assume that the perturbation types and are so that for any classifier and any , any if
then -smoothed class classifier satisfies for all .
A compression scheme of size is a pair of functions , where the compression function maps samples of arbitrary size to sub-samples of of size at most , and is a decompression function that maps samples to classifiers. The pair is a sample compression scheme for loss and class , if for any samples realizable by , we recover the correct labels for all , that is, implies that .
For tolerant learning, we introduce the following generalization of compression schemes:
Definition 6.1 (Tolerant sample compression scheme).
A sample compression scheme is a -tolerant sample compression scheme for class , if for any samples that are realizable by , that is , we have .
The next lemma establishes that the existence of a sufficiently small tolerant compression scheme for the class yields bounds on the sample complexity of tolerantly learning . The proof of the lemma is based on a modification of a standard compression based generalization bound. Appendix Section D provides more details.
Lemma 6.2.
Let be a hypothesis class and and be perturbation types with included in . If the class admits a -tolerant compression scheme of size bounded by for sample of size , then the class is -tolerantly learnable in the realizable case with sample complexity bounded by .
We next establish a bound on the tolerant compression size for general VC-classes, which will then immediately yield the improved sample complexity bounds for tolerant learning in the realizable case. The proof is sketched here; its full version has been moved to the Appendix.
Lemma 6.3.
Let be some hypothesis class with finite VC-dimension , and let satisfy the conditions in Property 1 for some . Then there exists a -tolerant sample compression scheme for of size .
Proof Sketch.
We will employ a boosting-based approach to establish the claimed compression sizes. Let be a data-set that is -realizable with respect to . We let denote an “inflated data-set” that contains all domain points in the -perturbation sets of the , that is . Every point is assigned the label of the minimally-indexed with , and we set to be the resulting collection of labeled data-points.
We then use the boost-by-majority method to encode a classifier that (roughly speaking) has error bounded by over (a suitable measure over) . This boosting method outputs a -majority vote over weak learners , which in our case will be hypotheses from . We prove that this error can be achieved with rounds of boosting. We prove that each weak learner that is used in the boosting procedure can be encoded with many sample points from . The resulting compression size is thus .
Finally, the error bound of over implies that the error in each perturbation set of a sample point is at most . Property 1 then implies for the -smoothed classifier , establishing the -tolerant correctness of the compression scheme. ∎
This yields the following result
Theorem 6.4.
Let be a hypothesis class of finite VC-dimension and be three perturbation types (actual, reference and smoothing) satisfying Property 1 for some . Then the sample complexity (omitting log-factors) of -tolerantly learning is bounded by
in the realizable case, and in the agnostic case by
Proof.
The bound for the realizable case follows immediately from Lemma 6.3 and the subsequent discussion (in the Appendix). For the agnostic case, we employ a reduction from agnostic robust learnabilty to realizable robust learnability (Montasser et al., 2019; Moran and Yehudayoff, 2016). The reduction is analogous to the one presented in Appendix C of Montasser et al. (2019) for usual (non-tolerant) robust learnablity with some minor modifications. Namely, for a sample , we choose the largest subsample that is -realizable (this will result in competitiveness with a -optimal classifier), and we will use the boosting procedure described there for the loss. For the sample sizes employed for the weak learners in that procedure, we can use the sample complexity for of an optimal -tolerant learner in the realizable case (note that each learning problem during the boosting procedure is a realizable -tolerant learning task). These modifications result in the stated sample complexity for agnostic tolerant learnability. ∎
In particular, for the doubling measure scenario (as considered in the previous section), we obtain
Corollary 6.5.
For metric tolerant learning with tolerance parameter in doubling dimension the sample complexity of adversarially robust learning with tolerance in the realizable case is bounded by and in the agnostic case by .
Discussion of linear dependence on
Earlier, general compression based sample complexity bounds for robust learning with arbitrary perturbation sets exhibit a dependence on the dual VC-dimension of the hypothesis class and therefore potentially an exponential dependence on (Montasser et al., 2019). In our setting, we show that it is possible to avoid the dependence on dual-VC by exploiting both the metric structure of the domain set and the tolerant framework. In the full proof of Lemma 6.3, we show that if we can encode a classifier with small error (exponentially small with respect to the doubling dimension for the metric case) on the perturbed distribution w.r.t. larger perturbation sets, then we can use smoothing to get a classifier that correctly classifies every point in the inner inflated sets. And, as for TPaS, the tolerant perspective is crucial to exploit a smoothing step in the compression approach (through the guarantee from Property 1 or Lemma 5.3).
More specifically, we define a tolerant compression scheme (Definition 12) that naturally extends the classic definition of compression to the tolerant framework. The compression scheme we establish in the proof of Lemma 6.3 then borrows ideas from our perturb-and-smooth algorithm. Within the compression argument, we define the perturbed distribution over the sample that we want to compress with respect to the larger perturbation sets. We then use boosting to build a classifier with very small error with respect to this distribution. The nice property of boosting is that its error decreases exponentially with the number of iterations. As a result, we also get linear dependence on the doubling dimension. This classifier can be encoded using samples ( rounds of boosting, and each weak classifier can be encoded using samples, since we can here use simple -approximations rather than invoking VC-theory in the dual space). Our decoder receives the description of these weak classifiers, combines them, and performs a final smoothing step. The smoothing step translates the exponentially small error with respect to the perturbed distribution to zero error with respect to the (inner) inflated set, thereby satisfying the requirement of a tolerant compression scheme.
References
- Ashtiani et al. (2020) Hassan Ashtiani, Vinayak Pathak, and Ruth Urner. Black-box certification and learning under adversarial perturbations. In International Conference on Machine Learning, pages 388–398. PMLR, 2020.
- Attias et al. (2019) Idan Attias, Aryeh Kontorovich, and Yishay Mansour. Improved generalization bounds for robust learning. In Algorithmic Learning Theory, ALT, pages 162–183, 2019.
- Awasthi et al. (2019) Pranjal Awasthi, Abhratanu Dutta, and Aravindan Vijayaraghavan. On robustness to adversarial examples and polynomial optimization. In Advances in Neural Information Processing Systems, NeurIPS, pages 13760–13770, 2019.
- Bhattacharjee et al. (2022) Robi Bhattacharjee, Max Hopkins, Akash Kumar, Hantao Yu, and Kamalika Chaudhuri. Robust empirical risk minimization with tolerance. arXiv preprint arXiv:2210.00635, 2022.
- Blum et al. (2020) Avrim Blum, Travis Dick, Naren Manoj, and Hongyang Zhang. Random smoothing might be unable to certify l_infinity robustness for high-dimensional images. Journal of machine learning research, 21(211), 2020.
- Blumer et al. (1989) Anselm Blumer, Andrzej Ehrenfeucht, David Haussler, and Manfred K Warmuth. Learnability and the vapnik-chervonenkis dimension. Journal of the ACM (JACM), 36(4):929–965, 1989.
- Bubeck and Sellke (2021) Sébastien Bubeck and Mark Sellke. A universal law of robustness via isoperimetry. Advances in Neural Information Processing Systems, 34:28811–28822, 2021.
- Cao and Gong (2017) Xiaoyu Cao and Neil Zhenqiang Gong. Mitigating evasion attacks to deep neural networks via region-based classification. In Proceedings of the 33rd Annual Computer Security Applications Conference, pages 278–287, 2017.
- Cohen et al. (2019) Jeremy M. Cohen, Elan Rosenfeld, and J. Zico Kolter. Certified adversarial robustness via randomized smoothing. In Proceedings of the 36th International Conference on Machine Learning, ICML, pages 1310–1320, 2019.
- Cullina et al. (2018) Daniel Cullina, Arjun Nitin Bhagoji, and Prateek Mittal. Pac-learning in the presence of adversaries. In Advances in Neural Information Processing Systems, NeurIPS, pages 230–241, 2018.
- Engstrom et al. (2019) Logan Engstrom, Brandon Tran, Dimitris Tsipras, Ludwig Schmidt, and Aleksander Madry. Exploring the landscape of spatial robustness. In International Conference on Machine Learning, pages 1802–1811. PMLR, 2019.
- Fawzi and Frossard (2015) Alhussein Fawzi and Pascal Frossard. Manitest: Are classifiers really invariant? In British Machine Vision Conference (BMVC), number CONF, 2015.
- Feige et al. (2015) Uriel Feige, Yishay Mansour, and Robert Schapire. Learning and inference in the presence of corrupted inputs. In Conference on Learning Theory, COLT, pages 637–657, 2015.
- Goldwasser et al. (2020) Shafi Goldwasser, Adam Tauman Kalai, Yael Tauman Kalai, and Omar Montasser. Beyond perturbations: Learning guarantees with arbitrary adversarial test examples. arXiv preprint arXiv:2007.05145, 2020.
- Goodfellow et al. (2018) Ian J. Goodfellow, Patrick D. McDaniel, and Nicolas Papernot. Making machine learning robust against adversarial inputs. Commun. ACM, 61(7):56–66, 2018.
- Hanneke (2016) Steve Hanneke. The optimal sample complexity of pac learning. The Journal of Machine Learning Research, 17(1):1319–1333, 2016.
- Haussler (1992) David Haussler. Decision theoretic generalizations of the pac model for neural net and other learning applications. Information and computation, 100(1):78–150, 1992.
- Haussler and Welzl (1987) David Haussler and Emo Welzl. epsilon-nets and simplex range queries. Discret. Comput. Geom., 2:127–151, 1987.
- Hendrycks et al. (2019) Dan Hendrycks, Norman Mu, Ekin Dogus Cubuk, Barret Zoph, Justin Gilmer, and Balaji Lakshminarayanan. Augmix: A simple data processing method to improve robustness and uncertainty. In International Conference on Learning Representations, 2019.
- Hosseini and Poovendran (2018) Hossein Hosseini and Radha Poovendran. Semantic adversarial examples. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pages 1614–1619, 2018.
- Inkawhich et al. (2019) Nathan Inkawhich, Wei Wen, Hai Helen Li, and Yiran Chen. Feature space perturbations yield more transferable adversarial examples. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 7066–7074, 2019.
- Kanbak et al. (2018) Can Kanbak, Seyed-Mohsen Moosavi-Dezfooli, and Pascal Frossard. Geometric robustness of deep networks: Analysis and improvement. In 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 4441–4449. IEEE, 2018.
- Kang et al. (2019) Daniel Kang, Yi Sun, Dan Hendrycks, Tom Brown, and Jacob Steinhardt. Testing robustness against unforeseen adversaries. arXiv preprint arXiv:1908.08016, 2019.
- Lecuyer et al. (2019) Mathias Lecuyer, Vaggelis Atlidakis, Roxana Geambasu, Daniel Hsu, and Suman Jana. Certified robustness to adversarial examples with differential privacy. In 2019 IEEE Symposium on Security and Privacy (SP), pages 656–672. IEEE, 2019.
- Levine and Feizi (2020) Alexander Levine and Soheil Feizi. Robustness certificates for sparse adversarial attacks by randomized ablation. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pages 4585–4593, 2020.
- Li et al. (2019) Bai Li, Changyou Chen, Wenlin Wang, and Lawrence Carin. Certified adversarial robustness with additive noise. Advances in Neural Information Processing Systems, 32:9464–9474, 2019.
- Liu et al. (2018) Xuanqing Liu, Minhao Cheng, Huan Zhang, and Cho-Jui Hsieh. Towards robust neural networks via random self-ensemble. In Proceedings of the European Conference on Computer Vision (ECCV), pages 369–385, 2018.
- Luukkainen and Saksman (1998) Jouni Luukkainen and Eero Saksman. Every complete doubling metric space carries a doubling measure. Proceedings of the American Mathematical Society, 126(2):531–534, 1998.
- Montasser et al. (2019) Omar Montasser, Steve Hanneke, and Nathan Srebro. VC classes are adversarially robustly learnable, but only improperly. In Conference on Learning Theory, COLT, pages 2512–2530, 2019.
- Montasser et al. (2020a) Omar Montasser, Surbhi Goel, Ilias Diakonikolas, and Nathan Srebro. Efficiently learning adversarially robust halfspaces with noise. arXiv preprint arXiv:2005.07652, 2020a.
- Montasser et al. (2020b) Omar Montasser, Steve Hanneke, and Nati Srebro. Reducing adversarially robust learning to non-robust pac learning. In NeurIPS, 2020b.
- Montasser et al. (2021a) Omar Montasser, Steve Hanneke, and Nathan Srebro. Adversarially robust learning with unknown perturbation sets. arXiv preprint arXiv:2102.02145, 2021a.
- Montasser et al. (2021b) Omar Montasser, Steve Hanneke, and Nathan Srebro. Transductive robust learning guarantees. arXiv preprint arXiv:2110.10602, 2021b.
- Moran and Yehudayoff (2016) Shay Moran and Amir Yehudayoff. Sample compression schemes for vc classes. Journal of the ACM (JACM), 63(3):1–10, 2016.
- Raman et al. (2022) VInod Raman, Unique Subedi, and Ambuj Tewari. Probabilistically robust pac learning. arXiv preprint arXiv:2211.05656, 2022.
- Sabour et al. (2016) Sara Sabour, Yanshuai Cao, Fartash Faghri, and David J Fleet. Adversarial manipulation of deep representations. In ICLR (Poster), 2016.
- Salman et al. (2019) Hadi Salman, Jerry Li, Ilya P. Razenshteyn, Pengchuan Zhang, Huan Zhang, Sébastien Bubeck, and Greg Yang. Provably robust deep learning via adversarially trained smoothed classifiers. In Advances in Neural Information Processing Systems 32, NeurIPS, pages 11289–11300, 2019.
- Schapire and Freund (2013) Robert E Schapire and Yoav Freund. Boosting: Foundations and algorithms. Kybernetes, 2013.
- Shalev-Shwartz and Ben-David (2014) Shai Shalev-Shwartz and Shai Ben-David. Understanding Machine Learning: From Theory to Algorithms. Cambridge University Press, 2014.
- Simon (2015) Hans U Simon. An almost optimal pac algorithm. In Conference on Learning Theory, pages 1552–1563. PMLR, 2015.
- Song et al. (2018) Yang Song, Rui Shu, Nate Kushman, and Stefano Ermon. Constructing unrestricted adversarial examples with generative models. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, pages 8322–8333, 2018.
- Szegedy et al. (2014) Christian Szegedy, Wojciech Zaremba, Ilya Sutskever, Joan Bruna, Dumitru Erhan, Ian J. Goodfellow, and Rob Fergus. Intriguing properties of neural networks. In 2nd International Conference on Learning Representations, ICLR, 2014.
- Valiant (1984) Leslie G. Valiant. A theory of the learnable. Commun. ACM, 27(11):1134–1142, 1984.
- Vapnik and Chervonenkis (1971) V. N. Vapnik and A. Ya. Chervonenkis. On the uniform convergence of relative frequencies of events to their probabilities. Theory of Probability & Its Applications, 16(2):264–280, 1971.
- Xiao et al. (2018) Chaowei Xiao, Jun-Yan Zhu, Bo Li, Warren He, Mingyan Liu, and Dawn Song. Spatially transformed adversarial examples. In International Conference on Learning Representations, 2018.
- Xu et al. (2020) Qiuling Xu, Guanhong Tao, Siyuan Cheng, and Xiangyu Zhang. Towards feature space adversarial attack. arXiv preprint arXiv:2004.12385, 2020.
- Yaeger et al. (1996) Larry Yaeger, Richard Lyon, and Brandyn Webb. Effective training of a neural network character classifier for word recognition. Advances in neural information processing systems, 9:807–816, 1996.
- Yang et al. (2020a) Greg Yang, Tony Duan, J Edward Hu, Hadi Salman, Ilya Razenshteyn, and Jerry Li. Randomized smoothing of all shapes and sizes. In Proceedings of the 37th International Conference on Machine Learning, pages 693–10705, 2020a.
- Yang et al. (2020b) Yao-Yuan Yang, Cyrus Rashtchian, Hongyang Zhang, Ruslan Salakhutdinov, and Kamalika Chaudhuri. A closer look at accuracy vs. robustness. In Proceedings of the 34th International Conference on Neural Information Processing Systems, pages 8588–8601, 2020b.
- Zheng et al. (2016) Stephan Zheng, Yang Song, Thomas Leung, and Ian Goodfellow. Improving the robustness of deep neural networks via stability training. In Proceedings of the ieee conference on computer vision and pattern recognition, pages 4480–4488, 2016.
Appendix A Standard results from VC theory
Let be a domain. For hypothesis and let .
Definition A.1 (VC-dimension).
We say shatters if . The VC-dimension of , denoted by , is defined to be the supremum of the size of the sets that are shattered by .
Theorem A.2 (Existence of Realizable PAC-learners Hanneke (2016); Simon (2015); Blumer et al. (1989)).
Let be a hypothesis class with bounded VC-dimension. Then is PAC-learnable in the realizable setting using samples.
Theorem A.3 (Existence of Agnostic PAC-learners Haussler (1992)).
Let be a hypothesis class with bounded VC-dimension. Then is PAC-learnable in the agnostic setting using samples.
Appendix B Metric spaces
Definition B.1.
A metric space is called a doubling metric if there exists a constant such that every ball of radius in it can be covered by at most balls of radius . The quantity is called the doubling dimension.
Definition B.2.
For a metric space , a measure defined on is called a doubling measure if there exists a constant , such that for all and , we have that . In this case, is called -doubling.
It can be shown (Luukkainen and Saksman, 1998) that every complete metric with doubling dimension has a -doubling measure for some where is a universal constant. For example, Euclidean spaces with an distance metric are complete and the Lesbesgue measure is a doubling measure.
The following lemmas follow straightforwardly from the definitions doubling metric and measures.
Lemma B.3.
Let be a doubling metric equipped with a -doubling measure . Then for all , , and , we have that
Proof.
Since is a measure, if such that , then . Let . It’s clear that . Therefore . Expanding by a factor of two times, we get , which means . But since , we get the desired result. ∎
Lemma B.4.
Let be a doubling metric equipped with a -doubling measure . Let , , and be such that . Then .
Proof.
By Lemma B.3, all we need to show is that . Indeed, let be any point. Then, from triangle inequality, we have that
Moreover, since , we have that . Substituting into the equation above, we get , which means . ∎
Finally, we also get:
Lemma B.5.
For any family of complete, doubling metric spaces, there exist constants such that for any metric space with doubling dimension , there exists a measure such that if a ball of radius is completely contained inside a ball of radius (with potentially a different center) for any , then .
Proof.
Corollary B.6.
Suppose we have a constant such that we know that . Then the bound in Lemma B.5 can be further simplified to , where depends on and . Furthermore, if then we can set .
Proof.
for . If , then for all . ∎
Appendix C Proof of Lemma 5.3
Let . Then, we have that . Further, for all , we have .
Let . Since this implies that , the worst case happens when . Therefore,
(3) | ||||
where the last inequality is implied by Lemma B.5. Thus, implies that as claimed.
Appendix D Compression based bounds
D.1 Proof of Lemma 6.2
To prove the generalization bound for tolerant learning, we employ the following lemma that establishes generalization for compression schemes for adversarial losses:
Lemma D.1 (Lemma 11, (Montasser et al., 2019)).
For any and fixed function , for any distribution over and any , with probability at least over an i.i.d. sample : if there exist indices such that
then the robust loss of the decompression with respect to is bounded by
The above lemma implies that if is a compression scheme that compresses data-sets of size to at most data points, for class and robust loss , then the sample complexity (omitting logarithmic factors) of robustly learning in the realizable case is bounded by
For the tolerant setting, since every sample that is realizable with respect to is also realizable with respect to , if a -tolerant compression scheme compresses to at most data-points and decompresses all -realizable samples to functions that have -loss on , then the lemma implies the above bound for the -tolerant sample complexity of learning .
D.2 Proof of Lemma 6.3
The proof of this Lemma employs the notions of a sample being -net or and -approximation for a hypothesis class . A labeled data set is an -net for class with respect to distribution over if for every hypothesis with , there exists an index and with . is an -approximation for class with respect to distribution over if for every hypothesis we have . Standard VC-theory tells us that, for classes with bounded VC-dimension, sufficiently large samples from are -nets or -approximations with high probability (Haussler and Welzl, 1987).
Proof.
We will employ a boosting-based approach to establish the claimed compression sizes. Let be a data-set that is -realizable with respect to . We let denote an “inflated data-set” that contains all domain points in the perturbation sets of the , that is
Every point is assigned the label of the minimally-indexed with , and we set to be the resulting collection of labeled data-points. (Note that since the sample is assumed to be -realizable, assigning it the label of some other corresponding data point in case for , would not induce any inconsistencies). Now let be the probability measure over defined by first sampling an index uniformly from and then sampling a domain point from the -perturbation set around the -th sample point in . Note that this implies that if for some subset , then
(4) |
for all .
We will now show that, by means of a compression scheme, we can encode a hypothesis with binary loss
(5) |
Property 1 together with Equation 4 then implies that the resulting -smoothed function has -robust loss on the sample , . Since the smoothing is a deterministic operation once is fixed, this implies the existence of a -tolerant compression scheme.
Standard VC-theory tells us that, for a class of bounded VC-dimension, for any distribution over , and any , with probability at least an i.i.d. sample of size is an -approximation for the class (Haussler and Welzl, 1987). This implies in particular, that there exists a finite subset of size at most (for some constant ) with the property that any classifier with empirical (binary) loss at most on has loss with respect to the distribution . We will choose such a set for the class of -majority votes over for . That is
The VC-dimension of is bounded by (Shalev-Shwartz and Ben-David, 2014)
We will now show how to obtain the classifier by means of a boosting approach on the finite data-set . More specifically, we will use the boost-by-majority method. This method outputs a -majority vote over weak learners , which in our case will be hypotheses from . After iterations with -weak learners, the empirical loss over the sample is bounded by (see Section 13.1 in (Schapire and Freund, 2013)). Thus, with , and , we obtain
which, by the choice of implies
which is what we needed to show according to Equation 5.
It remains to argue that the weak learners to be employed in the boosting procedure can be encoded by a small number of sample points from the original sample . For this part, we will employ a technique introduced earlier for robust compression (Montasser et al., 2019). Recall that the set is -robustly realizable, which implies that the set is (binary loss-) realizable by . By standard VC-theory, for every distribution over , there exists an -net of size (Haussler and Welzl, 1987). Thus, for every distribution over (that may occur during the boosting procedure), there exists a subsample of , of size at most with the property that every hypothesis from that is consistent with has binary loss at most with respect to (thus can serve as a weak learner for margin in the above procedure). Now for every labeled point , there is a sample point in the original sample such that and . Let be the collection of these corresponding original sample points. Note that any hypothesis that is -robustly consistent with is consistent with . Therefore we can use the original data-points in to encode the weak learner (for the decoding any -robust ERM hypothesis can be chosen to obtain ).
To summarize, we will compress the sample to the sequence of sample points from . To decode, we obtain the function as a majority vote over the weak learner and proceed to obtain the -smoothed function . This function satisfies and by this we have established the existence of a -tolerant compression scheme of size as claimed. ∎