This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Defending SVMs Against Poisoning Attacks: The Hardness and DBSCAN Approach

Hu Ding Fan Yang Jiawei Huang
Abstract

Adversarial machine learning has attracted a great amount of attention in recent years. Due to the great importance of support vector machines (SVM) in machine learning, we consider defending SVM against poisoning attacks in this paper. We study two commonly used strategies for defending: designing robust SVM algorithms and data sanitization. Though several robust SVM algorithms have been proposed before, most of them either are in lack of adversarial-resilience, or rely on strong assumptions about the data distribution or the attacker’s behavior. Moreover, the research on the hardness of designing a quality-guaranteed adversarially-resilient SVM algorithm is still quite limited. We are the first, to the best of our knowledge, to prove that even the simplest hard-margin one-class SVM with adversarial outliers problem is NP-complete, and has no fully PTAS unless P=NP. For data sanitization, we explain the effectiveness of DBSCAN (as a density-based outlier removal method) for defending against poisoning attacks. In particular, we link it to the intrinsic dimensionality by proving a sampling theorem in doubling metrics. In our empirical experiments, we systematically compare several defenses including the DBSCAN and robust SVM methods, and investigate the influences from the intrinsic dimensionality and poisoned fraction to their performances.

1 Introduction

In the past decades we have witnessed enormous progress in machine learning. One driving force behind this is the successful applications of machine learning technologies to many different fields, such as data mining, networking, and bioinformatics. However, with its territory rapidly enlarging, machine learning has also imposed a number of new challenges. In particular, adversarial machine learning which concerns about the potential vulnerabilities of the algorithms, has attracted a great amount of attention [Barreno et al., 2006, Huang et al., 2011, Biggio and Roli, 2018, Goodfellow et al., 2018]. As mentioned in the survey paper [Biggio and Roli, 2018], the very first work of adversarial machine learning dates back to 2004, in which Dalvi et al. [2004] formulated the adversarial classification problem as a game between the classifier and the adversary. In general, the adversarial attacks against machine learning can be categorized to evasion attacks and poisoning attacks [Biggio and Roli, 2018]. An evasion attack happens at test time, where the adversary aims to evade the trained classifier by manipulating test examples. For example, Szegedy et al. [2014] observed that small perturbation to a test image can arbitrarily change the neural network’s prediction.

In this paper, we focus on poisoning attacks that happen at training time. Usually, the adversary injects a small number of specially crafted samples into the training data which can make the decision boundary severely deviate and cause unexpected misclassification. In particular, because open datasets are commonly used to train our machine learning algorithms nowadays, poisoning attack has become a key security issue that seriously limits real-world applications [Biggio and Roli, 2018]. For instance, even a small number of poisoning samples can significantly increase the test error of support vector machine (SVM) [Biggio et al., 2012, Mei and Zhu, 2015, Xiao et al., 2012]. Beyond linear classifiers, a number of works studied the poisoning attacks for other machine learning problems, such as clustering [Biggio et al., 2014], PCA [Rubinstein et al., 2009], and regression [Jagielski et al., 2018].

Though lots of works focused on constructing poisoning attacks, our ultimate goal is to design defenses. Poisoning samples can be regarded as outliers, and this leads to two natural approaches to defend: (1) data sanitization defense, i.e., first perform outlier removal and then run an existing machine learning algorithm on the cleaned data [Cretu et al., 2008], or (2) directly design a robust optimization algorithm that is resilient against outliers [Christmann and Steinwart, 2004, Jagielski et al., 2018].

Steinhardt et al. [2017] studied two basic methods of data sanitization defense, which remove the points outside a specified sphere or slab, for binary classification; they showed that high dimensionality gives attacker more room for constructing attacks to evade outlier removal. Laishram and Phoha [2016] applied the seminal DBSCAN (Density-Based Spatial Clustering of Applications with Noise) method [Ester et al., 1996] to remove outliers for SVM and showed that it can successfully identify most of the poisoning data. However, their DBSCAN approach is lack of theoretical analysis. Several other outlier removal methods for fighting poisoning attacks have also been studied recently [Paudice et al., 2018b, a]. Also, it is worth noting that outlier removal actually is a topic that has been extensively studied in various fields before [Chandola et al., 2009].

The other defense strategy, designing robust optimization algorithms, also has a long history in the machine learning community. A substantial part of robust optimization algorithms rely on the idea of regularization. For example, Xu et al. [2009] studied the relation between robustness and regularization for SVM; other robust SVM algorithms include [Tax and Duin, 1999, Xu et al., 2006, Suzumura et al., 2014, Natarajan et al., 2013, Xu et al., 2017, Kanamori et al., 2017]. However, as discussed in [Mei and Zhu, 2015, Jagielski et al., 2018], these approaches are not quite ideal to defend against poisoning attacks since the outliers can be located arbitrarily in the feature space by the adversary. Another idea for achieving the robustness guarantee is to add strong assumptions about the data distribution or the attacker’s behavior [Feng et al., 2014, Weerasinghe et al., 2019], but these assumptions are usually not well satisfied in practice. An alternative approach is to explicitly remove outliers during optimization, such as the “trimmed” method for robust regression [Jagielski et al., 2018]; but this approach often results in a challenging combinatorial optimization problem: if zz of the input nn data items are outliers (z<nz<n), we have to consider an exponentially large number (nz){n\choose z} of different possible cases in the adversarial setting.

1.1 Our Contributions

Due to the great importance in machine learning [Chang and Lin, 2011], we focus on defending SVM against poisoning attacks in this paper. Our contributions are twofold.

(i). First, we consider the robust optimization approach. To study its complexity, we only consider the hard-margin case (because the soft-margin case is more complicated and thus should have an even higher complexity). As mentioned above, we can formulate the SVM with outliers problem as a combinatorial optimization problem for achieving the adversarial-resilience: finding an optimal subset of nzn-z items from the poisoned input data to achieve the largest separating margin (the formal definition is shown in Section 2).

Though its local optimum can be obtained by using various methods, such as the alternating minimization approach [Jagielski et al., 2018], it is often very challenging to achieve a quality guaranteed solution for such adversarial-resilience optimization problem. For instance, Simonov et al. [2019] showed that unless the Exponential Time Hypothesis (ETH) fails, it is impossible not only to solve the PCA with outliers problem exactly but even to approximate it within a constant factor. A similar hardness result was also proved for linear regression with outliers by Mount et al. [2014]. But for SVM with outliers, we are unaware of any hardness-of-approximation result before. We try to bridge the gap in the current state of knowledge in Section 3. We prove that even the simplest one-class SVM with outliers problem is NP-complete, and has no fully polynomial-time approximation scheme (PTAS) unless P==NP. So it is quite unlikely that one can achieve a (nearly) optimal solution in polynomial time.

(ii). Second, we investigate the DBSCAN based data sanitization defense and explain its effectiveness in theory (Section 4). DBSCAN is one of the most popular density-based clustering methods and has been implemented for solving many real-world outlier removal problems [Ester et al., 1996, Schubert et al., 2017]; roughly speaking, the inliers are assumed to be located in some dense regions and the remaining points are recognized as the outliers. Actually, the intuition of using DBSCAN for data sanitization is straightforward [Laishram and Phoha, 2016]. We assume the original input training data (before poisoning attack) is large and dense enough in the domain Ω\Omega; thus the poisoning data should be the sparse outliers together with some small clusters located outside the dense regions, which can be identified by the DBSCAN. Obviously, if the attacker has a fixed budget zz (the number of poisoning points), the lager the data size nn is, the sparser the outliers appear to be (and the more efficiently the DBSCAN performs).

Thus, to guarantee the effectiveness of the DBSCAN approach, a fundamental question in theory is what about the lower bound of the data size nn (we can assume that the original input data is a set of i.i.d. samples drawn from the domain Ω\Omega). However, to achieve a favorable lower bound is a non-trivial task. The VC dimension [Li et al., 2001] of the range space induced by the Euclidean distance is high in a high-dimensional feature space, and thus the lower bound of the data size nn can be very large. Our idea is motivated by the recent observations on the link between the adversarial vulnerability and the intrinsic dimensionality [Khoury and Hadfield-Menell, 2019, Amsaleg et al., 2017, Ma et al., 2018]. We prove a lower bound of nn that depends on the intrinsic dimension of Ω\Omega and is independent of the feature space’s dimensionality.

Our result strengthens the observation from Steinhardt et al. [2017] who only considered the Euclidean space’s dimensionality: more precisely, it is the “high intrinsic dimensionality” that gives attacker more room to evade outlier removal. In particular, different from the previous results on evasion attacks [Khoury and Hadfield-Menell, 2019, Amsaleg et al., 2017, Ma et al., 2018], our result is the first one linking poisoning attacks to intrinsic dimensionality, to the best of our knowledge. In Section 5, we investigate several popular defending methods (including DBSCAN), where the intrinsic dimension of data demonstrates significant influence on their defending performances.

2 Preliminaries

Given two point sets P+P^{+} and PP^{-} in d\mathbb{R}^{d}, the problem of linear support vector machine (SVM) [Chang and Lin, 2011] is to find the maximum margin separating these two point sets (if they are separable). If P+P^{+} (or PP^{-}) is a single point, say the origin, the problem is called one-class SVM. The SVM can be formulated as a quadratic programming problem, and a number of efficient techniques have been developed in the past, such as the soft margin SVM [Cortes and Vapnik, 1995], ν\nu-SVM [Scholkopf et al., 2000, Crisp and Burges, 1999], and Core-SVM [Tsang et al., 2005]. If P+P^{+} and PP^{-} are not separable, we can apply the kernel method: each point pP+Pp\in P^{+}\cup P^{-} is mapped to be ϕ(p)\phi(p) in a higher dimensional space; the inner product ϕ(p1),ϕ(p2)\langle\phi(p_{1}),\phi(p_{2})\rangle is defined by a kernel function 𝒦(p1,p2)\mathcal{K}(p_{1},p_{2}). Many existing SVM algorithms can be adapted to handle the non-separable case by using kernel functions.

Poisoning attacks. The adversary usually injects some bad points to the original data set P+PP^{+}\cup P^{-}. For instance, the adversary can take a sample qq from the domain of P+P^{+}, and flip its label to be “-”; therefore, this poisoning sample qq can be viewed as an outlier of PP^{-}. Since poisoning attack is expensive, we often assume that the adversary can poison at most z+z\in\mathbb{Z}^{+} points (or the poisoned fraction z|P+P|\frac{z}{|P^{+}\cup P^{-}|} is a fixed small number in (0,1)(0,1)). We can formulate the defense against poisoning attacks as the following combinatorial optimization problem. As mentioned in Section 1.1, it is sufficient to consider only the simpler hard-margin case for studying its hardness.

Definition 1 (SVM with Outliers).

Let (P+,P)(P^{+},P^{-}) be an instance of SVM in d\mathbb{R}^{d}, and suppose |P+P|=n\big{|}P^{+}\cup P^{-}\big{|}=n. Given a positive integer z<nz<n, the problem of SVM with outliers is to find two subsets P1+P+P^{+}_{1}\subseteq P^{+} and P1PP^{-}_{1}\subseteq P^{-} with |P1+P1|=nz\big{|}P^{+}_{1}\cup P^{-}_{1}\big{|}=n-z, such that the width of the margin separating P1+P^{+}_{1} and P1P^{-}_{1} is maximized.

Suppose the optimal margin has the width hopt>0h_{opt}>0. If we achieve a solution with the margin width h(1ϵ)hopth\geq(1-\epsilon)h_{opt} where ϵ\epsilon is a small number in (0,1)(0,1), we say that it is a (1ϵ)(1-\epsilon)-approximation.

Remark 1.

The model proposed in Definition 1 follows the popular data trimming idea from robust statistics [Rousseeuw and Leroy, 1987]. As an example similar with Definition 1, Jagielski et al. [2018] proposed a data trimming based regression model to defend against poisoning attacks.

We also need to clarify the intrinsic dimensionality for our following analysis. Doubling dimension is a measure of intrinsic dimensionality that has been widely adopted in the learning theory community [Bshouty et al., 2009]. Given a point pp and r0r\geq 0, we use 𝔹(p,r)\mathbb{B}(p,r) to indicate the ball of radius rr around pp in the space.

Definition 2 (Doubling Dimension).

The doubling dimension of a point set PP from some metric space111The space can be a Euclidean space or an abstract metric space. is the smallest number ρ\rho, such that for any pPp\in P and r0r\geq 0, the set P𝔹(p,2r)P\cap\mathbb{B}(p,2r) can always be covered by the union of at most 2ρ2^{\rho} balls with radius rr in the space.

To understand doubling dimension, we consider the following simple case. If the points of PP distribute uniformly in a dd^{\prime}-dimensional flat in d\mathbb{R}^{d}, then it is easy to see that PP has the doubling dimension ρ=O(d)\rho=O(d^{\prime}), which is independent of the Euclidean dimension dd (e.g., dd can be much higher than ρ\rho). Intuitively, doubling dimension is used for describing the expansion rate of a given point set in the space. It is worth noting that the intrinsic dimensionality described in [Amsaleg et al., 2017, Ma et al., 2018] is quite similar to doubling dimension, which also measures expansion rate.

3 The Hardness of SVM with Outliers

Refer to caption
(a)
Refer to caption
(b)
Figure 1: (a) An illustration for the formula (4); (b) the ball B1B_{1} is enclosed by Ω\Omega and the ball B2B_{2} is not.

In this section, we prove that even the one-class SVM with outliers problem is NP-complete and has no fully PTAS unless P==NP (that is, we cannot achieve a polynomial time (1ϵ)(1-\epsilon)-approximation for any given ϵ(0,1)\epsilon\in(0,1)). Our idea is partly inspired by the result from Megiddo [1990]. Given a set of points in d\mathbb{R}^{d}, the “covering by two balls” problem is to determine that whether the point set can be covered by two unit balls. By the reduction from 33-SAT, Megiddo proved that the “covering by two balls” problem is NP-complete. In the proof of the following theorem, we modify Megiddo’s construction of the reduction to adapt the one-class SVM with outliers problem.

Theorem 1.

The one-class SVM with outliers problem is NP-complete, and has no fully PTAS unless P==NP.

Let Γ\Gamma be a 33-SAT instance with the literal set {u1,u¯1,,ul,u¯l}\{u_{1},\bar{u}_{1},\cdots,u_{l},\bar{u}_{l}\} and clause set {E1,,Em}\{E_{1},\cdots,E_{m}\}. We construct the corresponding instance PΓP_{\Gamma} of one-class SVM with outliers. First, let U={±eii=1,2,,l+1}U=\{\pm e_{i}\mid i=1,2,\cdots,l+1\} be the 2(l+1)2(l+1) unit vectors of l+1\mathbb{R}^{l+1}, where each eie_{i} has “11” in the ii-th position and “0” in other positions. Also, for each clause EjE_{j} with 1jm1\leq j\leq m, we generate a point qj=(qj,1,qj,2,,qj,l+1)q_{j}=(q_{j,1},q_{j,2},\cdots,q_{j,l+1}) as follows. For 1il1\leq i\leq l,

qj,i={α, if ui occurs in Ejα, else if u¯i occurs in Ej;0, otherwise.q_{j,i}=\left\{\begin{array}[]{ll}\alpha,&\text{ if $u_{i}$ occurs in $E_{j}$; }\\ -\alpha,&\text{ else if $\bar{u}_{i}$ occurs in $E_{j}$;}\\ 0,&\text{ otherwise.}\end{array}\right.

In addition, qj,l+1=3αq_{j,l+1}=3\alpha. For example, if Ej=ui1u¯i2ui3E_{j}=u_{i_{1}}\vee\bar{u}_{i_{2}}\vee u_{i_{3}}, the point

qj=(0,,0,αi1,0,,0,αi2,0,,\displaystyle q_{j}=(0,\cdots,0,\underset{i_{1}}{\alpha},0,\cdots,0,\underset{i_{2}}{-\alpha},0,\cdots,
0,αi3,0,,0,3α).\displaystyle 0,\underset{i_{3}}{\alpha},0,\cdots,0,3\alpha). (1)

The value of α\alpha will be determined later. Let QQ denote the set {q1,,qm}\{q_{1},\cdots,q_{m}\}. Now, we construct the instance PΓ=UQP_{\Gamma}=U\cup Q of one-class SVM with outliers, where the number of points n=2(l+1)+mn=2(l+1)+m and the number of outliers z=l+1z=l+1. Then we have the following lemma.

Lemma 1.

Let α>1/2\alpha>1/2. Γ\Gamma has a satisfying assignment if and only if PΓP_{\Gamma} has a solution with margin width 1l+1\frac{1}{\sqrt{l+1}}.

Proof.

First, we suppose there exists a satisfying assignment for Γ\Gamma. We define the set SPΓS\subset P_{\Gamma} as follows. If uiu_{i} is true in Γ\Gamma, we include eie_{i} in SS, else, we include ei-e_{i} in SS; we also include el+1e_{l+1} in SS. We claim that the set SQS\cup Q yields a solution of the instance PΓP_{\Gamma} with the margin width 1l+1\frac{1}{\sqrt{l+1}}, that is, the size |SQ|=nz|S\cup Q|=n-z and the margin separating the origin oo and SQS\cup Q has width 1l+1\frac{1}{\sqrt{l+1}}. It is easy to verify the size of SQS\cup Q. To compute the width, we consider the mean point of SS which is denoted as tt. For each 1il1\leq i\leq l, if uiu_{i} is true, the ii-th position of tt should be 1l+1\frac{1}{l+1}, else, the ii-th position of tt should be 1l+1-\frac{1}{l+1}; the (l+1)(l+1)-th position of tt is 1l+1\frac{1}{l+1}. Obviously, t=1l+1||t||=\frac{1}{\sqrt{l+1}}. Let t\mathcal{H}_{t} be the hyperplane that is orthogonal to the vector tot-o and passing through tt. So t\mathcal{H}_{t} separates SS and oo with the margin width t=1l+1||t||=\frac{1}{\sqrt{l+1}}. Furthermore, for any point qjQq_{j}\in Q, since there exists at least one true variable in EjE_{j}, we have the inner product

qj,tt\displaystyle\langle q_{j},\frac{t}{||t||}\rangle \displaystyle\geq 3αl+1+αl+12αl+1\displaystyle\frac{3\alpha}{\sqrt{l+1}}+\frac{\alpha}{\sqrt{l+1}}-\frac{2\alpha}{\sqrt{l+1}} (2)
=\displaystyle= 2αl+1>1l+1,\displaystyle\frac{2\alpha}{\sqrt{l+1}}>\frac{1}{\sqrt{l+1}},

where the last inequality comes from the fact α>1/2\alpha>1/2. Therefore, all the points from QQ lie on the same side of t\mathcal{H}_{t} as SS, and then the set SQS\cup Q can be separated from oo by a margin with width 1l+1\frac{1}{\sqrt{l+1}}.

Second, suppose the instance PΓP_{\Gamma} has a solution with margin width 1l+1\frac{1}{\sqrt{l+1}}. With a slight abuse of notations, we still use SS to denote the subset of UU that is included in the set of nzn-z inliers. Since the number of outliers is z=l+1z=l+1, we know that for any pair ±ei\pm e_{i}, there exists exactly one point belonging to SS; also, the whole set QQ should be included in the set of inliers so as to guarantee that there are nzn-z inliers in total. We still use tt to denote the mean point of SS (t=1l+1||t||=\frac{1}{\sqrt{l+1}}). Now, we design the assignment for Γ\Gamma: if eiSe_{i}\in S, we assign uiu_{i} to be true, else, we assign u¯i\bar{u}_{i} to be true. We claim that Γ\Gamma is satisfied by this assignment. For any clause EjE_{j}, if it is not satisfied, i.e., all the three variables in EjE_{j} are false, then we have the inner product

qj,tt3αl+13αl+1=0.\displaystyle\langle q_{j},\frac{t}{||t||}\rangle\leq\frac{3\alpha}{\sqrt{l+1}}-\frac{3\alpha}{\sqrt{l+1}}=0. (3)

That means the angle qjotπ/2\angle q_{j}ot\geq\pi/2. So any margin separating the origin oo and the set SQS\cup Q should has the width at most

qjtqj2+t2<t=1l+1.\displaystyle\frac{||q_{j}||\cdot||t||}{\sqrt{||q_{j}||^{2}+||t||^{2}}}<||t||=\frac{1}{\sqrt{l+1}}. (4)

See Figure 1a for an illustration. This is in contradiction to the assumption that PΓP_{\Gamma} has a solution with margin width 1l+1\frac{1}{\sqrt{l+1}}.

Overall, Γ\Gamma has a satisfying assignment if and only if PΓP_{\Gamma} has a solution with margin width 1l+1\frac{1}{\sqrt{l+1}}. ∎

Now we are ready to prove the theorem.

Proof.

(of Theorem 1) Since 3-SAT is NP-complete, Lemma 1 implies that the one-class SVM with outliers problem is NP-complete too; otherwise, we can determine that whether a given instance Γ\Gamma is satisfiable by computing the optimal solution of PΓP_{\Gamma}. Moreover, the gap between 1l+1\frac{1}{\sqrt{l+1}} and qjtqj2+t2\frac{||q_{j}||\cdot||t||}{\sqrt{||q_{j}||^{2}+||t||^{2}}} (from the formula (4)) is

1l+112α21l+112α2+1l+1\displaystyle\frac{1}{\sqrt{l+1}}-\sqrt{\frac{12\alpha^{2}\frac{1}{l+1}}{12\alpha^{2}+\frac{1}{l+1}}}
=\displaystyle= (1l+1)3/2112α2+1l+1(12α2+1l+1+23α)\displaystyle(\frac{1}{l+1})^{3/2}\frac{1}{\sqrt{12\alpha^{2}+\frac{1}{l+1}}(\sqrt{12\alpha^{2}+\frac{1}{l+1}}+2\sqrt{3}\alpha)}
=\displaystyle= Θ((1l+1)3/2),\displaystyle\Theta\big{(}(\frac{1}{l+1})^{3/2}\big{)}, (5)

if we assume α\alpha is a fixed constant. Therefore, if we set ϵ=O((1l+1)3/2(1l+1)1/2)=O(1l+1)\epsilon=O\big{(}\frac{(\frac{1}{l+1})^{3/2}}{(\frac{1}{l+1})^{1/2}}\big{)}=O(\frac{1}{l+1}), then Γ\Gamma is satisfiable if and only if any (1ϵ)(1-\epsilon)-approximation of the instance PΓP_{\Gamma} has width >12α21l+112α2+1l+1>\sqrt{\frac{12\alpha^{2}\frac{1}{l+1}}{12\alpha^{2}+\frac{1}{l+1}}}. That means if we have a fully PTAS for the one-class SVM with outliers problem, we can determine that whether Γ\Gamma is satisfiable or not in polynomial time. In other words, we cannot even achieve a fully PTAS for one-class SVM with outliers, unless P==NP. ∎

4 The Data Sanitization Defense

From Theorem 1, we know that it is extremely challenging to achieve the optimal solution even for one-class SVM with outliers. Therefore, we turn to consider the other approach, data sanitization defense, under some reasonable assumption in practice. First, we prove a general sampling theorem in Section 4.1. Then, we apply this theorem to explain the effectiveness of DBSCAN for defending against poisoning attacks in Section 4.2.

4.1 A Sampling Theorem

Let PP be a set of i.i.d. samples drawn from a connected and compact domain Ω\Omega who has the doubling dimension ρ>0\rho>0. For ease of presentation, we assume that Ω\Omega lies on a manifold \mathcal{F} in the space. Let Δ\Delta denote the diameter of Ω\Omega, i.e., Δ=supp1,p2Ωp1p2\Delta=\sup_{p_{1},p_{2}\in\Omega}||p_{1}-p_{2}||. Also, we let ff be the probability density function of the data distribution over Ω\Omega.

To measure the uniformity of ff, we define a value λ\lambda as follows. For any cΩc\in\Omega and any r>0r>0, we say “the ball 𝔹(c,r)\mathbb{B}(c,r) is enclosed by Ω\Omega” if 𝔹(c,r)Ω\partial\mathbb{B}(c,r)\cap\mathcal{F}\subset\Omega; intuitively, if the ball center cc is close to the boundary Ω\partial\Omega of Ω\Omega or the radius rr is too large, the ball will not be enclosed by Ω\Omega. See Figure 1b for an illustration. We define λsupc,c,r𝔹(c,r)f(x)dx𝔹(c,r)f(x)dx\lambda\coloneqq\sup_{c,c^{\prime},r}\frac{\int_{\mathbb{B}(c^{\prime},r)}\!f(x)\,\mathrm{d}x}{\int_{\mathbb{B}(c,r)}\!f(x)\,\mathrm{d}x}, where 𝔹(c,r)\mathbb{B}(c,r) and 𝔹(c,r)\mathbb{B}(c^{\prime},r) are any two equal-sized balls, and 𝔹(c,r)\mathbb{B}(c,r) is required to be enclosed by Ω\Omega. As a simple example, if Ω\Omega lies on a flat manifold and the data uniformly distribute over Ω\Omega, the value λ\lambda will be equal to 11. On the other hand, if the distribution is very imbalanced or the manifold \mathcal{F} is very rugged, the value λ\lambda can be high.

Theorem 2.

Let m+m\in\mathbb{Z}^{+}, ϵ(0,18)\epsilon\in(0,\frac{1}{8}), and δ(0,Δ)\delta\in(0,\Delta). If the sample size

|P|>max{Θ(m1ϵλ(1+ϵ1ϵΔδ)ρ),\displaystyle|P|>\max\Big{\{}\Theta\big{(}\frac{m}{1-\epsilon}\cdot\lambda\cdot(\frac{1+\epsilon}{1-\epsilon}\frac{\Delta}{\delta})^{\rho}\big{)},
Θ~(ρλ2(1+ϵ1ϵΔδ)2ρ(1ϵ)ρ+2)},\displaystyle\tilde{\Theta}\big{(}\rho\cdot\lambda^{2}\cdot(\frac{1+\epsilon}{1-\epsilon}\frac{\Delta}{\delta})^{2\rho}(\frac{1}{\epsilon})^{\rho+2}\big{)}\Big{\}}, (6)

then with constant probability, for any ball 𝔹(c,δ)\mathbb{B}\big{(}c,\delta\big{)} enclosed by Ω\Omega, the size |𝔹(c,δ)P|>m|\mathbb{B}\big{(}c,\delta\big{)}\cap P|>m. The asymptotic notation Θ~(f)=Θ(f𝚙𝚘𝚕𝚢𝚕𝚘𝚐(LΔδϵ))\tilde{\Theta}(f)=\Theta\big{(}f\cdot\mathtt{polylog}(\frac{L\Delta}{\delta\epsilon})\big{)}.

Remark 2.

(i) A highlight of Theorem 2 is that the lower bound of |P||P| is independent of the dimensionality of the input space (which could be much higher than the intrinsic dimension).

(ii) For the simplest case that Ω\Omega lies on a flat manifold and the data uniformly distribute over Ω\Omega, λ\lambda will be equal to 11 and thus the lower bound of |P||P| in Theorem 2 becomes max{Θ(m1ϵ(1+ϵ1ϵΔδ)ρ),Θ~(ρ(1+ϵ1ϵΔδ)2ρ(1ϵ)ρ+2)}\max\Big{\{}\Theta\big{(}\frac{m}{1-\epsilon}(\frac{1+\epsilon}{1-\epsilon}\frac{\Delta}{\delta})^{\rho}\big{)},\tilde{\Theta}\big{(}\rho(\frac{1+\epsilon}{1-\epsilon}\frac{\Delta}{\delta})^{2\rho}(\frac{1}{\epsilon})^{\rho+2}\big{)}\Big{\}}.

Before proving Theorem 2, we need to relate the doubling dimension ρ\rho to the VC dimension 𝚍𝚒𝚖\mathtt{dim} of the range space consisting of all balls with different radii [Li et al., 2001]. Unfortunately, Huang et al. [2018] recently showed that “although both dimensions are subjects of extensive research, to the best of our knowledge, there is no nontrivial relation known between the two”. For instance, they constructed a doubling metric having unbounded VC dimension, and the other direction cannot be bounded neither. However, if allowing a small distortion to the distance, we can achieve an upper bound on the VC dimension for a given metric space with bounded doubling dimension. For stating the result, they defined a distance function called “ϵ\epsilon-smoothed distance function”: g(p,q)(1±ϵ)pqg(p,q)\in(1\pm\epsilon)||p-q|| for any two data points pp and qq, where ϵ(0,18)\epsilon\in(0,\frac{1}{8}). Given a point pp and δ>0\delta>0, the ball defined by this distance function g(,)g(\cdot,\cdot) is denoted by 𝔹g(p,δ)={qthe input spaceg(p,q)δ}\mathbb{B}_{g}(p,\delta)=\{q\in\text{the input space}\mid g(p,q)\leq\delta\}.

Theorem 3 (Huang et al. [2018]).

Suppose the point set PP has the doubling dimension ρ>0\rho>0. There exists an ϵ\epsilon-smoothed distance function “g(,)g(\cdot,\cdot)” such that the VC dimension222Huang et al. [2018] used “shattering dimension” to state their result. Actually, the shattering dimension is another measure for the complexity of range space, which is tightly related to the VC dimension [Feldman and Langberg, 2011]. For example, if the shattering dimension is ρ0\rho_{0}, the VC dimension should be bounded by O(ρ0logρ0)O(\rho_{0}\log\rho_{0}). 𝚍𝚒𝚖ϵ\mathtt{dim}_{\epsilon} of the range space consisting of all balls with different radii is at most O~(ρϵρ)\tilde{O}(\frac{\rho}{\epsilon^{\rho}}), if replacing the Euclidean distance by g(,)g(\cdot,\cdot).

Proof.

(of Theorem 2) Let rr be any positive number. First, since the doubling dimension of Ω\Omega is ρ\rho, if recursively applying Definition 2 logΔr\log\frac{\Delta}{r} times, we know that Ω\Omega can be covered by at most Θ((Δr)ρ)\Theta\Big{(}\big{(}\frac{\Delta}{r}\big{)}^{\rho}\Big{)} balls with radius rr. Thus, if 𝔹(c,r)\mathbb{B}(c,r) is enclosed by Ω\Omega, we have

𝔹(c,r)f(x)dxΩf(x)dxΘ(1λ(rΔ)ρ).\displaystyle\frac{\int_{\mathbb{B}(c,r)}\!f(x)\,\mathrm{d}x}{\int_{\Omega}\!f(x)\,\mathrm{d}x}\geq\Theta\Big{(}\frac{1}{\lambda}\cdot(\frac{r}{\Delta})^{\rho}\Big{)}. (7)

Now we consider the size |𝔹(c,δ)P||\mathbb{B}(c,\delta)\cap P|. From Theorem 3, we know that the VC dimension 𝚍𝚒𝚖ϵ\mathtt{dim}_{\epsilon} with respect to the ϵ\epsilon-smoothed distance is O~(ρϵρ)\tilde{O}(\frac{\rho}{\epsilon^{\rho}}). Thus, for any ϵ0(0,1)\epsilon_{0}\in(0,1), if

|P|Θ(1ϵ02𝚍𝚒𝚖ϵlog𝚍𝚒𝚖ϵϵ0),\displaystyle|P|\geq\Theta\big{(}\frac{1}{\epsilon^{2}_{0}}\mathtt{dim}_{\epsilon}\log\frac{\mathtt{dim}_{\epsilon}}{\epsilon_{0}}\big{)}, (8)

the set PP will be an ϵ0\epsilon_{0}-sample of Ω\Omega; that is, for any point cc and δ0\delta^{\prime}\geq 0,

|𝔹g(c,δ)P||P|𝔹g(c,δ)f(x)dxΩf(x)dx±ϵ0\displaystyle\frac{|\mathbb{B}_{g}(c,\delta^{\prime})\cap P|}{|P|}\in\frac{\int_{\mathbb{B}_{g}(c,\delta^{\prime})}\!f(x)\,\mathrm{d}x}{\int_{\Omega}\!f(x)\,\mathrm{d}x}\pm\epsilon_{0} (9)

with constant probability333The exact probability comes from the success probability that PP is an ϵ0\epsilon_{0}-sample of Ω\Omega. Let η(0,1)\eta\in(0,1), and the size |P||P| in (8) should be at least Θ(1ϵ02(𝚍𝚒𝚖ϵlog𝚍𝚒𝚖ϵϵ0+log1η))\Theta\big{(}\frac{1}{\epsilon^{2}_{0}}(\mathtt{dim}_{\epsilon}\log\frac{\mathtt{dim}_{\epsilon}}{\epsilon_{0}}+\log\frac{1}{\eta})\big{)} to guarantee a success probability 1η1-\eta. For convenience, we assume η\eta is a fixed small constant and simply say “1η1-\eta” is a “constant probability”. [Li et al., 2001]. Because g(,)g(\cdot,\cdot) is an ϵ\epsilon-smoothed distance function of the Euclidean distance, we have

𝔹(c,δ1+ϵ)𝔹g(c,δ)𝔹(c,δ1ϵ).\displaystyle\mathbb{B}(c,\frac{\delta^{\prime}}{1+\epsilon})\subseteq\mathbb{B}_{g}(c,\delta^{\prime})\subseteq\mathbb{B}(c,\frac{\delta^{\prime}}{1-\epsilon}). (10)

So if we set ϵ0=ϵΘ(1λ(1ϵ1+ϵδΔ)ρ)\epsilon_{0}=\epsilon\cdot\Theta\Big{(}\frac{1}{\lambda}\cdot(\frac{1-\epsilon}{1+\epsilon}\frac{\delta}{\Delta})^{\rho}\Big{)} and δ=(1ϵ)δ\delta^{\prime}=(1-\epsilon)\delta, (7), (9), and (10) jointly imply |𝔹(c,δ)P||P|=\frac{|\mathbb{B}(c,\delta)\cap P|}{|P|}=

|𝔹(c,δ1ϵ)P||P|\displaystyle\frac{|\mathbb{B}(c,\frac{\delta^{\prime}}{1-\epsilon})\cap P|}{|P|} \displaystyle\geq |𝔹g(c,δ)P||P|\displaystyle\frac{|\mathbb{B}_{g}(c,\delta^{\prime})\cap P|}{|P|}
\displaystyle\geq 𝔹g(c,δ)f(x)dxΩf(x)dxϵ0\displaystyle\frac{\int_{\mathbb{B}_{g}(c,\delta^{\prime})}\!f(x)\,\mathrm{d}x}{\int_{\Omega}\!f(x)\,\mathrm{d}x}-\epsilon_{0}
\displaystyle\geq 𝔹(c,δ1+ϵ)f(x)dxΩf(x)dxϵ0\displaystyle\frac{\int_{\mathbb{B}(c,\frac{\delta^{\prime}}{1+\epsilon})}\!f(x)\,\mathrm{d}x}{\int_{\Omega}\!f(x)\,\mathrm{d}x}-\epsilon_{0}
\displaystyle\geq (1ϵ)Θ(1λ(1ϵ1+ϵδΔ)ρ).\displaystyle(1-\epsilon)\cdot\Theta\Big{(}\frac{1}{\lambda}\cdot(\frac{1-\epsilon}{1+\epsilon}\frac{\delta}{\Delta})^{\rho}\Big{)}. (11)

The last inequality comes from (7) (since we assume the ball 𝔹(c,δ)\mathbb{B}\big{(}c,\delta\big{)} is enclosed by Ω\Omega, the shrunk ball 𝔹(c,δ1+ϵ)=𝔹(c,1ϵ1+ϵδ)\mathbb{B}(c,\frac{\delta^{\prime}}{1+\epsilon})=\mathbb{B}(c,\frac{1-\epsilon}{1+\epsilon}\delta) should be enclosed as well). Moreover, if

|P|Θ(m1ϵλ(1+ϵ1ϵΔδ)ρ),\displaystyle|P|\geq\Theta\big{(}\frac{m}{1-\epsilon}\cdot\lambda\cdot(\frac{1+\epsilon}{1-\epsilon}\frac{\Delta}{\delta})^{\rho}\big{)}, (12)

we have |𝔹(c,δ)P|>m|\mathbb{B}\big{(}c,\delta\big{)}\cap P|>m from (11). Combining (8) and (12), we obtain the lower bound of |P||P|. ∎

4.2 The DBSCAN Approach

For the sake of completeness, we briefly introduce the method of DBSCAN [Ester et al., 1996]. Given two parameters r>0r>0 and 𝙼𝚒𝚗𝙿𝚝𝚜+\mathtt{MinPts}\in\mathbb{Z}^{+}, the DBSCAN divides the set PP into three classes: (1) pp is a core point, if |𝔹(p,r)P|>𝙼𝚒𝚗𝙿𝚝𝚜|\mathbb{B}(p,r)\cap P|>\mathtt{MinPts}; (2) pp is a border point, if pp is not a core point but p𝔹(q,r)p\in\mathbb{B}(q,r) of some core point qq; (3) all the other points are outliers. Actually, we can imagine that the set PP forms a graph where any pair of core or border points are connected if their pairwise distance is no larger than rr; then the set of core points and border points form several clusters where each cluster is a connected component (a border point may belong to multiple clusters, but we can arbitrarily assign it to only one cluster). The goal of DBSCAN is to identify these clusters and the remaining outliers. Several efficient implementations for DBSCAN can be found in [Gan and Tao, 2015, Schubert et al., 2017].

Following Section 4.1, we assume that PP is a set of i.i.d. samples drawn from the connected and compact domain Ω\Omega who has the doubling dimension ρ>0\rho>0. We let QQ be the set of zz poisoning data items injected by the attacker to PP, and suppose each qQq\in Q has distance larger than δ1>0\delta_{1}>0 to Ω\Omega. In an evasion attack, we often use the adversarial perturbation distance to evaluate the attacker’s capability; but in a poisoning attack, the attacker can easily achieve a large perturbation distance (e.g., in the SVM problem, if the attacker flips the label of some point pp, it will become an outlier having the perturbation distance larger than hopth_{opt} to its ground truth domain, where hopth_{opt} is the optimal margin width). Also, we assume the boundary Ω\partial\Omega is smooth and has curvature radius at least δ2>0\delta_{2}>0 everywhere. For simplicity, let δ=min{δ1,δ2}\delta=\min\{\delta_{1},\delta_{2}\}. The following theorem states the effectiveness of the DBSCAN with respect to the poisoned dataset PQP\cup Q. We assume the poisoned fraction |Q||P|=z|P|<1\frac{|Q|}{|P|}=\frac{z}{|P|}<1.

Theorem 4.

We let mm be any absolute constant number larger than 11, and assume that the size of PP satisfies the lower bound of Theorem 2. If we set r=δr=\delta and 𝙼𝚒𝚗𝙿𝚝𝚜=m\mathtt{MinPts}=m, and run DBSCAN on the poisoned dataset PQP\cup Q, the obtained largest cluster should be exactly PP. In other word, the set QQ should be formed by the outliers and the clusters except the largest one from the DBSCAN.

Proof.

Since δδ2\delta\leq\delta_{2}, for any pPp\in P, either the ball 𝔹(p,δ)\mathbb{B}(p,\delta) is enclosed by Ω\Omega, or pp is covered by some ball 𝔹(q,δ)\mathbb{B}(q,\delta) enclosed by Ω\Omega. We set r=δr=\delta and 𝙼𝚒𝚗𝙿𝚝𝚜=m\mathtt{MinPts}=m, and hence from Theorem 2 we know that all the points of PP will be core points or border points. Moreover, any point qq from QQ has distance larger than rr to the points of PP, that is, any two points qQq\in Q and pPp\in P should not belong to the same cluster of the DBSCAN. Also, because the domain Ω\Omega is connected and compact, the set PP must form the largest cluster. ∎

Remark 3.

(i) We often adopt the poisoned fraction z|P|\frac{z}{|P|} as the measure to indicate the attacker’s capability. If we fix the value of zz, the bound of |P||P| from Theorem 2 reveals that the larger the doubling dimension ρ\rho, the lower the poisoned fraction z|P|\frac{z}{|P|} (and the easier corrupting the DBSCAN defense). In addition, when δ\delta is large, i.e., each poisoning point has large perturbation distance and Ω\partial\Omega is sufficiently smooth, it will be relatively easy for DBSCAN to defend.

But we should point out that this theoretical bound probably is overly conservative, since it requires a “perfect” sanitization result that removes all the poisoning samples (this is not always a necessary condition for achieving a good defending performance in practice). In our experiments, we show that the DBSCAN method can achieve promising performance, even when the poisoned fraction is higher than the threshold.

(ii) In practice, we cannot obtain the exact values of δ\delta and mm. We follow the strategy that was commonly used in DBSCAN implementations before [Gan and Tao, 2015, Schubert et al., 2017]; we set 𝙼𝚒𝚗𝙿𝚝𝚜\mathtt{MinPts} to be a small constant and tune the value of rr until the largest cluster has |PQ|z|P\cup Q|-z points.

Putting it all together. Let (P+,P)(P^{+},P^{-}) be an instance of SVM with zz outliers, where zz is the number of poisoning points. We assume that the original input point sets P+P^{+} and PP^{-} (before the poisoning attack) are i.i.d. samples drawn respectively from the connected and compact domains Ω+\Omega^{+} and Ω\Omega^{-} with doubling dimension ρ\rho. Then, we perform the DBSCAN procedure on P+P^{+} and PP^{-} respectively (as Remark 3 (ii)). Suppose the obtained largest clusters are P~+\tilde{P}^{+} and P~\tilde{P}^{-}. Finally, we run an existing SVM algorithm on the cleaned instance (P~+,P~)(\tilde{P}^{+},\tilde{P}^{-}).

5 Empirical Experiments

Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Refer to caption
(d)
Refer to caption
(e)
Refer to caption
(f)
Refer to caption
(g)
Refer to caption
(h)
Refer to caption
(i)
Figure 2: The classification accuracy on the Synthetic datasets of Linear SVM (the first line) and SVM with RBF kernel (the second line) under Min-Max attack. The third line are the average F1F_{1} scores.
Refer to caption
(a) Letter
Refer to caption
(b) Mushrooms
Refer to caption
(c) Satimage
Refer to caption
(d) Letter
Refer to caption
(e) Mushrooms
Refer to caption
(f) Satimage
Figure 3: The classification accuracy on the real datasets of linear SVM (the first line) and SVM with RBF kernel (the second line) under Min-Max attack.

All the experiments were repeated 2020 times on a Windows 1010 workstation equipped with an Intel core i5i5-84008400 processor and 88GB RAM. To generate the poisoning attacks, we use the Min-Max attack from [Koh et al., 2018] and the adversarial label-flipping attack ALFA from ALFASVMLib [Xiao et al., 2015]. We evaluate the defending performances of the basic SVM algorithms and several different defenses by using their publicly available implementations.

  1. 1.

    We consider both the cases that not using and using kernel. For SVM without kernel, we directly use linear SVM as the basic SVM algorithm; for SVM with kernel, we consider RBF kernel (rbf SVM). Both the implementations are from [Chang and Lin, 2011].

  2. 2.

    The recently proposed robust SVM algorithm RSVM-SS based on the rescaled hinge loss function [Xu et al., 2017]. The parameter “SS” indicates the iteration number of the half-quadratic optimization (e.g., we set S=3S=3 and 1010 following their paper’s setting). The algorithm also works fine when using a kernel.

  3. 3.

    The DBSCAN method [Schubert et al., 2017] implemented as Remark 3 (ii). We set 𝙼𝚒𝚗𝙿𝚝𝚜=5\mathtt{MinPts}=5 (our empirical study finds that the difference is minor within the range [3,10][3,10]).

  4. 4.

    The data sanitization defenses from [Koh et al., 2018] based on the spatial distribution of input data, which include Slab, L2, Loss, and k-NN.

For the data sanitization defenses, we run them on the poisoned data in the original input space; then, apply the basic SVM algorithm, linear SVM or rbf SVM (if using RBF kernel), on the cleaned data to compute their final solutions.

Table 1: Datasets
Dataset Size Dimension
Synthetic 1000010000 5050-200200
letter 15201520 1616
mushrooms 81248124 112112
satimage 22362236 3636

Datasets. We consider both the synthetic and real-world datasets in our experiments. For each synthetic dataset, we generate two manifolds in d\mathbb{R}^{d}, and each manifold is represented by a random polynomial function with degree dd^{\prime} (the values of dd and dd^{\prime} will be varied in the experiments). Note that it is challenging to achieve the exact doubling dimensions of the datasets, and thus we use the degree of the polynomial function as a “rough indicator” for the doubling dimension (the higher the degree, the larger the doubling dimension). In each of the manifolds, we randomly sample 50005000 points; the data is randomly partitioned into 30%30\% and 70%70\% respectively for training and testing, and we report the classification accuracy on the test data. We also consider three real-world datasets from [Chang and Lin, 2011]. The details are shown in Table 1.

Results. First, we study the influence from the intrinsic dimensionality. We set the Euclidean dimensionality dd to be 100100 and vary the polynomial function’s degree dd^{\prime} from 2525 to 6565 in Figure 2a and 2d. Then, we fix the degree dd^{\prime} to be 4040 and vary the Euclidean dimensionality dd in Figure 2b and 2e. We can observe that the accuracies of most methods dramatically decrease when the degree dd^{\prime} (intrinsic dimension) increases, and the influence from the intrinsic dimension is more significant than that from the Euclidean dimension, which is in agreement with our theoretical analysis.

We also study their classification performances under different poisoned fraction in Figure 2c and 2f. We can see that all the defenses yield lower accuracies when the poisoned fraction increases, while the performance of DBSCAN keeps much more stable compared with other defenses. Moreover, we calculate the widely used F1F_{1} scores from the sanitization defenses for identifying the outliers. Loss and k-NN both yield very low F1F_{1} scores (<0.1<0.1); that means they are not quite capable to identify the real poisoning data items. The F1F_{1} scores yielded by DBSCAN, L22 and Slab are shown in Figure 2g-2i, where DBSCAN in general outperforms the other two sanitization defenses for most cases.

We also perform the experiments on the real datasets under Min-Max attack with the poisoned fraction ranging from 4%4\% to 10%10\%. The experimental results (Figure 3) reveal the similar trends as the results for the synthetic datasets, and DBSCAN keeps considerably better performance compared with other defenses. Due to the space limit, the results under ALFA attack are shown in Appendix.

6 Discussion

In this paper, we study two different strategies for protecting SVM against poisoning attacks. We also have several open questions to study in future. For example, what about the complexities of other machine learning problems under the adversarially-resilient formulations as Definition 1? For many other adversarial machine learning problems, the study on their complexities is still in its infancy.

References

  • Amsaleg et al. [2017] Laurent Amsaleg, James Bailey, Dominique Barbe, Sarah M. Erfani, Michael E. Houle, Vinh Nguyen, and Milos Radovanovic. The vulnerability of learning to adversarial perturbation increases with intrinsic dimensionality. In 2017 IEEE Workshop on Information Forensics and Security, WIFS, pages 1–6. IEEE, 2017.
  • Barreno et al. [2006] Marco Barreno, Blaine Nelson, Russell Sears, Anthony D. Joseph, and J. D. Tygar. Can machine learning be secure? In Proceedings of the 2006 ACM Symposium on Information, Computer and Communications Security, ASIACCS, pages 16–25. ACM, 2006.
  • Biggio and Roli [2018] Battista Biggio and Fabio Roli. Wild patterns: Ten years after the rise of adversarial machine learning. Pattern Recognition, 84:317–331, 2018.
  • Biggio et al. [2012] Battista Biggio, Blaine Nelson, and Pavel Laskov. Poisoning attacks against support vector machines. In Proceedings of the 29th International Conference on Machine Learning, ICML, 2012.
  • Biggio et al. [2014] Battista Biggio, Samuel Rota Bulò, Ignazio Pillai, Michele Mura, Eyasu Zemene Mequanint, Marcello Pelillo, and Fabio Roli. Poisoning complete-linkage hierarchical clustering. In Structural, Syntactic, and Statistical Pattern Recognition - Joint IAPR, 2014.
  • Bshouty et al. [2009] Nader H. Bshouty, Yi Li, and Philip M. Long. Using the doubling dimension to analyze the generalization of learning algorithms. J. Comput. Syst. Sci., 75(6):323–335, 2009.
  • Chandola et al. [2009] Varun Chandola, Arindam Banerjee, and Vipin Kumar. Anomaly detection: A survey. ACM Computing Surveys, 41(3):15, 2009.
  • Chang and Lin [2011] Chih-Chung Chang and Chih-Jen Lin. LIBSVM: A library for support vector machines. ACM TIST, 2(3), 2011.
  • Christmann and Steinwart [2004] Andreas Christmann and Ingo Steinwart. On robustness properties of convex risk minimization methods for pattern recognition. J. Mach. Learn. Res., 5:1007–1034, 2004.
  • Cortes and Vapnik [1995] Corinna Cortes and Vladimir Vapnik. Support-vector networks. Machine Learning, 20:273, 1995.
  • Cretu et al. [2008] Gabriela F. Cretu, Angelos Stavrou, Michael E. Locasto, Salvatore J. Stolfo, and Angelos D. Keromytis. Casting out demons: Sanitizing training data for anomaly sensors. In 2008 IEEE Symposium on Security and Privacy, pages 81–95. IEEE Computer Society, 2008.
  • Crisp and Burges [1999] David J. Crisp and Christopher J. C. Burges. A geometric interpretation of v-SVM classifiers. In NIPS, pages 244–250. The MIT Press, 1999.
  • Dalvi et al. [2004] Nilesh N. Dalvi, Pedro M. Domingos, Mausam, Sumit K. Sanghai, and Deepak Verma. Adversarial classification. In Proceedings of the Tenth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pages 99–108, 2004.
  • Ester et al. [1996] Martin Ester, Hans-Peter Kriegel, Jörg Sander, and Xiaowei Xu. A density-based algorithm for discovering clusters in large spatial databases with noise. In Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pages 226–231, 1996.
  • Feldman and Langberg [2011] Dan Feldman and Michael Langberg. A unified framework for approximating and clustering data. In Proceedings of the 43rd ACM Symposium on Theory of Computing, STOC, pages 569–578. ACM, 2011.
  • Feng et al. [2014] Jiashi Feng, Huan Xu, Shie Mannor, and Shuicheng Yan. Robust logistic regression and classification. In Advances in Neural Information Processing Systems, pages 253–261, 2014.
  • Gan and Tao [2015] Junhao Gan and Yufei Tao. DBSCAN revisited: mis-claim, un-fixability, and approximation. In Proceedings of the 2015 ACM SIGMOD International Conference on Management of Data, pages 519–530, 2015.
  • Goodfellow et al. [2018] Ian J. Goodfellow, Patrick D. McDaniel, and Nicolas Papernot. Making machine learning robust against adversarial inputs. Commun. ACM, 61(7):56–66, 2018.
  • Huang et al. [2011] Ling Huang, Anthony D Joseph, Blaine Nelson, Benjamin IP Rubinstein, and J Doug Tygar. Adversarial machine learning. In Proceedings of the 4th ACM workshop on Security and artificial intelligence, pages 43–58, 2011.
  • Huang et al. [2018] Lingxiao Huang, Shaofeng Jiang, Jian Li, and Xuan Wu. Epsilon-coresets for clustering (with outliers) in doubling metrics. In 59th IEEE Annual Symposium on Foundations of Computer Science, FOCS, pages 814–825, 2018.
  • Jagielski et al. [2018] Matthew Jagielski, Alina Oprea, Battista Biggio, Chang Liu, Cristina Nita-Rotaru, and Bo Li. Manipulating machine learning: Poisoning attacks and countermeasures for regression learning. In Symposium on Security and Privacy, SP, pages 19–35, 2018.
  • Kanamori et al. [2017] Takafumi Kanamori, Shuhei Fujiwara, and Akiko Takeda. Breakdown point of robust support vector machines. Entropy, 19(2):83, 2017.
  • Khoury and Hadfield-Menell [2019] Marc Khoury and Dylan Hadfield-Menell. Adversarial training with voronoi constraints. CoRR, abs/1905.01019, 2019.
  • Koh et al. [2018] Pang Wei Koh, Jacob Steinhardt, and Percy Liang. Stronger data poisoning attacks break data sanitization defenses. CoRR, abs/1811.00741, 2018.
  • Laishram and Phoha [2016] Ricky Laishram and Vir Virander Phoha. Curie: A method for protecting SVM classifier from poisoning attack. CoRR, abs/1606.01584, 2016.
  • Li et al. [2001] Yi Li, Philip M. Long, and Aravind Srinivasan. Improved bounds on the sample complexity of learning. J. Comput. Syst. Sci., 62(3):516–527, 2001.
  • Ma et al. [2018] Xingjun Ma, Bo Li, Yisen Wang, Sarah M. Erfani, Sudanthi N. R. Wijewickrema, Grant Schoenebeck, Dawn Song, Michael E. Houle, and James Bailey. Characterizing adversarial subspaces using local intrinsic dimensionality. In 6th International Conference on Learning Representations, ICLR. OpenReview.net, 2018.
  • Megiddo [1990] Nimrod Megiddo. On the complexity of some geometric problems in unbounded dimension. J. Symb. Comput., 10(3/4):327–334, 1990.
  • Mei and Zhu [2015] Shike Mei and Xiaojin Zhu. Using machine teaching to identify optimal training-set attacks on machine learners. In AAAI, pages 2871–2877. AAAI Press, 2015.
  • Mount et al. [2014] David M. Mount, Nathan S. Netanyahu, Christine D. Piatko, Ruth Silverman, and Angela Y. Wu. On the least trimmed squares estimator. Algorithmica, 69(1):148–183, 2014.
  • Natarajan et al. [2013] Nagarajan Natarajan, Inderjit S. Dhillon, Pradeep Ravikumar, and Ambuj Tewari. Learning with noisy labels. In Advances in Neural Information Processing Systems, pages 1196–1204, 2013.
  • Paudice et al. [2018a] Andrea Paudice, Luis Muñoz-González, András György, and Emil C. Lupu. Detection of adversarial training examples in poisoning attacks through anomaly detection. CoRR, abs/1802.03041, 2018a.
  • Paudice et al. [2018b] Andrea Paudice, Luis Muñoz-González, and Emil C. Lupu. Label sanitization against label flipping poisoning attacks. In ECML PKDD 2018 Workshops, pages 5–15, 2018b.
  • Rousseeuw and Leroy [1987] Peter J. Rousseeuw and Annick Leroy. Robust Regression and Outlier Detection. Wiley, 1987.
  • Rubinstein et al. [2009] Benjamin I. P. Rubinstein, Blaine Nelson, Ling Huang, Anthony D. Joseph, Shing-hon Lau, Satish Rao, Nina Taft, and J. D. Tygar. ANTIDOTE: understanding and defending against poisoning of anomaly detectors. In Proceedings of the 9th ACM SIGCOMM Internet Measurement Conference, IMC, pages 1–14. ACM, 2009.
  • Scholkopf et al. [2000] B. Scholkopf, A. J. Smola, K. R. Muller, and P. L. Bartlett. New support vector algorithms. Neural Computation, 12:1207–1245, 2000.
  • Schubert et al. [2017] Erich Schubert, Jörg Sander, Martin Ester, Hans Peter Kriegel, and Xiaowei Xu. DBSCAN revisited, revisited: why and how you should (still) use dbscan. ACM Transactions on Database Systems (TODS), 42(3):19, 2017.
  • Simonov et al. [2019] Kirill Simonov, Fedor V. Fomin, Petr A. Golovach, and Fahad Panolan. Refined complexity of PCA with outliers. In Proceedings of the 36th International Conference on Machine Learning, ICML, pages 5818–5826, 2019.
  • Steinhardt et al. [2017] Jacob Steinhardt, Pang Wei Koh, and Percy Liang. Certified defenses for data poisoning attacks. In Neural Information Processing System, pages 3517–3529, 2017.
  • Suzumura et al. [2014] Shinya Suzumura, Kohei Ogawa, Masashi Sugiyama, and Ichiro Takeuchi. Outlier path: A homotopy algorithm for robust svm. In Proceedings of the 31st International Conference on Machine Learning (ICML), pages 1098–1106, 2014.
  • Szegedy et al. [2014] Christian Szegedy, Wojciech Zaremba, Ilya Sutskever, Joan Bruna, Dumitru Erhan, Ian J. Goodfellow, and Rob Fergus. Intriguing properties of neural networks. In ICLR, 2014.
  • Tax and Duin [1999] David M. J. Tax and Robert P. W. Duin. Support vector domain description. Pattern Recognit. Lett., 20(11-13):1191–1199, 1999.
  • Tsang et al. [2005] Ivor W. Tsang, James T. Kwok, and Pak-Ming Cheung. Core vector machines: Fast SVM training on very large data sets. Journal of Machine Learning Research, 6:363–392, 2005.
  • Weerasinghe et al. [2019] Sandamal Weerasinghe, Sarah M. Erfani, Tansu Alpcan, and Christopher Leckie. Support vector machines resilient against training data integrity attacks. Pattern Recognit., 96, 2019.
  • Xiao et al. [2012] Han Xiao, Huang Xiao, and Claudia Eckert. Adversarial label flips attack on support vector machines. In 20th European Conference on Artificial Intelligence. Including Prestigious Applications of Artificial Intelligence, volume 242, pages 870–875. IOS Press, 2012.
  • Xiao et al. [2015] Huang Xiao, Battista Biggio, Blaine Nelson, Han Xiao, Claudia Eckert, and Fabio Roli. Support vector machines under adversarial label contamination. Neurocomputing, 160:53–62, 2015.
  • Xu et al. [2017] Guibiao Xu, Zheng Cao, Bao-Gang Hu, and José C. Príncipe. Robust support vector machines based on the rescaled hinge loss function. Pattern Recognit., 63:139–148, 2017.
  • Xu et al. [2009] Huan Xu, Constantine Caramanis, and Shie Mannor. Robustness and regularization of support vector machines. J. Mach. Learn. Res., 10:1485–1510, 2009.
  • Xu et al. [2006] Linli Xu, Koby Crammer, and Dale Schuurmans. Robust support vector machine training via convex outlier ablation. In AAAI, pages 536–542. AAAI Press, 2006.
Refer to caption
(a) Letter
Refer to caption
(b) Mushrooms
Refer to caption
(c) Satimage
Figure 4: The classification accuracy of linear SVM under ALFA attack.
Refer to caption
(a) Letter
Refer to caption
(b) Mushrooms
Refer to caption
(c) Satimage
Figure 5: The classification accuracy of SVM with RBF kernel under ALFA attack.
Table 2: F1F_{1} scores on Letter dataset.
ALFA Min-Max
4%4\% 6%6\% 8%8\% 10%10\% 4%4\% 6%6\% 8%8\% 10%10\%
DBSCAN 1.00\bf{1.00} 1.00\bf{1.00} 1.00\bf{1.00} 1.00\bf{1.00} 1.00\bf{1.00} 1.00\bf{1.00} 1.00\bf{1.00} 1.00\bf{1.00}
Slab <0.1<0.1 <0.1<0.1 <0.1<0.1 <0.1<0.1 <0.1<0.1 <0.1<0.1 <0.1<0.1 <0.1<0.1
L22 0.400.40 0.460.46 0.390.39 0.600.60 <0.1<0.1 <0.1<0.1 <0.1<0.1 <0.1<0.1
Loss <0.1<0.1 0.540.54 0.700.70 0.590.59 <0.1<0.1 <0.1<0.1 0.190.19 <0.1<0.1
Knn <0.1<0.1 <0.1<0.1 <0.1<0.1 <0.1<0.1 <0.1<0.1 <0.1<0.1 <0.1<0.1 <0.1<0.1
Table 3: F1F_{1} scores on Mushroom dataset.
ALFA Min-Max
4%4\% 6%6\% 8%8\% 10%10\% 4%4\% 6%6\% 8%8\% 10%10\%
DBSCAN 0.72\bf{0.72} 0.79\bf{0.79} 0.84\bf{0.84} 0.86\bf{0.86} 0.72\bf{0.72} 0.79\bf{0.79} 0.84\bf{0.84} 0.87\bf{0.87}
Slab <0.1<0.1 <0.1<0.1 <0.1<0.1 <0.1<0.1 0.170.17 0.220.22 0.270.27 0.300.30
L22 0.340.34 0.370.37 0.400.40 0.400.40 0.670.67 0.660.66 0.650.65 0.690.69
Loss 0.110.11 0.600.60 0.370.37 <0.1<0.1 0.140.14 0.280.28 0.370.37 <0.1<0.1
Knn <0.1<0.1 <0.1<0.1 <0.1<0.1 <0.1<0.1 0.170.17 0.110.11 <0.1<0.1 <0.1<0.1
Table 4: F1F_{1} scores on Satimage dataset.
ALFA Min-Max
4%4\% 6%6\% 8%8\% 10%10\% 4%4\% 6%6\% 8%8\% 10%10\%
DBSCAN 1.00\bf{1.00} 1.00\bf{1.00} 1.00\bf{1.00} 1.00\bf{1.00} 0.95\bf{0.95} 0.97\bf{0.97} 0.98\bf{0.98} 0.98\bf{0.98}
Slab <0.1<0.1 <0.1<0.1 <0.1<0.1 <0.1<0.1 <0.1<0.1 <0.1<0.1 <0.1<0.1 <0.1<0.1
L22 1.001.00 1.001.00 1.001.00 0.980.98 0.630.63 0.750.75 0.880.88 0.850.85
Loss 0.710.71 0.520.52 0.450.45 0.410.41 0.340.34 0.410.41 0.520.52 0.440.44
Knn <0.1<0.1 <0.1<0.1 <0.1<0.1 <0.1<0.1 <0.1<0.1 0.330.33 0.520.52 0.180.18