This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

oddsidemargin has been altered.
textheight has been altered.
marginparsep has been altered.
textwidth has been altered.
marginparwidth has been altered.
marginparpush has been altered.
The page layout violates the UAI style. Please do not change the page layout, or include packages like geometry, savetrees, or fullpage, which change it for you. We’re not able to reliably undo arbitrary changes to the style. Please remove the offending package(s), or layout-changing commands and try again.

Robust Collective Classification against Structural Attacks

Kai Zhou
Computer Science and Engineering
Washington University in St. Louis
Saint Louis, MO 63105
zhoukai@wustl &Yevgeniy Vorobeychik
Computer Science and Engineering
Washington University in St. Louis
Saint Louis, MO 63105
[email protected]
Abstract

Collective learning methods exploit relations among data points to enhance classification performance. However, such relations, represented as edges in the underlying graphical model, expose an extra attack surface to the adversaries. We study adversarial robustness of an important class of such graphical models, Associative Markov Networks (AMN), to structural attacks, where an attacker can modify the graph structure at test time. We formulate the task of learning a robust AMN classifier as a bi-level program, where the inner problem is a challenging non-linear integer program that computes optimal structural changes to the AMN. To address this technical challenge, we first relax the attacker problem, and then use duality to obtain a convex quadratic upper bound for the robust AMN problem. We then prove a bound on the quality of the resulting approximately optimal solutions, and experimentally demonstrate the efficacy of our approach. Finally, we apply our approach in a transductive learning setting, and show that robust AMN is much more robust than state-of-the-art deep learning methods, while sacrificing little in accuracy on non-adversarial data.

1 INTRODUCTION

Data from various domains can be compactly represented as graphs, such as social networks, citation networks, and protein interaction networks. In such graphical models, nodes, associated with attributes, represent the entities and edges indicate their relations. A common task is to classify nodes into different categories, e.g., classifying an account in a social network as malicious or benign. Collective learning methods (Sen et al., 2008; Taskar et al., 2002) exploit such relational structure in the data for classification tasks. For instance, in hypertext classification, the linked Web-pages tend to possess the same label, as they often share similar contents. Graphical models (Koller et al., 2009), such as Markov networks (Richardson and Domingos, 2006; Taskar et al., 2004b, a; Kok and Domingos, 2005), take into account such linking information in classifying relational data, exhibiting considerable performance improvement on a wide range of classification tasks (Li, 1994; Taskar et al., 2004c; Munoz et al., 2009).

However, making use of relational information in classification also exposes a new vulnerability in adversarial settings. Consider the task of classifying nodes in a social network as malicious or benign. If connectivity is used for this task, malicious parties have an incentive to manipulate information obtained about the network structure to reduce classification accuracy. For example, malicious nodes may ostensibly cut ties among one another, or add links to benign nodes, with the purpose of remaining undetected. While such structural attacks on collective learning have recently emerged (Dai et al., 2018; Zügner et al., 2018; Zügner and Günnemann, 2019), they have focused primarily on defeating neural network embedding-based approaches and transductive learning scenarios.

Our goal is to learn robust Associative Markov Networks (AMN) (Taskar et al., 2004a) that jointly classify the nodes in a given graph, where edges indicate that the connected nodes are likely to have the same label. We formalize the problem of learning robust AMN as a bi-level program in which the inner optimization problem involves optimal structural attacks (adding or deleting edges). The key technical challenge is that even the inner problem is a non-linear integer program, rendering the full optimization problem intractable. We address this challenge by first relaxing the inner adversarial optimization, which allows us to approximate the full bi-level program by a convex quadratic upper bound that we can now efficiently minimize. Our subsequent analysis first exhibits an approximation bound for the adversarial problem, and then an approximation bound on the solutions of our approximate robust AMN approach.

We test our approach on real datasets in several different settings.

First, we show that structural attacks can effectively degrade the accuracy of AMN, with relational information now becoming an Achilles heel instead of an advantage. In contrast, robust AMN degrades gracefully, preserving the advantages of the use of relational information in classification, even with a relatively large adversarial modification of graph structure. In addition, we compared robust AMN to a Graph Convolutional Network (GCN) classifier in a transductive learning setting under a structural attack, and show that robust AMN is significantly more robust than GCN, and is nearly as accurate as GCN on non-adversarial data. This observation is particularly noteworthy given the fact that robust AMN was not specifically designed to be robust to transductive learning attacks, which effectively pollute both the training and test data.

Related Work Our work falls into the realm of learning robust classifiers against decision-time reliability attacks (Vorobeychik and Kantarcioglu, 2018). There is a rich collection of prior work on decision-time attacks targeting a variety of classification approaches, ranging from classical models (Globerson and Roweis, 2006; Torkamani and Lowd, 2013) to deep-learning methods (Eykholt et al., 2018; Grosse et al., 2017). As their countermeasures, several efforts (Li and Vorobeychik, 2014, 2018; Goodfellow et al., 2015; Madry et al., 2018) have been devoted to enhancing the robustness of classifiers in adversarial settings. A fundamental difference in our work is that we are defending against structural attacks that exploit the relations among data points for the purpose of attacking, while most prior work on robust learning considers settings that treat data independently.

More closely related to our focus are several prior studies of the vulnerability and robustness of collective learning models. Torkamani and Lowd (2013) also considered the robustness of AMN to prediction-time attacks, but consider attacks that modify node features, leaving robustness to structural attacks as an open problem. Recently there have been a number of demonstrations of vulnerabilities of graph neural networks to attacks in semi-supervised (transductive learning) settings (Dai et al., 2018; Zügner et al., 2018; Zügner and Günnemann, 2019). Specifically, Dai et al. (2018) and Zügner et al. (2018) proposed targeted attacks aiming at misclassifying a subset of target nodes while Zügner and Günnemann (2019) focused on reliability attacks at training time. While our focus is not on such transductive learning problems (in which attacks also poison the training data), we explore the robustness of our approach in such settings in the experiments below. Other than collective learning, there are a host of works studying structural attacks as well as defense approaches in a more general setting of network analysis tasks, such as link prediction (Zhou et al., 2019b, a), community detection (Waniek et al., 2018) and so on.

2 BACKGROUND

A Markov network is defined over an undirected graph 𝒢=(V,E)\mathcal{G}=(V,E), where a node viVv_{i}\in V, i=1,2,,Ni=1,2,\cdots,N, is associated with a node variable YiY_{i}, representing an object to be classified. We assume there are KK discrete labels, i.e., Yi{1,2,,K}Y_{i}\in\{1,2,\cdots,K\}. At a high level, a Markov network defines a joint distribution of Y={Y1,Y2,,YN}Y=\{Y_{1},Y_{2},\cdots,Y_{N}\}, which is a normalized product of potentials: Pϕ(𝐲)=1Zc𝒞ϕc(𝐲c)P_{\phi}(\mathbf{y})=\frac{1}{Z}\prod_{c\in\mathcal{C}}\phi_{c}(\mathbf{y}_{c}), where ϕc(𝐲c)\phi_{c}({\mathbf{y}_{c}}) is a potential function associated with each clique c𝒞c\in\mathcal{C} in 𝒢\mathcal{G}, and ZZ is a normalization factor. The potential function ϕc(𝐲c)\phi_{c}(\mathbf{y}_{c}) maps a label assignment 𝐲c\mathbf{y}_{c} of the nodes in the clique cc to a non-negative value.

Our focus is the pairwise Associative Markov Networks (AMN) (Taskar et al., 2004a), where each clique is either a node or a pair of nodes. Thus, the joint distribution of YY can be written as

Pϕ(𝐲|𝒢)=1Zi=1Nϕi(yi)(i,j)Eϕij(yi,yj),\displaystyle P_{\phi}(\mathbf{y}|\mathcal{G})=\frac{1}{Z}\prod_{i=1}^{N}\phi_{i}(y_{i})\prod_{(i,j)\in E}\phi_{ij}(y_{i},y_{j}), (1)

where we make explicit the dependency of the probability on the network structure 𝒢\mathcal{G}. In detail, let yiky_{i}^{k} be a binary indicator where yik=1y_{i}^{k}=1 means that node ii is assigned label kk. The node and edge potentials then join such label assignments with the node and edge features, respectively. Specifically, let 𝐱idn\mathbf{x}_{i}\in\mathbb{R}^{d_{n}} and 𝐱ijde\mathbf{x}_{ij}\in\mathbb{R}^{d_{e}} be the feature vectors of node ii and edge (i,j)(i,j). In the log-linear model, the potential functions are defined as logϕi(yik)=𝐰nk𝐱i\log\phi_{i}(y_{i}^{k})=\mathbf{w}_{n}^{k}\cdot\mathbf{x}_{i} and logϕij(yik,yjk)=𝐰𝐞k,k𝐱ij\log\phi_{ij}(y_{i}^{k},y_{j}^{k^{\prime}})=\mathbf{w_{e}}^{k,k^{\prime}}\cdot\mathbf{x}_{ij}, where 𝐰nkdn\mathbf{w}_{n}^{k}\in\mathbb{R}^{d_{n}} and 𝐰ek,kde\mathbf{w}_{e}^{k,k^{\prime}}\in\mathbb{R}^{d_{e}} are node and edge parameters. Note that such parameters are label-specific in that they are different with respect to labels k,kk,k^{\prime}, and the same for the nodes and edges, respectively.

In associative MN, the link (i,j)(i,j) indicates an associative relation between nodes ii and jj, meaning that ii and jj tend to be classified as the same label. Reflected on the edge potentials, it is assumed that ϕij(yik,yjk)=1\phi_{ij}(y_{i}^{k},y_{j}^{k^{\prime}})=1 for any different labels k,k{1,2,,K}k,k^{\prime}\in\{1,2,\cdots,K\} while ϕij(yik,yjk)\phi_{ij}(y_{i}^{k},y_{j}^{k}) equals some value greater than 11. Consequently, the edge potentials associated with those edges connecting differently labeled nodes are 0 in the log space. For simplicity, we write the edge parameters for label kk as 𝐰ek\mathbf{w}_{e}^{k}. Putting everything together, the AMN defines the log of conditional probability as

logP𝐰(𝐲|𝐱,𝒢)\displaystyle\log P_{\mathbf{w}}(\mathbf{y}|\mathbf{x},\mathcal{G}) (2)
=\displaystyle= i=1Nk=1K(𝐰nk𝐱i)yik+(i,j)Ek=1K(𝐰ek𝐱ij)yikyjk\displaystyle\sum_{i=1}^{N}\sum_{k=1}^{K}(\mathbf{w}_{n}^{k}\cdot\mathbf{x}_{i})y_{i}^{k}+\sum_{(i,j)\in E}\sum_{k=1}^{K}(\mathbf{w}_{e}^{k}\cdot\mathbf{x}_{ij})y_{i}^{k}y_{j}^{k}
logZ𝐰(𝐱),\displaystyle-\log Z_{\mathbf{w}}(\mathbf{x}),

where 𝐰\mathbf{w} and 𝐱\mathbf{x} represent all the node and edge parameters and features, and 𝐲\mathbf{y} represents a label assignment. Importantly, the last term Z𝐰(𝐱)Z_{\mathbf{w}}(\mathbf{x}) does not depend on the assignment 𝐲\mathbf{y}.

Two essential tasks of AMN are inference and learning. In inference, one seeks the optimal assignment 𝐲\mathbf{y} that maximizes the log conditional probability logP𝐰(𝐲|𝐱,𝒢)\log P_{\mathbf{w}}(\mathbf{y}|\mathbf{x},\mathcal{G}) (excluding the term Z𝐰(𝐱)Z_{\mathbf{w}}(\mathbf{x})), given observed features 𝐱\mathbf{x} and learned parameters 𝐰\mathbf{w}. Taskar et al. (2004a) showed that in AMN, such an inference problem can be (approximately) solved in polynomial time on arbitrary graph structures (when K=2K=2, the solution is optimal). To learn the weights 𝐰\mathbf{w}, Taskar et al. (2004b) proposed a maximum margin approach by maximizing the gap between the confidence in the true labeling 𝐲^\hat{\mathbf{y}} and any alternative labeling 𝐲\mathbf{y}, denoted by ΔP𝐰(𝐲^,𝐲|𝐱,𝒢)=logP𝐰(𝐲^|𝐱,𝒢)logP𝐰(𝐲|𝐱,𝒢)\Delta P_{\mathbf{w}}(\hat{\mathbf{y}},\mathbf{y}|\mathbf{x},\mathcal{G})=\log P_{\mathbf{w}}(\hat{\mathbf{y}}|\mathbf{x},\mathcal{G})-\log P_{\mathbf{w}}(\mathbf{y}|\mathbf{x},\mathcal{G}). Specifically, they formulate the learning problem as follows:

min\displaystyle\min\quad 12𝐰2+Cξ,\displaystyle\frac{1}{2}||\mathbf{w}||^{2}+C\xi, (3)
s.t. ξmax𝐲𝒴(Δ(𝐲^,𝐲)ΔP𝐰(𝐲^,𝐲|𝐱,𝒢)).\displaystyle\xi\geq\max_{\mathbf{y}\in\mathcal{Y}}(\Delta(\hat{\mathbf{y}},\mathbf{y})-\Delta P_{\mathbf{w}}(\hat{\mathbf{y}},\mathbf{y}|\mathbf{x},\mathcal{G})).

By relaxing the integrality constraints and using strong duality of linear programming, the inner maximization problem is replaced with its dual minimization problem. As a result, the weights 𝐰\mathbf{w} can be efficiently learned through solving the quadratic program with linear constraints.

3 LEARNING ROBUST AMN

3.1 MODEL

In max-margin AMN learning, the exponential set of constraints is replaced by a single most-violated constraint. In the adversarial setting, the attacker is also modifying the structure of 𝒢\mathcal{G}, potentially strengthening the constraint in Eqn. (3). Our robust formulation extends the max-margin learning formulation, taking into account the change caused by modifications of 𝒢\mathcal{G}.

We begin by considering an attacker who can delete existing edges from 𝒢\mathcal{G}, and subsequently show that our model can be extended to the case where an attacker can both add and delete edges. To formalize, we assign a binary decision variable eije_{ij} for each edge (i,j)E(i,j)\in E, where eij=0e_{ij}=0 means that the attacker decides to delete that edge. Then 𝐞=(eij)(i,j)E\mathbf{e}=(e_{ij})_{(i,j)\in E} is the decision vector of the attacker. Let ={𝐞:eij{0,1};(i,j)Eeij|E|D}\mathcal{E}=\{\mathbf{e}:e_{ij}\in\{0,1\};\sum_{(i,j)\in E}e_{ij}\geq|E|-D^{-}\} be the space of all the possible decision vectors, where DD^{-} is a budget on the number of edges that the attacker can delete. Then robust learning can be formulated as

min\displaystyle\min\quad 12𝐰2+Cξ,\displaystyle\frac{1}{2}||\mathbf{w}||^{2}+C\xi, (4)
s.t. ξmax𝐲𝒴,𝐞(Δ(𝐲^,𝐲)ΔP𝐰(𝐲^,𝐲|𝐱,𝒢(𝐞))),\displaystyle\xi\geq\max_{\mathbf{y}\in\mathcal{Y},\mathbf{e}\in\mathcal{E}}(\Delta(\hat{\mathbf{y}},\mathbf{y})-\Delta P_{\mathbf{w}}(\hat{\mathbf{y}},\mathbf{y}|\mathbf{x},\mathcal{G}(\mathbf{e}))),

where 𝒢(𝐞)\mathcal{G}(\mathbf{e}) denotes the modified graph obtained by deleting edges as indicated in 𝐞\mathbf{e} from 𝒢\mathcal{G}. Henceforth, we omit the edge features 𝐱ij\mathbf{x}_{ij} and write 𝐰ek𝐱ij\mathbf{w}_{e}^{k}\mathbf{x}_{ij} as wekw_{e}^{k} to simplify notation. From Eqn. (2), we can rewrite Eqn. (4) as

min\displaystyle\min\quad 12𝐰2+Cξ,\displaystyle\frac{1}{2}||\mathbf{w}||^{2}+C\xi, (5a)
s.t. ξmax𝐲𝒴,𝐞i=1Nk=1K(𝐰nk𝐱i)(yiky^ik)\displaystyle\xi\geq\max_{\mathbf{y}\in\mathcal{Y},\mathbf{e}\in\mathcal{E}}\sum_{i=1}^{N}\sum_{k=1}^{K}(\mathbf{w}_{n}^{k}\mathbf{x}_{i})(y_{i}^{k}-\hat{y}_{i}^{k}) (5b)
+(i,j)Ek=1Kwek(yikyjky^iky^jk)eij\displaystyle+\sum_{(i,j)\in E}\sum_{k=1}^{K}w_{e}^{k}(y_{i}^{k}y_{j}^{k}-\hat{y}_{i}^{k}\hat{y}_{j}^{k})\cdot e_{ij}
+Ni=1Nk=1Ky^ikyik.\displaystyle+N-\sum_{i=1}^{N}\sum_{k=1}^{K}\hat{y}_{i}^{k}y_{i}^{k}.

In this formulation, the attacker’s choice of which edges to remove is captured by the inner maximization problem over 𝐞\mathbf{e} inside Constraint (5b). We call this attack, where edges can only be deleted, Struct-D.

A natural extension is to consider a strong attacker who can simultaneously delete and add edges. We term such an attack Struct-AD (AD for adding and deleting edges). A critical difference between Struct-D and Struct-AD is that the search space of which “non-edges” should be added is significantly larger than that of edges to be removed, since graphs tend to be sparse. This dramatically increases the complexity of solving Eqn. (5) (in particular, of enforcing the constraint associated with computing the optimal attack). To address this, we restrict the attacker to only add edges between two data points that have different labels, resulting in a reduced set of non-edges denoted by E¯\bar{E}. The intuition behind this restriction is that adding edges between nodes with the same label provides useful information for classification, and is unlikely to be a part of an optimal attack; rather, the attacker would focus on adding edges between pairs of nodes with different labels to increase classification noise.

For the Struct-AD attack, we use a binary decision variable e¯ij\bar{e}_{ij} for each non-edge (i,j)E¯(i,j)\in\bar{E}, where e¯ij=1\bar{e}_{ij}=1 means that the attacker chooses to add an edge between ii and jj. As a by-product, each term weky^iky^jke¯ijw_{e}^{k}\hat{y}_{i}^{k}\hat{y}_{j}^{k}\bar{e}_{ij} becomes 0. Then in the formulation of Struct-AD, we only need to add terms (i,j)E¯k=1Kwekyikyjke¯ij\sum_{(i,j)\in\bar{E}}\sum_{k=1}^{K}w_{e}^{k}y_{i}^{k}y_{j}^{k}\bar{e}_{ij} in the attacker’s objective and an extra linear constraint (i,j)E¯e¯ijD+\sum_{(i,j)\in\bar{E}}\bar{e}_{ij}\leq D^{+}, where D+D^{+} is the constraint on the number of edges that the attacker can add. We can further extend the formulation to allow additional restrictions on the attacker, such as limiting the change to node degrees; indeed, we can accommodate the addition of any linear constraints on the attack.

The weights of the robust AMN are learned through solving the bi-level optimization problem above. However, this is a challenging task: first, even the inner maximization problem (optimal structural attack) is a combinatorial optimization problem, and the bi-level nature of the underlying formulation makes it all the more intractable. Next, we present an efficient approximate solution to robust AMN learning with provable guarantees both for the attack subproblem, and to the overall robust learning problem. We focus on formulation (5) to simplify exposition, but all our results generalize to the setting with Struct-AD attacks.

3.2 APPROXIMATE SOLUTION

Our solution is based on approximating the inner-layer non-linear integer program by a linear program (LP). To this end, we first linearize the non-linear terms (product of three binary variables) using standard techniques, and then relax the integrality constraints. Finally, we use LP duality to obtain a single convex quadratic program for robust AMN learning, thereby minimizing an upper bound on the original bi-level program in Eqn. (5). Subsequently, we provide approximation guarantees for the resulting solutions.

To begin, we replace each non-linear term yikyjkeijy_{i}^{k}y_{j}^{k}e_{ij} in Eqn. (5) with a non-negative continuous variable zijkz_{ij}^{k} and add three linear constraints zijkyikz_{ij}^{k}\leq y_{i}^{k}, zijkyjkz_{ij}^{k}\leq y_{j}^{k}, and zijkeijz_{ij}^{k}\leq e_{ij}. We omit the constraint zijkyik+yjk+eij2z_{ij}^{k}\geq y_{i}^{k}+y_{j}^{k}+e_{ij}-2, since we are maximizing the objective (in the inner optimization problem) and the weights wekw_{e}^{k} are non-negative; consequently, the optimal solution takes the value zijk=min{yik,yjk,eij}z_{ij}^{k}=\min\{y_{i}^{k},y_{j}^{k},e_{ij}\}, which is equivalent to yikyjkeijy_{i}^{k}y_{j}^{k}e_{ij} in the binary case. We further relax the integrality constraints on the binary variables 𝐲\mathbf{y} and 𝐞\mathbf{e}, resulting in a linear program to approximate the attacker’s problem, omitting the constant terms in 𝐲\mathbf{y} and 𝐞\mathbf{e}:

max𝐲,𝐞\displaystyle\max_{\mathbf{y},\mathbf{e}}\quad i=1Nk=1K(𝐰nk𝐱iy^ik)yik+(i,j)Ek=1Kwekzijk\displaystyle\sum_{i=1}^{N}\sum_{k=1}^{K}(\mathbf{w}_{n}^{k}\mathbf{x}_{i}-\hat{y}_{i}^{k})y_{i}^{k}+\sum_{(i,j)\in E}\sum_{k=1}^{K}w_{e}^{k}z_{ij}^{k} (6)
(i,j)Ek=1Kweky^iky^jkeij\displaystyle-\sum_{(i,j)\in E}\sum_{k=1}^{K}w_{e}^{k}\hat{y}_{i}^{k}\hat{y}_{j}^{k}e_{ij}
s.t. i,k=1Kyik=1,yik0;(i,j)E,eij[0,1];\displaystyle\forall i,\sum_{k=1}^{K}y_{i}^{k}=1,y_{i}^{k}\geq 0;\forall(i,j)\in E,e_{ij}\in[0,1];
|E|(i,j)EeijD;\displaystyle|E|-\sum_{(i,j)\in E}e_{ij}\leq D^{-};
(i,j)E,k,zijkyik,zijkyjk,zijkeij\displaystyle\forall(i,j)\in E,\forall k,z_{ij}^{k}\leq y_{i}^{k},z_{ij}^{k}\leq y_{j}^{k},z_{ij}^{k}\leq e_{ij}

By LP duality, we can replace the attacker’s maximization problem using its dual minimization problem, which is further integrated into Eqn. (5). Consequently, we can approximate Eqn. (5) by a convex quadratic program (QP), which is presented in the appendix. We can use the same techniques to formulate a corresponding quadratic program for Struct-AD.

When the LP relaxation defined in Eqn. (6) produces integral solutions of 𝐲\mathbf{y} and 𝐞\mathbf{e}, the QP will produce optimal weights. In the case where the LP’s solutions are fractional, we obtain an upper bound on the true objective function. Thus, the approximation quality of LP determines the gap between the true and approximate objective values for the defender and, consequently, the gap between approximate and optimal solutions to robust AMN. In Section 4 we bound this gap.

4 BOUND ANALYSIS

The key to analyzing the bound of the defender’s objective is to devise a well-approximated integral solution to the attacker’s problem as defined in Eqn. (5). On the one hand, such an integral solution bridges the gap between the optimal integral solution of the attacker’s problem and the optimal fractional solution of its LP relaxation. On the other hand, it generates effective structural attacks to test our robust AMN model. We first focus on approximating structural attacks and then analyze how to transfer the bound on the attack performance to the bound on the defender’s objective.

4.1 APPROXIMATING STRUCTURAL ATTACKS

Given fixed weights of the AMN model, the attacker solves the LP in (6) and determine which edges to delete. Unfortunately, Eqn. (6) will produce fractional solutions, meaning that the attacker needs to round the results to obtain a feasible (but not necessarily optimal) attack. Kleinberg and Tardos (2002) proposed a randomized rounding scheme to assign labels to nodes in a class of classification problems with pairwise relationships. We follow the idea and apply it to our case where we are simultaneously assigning KK labels to the nodes and two labels (delete or not) to the edges.

Given the optimal solution (𝐲,𝐞,𝐳)(\mathbf{y}^{*},\mathbf{e}^{*},\mathbf{z}^{*}) of LP (6), the randomized rounding procedure (termed RRound) produces a corresponding integral solution (𝐲I,𝐞I)(\mathbf{y}^{I},\mathbf{e}^{I}). Specifically, RRound rounds 𝐲\mathbf{y}^{*} and 𝐞\mathbf{e}^{*} in phases. In a single phase, we independently draw a label k{1,2,,K}k\in\{1,2,\cdots,K\} and a label b{0,1}b\in\{0,1\} (where b=0b=0 means the corresponding edge is chosen to be deleted) at the beginning. We assign this specific label kk to each node and bb to each edge in a probabilistic way. Specifically, we generate a continuous random number β\beta uniformly from [0,1][0,1]. For each node ii, if βyik\beta\leq y_{i}^{k*}, we assign the label kk to ii (i.e., set yik=1y_{i}^{k}=1). For each edge (i,j)E(i,j)\in E, if βeijb\beta\leq e_{ij}^{b*}, where eij1=eije_{ij}^{1*}=e_{ij}^{*} and eij0=1eije_{ij}^{0*}=1-e_{ij}^{*}, we assign the label bb to (i,j)(i,j), i.e., eij=be_{ij}=b. RRound stops when all nodes and edges are assigned labels. For the auxiliary variable zijkz_{ij}^{k}, it takes the minimum of the rounded yiky_{i}^{k}, yjky_{j}^{k}, and eije_{ij}, which is equivalent to yikyjkeijy_{i}^{k}\cdot y_{j}^{k}\cdot e_{ij} in the binary space. We thus omit 𝐳I\mathbf{z}^{I} and specify the output of RRound as (𝐲I,𝐞I)(\mathbf{y}^{I},\mathbf{e}^{I}).

To simplify presentation, we write the attacker’s approximate objective in Eqn. (6) as A(𝐲,𝐞,𝐳;𝐰)=A1(𝐲,𝐞;𝐰)+A2(𝐳;𝐰)A(\mathbf{y},\mathbf{e},\mathbf{z};\mathbf{w})=A_{1}(\mathbf{y},\mathbf{e};\mathbf{w})+A_{2}(\mathbf{z};\mathbf{w}), where A1(𝐲,𝐞;𝐰)=i=1Nk=1K(𝐰nk𝐱iy^ik)yik(i,j)Ek=1Kweky^iky^jkeijA_{1}(\mathbf{y},\mathbf{e};\mathbf{w})=\sum_{i=1}^{N}\sum_{k=1}^{K}(\mathbf{w}_{n}^{k}\mathbf{x}_{i}-\hat{y}_{i}^{k})y_{i}^{k}-\sum_{(i,j)\in E}\sum_{k=1}^{K}w_{e}^{k}\hat{y}_{i}^{k}\hat{y}_{j}^{k}e_{ij} and A2(𝐳;𝐰)=(i,j)Ek=1KwekzijkA_{2}(\mathbf{z};\mathbf{w})=\sum_{(i,j)\in E}\sum_{k=1}^{K}w_{e}^{k}z_{ij}^{k}. Similarity, the attacker’s true objective specified in (5) (inner problem) is denoted as AI(𝐲,𝐞)=A1I(𝐲,𝐞;𝐰)+A2I(𝐲,𝐞;𝐰)A^{I}(\mathbf{y},\mathbf{e})=A_{1}^{I}(\mathbf{y},\mathbf{e};\mathbf{w})+A_{2}^{I}(\mathbf{y},\mathbf{e};\mathbf{w}), where A1I(𝐲,𝐞;𝐰)=i=1Nk=1K(𝐰nk𝐱iy^ik)yik(i,j)Ek=1Kweky^iky^jkeijA_{1}^{I}(\mathbf{y},\mathbf{e};\mathbf{w})=\sum_{i=1}^{N}\sum_{k=1}^{K}(\mathbf{w}_{n}^{k}\mathbf{x}_{i}-\hat{y}_{i}^{k})y_{i}^{k}-\sum_{(i,j)\in E}\sum_{k=1}^{K}w_{e}^{k}\hat{y}_{i}^{k}\hat{y}_{j}^{k}e_{ij} and A2(𝐳;𝐰)=(i,j)Ek=1KwekyikyjkeijkA_{2}(\mathbf{z};\mathbf{w})=\sum_{(i,j)\in E}\sum_{k=1}^{K}w_{e}^{k}y_{i}^{k}y_{j}^{k}e_{ij}^{k}, with the constraint that 𝐲\mathbf{y} and 𝐞\mathbf{e} take binary values.

Our primary interest is then to analyze the gap between AI(𝐲I,𝐞I;𝐰)A^{I}(\mathbf{y}^{I},\mathbf{e}^{I};\mathbf{w}) and A(𝐲,𝐞,𝐳;𝐰)A(\mathbf{y}^{*},\mathbf{e}^{*},\mathbf{z}^{*};\mathbf{w}). The following lemma is a direct extension of Lemma B.1 of (Taskar et al., 2004a).

Lemma 4.1 ((Taskar et al., 2004a)).

RRound assigns label kk to node ii with probability yiky_{i}^{k*} and assigns label 11 to edge (i,j)(i,j) with probability eije_{ij}^{*}.

Next, Lemma 4.2 gives a bound on the probability that edge (i,j)(i,j) is not deleted and node ii and jj are assigned the same label.

Lemma 4.2.

RRound assigns label kk to node ii and node jj and assigns label 11 to edge (i,j)(i,j) simultaneously with probability at least 1K+4zijk\frac{1}{K+4}z_{ij}^{k*}.

Proof.

We consider the assignment in a single phase when nodes ii, jj and edge (i,j)(i,j) are not assigned any labels at the beginning of this phase. Note that in every phase, there are a total 2K2K combinations of the random draws (k,b)(k,b) with equal probabilities. Then the probability that ii and jj are assigned kk and (i,j)(i,j) is assigned 11 is 12Kmin{yik,yjk,eij1}=12Kzijk\frac{1}{2K}\min\{y_{i}^{k*},y_{j}^{k*},e_{ij}^{1*}\}=\frac{1}{2K}z_{ij}^{k*}, since the optimal solution of LP satisfies zijk=min{yik,yjk,eij1}z_{ij}^{k*}=\min\{y_{i}^{k*},y_{j}^{k*},e_{ij}^{1*}\}. Consider a specific draw (k,b)(k,b), the probability that at least one of ii and jj is assigned label kk or edge (i,j)(i,j) is assigned a label is 12Kmax{yik,yjk,eijb}\frac{1}{2K}\max\{y_{i}^{k*},y_{j}^{k*},e_{ij}^{b*}\}. By summing over all combination of (k,b)(k,b), we have the probability that none of ii, jj, and (i,j)(i,j) is assigned any label is 1(k,b)12Kmax{yik,yjk,eijb}=112Kk=1K(max{yik,yjk,eij0}+max{yik,yjk,eij1})1-\sum_{(k,b)}\frac{1}{2K}\max\{y_{i}^{k*},y_{j}^{k*},e_{ij}^{b*}\}=1-\frac{1}{2K}\sum_{k=1}^{K}(\max\{y_{i}^{k*},y_{j}^{k*},e_{ij}^{0*}\}+\max\{y_{i}^{k*},y_{j}^{k*},e_{ij}^{1*}\}).

Now consider all the phases. Nodes ii and jj and edge (i,j)(i,j) can be assigned label kk and 11 simultaneously in a single phase or in separate phases. For the purpose of deriving the lower-bound, we only consider the probability that ii and jj are assigned kk and (i,j)(i,j) is assigned 11 in a single phase. Summing over all phases, the probability that ii and jj are assigned kk and (i,j)(i,j) is assigned 11 is at least

p=112Kzijk[112Kk=1K(max{yik,yjk,eij0}\displaystyle\sum_{p=1}^{\infty}\frac{1}{2K}z_{ij}^{k*}[1-\frac{1}{2K}\sum_{k=1}^{K}(\max\{y_{i}^{k*},y_{j}^{k*},e_{ij}^{0*}\}
+max{yik,yjk,eij1})]p1\displaystyle+\max\{y_{i}^{k*},y_{j}^{k*},e_{ij}^{1*}\})]^{p-1}
=\displaystyle= zijkk=1K(max{yik,yjk,eij0}+max{yik,yjk,eij1}\displaystyle\frac{z_{ij}^{k*}}{\sum_{k=1}^{K}(\max\{y_{i}^{k*},y_{j}^{k*},e_{ij}^{0*}\}+\max\{y_{i}^{k*},y_{j}^{k*},e_{ij}^{1*}\}}
\displaystyle\geq zijkk=1K(yik+yjk+eij0+yik+yjk+eij1)\displaystyle\frac{z_{ij}^{k*}}{\sum_{k=1}^{K}(y_{i}^{k*}+y_{j}^{k*}+e_{ij}^{0*}+y_{i}^{k*}+y_{j}^{k*}+e_{ij}^{1*})}
=\displaystyle= zijkK+4.\displaystyle\frac{z_{ij}^{k*}}{K+4}.

The first equality comes from i=0di=11d\sum_{i=0}^{\infty}d^{i}=\frac{1}{1-d} for d<1d<1. ∎

Finally, we can use both of these lemmas to prove a lower bound on the expected value of AI(𝐲I,𝐞I;𝐰)A^{I}(\mathbf{y}^{I},\mathbf{e}^{I};\mathbf{w}):

Theorem 4.3.

Let 𝔼[AI(𝐲I,𝐞I;𝐰)]\mathbb{E}[A^{I}(\mathbf{y}^{I},\mathbf{e}^{I};\mathbf{w})] be the expected value of AI(𝐲I,𝐞I;𝐰)A^{I}(\mathbf{y}^{I},\mathbf{e}^{I};\mathbf{w}). Then 𝔼[AI(𝐲I,𝐞I;𝐰)]A1(𝐲,𝐞;𝐰)+1K+4A2(𝐳;𝐰)\mathbb{E}[A^{I}(\mathbf{y}^{I},\mathbf{e}^{I};\mathbf{w})]\geq A_{1}(\mathbf{y}^{*},\mathbf{e}^{*};\mathbf{w})+\frac{1}{K+4}A_{2}(\mathbf{z}^{*};\mathbf{w})

Proof.

From Lemma 4.1 and Lemma 4.2, we have 𝔼[A1I(𝐲I,𝐞I;𝐰)]=A1(𝐲,𝐞;𝐰)\mathbb{E}[A_{1}^{I}(\mathbf{y}^{I},\mathbf{e}^{I};\mathbf{w})]=A_{1}(\mathbf{y}^{*},\mathbf{e}^{*};\mathbf{w}) and 𝔼[A2I(𝐲I,𝐞I;𝐰)]1K+4A2(𝐳;𝐰)\mathbb{E}[A_{2}^{I}(\mathbf{y}^{I},\mathbf{e}^{I};\mathbf{w})]\geq\frac{1}{K+4}A_{2}(\mathbf{z}^{*};\mathbf{w}), giving the claimed result. ∎

Derandomization

We propose a semi-derandomized algorithm (termed Semi-RRound) to select the edges. The key observation is that once 𝐲\mathbf{y} are determined, the objective is a linear function with respect to 𝐞\mathbf{e}. Thus, by selecting those edges in ascending order of their coefficients, the objective is maximized. Let 𝐞I\mathbf{e}^{I^{\prime}} be the decision vector for the edges output by Semi-RRound. Specifically, Semi-RRound takes an output (𝐲I,𝐞I)(\mathbf{y}^{I},\mathbf{e}^{I}) of RRound and compute the objective as a linear function of 𝐞\mathbf{e}. It then deletes edges (i,j)(i,j) (sets eij=0e_{ij}=0) in ascending order of the coefficients of eije_{ij} until DD^{-} edges are deleted. Let 𝐞I\mathbf{e}^{I^{\prime}} be the decision vector for the edges output by Semi-RRound and AI(𝐲I,𝐞I;𝐰)A^{I}(\mathbf{y}^{I},\mathbf{e}^{I^{\prime}};\mathbf{w}) be the corresponding objective. The following Corollary shows that the expected value of AI(𝐲I,𝐞I;𝐰)A^{I}(\mathbf{y}^{I},\mathbf{e}^{I^{\prime}};\mathbf{w}) retains the lower bound in Theorem 4.3.

Corollary 4.4.

𝔼[AI(𝐲I,𝐞I;𝐰)]A1(𝐲,𝐞;𝐰)+1K+4A2(𝐳;𝐰)\mathbb{E}[A^{I}(\mathbf{y}^{I},\mathbf{e}^{I^{\prime}};\mathbf{w})]\geq A_{1}(\mathbf{y}^{*},\mathbf{e}^{*};\mathbf{w})+\frac{1}{K+4}A_{2}(\mathbf{z}^{*};\mathbf{w}).

Proof.

For each (𝐲I,𝐞I)(\mathbf{y}^{I},\mathbf{e}^{I}), Semi-RRound finds the optimal 𝐞I\mathbf{e}^{I^{\prime}} that maximizes AI(𝐲I,𝐞;𝐰)A^{I}(\mathbf{y}^{I},\mathbf{e};\mathbf{w}). Thus, AI(𝐲I,𝐞I;𝐰)AI(𝐲I,𝐞I;𝐰)A^{I}(\mathbf{y}^{I},\mathbf{e}^{I^{\prime}};\mathbf{w})\geq A^{I}(\mathbf{y}^{I},\mathbf{e}^{I};\mathbf{w}) for every rounded output (𝐲I,𝐞I)(\mathbf{y}^{I},\mathbf{e}^{I}). By the definition of expectation, we have 𝔼[AI(𝐲I,𝐞I;𝐰)]A1(𝐲,𝐞;𝐰)+1K+4A2(𝐳;𝐰)\mathbb{E}[A^{I}(\mathbf{y}^{I},\mathbf{e}^{I^{\prime}};\mathbf{w})]\geq A_{1}(\mathbf{y}^{*},\mathbf{e}^{*};\mathbf{w})+\frac{1}{K+4}A_{2}(\mathbf{z}^{*};\mathbf{w}). ∎

4.2 BOUNDING THE DEFENDER’S OBJECTIVE

We rewrite the defender’s objective in Eqn. (5) as

minFD(𝐰)=D(𝐰)+Cmax𝐲𝒴,𝐞AI(𝐲,𝐞;𝐰),\min\ F_{D}(\mathbf{w})=D(\mathbf{w})+C\cdot\max_{\mathbf{y}\in\mathcal{Y},\mathbf{e}\in\mathcal{E}}A^{I}(\mathbf{y},\mathbf{e};\mathbf{w}), (7)

where D(𝐰)=12𝐰2+C(Ni=1Nk=1K𝐰nk𝐱iy^ik)D(\mathbf{w})=\frac{1}{2}||\mathbf{w}||^{2}+C(N-\sum_{i=1}^{N}\sum_{k=1}^{K}\mathbf{w}_{n}^{k}\mathbf{x}_{i}\hat{y}_{i}^{k}). Note that the inner maximization problem implicitly defines a function of 𝐰\mathbf{w}, which we denote by FA(𝐰)=AI(𝐲I,𝐞I;𝐰)F_{A}(\mathbf{w})=A^{I}(\mathbf{y}^{I*},\mathbf{e}^{I*};\mathbf{w}). In solving Eqn. (7), we are approximating AI(𝐲,𝐞;𝐰)A^{I}(\mathbf{y},\mathbf{e};\mathbf{w}) using its LP relaxation; that is, the approximated inner-layer maximization defines a new function F^A(𝐰)=A(𝐲,𝐞,𝐳;𝐰)\hat{F}_{A}(\mathbf{w})=A(\mathbf{y}^{*},\mathbf{e}^{*},\mathbf{z}^{*};\mathbf{w}). Thus, instead of directly minimizing the actual objective FD(𝐰)=D(𝐰)+CFA(𝐰)F_{D}(\mathbf{w})=D(\mathbf{w})+CF_{A}(\mathbf{w}), the defender is minimizing an approximated (upper-bound) objective F^D(𝐰)=D(𝐰)+CF^A(𝐰)\hat{F}_{D}(\mathbf{w})=D(\mathbf{w})+C\hat{F}_{A}(\mathbf{w}).

Let 𝐰=argminFD(𝐰)\mathbf{w}^{*}=\operatorname*{arg\,min}F_{D}(\mathbf{w}) and 𝐰=argminF^D(𝐰)\mathbf{w}^{\diamond}=\operatorname*{arg\,min}\hat{F}_{D}(\mathbf{w}). The following theorem bounds the difference between F^D(𝐰)\hat{F}_{D}(\mathbf{w}^{\diamond}) and FD(𝐰)F_{D}(\mathbf{w}^{*}).

Theorem 4.5.

Let ϵ=max{we1,we2,,weK}\epsilon=\max\{w_{e}^{*1},w_{e}^{*2},\cdots,w_{e}^{*K}\}. Then, F^D(𝐰)FD(𝐰)(K+3)C|E|K+4ϵ\hat{F}_{D}(\mathbf{w}^{\diamond})-F_{D}(\mathbf{w}^{*})\leq\frac{(K+3)C|E|}{K+4}\epsilon.

Proof.

Since 𝐰\mathbf{w}^{\diamond} minimizes F^D(𝐰)\hat{F}_{D}(\mathbf{w}), we have F^D(𝐰)F^D(𝐰)\hat{F}_{D}(\mathbf{w}^{\diamond})\leq\hat{F}_{D}(\mathbf{w}^{*}). Then

F^D(𝐰)FD(𝐰)\displaystyle\hat{F}_{D}(\mathbf{w}^{\diamond})-F_{D}(\mathbf{w}^{*}) (8)
\displaystyle\leq F^D(𝐰)FD(𝐰)\displaystyle\ \hat{F}_{D}(\mathbf{w}^{*})-F_{D}(\mathbf{w}^{*})
=\displaystyle= CF^A(𝐰)CFA(𝐰)\displaystyle\ C\cdot\hat{F}_{A}(\mathbf{w}^{*})-C\cdot F_{A}(\mathbf{w}^{*})
=\displaystyle= C(A(𝐲,𝐞,𝐳;𝐰)AI(𝐲I,𝐞I;𝐰))\displaystyle\ C(A(\mathbf{y}^{*},\mathbf{e}^{*},\mathbf{z}^{*};\mathbf{w}^{*})-A^{I}(\mathbf{y}^{I*},\mathbf{e}^{I*};\mathbf{w}^{*}))
\displaystyle\leq CK+3K+4A2(𝐳;𝐰).\displaystyle\ C\frac{K+3}{K+4}A_{2}(\mathbf{z}^{*};\mathbf{w}^{*}). (9)

The second inequality is from the result of Corollary 4.4 and the fact that (𝐲I,𝐞I)(\mathbf{y}^{I*},\mathbf{e}^{I*}) is the integral optimal solution. Then A2(𝐳;𝐰)=(i,j)Ek=1KwekzijkA_{2}(\mathbf{z}^{*};\mathbf{w}^{*})=\sum_{(i,j)\in E}\sum_{k=1}^{K}w_{e}^{*k}z_{ij}^{*k}. As zijkyikz_{ij}^{*k}\leq y_{i}^{*k} and k=1Kyik=1\sum_{k=1}^{K}y_{i}^{*k}=1 in the optimal fractional solution, we have A2(𝐳;𝐰)(i,j)Eϵk=1Kyik=|E|ϵA_{2}(\mathbf{z}^{*};\mathbf{w}^{*})\leq\sum_{(i,j)\in E}\epsilon\sum_{k=1}^{K}y_{i}^{*k}=|E|\epsilon. Thus F^D(𝐰)FD(𝐰)(K+3)C|E|K+4ϵ\hat{F}_{D}(\mathbf{w}^{\diamond})-F_{D}(\mathbf{w}^{*})\leq\frac{(K+3)C|E|}{K+4}\epsilon. ∎

We note that the bound analysis can be extended to the case where the attacker can delete and add links simultaneously. One difference is that instead of using the inequality zijyikz_{ij}^{*}\leq y_{i}^{*k}, we can use z¯ijke¯ij\bar{z}_{ij}^{*k}\leq\bar{e}_{ij}^{*}, where z¯ijk\bar{z}_{ij}^{k} is the auxiliary variable to represent the product yikyjke¯ijy_{i}^{k}y_{j}^{k}\bar{e}_{ij}. This leads to a bound (K+3)C(|E|+KD+)K+4ϵ\frac{(K+3)C(|E|+K\cdot D^{+})}{K+4}\epsilon, where D+D^{+} is the budget on the number of added edges.

5 EXPERIMENTS

We test the performance of the robust AMN classifier (henceforth R-AMN) under structural attacks as well as a recently proposed attack on Graph Convolutional Networks (GCN). We focus on the binary classification task as AMN will in this case learn the optimal weights.

5.1 DATASETS

We consider four real-world datasets: Reuters, WebKB, Cora and CiteSeer. For the Reuters dataset (Taskar et al., 2004a; Torkamani and Lowd, 2013), we follow the procedure in (Taskar et al., 2004a) to extract four categories (“trade”, “crude”, “grain”, and “money-fx”) of documents where each document belongs to a unique category. We create a binary dataset by fixing “trade” as the positive class and randomly drawing an equal number of documents from the other three categories. We extract the bag-of-words representation for each document. To construct the links, we connect each document to its three closest neighbors in terms of the cosine distance of TF-IDF representations, resulting in a graph of 862862 nodes and 18601860 edges. The WebKB dataset (Craven et al., 1998) contains webpages (represented as binary vectors of dimension 17031703) from four universities, which are classified into five classes. We fix “student” as the positive class and the rest as the negative class. We connect each webpage to its 33 closest neighbors based on cosine distance, resulting in a network of 877877 nodes and 22822282 edges. Cora (McCallum et al., 2000) and CiteSeer (Giles et al., 1998) are two citation networks, where the nodes in the graph are the bag-of-words representations of papers and edges are the citation relations among them. For Cora, we use the “Prob-methods” as the positive class and randomly draw an equal number of papers from the other categories to form the negative class, resulting in a subgraph of 852852 nodes and 838838 edges. Similarly for CiteSeer, we fix “agents” as the positive class, resulting in a subgraph of 11921192 nodes and 953953 edges. Cora and CiteSeer have more sparse graph structures than the Reuters and WebKB. Moreover, Reuters and WebKB tend to be regular graphs while a few nodes in Cora and CiteSeer have large degrees. We split each dataset evenly into training and test data for our experiments.

Refer to caption
(a) Reuters-H
Refer to caption
(b) Reuters-L
Refer to caption
(c) Cora-H
Refer to caption
(d) Cora-L
Refer to caption
(e) WebKB-H
Refer to caption
(f) WebKB-L
Refer to caption
(g) CiteSeer-H
Refer to caption
(h) CiteSeer-L
Figure 1: Accuracies of AMN (dotted lines) and R-AMN (solid lines) under edge deletion attacks.

5.2 ROBUSTNESS AGAINST STRUCTURAL ATTACKS

The AMN model jointly uses node features and structural information to do classification. The impact of structural attacks depends on the importance of structure in classification, which in turn depends on the quality of information captured by node features.

Refer to caption
(a) Reuters-H
Refer to caption
(b) Reuters-L
Refer to caption
(c) Cora-H
Refer to caption
(d) Cora-L
Refer to caption
(e) WebKB-H
Refer to caption
(f) WebKB-L
Refer to caption
(g) CiteSeer-H
Refer to caption
(h) CiteSeer-L
Figure 2: Accuracies of AMN (dotted lines) and R-AMN (solid lines) under edge addition and deletion attacks

To study the impact of such node-specific information, we consider two settings: high-discriminative (H-Dis), which uses more features, and low-discriminative (L-Dis), which uses a smaller subset of features. Specifically, for the Reuters dataset, we select 200200 top (in terms of frequencies) features in the H-Dis case (termed Reuters-H henceforth) and randomly selected 200200 out of the top 600600 features in the L-Dis case (Reuters-L). For WebKB, we randomly select 200200 in L-Dis case (WebKB-L) and 300300 in H-Dis case (WebKB-H) out of the 17031703 features. For Cora, we use all the 14331433 features in H-Dis case (Cora-H) and randomly select 600600 features in the L-Dis case (Cora-L). For CiteSeer, we randomly select 500500 in L-Dis case (CiteSeer-L) and 10001000 in H-Dis case (CiteSeer-H) out of the 37033703 features. In our experiments, we use cross-validation on the training set to tune the two parameters of the R-AMN: the trade-off parameter CC and the adversarial budget bb. We note that when tuning R-AMN, the defender has no knowledge of the strength of the attacker. We use a simulated attacker that can modify the validation set (in cross-validation) with b=0.1b=0.1, meaning that the attacker can change 10%10\% of the edges.

We consider a baseline attacker that can randomly add links between nodes belonging to different classes and delete links between nodes with the same labels. We term such an attack Struct-RSAD (Remove Same and Add Different). In the case where the attacker is only allowed to delete links, we term the attack as Struct-RS. We test R-AMN under four structural attacks: Struct-D and Struct-AD (our attacks, deleting links in the former, and adding or deleting in the latter), and Struct-RS and Struct-RSAD (heuristic baseline attacks above). We denote the performance of R-AMN exposed to these attacks as Robust-D, Robust-AD, Robust-RS, Robust-RSAD, respectively. We overload the notations by denoting the performance of AMN under the four attacks as Struct-D, Struct-AD, Struct-RS, Struct-RSAD, respectively.

Fig. 1 and Fig. 2 show the average accuracy (over 2020 independent data splits) of AMN (dotted lines) and R-AMN (solid lines) under structural attacks as well as the accuracy of a linear SVM classifier. First, by modifying a small portion of the links in the graph, the accuracy of AMN drops below that of SVM (which does not exploit relations), meaning that relations among data points indeed introduce extra vulnerabilities. Moreover, structural attacks tend to be more severe when the node features are less discriminative (the L-Dis case), where linking information plays a relatively more important role in classification. Notably, the accuracy of R-AMN drops significantly slower than that of AMN under structural attacks in all settings and stays above that of the SVM, even when a relatively large fraction of the node connections are modified. These show that robust AMN preserves the benefits of using structural information even if network structure is maliciously modified.

Refer to caption
(a) Reuters-H
Refer to caption
(b) Reuters-L
Refer to caption
(c) WebKB-H
Refer to caption
(d) WebKB-L
Figure 3: R-AMN and R-GCN under deep-attack; GCN under deep-attack and Struct-AD attack.
Refer to caption
(a) Reuters-H
Refer to caption
(b) Reuters-L
Refer to caption
(c) WebKB-H
Refer to caption
(d) WebKB-L
Figure 4: R-AMN and GCN on non-adversary data as graphs are purified, e.g. R-AMN(0.50.5) stands for R-AMN when noisy edges are deleted with probability 0.50.5.

5.3 ROBUSTNESS AGAINST DEEP LEARNING BASED ATTACKS

Having observed that our approach for robust AMN is indeed robust against our newly designed attacks on AMN, it is natural to wonder whether robust AMN remains robust against recent attacks on graph convolutional networks (GCNs). The answer is non-obvious: on the one hand, attacks on GCN may not transfer to AMN—although there is ample prior evidence that attacks do often transfer from one learning approach to another (Vorobeychik and Kantarcioglu, 2018); on the other hand, attacks on GCN target transductive learning, and as such also poison the data on which the AMN would be trained. To this end, we use a recent structural attack on GCN (termed deep-attack) proposed in (Zügner and Günnemann, 2019), which has demonstrated impressive attack performance and transferability to neural network embedding-based approaches. We thus test our R-AMN against this deep-attack, compared to GCN. In addition, we compare the performance of R-AMN with a GCN based classifier (Kipf and Welling, 2016) on non-adversarial data.

GCN classifies nodes in a transductive setting, where labeled and unlabeled nodes reside in the same graph; while R-AMN is trained over a training graph, and makes predictions over an unseen test graph. To adapt R-AMN to the transductive setting, we use deep-attack (with the same configurations as in (Zügner and Günnemann, 2019), e.g., strongest “Meta-Self” mode and 0.1/0.1/0.80.1/0.1/0.8 train/validation/test split) to modify the training graph and test graph separately (these can be subsets of the same larger graph, as is the case in transductive learning). Then we train R-AMN on the attacked training graph and test it on the attacked test graph. We also test the performance of GCN under deep-attack and our proposed structural attack Struct-AD on the test graph. In addition, we test a robust GCN model (termed R-GCN) under deep-attack, which is based on adversarial training approach proposed in (Xu et al., 2019) on the test graph. The accuracies of R-AMN, GCN, and R-GCN under these attacks are presented in Fig. 5 for the Reuters and WebKB datasets (the appendix presents similar results for the Cora and CiteSeer datasets). The main observation is that where R-GCN cannot defend against such deep-attack in a transductive setting, our proposed R-AMN is essentially invariant under deep-attack, in contrast to GCN, which is highly vulnerable to this attack, and also quite vulnerable to our proposed structural attacks aimed at AMN.

Finally, we compare R-AMN and GCN on non-adversarial data, which are evenly split into training and test graphs. We note that R-AMN ignores the links between the training and test (sub-)graphs. The results are shown in Fig. 4. To interpret these, consider first GCN/0 and R-AMN/0 bars, which correspond to the direct performance comparison on the given data. We can observe that the difference in accuracy between R-AMN and GCN in these cases is either small (on WebKB data) or, in fact, R-AMN actually outperforms GCN (on Reuters data). It’s this latter observation that is surprising. The reason is that Reuters data contains 10%\sim 10\% of noisy links, that is, links connecting pairs of nodes with different labels. This can be viewed as another symptom of ragility of GCN, but in any case, we next consider what happens when we remove each noisy link with some probability (p=0.5p=0.5 or p=0.8p=0.8). The results are presented as bars with R-AMN/pp and GCN/pp, where pp corresponds to this probability of removed noisy links, and, as expected, GCN performance improves as we improve data quality. This improvement is significant when the graph is not particularly noisy (WebKB), but the gap between R-AMN and GCN remains relatively small when enough noisy links remain (Reuters, as well as Cora and CiteSeer; see the appendix, Fig. 6). This further attests to greater robustness of R-AMN, but does exhibit some cost in terms of accuracy on non-adversarial data, if this data is sufficiently high quality.

6 CONCLUSION

We study robustness of the associative Markov network classifier under test-time attacks on network structure, where an attacker can delete and/or add links in the underlying graph. We formulate the task of robust learning as a bi-level program and propose an approximation algorithm to efficiently solve it. Our experiments on real-world datasets demonstrate that the performance of robust AMN degrades gracefully even under large adversarial modifications of the graph structure, preserving the advantages of using structural information in classifying relational data. We additionally compare robust AMN with the state-of-the-art deep learning based approaches in the transductive setting and demonstrate that robust AMN is significantly more robust to structural perturbations compared to deep graph embedding methods while sacrificing little performance on non-adversarial data, except when network data is of extremely high quality (a rarity in practice).

Acknowledgements

This work was partially supported by the National Science Foundation (grants IIS-1905558 (CAREER) and IIS-1903207) and Army Research Office (grants W911NF1810208 (MURI) and W911NF1910241).

References

  • Craven et al. [1998] M. Craven, A. McCallum, D. PiPasquo, T. Mitchell, and D. Freitag. Learning to extract symbolic knowledge from the world wide web. Technical report, Carnegie-mellon univ pittsburgh pa school of computer Science, 1998.
  • Dai et al. [2018] H. Dai, H. Li, T. Tian, X. Huang, L. Wang, J. Zhu, and L. Song. Adversarial attack on graph structured data. In International Conference on Machine Learning, pages 1115–1124, 2018.
  • Eykholt et al. [2018] K. Eykholt, I. Evtimov, E. Fernandes, B. Li, A. Rahmati, C. Xiao, A. Prakash, T. Kohno, and D. Song. Robust physical-world attacks on deep learning visual classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 1625–1634, 2018.
  • Giles et al. [1998] C. L. Giles, K. D. Bollacker, and S. Lawrence. Citeseer: An automatic citation indexing system. In ACM DL, pages 89–98, 1998.
  • Globerson and Roweis [2006] A. Globerson and S. Roweis. Nightmare at test time: robust learning by feature deletion. In Proceedings of the 23rd international conference on Machine learning, pages 353–360. ACM, 2006.
  • Goodfellow et al. [2015] I. Goodfellow, J. Shlens, and C. Szegedy. Explaining and harnessing adversarial examples. In International Conference on Learning Representations, 2015.
  • Grosse et al. [2017] K. Grosse, N. Papernot, P. Manoharan, M. Backes, and P. McDaniel. Adversarial examples for malware detection. In European Symposium on Research in Computer Security, pages 62–79. Springer, 2017.
  • Kipf and Welling [2016] T. N. Kipf and M. Welling. Semi-supervised classification with graph convolutional networks. In International Conference on Learning Representations, 2016.
  • Kleinberg and Tardos [2002] J. Kleinberg and E. Tardos. Approximation algorithms for classification problems with pairwise relationships: Metric labeling and markov random fields. Journal of the ACM (JACM), 49(5):616–639, 2002.
  • Kok and Domingos [2005] S. Kok and P. Domingos. Learning the structure of markov logic networks. In Proceedings of the 22nd international conference on Machine learning, pages 441–448. ACM, 2005.
  • Koller et al. [2009] D. Koller, N. Friedman, and F. Bach. Probabilistic graphical models: principles and techniques. MIT press, 2009.
  • Li and Vorobeychik [2014] B. Li and Y. Vorobeychik. Feature cross-substitution in adversarial classification. In Advances in neural information processing systems, pages 2087–2095, 2014.
  • Li and Vorobeychik [2018] B. Li and Y. Vorobeychik. Evasion-robust classification on binary domains. ACM Transactions on Knowledge Discovery from Data (TKDD), 12(4):50, 2018.
  • Li [1994] S. Z. Li. Markov field models in computer vision. In European conference on computer vision, pages 361–370. Springer, 1994.
  • Madry et al. [2018] A. Madry, A. Makelov, L. Schmidt, D. Tsipras, and A. Vladu. Towards deep learning models resistant to adversarial attacks. In International Conference on Learning Representations, 2018.
  • McCallum et al. [2000] A. K. McCallum, K. Nigam, J. Rennie, and K. Seymore. Automating the construction of internet portals with machine learning. Information Retrieval, 3(2):127–163, 2000.
  • Munoz et al. [2009] D. Munoz, J. A. Bagnell, N. Vandapel, and M. Hebert. Contextual classification with functional max-margin markov networks. In 2009 IEEE Conference on Computer Vision and Pattern Recognition, pages 975–982. IEEE, 2009.
  • Richardson and Domingos [2006] M. Richardson and P. Domingos. Markov logic networks. Machine learning, 62(1-2):107–136, 2006.
  • Sen et al. [2008] P. Sen, G. Namata, M. Bilgic, L. Getoor, B. Galligher, and T. Eliassi-Rad. Collective classification in network data. AI magazine, 29(3):93–93, 2008.
  • Taskar et al. [2002] B. Taskar, P. Abbeel, and D. Koller. Discriminative probabilistic models for relational data. In Proceedings of the Eighteenth conference on Uncertainty in artificial intelligence, pages 485–492. Morgan Kaufmann Publishers Inc., 2002.
  • Taskar et al. [2004a] B. Taskar, V. Chatalbashev, and D. Koller. Learning associative markov networks. In Proceedings of the twenty-first international conference on Machine learning, page 102. ACM, 2004a.
  • Taskar et al. [2004b] B. Taskar, C. Guestrin, and D. Koller. Max-margin markov networks. In Advances in neural information processing systems, pages 25–32, 2004b.
  • Taskar et al. [2004c] B. Taskar, M.-F. Wong, P. Abbeel, and D. Koller. Link prediction in relational data. In Advances in neural information processing systems, pages 659–666, 2004c.
  • Torkamani and Lowd [2013] M. Torkamani and D. Lowd. Convex adversarial collective classification. In International Conference on Machine Learning, pages 642–650, 2013.
  • Vorobeychik and Kantarcioglu [2018] Y. Vorobeychik and M. Kantarcioglu. Adversarial machine learning. Synthesis Lectures on Artificial Intelligence and Machine Learning, 12(3):1–169, 2018.
  • Waniek et al. [2018] M. Waniek, T. P. Michalak, M. J. Wooldridge, and T. Rahwan. Hiding individuals and communities in a social network. Nature Human Behaviour, 2(2):139–147, 2018.
  • Xu et al. [2019] K. Xu, H. Chen, S. Liu, P.-Y. Chen, T.-W. Weng, M. Hong, and X. Lin. Topology attack and defense for graph neural networks: An optimization perspective. In International Joint Conference on Artificial Intelligence, 2019.
  • Zhou et al. [2019a] K. Zhou, T. Michalak, and Y. Vorobeychik. Adversarial robustness of similarity-based link prediction. In IEEE International Conference on Data Mining, 2019a.
  • Zhou et al. [2019b] K. Zhou, T. P. Michalak, M. Waniek, T. Rahwan, and Y. Vorobeychik. Attacking similarity-based link prediction in social networks. In Proceedings of the 18th International Conference on Autonomous Agents and MultiAgent Systems, pages 305–313. International Foundation for Autonomous Agents and Multiagent Systems, 2019b.
  • Zügner and Günnemann [2019] D. Zügner and S. Günnemann. Adversarial attacks on graph neural networks via meta learning. In International Conference on Learning Representations (ICLR), 2019.
  • Zügner et al. [2018] D. Zügner, A. Akbarnejad, and S. Günnemann. Adversarial attacks on neural networks for graph data. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pages 2847–2856. ACM, 2018.

Appendix

Appendix A Formulation of Convex Quadratic Program

We explicitly write out the convex quadratic program for learning robsut AMN, which is omitted in the main paper. By LP duality, we can replace the attacker’s maximization problem using its dual minimization problem, which is further integrated into Eqn. (5). Consequently, we can approximate Eqn. (5) by the following convex quadratic program:

min12𝐰2+C(Ni=1Nk=1K𝐰nk𝐱iy^ik+i=1Nti+(i,j)EpijDtD)\displaystyle\min\ \frac{1}{2}||\mathbf{w}||^{2}+C(N-\sum_{i=1}^{N}\sum_{k=1}^{K}\mathbf{w}_{n}^{k}\mathbf{x}_{i}\hat{y}_{i}^{k}+\sum_{i=1}^{N}t_{i}+\sum_{(i,j)\in E}p_{ij}-D^{-}\cdot t_{D})
s.t.i,k,ti(i,j),(j,i)Etijk𝐰nk𝐱i+y^ik0,\displaystyle\text{s.t.}\quad\forall i,k,\quad t_{i}-\sum_{(i,j),(j,i)\in E}t_{ij}^{k}-\mathbf{w}_{n}^{k}\mathbf{x}_{i}+\hat{y}_{i}^{k}\geq 0,
(i,j)E,k,sijk+tijk+tjikwek0,sijk,tijk,tjik0,\displaystyle\forall(i,j)\in E,k,\quad s_{ij}^{k}+t_{ij}^{k}+t_{ji}^{k}-w_{e}^{k}\geq 0,\ s_{ij}^{k},t_{ij}^{k},t_{ji}^{k}\geq 0,
(i,j)E,pijk=1KsijktD+(i,j)Ek=1Kweky^iky^jk0,pij,tD0.\displaystyle\forall(i,j)\in E,\quad p_{ij}-\sum_{k=1}^{K}s_{ij}^{k}-t_{D}+\sum_{(i,j)\in E}\sum_{k=1}^{K}w_{e}^{k}\hat{y}_{i}^{k}\hat{y}_{j}^{k}\geq 0,\ p_{ij},t_{D}\geq 0. (10)

The minimization is over the weights 𝐰\mathbf{w} and the dual variables ti,pij,sijk,tijk,tjik,tDt_{i},p_{ij},s_{ij}^{k},t_{ij}^{k},t_{ji}^{k},t_{D}.

Appendix B Additional Experiment Results

We compare R-AMN and GCN under the deep-attack as well as on non-adversarial data on the Cora and CiteSeer datasets in the same experiment settings as in the main paper. Specifically, in Fig. 5, ”R-AMN/deep-attack” shows the accuracies of R-AMN under deep-attack with various degrees of graph perturbations, where the train graph and test graph are attacked by deep-attack separately. It demonstrates that R-AMN is robust to deep-attack even with relatively large structural perturbations. ”GCN/deep-attack” and ”GCN/Struct-AD” show the accuracies of GCN under deep-attack and our proposed Struct-AD attack, respectively. Generally, deep-attack is a much more effective method to attack GCN models. Fig. 6 demonstrated that on non-adversarial data, the performances of R-AMN and GCN are comparable.

Refer to caption
(a) Cora-H
Refer to caption
(b) Cora-L
Refer to caption
(c) CiteSeer-H
Refer to caption
(d) CiteSeer-L
Figure 5: R-AMN and R-GCN under deep attack; GCN under deep attack and Struct-AD attack.
Refer to caption
(a) Cora-H
Refer to caption
(b) Cora-L
Refer to caption
(c) CiteSeer-H
Refer to caption
(d) CiteSeer-L
Figure 6: R-AMN and GCN on non-adversary data as graphs are purified, e.g. R-AMN(0.5) stands for R-AMN when noisy edges are deleted with probability 0.5.