# adversarial_attacks_on_deep_graph_matching__483d155c.pdf Adversarial Attacks on Deep Graph Matching Zijie Zhang Auburn University zzz0092@auburn.edu Zeru Zhang Auburn University zzz0054@auburn.edu Yang Zhou Auburn University yangzhou@auburn.edu Yelong Shen Microsoft Dynamics 365 AI yeshe@microsoft.com Ruoming Jin Kent State University rjin1@kent.edu Dejing Dou University of Oregon, Baidu Research dou@cs.uoregon.edu, doudejing@baidu.com Despite achieving remarkable performance, deep graph learning models, such as node classification and network embedding, suffer from harassment caused by small adversarial perturbations. However, the vulnerability analysis of graph matching under adversarial attacks has not been fully investigated yet. This paper proposes an adversarial attack model with two novel attack techniques to perturb the graph structure and degrade the quality of deep graph matching: (1) a kernel density estimation approach is utilized to estimate and maximize node densities to derive imperceptible perturbations, by pushing attacked nodes to dense regions in two graphs, such that they are indistinguishable from many neighbors; and (2) a meta learning-based projected gradient descent method is developed to well choose attack starting points and to improve the search performance for producing effective perturbations. We evaluate the effectiveness of the attack model on real datasets and validate that the attacks can be transferable to other graph learning models. 1 Introduction Graph matching is one of the most important research topics in the graph domain, which aims to match the same entities (i.e., nodes) across two or more graphs [91, 98, 43, 46, 48, 72, 54, 105, 13, 75]. It has been widely applied to many real-world applications ranging from protein network matching in bioinformatics [33, 63], user account linking in different social networks [62, 51, 100, 37, 101, 21, 38], and knowledge translation in multilingual knowledge bases [87, 124], to geometric keypoint matching in computer vision [22]. Existing research efforts on graph matching can be classified into three broad categories: (1) structure-based techniques, which rely only upon the topological information to match two or multiple input graphs [43, 49, 95, 46, 54, 105, 13, 96, 84, 40, 67, 57, 38, 24]; (2) attribute-based approaches, which utilize highly discriminative structure and/or attribute features for ensuring the matching effectiveness [93, 94, 51, 10, 65, 16, 88, 100, 28, 39, 18, 97, 50, 52, 22]; and (3) heterogeneous methods, which employ heterogeneous structural, content, spatial, and temporal features to further improve the matching performance [92, 34, 44, 98, 83, 99, 77, 59, 102, 103, 21]. Recent literature has shown that both traditional and deep graph learning algorithms remain highly sensitive to adversarial attacks, i.e., carefully designed small perturbations in graph structure and attributes can cause the models to produce wrong prediction results [14, 126, 64, 125, 123, 69, 90, 74, 45, 85, 80, 127]. We have witnessed various effective attack models to cause failures of 34th Conference on Neural Information Processing Systems (Neur IPS 2020), Vancouver, Canada. node classification [14, 126, 74, 86, 125, 71, 19, 70], community detection [9, 78, 7, 41], network embedding [6, 4, 5], link prediction [104], similarity search [15], malware detection [30], and knowledge graph embedding [90]. However, there is still a paucity of analyses of the vulnerability of graph matching under adversarial attacks, which is much more difficult to study. Most of the existing models to fool other graph learning tasks conduct the adversarial attacks on a single graph but the graph matching task analyzes both intra-graph and inter-graph interactions of multiple graphs. In this work, we aim to answer the following questions: (1) Are graph matching algorithms sensitive to small perturbation of graph structure? (2) How do we develop effective and imperceptible perturbations for degrading the performance of deep graph matching models? A large number of research advances in adversarial attacks on graph data utilize iterative gradientbased methods to produce effective adversarial perturbations that fool a graph learning model [68, 76, 14, 125, 86, 71, 89, 8]. However, a recent study reports that the iterative gradient-based methods, such as Fast Gradient Sign Method (FGSM) [26] and Projected Gradient Descent (PGD) [47], start the attacks from original examples and add perturbations monotonically along the direction of gradient descent, resulting in a lack of diversity and adaptability of generated iterative trajectories [61]. This often leads to invalid attacks since the iterative trajectories have difficulties crossing decision boundary of target learning model with small perturbation. Can we find a shortcut across the decision boundary to derive more effective attacks by beginning from good attack starting points in the graph matching? Traditionally, graph matching techniques are based on the assumption of feature consistency across graphs: Two nodes in different graphs are more likely to be found to be matching if they have similar topological and/or attribute features in respective graphs [98, 93, 10, 17, 28, 96]. These methods compute the similarity (or distance) scores between pairwise nodes across graphs and choose the node pairs with largest similarity (or smallest distance) as matching results [101, 88, 39, 38]. Intuitively, if an attacker perturbs a node by throwing it into a dense region in the graph with many similar nodes, i.e., a pile of nodes similar to each other, such that this attacked node is similar to many neighbors, then it is hard for humans or defender programs to recognize it from the node pile. In addition, if two matched nodes are simultaneously moved to such dense regions in respective graphs, then this dramatically increases the difficulty in matching them correctly among many similar candidate nodes. To our best knowledge, this work is the first to study adversarial attacks on graph matching. We propose to utilize kernel density estimation (KDE) technique to estimate the probability density function of nodes in two graphs, to understand the intrinsic distribution of graphs. By maximizing the estimated densities of nodes to be attacked, we push them to dense regions in respective graphs to generate adversarial nodes that are indistinguishable from many neighbors in dense regions. This increases the chance of producing wrong matching results as well as reduces the risk of perturbations being detected by humans or by defender programs. Our analysis is the first to introduce the KDE technique to conduct imperceptible attacks on graph data. Searching for good attack starting points in large graphs is computationally inefficient. We develop a meta learning-based projected gradient descent (MLPGD) model to quickly adapt to a variety of new search tasks on multiple batches of target nodes for deriving effective attacks. However, the MLPGD model is non-smooth and non-differential, as the perturbation is a multi-step process and the projection at each step is non-differential. A Gaussian smoothing method is designed to approximate a smoothed model, and a Monte Carlo REINFORCE method is used to estimate the model gradient. Empirical evaluation on real datasets demonstrates the superior performance of the GMA model against several state-of-the-art adversarial attack methods on graph data. Moreover, we validate that the attack strategies are transferable to other popular graph learning models in Appendix A.2. 2 Problem Definition Given two graphs G1 and G2 to be matched, each is denoted as Gs = (V s, Es) (s = 1 or 2), where V s = {vs 1, , vs Ns} is the set of N s nodes and Es = {(vs i , vs j) : 1 i, j N s} is the set of edges. Each Gs has an N s N s binary adjacency matrix As, where each entry As ij = 1 if there exists an edge (vs i , vs j) Es; otherwise As ij = 0. As i: specifies the ith row vector of As. In this paper, if there are no specific descriptions, we use vs i to denote a node vs i itself and its representation As i:, i.e., vs i = As i: and we utilize vs ij to specify the jth dimension of vs i , i.e., vs ij = As ij. The dataset is divided into two disjoint sets D and D. The former denotes a set of known matched node pairs D = {(v1 i , v2 k)|v1 i v2 k, v1 i V 1, v2 k V 2}, where v1 i v2 k indicates that two nodes v1 i and v2 k belong to the same entity. The latter, denoted by D = {(v1 i , v2 k)|v1 i v2 k, v1 i V 1, v2 k V 2}, is used to evaluate the graph matching performance, where the nodes (but not their matchings) are also observed during training. The goal of graph matching is to utilize D as the training data to identify the one-to-one matching relationships between nodes v1 i and v2 k in the test data D. By following the same idea in existing efforts [101, 88, 39, 38], this paper aims to minimize the distances between projected source nodes M(v1 i ) D and target ones v2 k D . The node pairs (v1 i , v2 k) D with the smallest distances are selected as the matching results. min M L where L = E(v1 i ,v2 k) D M(v1 i ) v2 k 2 2 (1) where M denotes an injective one-to-one matching function M : v1 i V 1 7 v2 k V 2. The adversarial attack problem is defined as maximally degrading the matching performance of M on the test data D by injecting edge perturbations (including edge insertion and deletion) into Gs = (V s, Es) (s = 1 or 2), leading to two adversarial graphs ˆGs = ( ˆV s, ˆEs). We assume the attacker has limited capability, so that he/she can only make small perturbations. 3 Imperceptible Attacks with Node Density Estimation and Maximization PD[ f Y i PD[ g Y k :URQJ 0DWFKLQJ $FWXDO 0DWFKLQJ Figure 1: Imperceptible Attacks Intuitively, in Eq.(1), if there exist nodes v1 j similar to v1 i , i.e., v1 j v1 i , such that M(v1 j) v2 k 2 2 < M(v1 i ) v2 k 2 2, then a wrong matching (v1 j, v2 k) will be generated. In addition, if there are many such v1 js around v1 i , then it is hard to recognize v1 i from a pile of similar nodes. Thus, if we move v1 i to dense regions that contain many similar v1 js, then this dramatically increases the possibility of deriving the wrong matching (v1 j, v2 k) among many similar candidate nodes. Also, as many v1 js are around the adversarial node ˆv1 i , it is difficult for humans or defender programs to detect ˆv1 i , as shown in a toy example in Figure 1. Motivated by this, we propose to employ kernel density estimation (KDE) method to generate imperceptible perturbations. In statistics, the KDE is to estimate the probability density function f(x) of a random variable x with unknown distribution [55]. It helps reveal the intrinsic distribution. Concretely, let v1 be a N 1-dimensional random variable to denote all nodes {v1 i , , v1 N 1} in graph G1 with an unknown density f. A function ˆf(x) is estimated to best approximate f(x). ˆf(v1) = 1 N 1det(B) i=1 K B 1 v1 v1 i (2) where det( ) denotes the determinant operation. B > 0 is a bandwidth to be estimated. It is an N 1 N 1 diagonal matrix B = diag(b1, , b N1), which has strong influence on the density estimation ˆf(v1). A good B should be as small as the data can allow. K is a product symmetric kernel that satisfies R K(x)dx = 1 and R x K(x)dx = 0. The above vector-wise form ˆf(v1) can be rewritten as an element-wise form, where v1 j represents the jth dimension in v1. ˆf(v1) = 1 N 1 1 bj K v1 j v1 ij bj The derivative ˆ f(v1) bj w.r.t. each bandwidth bj in B is computed as follows, where K(x) = d log K(x) l=1 1 bl K v1 l v1 il bl i bj +v1 l v1 il b2 j K v1 l v1 il bj 1 bl K v1 l v1 il bl Traditional KDE methods often fail on high-dimensional data [29, 60, 32, 36], when bandwidths need to be selected for each dimension. A greedy search method is utilized to select bandwidths in the KDE: If a dimension j is insignificant, then changing the bandwidth bj for that dimension should have a weak impact on ˆf(v1), while the changing bj for an important j should cause a large change in ˆf(v1). Fortunately, ˆ f(v1) bj can differentiate these two types of dimensions. Based on the above analysis, we greedily decrease bj with a sequence b0, b0s, b0s2, for a parameter 0 < s < 1, until bj is smaller than a certain threshold τj, to see if a small change in bj can result in a large change in ˆf(v1). The method also offers a good way to estimate ˆ f(v1) b1 , , ˆ f(v1) b N1 along a sparse path. Concretely, ˆf(v1) is estimated by beginning with an initial B = diag(b0, , b0) for a large b0, and then estimate ˆ f(v1) bj as follows and decrease bj if ˆ f(v1) bj is large. l=1 1 bl K v1 l v1 il bl i K v1 j v1 ij bj K v1 j v1 ij bj l=1 K v1 l v1 il bl ˆf(v1 i ) bj (5) The corresponding variance Var ˆ f(v1) is given below. ˆf(v1 i ) bj Theorems 1-5 in Appendix A.5 provide the theoretical analysis about the density estimation, derivatives, and variances for well understanding the KDE technique. In this work, assuming that the graph data follow the Gaussian distribution, a product Gaussian kernel K is used to estimate the node density ˆf(v1). ˆ f(v1) bj is accordingly updated as follows. v1 j v1 ij 2 b2 j N1 Y l=1 K v1 l v1 il bl v1 j v1 ij 2 b2 j N1 Y l=1 K v1 l v1 il bl v1 j v1 ij 2 b2 j exp v1 l v1 il 2 where C denotes a proportionality constant C = 1 b3 j QN1 l=1 1 bl . It can be safely ignored to avoid com- putation overflow when bl 0 for large N 1. The bandwidth estimation is presented in Algorithm 1. Based on the estimated B and the Gaussian kernel K, the closed form of ˆf(v1) is derived below. ˆf(v1) = 1 N 1 j=1 K v1 j v1 ij bj |Σ| exp (v1 µ)T Σ 1 (B + Σ) 1 (v1 µ) 2 (8) where µ and Σ are the maximum likelihood estimation of the mean vector and covariance matrix of the Gaussian distribution. Please refer to Appendices A.6 and A.7 for detailed derivation of ˆf(v1). As two graphs G1 and G2 often have different structures and distributions and thus the same KDE method as Algorithm 1 is utilized to estimate the density ˆg(v2) of v2. Based on the estimations ˆf(v1) and ˆg(v2), the attacker aims to maximize the following loss LD with imperceptible perturbations. (ˆv1 i ,ˆv2 k) D L(ˆv1 i , ˆv2 k) where L(ˆv1 i , ˆv2 k) = M(ˆv1 i ) ˆv2 k) 2 2 + ˆf(ˆv1 i ) + ˆg(ˆv2 k) (9) where ˆv1 i = v1 i +δ1 i (and ˆv2 k = v2 k +δ2 k) denote adversarial versions of clean nodes v1 i (and v2 k) in G1 (and G2) by adding a small amount of edge perturbations δ1 i (and δ2 k) through our proposed MLPGD method in the next section, such that M(ˆv1 i ) is far away from ˆv2 k and thus the matching accuracy is decreased. In addition, we push v1 i and v2 k to dense regions to generate ˆv1 i and ˆv2 k, by maximizing ˆf(ˆv1 i ) and ˆg(ˆv2 k), such that ˆv1 i and ˆv2 k are indistinguishable from their neighbors in perturbed graphs. This reduces the possibility of perturbation detection by humans or defender programs. Algorithm 1 Bandwidth Matrix Estimation Input: graph G1 = (V 1, E1), parameter 0 < s < 1, initial bandwidth b0, and parameter c. Output: Bandwidth matrix B. 1: Initialize all b1, , b N1 with b0; 2: for each j = 1 to N 1 3: do 4: Estimate the derivative f(v1) bj and variance Var( f(v1) bj ) in Eqs.(6)-(7); 5: Compute the threshold τj = q 2 Var( f(v1) bj ) log(c N 1); 6: if f(v1) bj > τj, then Update bj = bjs; 7: while f(v1) bj > τj 8: Return B. Algorithm 2 Meta Learning-based Projected Gradient Descent (MLPGD) Input: Batches D1, , DC in a set D of node pairs, initial general policy parameters {θ1, θ2}, adaptation step size α, meta step size β. Output: Optimized {θ1, θ2}. 1: Repeat until convergence 2: Sample C batches of anchor node pairs D1, , DC; 3: for c = 1 to C 4: Estimate gradient ec = e Dc, {θ1, θ2} ; 5: Compute adapted parameters {θ1 c, θ2 c} = {θ1, θ2} + αec; 6: Update parameters {θ1, θ2} = {θ1, θ2} + β C PC c=1 e Dc, {θ1 c, θ2 c} ; 7: Return {θ1, θ2}. 4 Effective Attacks via Meta Learning-based Projected Gradient Descent 'HFLVLRQ %RXQGDU\ 2ULJLQDO 6WDUWLQJ 3RLQW *RRG 6WDUWLQJ 3RLQW $WWDFN 7UDMHFWRU\ Figure 2: Effective Attacks In Figure 2, two dashed purple curves denote the decision boundary of graph matching. If we move a clean node v1 i across the decision boundary to generate an adversarial node ˆv1 i , then we have other nodes v1 j to make M(v1 j) and v2 k become more similar than M(ˆv1 i ) and v2 k, and thus a wrong matching (v1 j, v2 k) will be produced. Blue and green polylines denote attack trajectories starting from original and good starting pints with gradient descent method respectively. A shortcut from good starting points (v1 i )0 or (v2 k)0 is able to cr- oss the peak of the decision boundary and converge quickly, while the trajectories from the original nodes v1 i or v2 k take long walks to cross the non-peak boundary. Based on the attack loss in Eq.(9), we propose to integrate meta learning and PGD into an MLPGD model, to produce more effective adversarial nodes with good starting points towards graph matching. (v1 i )(t+1) = Π 1 i sgn Re LU (v1 i )t L((v1 i )t, (v2 k)t) (v2 k)(t+1) = Π 2 ksgn Re LU (v2 k)t L((v1 i )t, (v2 k)t) , t = 1, , T (10) where (v1 i )t and (v2 k)t denotes the adversarial nodes of v1 i and v2 k derived at step t. ϵ specifies the budget of allowed perturbed edges for each attacked node. 1 i = {(δ1 i )t|1T (δ1 i )t ϵ, (δ1 i )t {0, 1}N1}, where (δ1 i )t = (v1 i )t v1 i 2 2, represents the constraint set of the projection operator Π, i.e., it encodes whether an edge of v1 i is modified or not. 2 k has the similar definition for v2 k. The composition of the Re LU and sign operators guarantees (v1 i )t {0, 1}N1 and (v2 k)t {0, 1}N2, as it adds (or removes) an edge or keeps it unchanged when an derivate in the gradient is positive (or negative). The outputs (v1 i )T and (v2 k)T at final step T are used as the adversarial nodes ˆv1 i and ˆv2 k. Searching for attack starting points for each (v1 i , v2 k) in large graphs is computationally inefficient. Meta learning techniques aim to train a general model with general parameters that can quickly adapt to a variety of new learning tasks with refined parameters [23, 3, 42, 56]. This offers a great opportunity to find good attack starting points (v1 i )0 and (v2 k)0 for all (v1 i , v2 k) D with lower cost, such that the generated ˆv1 i and ˆv2 k by the PGD model can maximize the attack loss LD in Eq.(9). Algorithm 3 Gradient Estimation e Dc, {θ1, θ2}, N, λ Input: Batch Dc, general parameters {θ1, θ2}, number of samples N in Monte Carlo REINFORCE, smoothing parameter λ. Output: Gradient estimation of a. 1: Sample N i.i.d. Gaussian matrices g1, , g N N(0, I); 2: Return gradient estimation 1 Nλ PN i=1 a Dc, {θ1, θ2} + λgi gi. Algorithm 4 Adversarial Attack a Dc, {θ1 c, θ2 c} Input: Batch Dc, perturbation budget ϵ, specific parameters {θ1 c, θ2 c} Output: Attack loss LDc on Dc. 1: LDc = 0; 2: for each (v1 i , v2 k) Dc 3: Generate attack starting points (v1 i )0 = h1 v1 i |θ1 c and (v2 k)0 = h2 v2 k|θ2 c ; 4: Utilize PGD attack to generate adversarial nodes (v1 i )T and (v2 k)T in Eq.(10); 5: Aggregate attack loss LDc+ = L (v1 i )T , (v2 k)T in Eq.(9); 6: Return LDc. Algorithm 2 presents the pseudo code of our MLPGD model. D is partitioned into C batches D1, , DC, each with equal size of |D|/C. The search process on each batch Dc (1 c C) is treated as a single task, which aims to find good (v1 i )0 and (v2 k)0 for Dc to maximize the attack loss LDc = P (ˆv1 i ,ˆv2 k) Dc L(ˆv1 i , ˆv2 k). A general model that has general parameters θ1, θ2 is learnt to quickly adapt to search tasks on multiple batches. The learnt θ1, θ2 should be sensitive to changes of each Dc, such that small changes in θ1, θ2 will produce high rise on LDc over any of D1, , DC. Line 4 estimates the gradient of LDc by calling Algorithm 3. In Line 5, when adapting to the task on a new Dc, θ1, θ2 become specific parameters θ1 c, θ2 c for Dc. Here, we use {θ1 c, θ2 c} to denote the concatenation matrix of θ1 c and θ2 c. The parameters are trained by maximizing the attack loss a Dc, {θ1 c, θ2 c} w.r.t. general parameters θ1, θ2 across batches. The meta objective is given below. max LDc = max θ1,θ2 c=1 a Dc, {θ1 c, θ2 c} = c=1 a Dc, {θ1, θ2} + αec (11) In Line 6, the meta optimization is performed over the general θ1, θ2, while the objective is computed using the specific θ1 c, θ2 c. The general θ1, θ2 are updated in terms of the attack loss on each batch. {θ1, θ2} = {θ1, θ2} + β c=1 e Dc, {θ1 c, θ2 c} (12) Algorithm 4 exhibits the adversarial attack module a Dc, {θ1 c, θ2 c} on a batch Dc (1 c C). In Line 3, two neural networks h1 and h2 with specific parameters θ1 c and θ2 c are designed to generate the attack starting points (v1 i )0 and (v2 k)0 of each (v1 i , v2 k) Dc. The last layers of h1 and h2 use the composition of the Re LU [53] and Softsign [25] as activation function to ensure (v1 i )0 {0, 1}N 1 and (v2 k)0 {0, 1}N2. In Line 4, the PGD attack in Eq.(10) is utilized to generate the adversarial nodes ˆv1 i and ˆv2 k. Line 5 calculates the attack loss LDc on Dc to provide task-specific feedback. Standard meta learning models utilizes gradient ascent/descent techniques to compute the updated parameters on new tasks [23, 3, 42, 56]. However, the attack module in Algorithm 4 is non-smooth and non-differential w.r.t. parameters θ1, θ2, θ1 c, and θ2 c, since the perturbation is a multi-step process as well as the projection at each step is non-differential. Therefore, Algorithm 3 is proposed to employ Gaussian smoothing technique to approximate a smoothed attack module. ˆa Dc, {θ1, θ2} (2π) d 2 Z a Dc, {θ1, θ2} + λg exp 1 = Eg N (0,I)a Dc, {θ1, θ2} + λg (13) where ˆa is the Gaussian smoothing of a and differentiable everywhere. λ is a smoothing parameter, and d is the number of entries in {θ1, θ2}. g N(0, I) that has the same size as {θ1 c, θ2 c} is interpreted as policy exploration directions, i.e., as perturbations in policy space to be explored. Thus, the policy perturbations in g are introduced to θ1 c and θ2 c respectively. ˆa is obtained by perturbing a at a given point along Gaussian directions and averaging the evaluations. And then, Algorithm 3 estimates the gradient of ˆa via Monte Carlo REINFORCE method [79]. Table 2: Mismatching rate (%) with 5% perturbed edges AS SNS DBLP Attack Model SNNA Cross MNA DGMC SNNA Cross MNA DGMC SNNA Cross MNA DGMC Clean 53.9 46.6 34.7 45.2 50.4 41.6 56.1 51.9 63.2 Random 57.5 49.9 37.6 48.8 52.0 46.8 59.8 54.0 68.8 RL-S2V 56.5 51.8 36.5 51.3 53.2 45.8 62.6 56.7 69.3 Meta-Self 63.1 55.1 45.0 55.1 64.8 51.3 65.7 63.7 73.3 CW-PGD 61.7 59.1 49.6 54.9 63.0 49.6 68.7 66.6 75.4 GF-Attack 57.9 53.7 39.5 52.9 59.6 47.9 64.9 61.1 69.1 CD-ATTACK 59.0 51.7 42.7 54.0 59.8 50.2 64.0 61.8 72.0 GMA 64.2 62.9 54.9 61.2 69.6 55.7 74.2 74.3 80.7 e Dc, {θ1, θ2} θ1,θ2ˆa Dc, {θ1, θ2} (2π) d 2 Z a Dc, {θ1, θ2} + λg exp 1 2 g 2 2 gdg λEg N (0,I)a Dc, {θ1, θ2} + λg g 1 Nλ i=1 a Dc, {θ1, θ2} + λgi gi, gi N(0, I) (14) 5 Experimental Evaluation Table 1: Experiment Datasets Dataset AS SNS DBLP Graph v1 v2 Last.fm Live Journal 2013 2014 #Nodes 10,900 11,113 5,682 17,828 28,478 26,455 #Edges 31,180 31,434 23,393 244,496 128,073 114,588 #Matched Nodes 6,462 2,138 4,000 In this section, we will show the effectiveness of the GMA model in this work for deep graph matching tasks over three groups of datasets: social networks (SNS) [98], autonomous systems (AS) [2], and D- BLP coauthor graphs [1], as shown in Table 1. Baselines. We compare the GMA model with six state-of-the-art graph attack models. Random Attack randomly adds and removes edges to generate perturbed graphs. RL-S2V [14, 123] generates adversarial attacks on graph data based on reinforcement learning. Meta-Self [125] is a poisoning attack model for node classification by using meta-gradients to solve the bilevel optimization problem. CW-PGD [86] developed a PGD topology attack to attack a predefined or a retrainable GNN. GFAttack [5] attacks general learning methods by devising new loss and approximating the spectrum. CD-ATTACK [41] hides nodes in the community by attacking the graph autoencoder model. The majority of existing efforts focus on adversarial attacks on single graph learning. To our best knowledge, there are no other attack baselines on graph matching available. We replace the original losses in the baselines with the matching loss for fair comparison in the experiments. Variants of GMA model. We evaluate four variants of GMA to show the strengths of different components. GMA-KDE only uses the KDE and density maximization to generate imperceptible attacks. GMA-PGD only utilizes the basic PGD [47] to produce effective attacks. GMA-MLPGD employs our proposed MLPGD model to well choose good attack starting points in the PGD. GMA operates with the full support of both KDE and MLPGD components. Graph matching algorithms. We validate the effectiveness of the above attack models with three representative deep graph matching methods. SNNA [39] is an adversarial learning framework to solve the weakly-supervised identity matching problem by minimizing the distribution distance. Cross MNA [13] is a cross-network embedding-based supervised network alignment method by learning inter/intra-embedding vectors for each node and by computing pairwise node similarity scores across networks. Deep graph matching consensus (DGMC) [22] is a supervised graph matching method that reaches a data-driven neighborhood consensus between matched node pairs. Evaluation metrics. We use two popular measures in graph matching to verify the attack quality: Accuracy [93, 10, 95] and Precision@K [101, 13, 97]. A larger mismatching rate (i.e., 1 - Accuracy on test data) or a smaller Precision@K shows a better attack. K is fixed to 30 in all tests. Attack performance on various datasets with different matching algorithms. Table 2 exhibits the mismatching rates of three deep graph matching algorithms on test data by eight attack models over three groups of datasets. We randomly sample 10% of known matched node pairs as training data and the rest as test data. For all attack models, the number of perturbed edges is fixed to 5% in these experiments. It is observed that among eight attack methods, no matter how strong the attacks are, the GMA method achieve the highest mismatching rates on perturbed graphs in most experiments, showing the effectiveness of GMA to the adversarial attacks. Compared to the graph matching results under other attack models, GMA, on average, achieves 21.3%, 18.8%, and 19.2% 2 5 10 15 20 25 0.2 Perturbed edges (%) Precision@30 Clean Random RL S2V Meta Self CW PGD GF Attack CD ATTACK GMA (a) SNNA on SNS 2 5 10 15 20 25 Perturbed edges (%) Precision@30 Clean Random RL S2V Meta Self CW PGD GF Attack CD ATTACK GMA (b) DGMC on SNS Figure 3: Precision with varying perturbed edges 2 5 10 15 20 25 Training data (%) Precision@30 Clean Random RL S2V Meta Self CW PGD GF Attack CD ATTACK GMA (a) SNNA on SNS 2 5 10 15 20 25 Training data (%) Precision@30 Clean Random RL S2V Meta Self CW PGD GF Attack CD ATTACK GMA (b) DGMC on SNS Figure 4: Precision with varying training ratios improvement of mismatching rates on AS, SNS, and DBLP respectively. In addition, the promising performance of GMA with all three graph matching models implies that GMA has great potential as a general attack solution to other graph matching methods, which is desirable in practice. Attack performance with varying perturbation edges. Figure 3 presents the graph matching quality under eight attack models by varying the ratios of perturbed edges from 2% to 25%. It is obvious that the attacking performance improves for each attacker with an increase in the number of perturbed edges. This phenomenon indicates that current deep graph matching methods are very sensitive to adversarial attacks. GMA achieves the lowest Precision values (< 0.488), which are still better than the other seven methods in most tests. Especially, when the perturbation ratio is large than 10%, the Precision values drop quickly. Impact of training data ratios. Figure 4 shows the quality of two graph matching algorithms on SNS by varying the ratio of training data from 2% to 25%. Here, the number of perturbed edges is fixed to 5%. We make the following observations on the performances by eight attack models. (1) The performance curves keep increasing when the training data ratio increases. (2) GMA outperforms other methods in most experiments with the lowest Precision: < 0.482 with SNNA and < 0.571 with DGMC respectively. Even when there are many training data available ( 20%), the quality degradation by GMA is still obvious, although more training data makes the graph matching models be resilient to poisoning attacks under a small perturbation budget. Ablation study. Figure 5 presents the mismatching rates of graph matching on SNS with four variants of the GMA attack model. We have observed the complete GMA achieves the highest mismatching rates (> 54.9%) on AS, (> 55.7%) over SNS, and (> 74.2%) on DBLP, which are obviously better than other versions. Notice that GMA-MLPGD performs quite well in most experiments, compared with GMA-PGD. A reasonable explanation is that searching from good attack starting points can help the MLPGD converge quickly by crossing the peak of the decision boundary. In addition, GMA-KDE achieves the better attack performance than GMA-MLPGD. A rational guess is that it is difficult to correctly match two nodes results when they lie in dense regions with many similar nodes, although the main goal of KDE is to generate imperceptible attacks. These results illustrate both KDE and MLPGD models are important in producing effective attacks in graph matching. Impact of perturbation budget ϵ. Figure 6 (a) measures the performance effect of ϵ in the MLPGD model for the graph matching by varying ϵ from 1 to 5. It is observed that when increasing ϵ, the Precision of the GMA model decreases substantially. This demonstrates that it is difficult to train a robust graph matching model under large ϵ constraint. However, a large ϵ can be easily detected by humans or by defender programs. Notice that the average node degree of three groups of datasets is between 2.9 and 13.9. Thus we suggest generating both imperceptible and effective attacks for the graph matching task under ϵ between 2 and 3, such that ϵ is smaller than the average node degree. Time complexity analysis Based on [20], the complexity of meta learning is O(d2), where d is the problem dimension. In the context of graph matching, it is the number of nodes in two graphs (N s, s = 1 or 2). Both density estimation and PGD have complexity of O((N s)2). Thus, the overall complexity is O((N s)2), which is the same as most existing attack methods that search the entire graphs to find weak edges to attack. Impact of meta step size α. Figure 6 (b) shows the impact of α in our MLPGD model over three groups of datasets. The performance curves initially raise when α increases. Intuitively, the MLPGD with large α can help the meta learning converge quickly. Later on, the performance curves keep relatively stable or even decreasing when α continuously increases. A reasonable explanation is that AS SNS DBLP 50 Mismatching rate (%) GMA KDE GMA PGD GMA MLPGD GMA AS SNS DBLP 45 Mismatching rate (%) GMA KDE GMA PGD GMA MLPGD GMA (b) DGMC Figure 5: Mismatching rate (%) of GMA variants 1 2 3 4 5 0.1 Precision@30 AS SNS DBLP (a) Perturbation budget ϵ 0.01 0.05 0.1 0.2 0.5 Precision@30 AS SNS DBLP (b) Meta step size α Figure 6: Precision with varying parameters the too large α makes the meta learner take a big walk with rapid pace, such that it may miss the optimal meta parameters. Thus, it is important to determine the optimal α for the MLPGD model. 6 Related Work Adversarial Attacks on Graph Data. Several recent studies have presented that graph learning models, especially deep learning-based models, are highly sensitive to adversarial attacks, i.e., carefully designed small deliberate perturbations in graph structure and attributes can cause the models to produce incorrect prediction results [64, 69, 90, 74, 45, 85, 80, 127]. The current graph adversarial attack techniques mainly fall into two categories in terms of the attack surface: (1) evasion attacks occur after the target model is well trained in clean graphs, i.e., the learned model parameters are fixed during evasion attacks. The attacker tries to evade the graph learning models by generating malicious samples during testing phase [14, 126]; and (2) poisoning attacks, known as contamination of the training data, take place during the training time of deep learning models. An adversary tries to poison the training data by injecting carefully designed examples to cause failures of the target model on some given test samples [126, 69, 125, 4, 123]. Since transductive learning is widely used in most graph analysis tasks, the test samples (but not their labels) are participated in the training stage, which leads to the popularity of poisoning attacks. Various adversarial attack models have been developed to show the vulnerability of graph learning models in node classification [14, 126, 74, 86, 125, 71, 19, 70], community detection [9, 78, 7, 41], network embedding [6, 4, 5], link prediction [104], similarity search [15], malware detection [30], and knowledge graph embedding [90]. Graph Matching. Graph data analysis has attracted active research in the last decade [110, 111, 11, 106, 12, 107, 108, 66, 113, 109, 112, 114, 35, 115, 117, 116, 119, 118, 81, 82, 120, 121]. Graph matching is one of the most important research topics in the graph domain, which aims to match the same entities (i.e., nodes) across two or more graphs and has been a heated topic in recent years [91, 98, 62, 46, 48, 101, 13]. Research activities can be classified into three broad categories. (1) Topological structure-based techniques, which rely on only the structural information of nodes to match multiple or two input networks, including IONE [43], Geo Align [46], Lowrank Eigen Align [54], FRUI-P [105], Cross MNA [13], MOANA [96], GWL [84], MSUIL [38], Deep MGGE [24], and KEMINA [122]; (2) Structure and/or attribute-based approaches, which utilize highly discriminative structure and attribute features for ensuring the matching effectiveness, such as FINAL [93, 94], ULink [51], CAlign [10], MASTER [65], gsa NA [88], Co Link [100], REGAL [28], UUIL [37], SNNA [39], RANA [58], CENALP [18], ORIGIN [97], OPTANE [50], and Deep Graph Matching Consensus [22]; (3) Heterogeneous methods employ heterogeneous structural, content, spatial, and temporal features to further improve the matching performance, including COSNET [98], Factoid Embedding [83], HEP [99], LHNE [77], and DPLink [21]. Several papers review key achievements of graph matching across online information networks including state-ofthe-art algorithms, evaluation metrics, representative datasets, and empirical analysis [62, 27, 31, 73]. 7 Conclusions In this work, we have studied the graph matching adversarial attack problem. First, we proposed to utilize kernel density estimation technique to estimate and maximize the densities of attacked nodes and generate imperceptible perturbations, by pushing attacked nodes to dense regions in two graphs. Second, we developed a meta learning based projected gradient descent method to well choose attack starting points and improve the search performance of PGD for producing effective perturbations. The GMA model achieves superior attack performance against several representative attack models. Broader Impact Graph data are ubiquitous in the real world, ranging from biological, communication, and transportation graphs, to knowledge, social, and collaborative networks. Many real-world graphs are essentially crowdsourced projects, such as social and knowledge networks, where information and knowledge are produced by internet users who came to the sites. Thus, the quality of crowdsourced graph data is not stable, depending on human knowledge and expertise. In addition, it is well known that the openness of crowdsourced websites makes them vulnerable to malicious behaviors of interested parties to gain some level of control of the websites and steal users sensitive information, or deliberately influence public opinion by injecting misleading information and knowledge into crowdsourced graphs. Graph matching is one of the most important research topics in the graph domain, which aims to match the same entities (i.e., nodes) across two or more graphs [91, 98, 43, 46, 48, 72, 54, 105, 13, 75]. It has been widely applied to many real-world applications ranging from protein network matching in bioinformatics [33, 63], user account linking in different social networks [62, 51, 100, 37, 101, 21, 38], and knowledge translation in multilingual knowledge bases [87, 124], to geometric keypoint matching in computer vision [22]. Owing to the openness of crowdsourced graphs, more work is needed to analyze the vulnerability of graph matching under adversarial attacks and to future develop robust solutions that are readily applicable in production systems. A potential downside of this research is about the application of user account linking in different social networks due to the user privacy issues. Recent advances in differential privacy and privacy preserving graph analytics have shown the superior performance of protecting sensitive information about individuals in the datasets. Therefore, these techniques offer a great opportunity to integrate them into the vulnerability analysis of graph matching, for alleviating the user privacy threats. [1] http://dblp.uni-trier.de/xml/. [2] http://snap.stanford.edu/data/. [3] L. Bertinetto, J. F. Henriques, P. H. S. Torr, and A. Vedaldi. Meta-learning with differentiable closed-form solvers. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019, 2019. [4] A. Bojchevski and S. Günnemann. Adversarial attacks on node embeddings via graph poisoning. In Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 9-15 June 2019, Long Beach, California, USA, pages 695 704, 2019. [5] H. Chang, Y. Rong, T. Xu, W. Huang, H. Zhang, P. Cui, W. Zhu, and J. Huang. A restricted black-box adversarial framework towards attacking graph embedding models. In The Thirty Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, New Yrok, NY, USA, February 7 - 12, 2020, 2020. [6] J. Chen, Y. Wu, X. Xu, Y. Chen, H. Zheng, and Q. Xuan. Fast gradient attack on network embedding. Co RR, abs/1809.02797, 2018. [7] J. Chen, L. Chen, Y. Chen, M. Zhao, S. Yu, Q. Xuan, and X. Yang. Ga-based q-attack on community detection. IEEE Trans. Comput. Social Systems, 6(3):491 503, 2019. [8] J. Chen, Y. Chen, H. Zheng, S. Shen, S. Yu, D. Zhang, and Q. Xuan. MGA: momentum gradient attack on network. Co RR, abs/2002.11320, 2020. [9] Y. Chen, Y. Nadji, A. Kountouras, F. Monrose, R. Perdisci, M. Antonakakis, and N. Vasiloglou. Practical attacks against graph-based clustering. In Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security, CCS 2017, Dallas, TX, USA, October 30 - November 03, 2017, pages 1125 1142, 2017. [10] Z. Chen, X. Yu, B. Song, J. Gao, X. Hu, and W. Yang. Community-based network alignment for large attributed network. In Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, CIKM 2017, Singapore, November 06 - 10, 2017, pages 587 596, 2017. [11] H. Cheng, Y. Zhou, and J. X. Yu. Clustering large attributed graphs: A balance between structural and attribute similarities. ACM Transactions on Knowledge Discovery from Data (TKDD), 5(2):1 33, 2011. [12] H. Cheng, Y. Zhou, X. Huang, and J. X. Yu. Clustering large attributed information networks: An efficient incremental computing approach. Data Mining and Knowledge Discovery (DMKD), 25(3):450 477, 2012. [13] X. Chu, X. Fan, D. Yao, Z. Zhu, J. Huang, and J. Bi. Cross-network embedding for multinetwork alignment. In The World Wide Web Conference, WWW 2019, San Francisco, CA, USA, May 13-17, 2019, pages 273 284, 2019. [14] H. Dai, H. Li, T. Tian, X. Huang, L. Wang, J. Zhu, and L. Song. Adversarial attack on graph structured data. In Proceedings of the 35th International Conference on Machine Learning, ICML 2018, Stockholmsmässan, Stockholm, Sweden, July 10-15, 2018, pages 1123 1132, 2018. [15] P. Dey and S. Medya. Manipulating node similarity measures in networks. In Proceedings of the 19th International Conference on Autonomous Agents and Multi Agent Systems, AAMAS 20, Auckland, New Zealand, May 9-13, 2020, 2019. [16] B. Du and H. Tong. FASTEN: fast sylvester equation solver for graph mining. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, KDD 2018, London, UK, August 19-23, 2018, pages 1339 1347, 2018. [17] B. Du, S. Zhang, N. Cao, and H. Tong. FIRST: fast interactive attributed subgraph matching. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Halifax, NS, Canada, August 13 - 17, 2017, pages 1447 1456, 2017. [18] X. Du, J. Yan, and H. Zha. Joint link prediction and network alignment via cross-graph embedding. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI 2019, Macao, China, August 10-16, 2019, pages 2251 2257, 2019. [19] N. Entezari, S. Al-Sayouri, A. Darvishzadeh, and E. Papalexakis. All you need is low (rank): Defending against adversarial attacks on graphs. In Proceedings of the 13th ACM International Conference on Web Search and Data Mining, WSDM 2020, Houston, TX, February 3-7, 2020, 2020. [20] A. Fallah, A. Mokhtari, and A. E. Ozdaglar. On the convergence theory of gradient-based model-agnostic meta-learning algorithms. In The 23rd International Conference on Artificial Intelligence and Statistics, AISTATS 2020, 26-28 August 2020, Online [Palermo, Sicily, Italy], pages 1082 1092, 2020. [21] J. Feng, M. Zhang, H. Wang, Z. Yang, C. Zhang, Y. Li, and D. Jin. Dplink: User identity linkage via deep neural network from heterogeneous mobility data. In The World Wide Web Conference, WWW 2019, San Francisco, CA, USA, May 13-17, 2019, pages 459 469, 2019. [22] M. Fey, J. E. Lenssen, C. Morris, J. Masci, and N. M. Kriege. Deep graph matching consensus. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020, 2020. [23] C. Finn, P. Abbeel, and S. Levine. Model-agnostic meta-learning for fast adaptation of deep networks. In Proceedings of the 34th International Conference on Machine Learning, ICML 2017, Sydney, NSW, Australia, 6-11 August 2017, pages 1126 1135, 2017. [24] S. Fu, G. Wang, S. Xia, and L. Liu. Deep multi-granularity graph embedding for user identity linkage across social networks. Knowl. Based Syst., 193:105301, 2020. [25] X. Glorot and Y. Bengio. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, AISTATS 2010, Chia Laguna Resort, Sardinia, Italy, May 13-15, 2010, pages 249 256, 2010. [26] I. J. Goodfellow, J. Shlens, and C. Szegedy. Explaining and harnessing adversarial examples. In 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings, 2015. [27] P. H. Guzzi and T. Milenkovic. Survey of local and global biological network alignment: the need to reconcile the two sides of the same coin. Briefings in Bioinformatics, 19(3):472 481, 2018. [28] M. Heimann, H. Shen, T. Safavi, and D. Koutra. REGAL: representation learning-based graph alignment. In Proceedings of the 27th ACM International Conference on Information and Knowledge Management, CIKM 2018, Torino, Italy, October 22-26, 2018, pages 117 126, 2018. [29] N. L. Hjort and I. K. Glad. Nonparametric density estimation with a parametric start. The Annals of Statistics, 23(3):882 904, 1995. [30] S. Hou, Y. Fan, Y. Zhang, Y. Ye, J. Lei, W. Wan, J. Wang, Q. Xiong, and F. Shao. αCyber: Enhancing robustness of android malware detection system against adversarial attacks on heterogeneous graph based model. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management, CIKM 2019, Beijing, China, November 3-7, 2019, pages 609 618, 2019. [31] T. T. Huynh, N. T. Toan, V. V. Tong, T. D. Hoang, D. C. Thang, N. Q. V. Hung, and A. Sattar. A comparative study on network alignment techniques. Expert Syst. Appl., 140, 2020. [32] H. Jiang. Uniform convergence rates for kernel density estimation. In Proceedings of the 34th International Conference on Machine Learning, ICML 2017, Sydney, NSW, Australia, 6-11 August 2017, pages 1694 1703, 2017. [33] B. P. Kelley, R. Sharan, R. M. Karp, T. Sittler, D. E. Root, B. R. Stockwell, and T. Ideker. Conserved pathways within bacteria and yeast as revealed by global protein network alignment. Proceedings of the National Academy of Sciences, 100(20):11394 11399, 2003. [34] X. Kong, J. Zhang, and P. S. Yu. Inferring anchor links across multiple heterogeneous social networks. In 22nd ACM International Conference on Information and Knowledge Management, CIKM 13, San Francisco, CA, USA, October 27 - November 1, 2013, pages 179 188, 2013. [35] K. Lee, L. Liu, K. Schwan, C. Pu, Q. Zhang, Y. Zhou, E. Yigitoglu, and P. Yuan. Scaling iterative graph computations with graphmap. In Proceedings of the 27th IEEE international conference for High Performance Computing, Networking, Storage and Analysis (SC 15), pages 57:1 57:12, Austin, TX, November 15-20 2015. [36] R. Leibrandt and S. Günnemann. Making kernel density estimation robust towards missing values in highly incomplete multivariate data without imputation. In Proceedings of the 2018 SIAM International Conference on Data Mining, SDM 2018, May 3-5, 2018, San Diego Marriott Mission Valley, San Diego, CA, USA, pages 747 755, 2018. [37] C. Li, S. Wang, P. S. Yu, L. Zheng, X. Zhang, Z. Li, and Y. Liang. Distribution distance minimization for unsupervised user identity linkage. In Proceedings of the 27th ACM International Conference on Information and Knowledge Management, CIKM 2018, Torino, Italy, October 22-26, 2018, pages 447 456, 2018. [38] C. Li, S. Wang, H. Wang, Y. Liang, P. S. Yu, Z. Li, and W. Wang. Partially shared adversarial learning for semi-supervised multi-platform user identity linkage. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management, CIKM 2019, Beijing, China, November 3-7, 2019, pages 249 258, 2019. [39] C. Li, S. Wang, Y. Wang, P. S. Yu, Y. Liang, Y. Liu, and Z. Li. Adversarial learning for weaklysupervised social network alignment. In The Thirty-Third AAAI Conference on Artificial Intelligence, AAAI 2019, The Thirty-First Innovative Applications of Artificial Intelligence Conference, IAAI 2019, The Ninth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2019, Honolulu, Hawaii, USA, January 27 - February 1, 2019, pages 996 1003, 2019. [40] G. Li, L. Sun, Z. Zhang, P. Ji, S. Su, and P. S. Yu. Mc2: Unsupervised multiple social network alignment. In 2019 IEEE International Conference on Big Data (Big Data), Los Angeles, CA, USA, December 9-12, 2019, pages 1151 1156, 2019. [41] J. Li, H. Zhang, Z. Han, Y. Rong, H. Cheng, and J. Huang. Adversarial attack on community detection by hiding individuals. In WWW 20: The Web Conference 2020, Taipei, Taiwan, April 20-24, 2020, pages 917 927, 2020. [42] H. Liu, R. Socher, and C. Xiong. Taming MAML: efficient unbiased meta-reinforcement learning. In Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 9-15 June 2019, Long Beach, California, USA, pages 4061 4071, 2019. [43] L. Liu, W. K. Cheung, X. Li, and L. Liao. Aligning users across social networks using network embedding. In Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence, IJCAI 2016, New York, NY, USA, 9-15 July 2016, pages 1774 1780, 2016. [44] S. Liu, S. Wang, F. Zhu, J. Zhang, and R. Krishnan. HYDRA: large-scale social identity linkage via heterogeneous behavior modeling. In International Conference on Management of Data, SIGMOD 2014, Snowbird, UT, USA, June 22-27, 2014, pages 51 62, 2014. [45] X. Liu, S. Si, X. Zhu, Y. Li, and C. Hsieh. A unified framework for data poisoning attack to graph-based semi-supervised learning. In Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2018, Neur IPS 2019, 8-14 December 2019, Vancouver, Canada, pages 9777 9787, 2019. [46] Y. Liu, H. Ding, D. Chen, and J. Xu. Novel geometric approach for global alignment of PPI networks. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, AAAI 2017, February 4-9, 2017, San Francisco, California, USA., pages 31 37, 2017. [47] A. Madry, A. Makelov, L. Schmidt, D. Tsipras, and A. Vladu. Towards deep learning models resistant to adversarial attacks. In 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings, 2018. [48] E. Malmi, A. Gionis, and E. Terzi. Active network alignment: A matching-based approach. In Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, CIKM 2017, Singapore, November 06 - 10, 2017, pages 1687 1696, 2017. [49] T. Man, H. Shen, S. Liu, X. Jin, and X. Cheng. Predict anchor links across social networks via an embedding approach. In Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence, IJCAI 2016, New York, NY, USA, 9-15 July 2016, pages 1823 1829, 2016. [50] F. Masrour, P. Tan, and A. Esfahanian. OPTANE: an optimal transport algorithm for network alignment. In ASONAM 19: International Conference on Advances in Social Networks Analysis and Mining, Vancouver, British Columbia, Canada, 27-30 August, 2019, pages 448 451, 2019. [51] X. Mu, F. Zhu, E. Lim, J. Xiao, J. Wang, and Z. Zhou. User identity linkage by latent user space modelling. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, August 13-17, 2016, pages 1775 1784, 2016. [52] X. Mu, W. Xie, R. K. Lee, F. Zhu, and E. Lim. Ad-link: An adaptive approach for user identity linkage. In 2019 IEEE International Conference on Big Knowledge, ICBK 2019, Beijing, China, November 10-11, 2019, pages 183 190, 2019. [53] V. Nair and G. E. Hinton. Rectified linear units improve restricted boltzmann machines. In Proceedings of the 27th International Conference on Machine Learning (ICML-10), June 21-24, 2010, Haifa, Israel, pages 807 814, 2010. [54] H. Nassar, N. Veldt, S. Mohammadi, A. Grama, and D. F. Gleich. Low rank spectral network alignment. In Proceedings of the 2018 World Wide Web Conference on World Wide Web, WWW 2018, Lyon, France, April 23-27, 2018, pages 619 628, 2018. [55] E. Parzen. On estimation of a probability density function and mode. The Annals of Mathematical Statistics, 33(3):1065 1076, 1962. [56] A. Raghu, M. Raghu, S. Bengio, and O. Vinyals. Rapid learning or feature reuse? towards understanding the effectiveness of MAML. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020, 2020. [57] V. Ravindra, H. Nassar, D. F. Gleich, and A. Grama. Rigid graph alignment. In Complex Networks and Their Applications VIII - Volume 1 Proceedings of the Eighth International Conference on Complex Networks and Their Applications COMPLEX NETWORKS 2019, Lisbon, Portugal, December 10-12, 2019, pages 621 632, 2019. [58] J. Ren, Y. Zhou, R. Jin, Z. Zhang, D. Dou, and P. Wang. Dual adversarial learning based network alignment. In Proceedings of the 19th IEEE International Conference on Data Mining (ICDM 19), pages 1288 1293, Beijing, China, November 8-11 2019. [59] Y. Ren, C. C. Aggarwal, and J. Zhang. Meta diagram based active social networks alignment. In 35th IEEE International Conference on Data Engineering, ICDE 2019, Macao, China, April 8-11, 2019, pages 1690 1693, 2019. [60] S. R. Sain and D. W. Scott. On locally adaptive density estimation. Journal of the American Statistical Association, 91(436):1525 1534, 1996. [61] Y. Shi, S. Wang, and Y. Han. Curls & whey: Boosting black-box adversarial attacks. In IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, June 16-20, 2019, pages 6519 6527, 2019. [62] K. Shu, S. Wang, J. Tang, R. Zafarani, and H. Liu. User identity linkage across online social networks: A review. SIGKDD Explorations, 18(2):5 17, 2016. [63] R. Singh, J. Xu, and B. Berger. Global alignment of multiple protein interaction networks with application to functional orthology detection. Proceedings of the National Academy of Sciences, 105(35):12763 12768, 2008. ISSN 0027-8424. [64] W. Song, S. Wang, B. Yang, Y. Lu, X. Zhao, and X. Liu. Learning node and edge embeddings for signed networks. Neurocomputing, 319:42 54, 2018. [65] S. Su, L. Sun, Z. Zhang, G. Li, and J. Qu. MASTER: across multiple social networks, integrate attribute and structure embedding for reconciliation. In Proceedings of the Twenty Seventh International Joint Conference on Artificial Intelligence, IJCAI 2018, July 13-19, 2018, Stockholm, Sweden., pages 3863 3869, 2018. [66] Z. Su, L. Liu, M. Li, X. Fan, and Y. Zhou. Reliable and resilient trust management in distributed service provision networks. ACM Transactions on the Web (TWEB), 9(3):1 37, 2015. [67] L. Sun, Z. Zhang, P. Ji, J. Wen, S. Su, and P. S. Yu. DNA: dynamic social network alignment. In 2019 IEEE International Conference on Big Data (Big Data), Los Angeles, CA, USA, December 9-12, 2019, pages 1224 1231, 2019. [68] M. Sun, J. Tang, H. Li, B. Li, C. Xiao, Y. Chen, and D. Song. Data poisoning attack against unsupervised node embedding methods. Co RR, abs/1810.12881, 2018. [69] Y. Sun, S. Wang, X. Tang, T. Hsieh, and V. G. Honavar. Node injection attacks on graphs via reinforcement learning. Co RR, abs/1909.06543, 2019. [70] Y. Sun, S. Wang, X. Tang, T. Hsieh, and V. G. Honavar. Adversarial attacks on graph neural networks via node injections: A hierarchical reinforcement learning approach. In WWW 20: The Web Conference 2020, Taipei, Taiwan, April 20-24, 2020, pages 673 683, 2020. [71] T. Takahashi. Indirect adversarial attacks via poisoning neighbors for graph convolutional networks. In 2019 IEEE International Conference on Big Data (Big Data), Los Angeles, CA, USA, December 9-12, 2019, pages 1395 1400, 2019. [72] V. Vijayan and T. Milenkovic. Multiple network alignment via multimagna++. IEEE/ACM Trans. Comput. Biology Bioinform., 15(5):1669 1682, 2018. [73] V. Vijayan, S. Gu, E. T. Krebs, L. Meng, and T. Milenkovic. Pairwise versus multiple global network alignment. IEEE Access, 8:41961 41974, 2020. [74] B. Wang and N. Z. Gong. Attacking graph-based classification via manipulating the graph structure. In Proceedings of the 2019 ACM SIGSAC Conference on Computer and Communications Security, CCS 2019, London, UK, November 11-15, 2019, pages 2023 2040, 2019. [75] S. Wang, X. Li, Y. Ye, S. Feng, R. Y. K. Lau, X. Huang, and X. Du. Anchor link prediction across attributed networks via network embedding. Entropy, 21(3):254, 2019. [76] X. Wang, M. Cheng, J. Eaton, C. Hsieh, and S. F. Wu. Fake node attacks on graph convolutional networks. Co RR, abs/1810.10751, 2018. [77] Y. Wang, C. Feng, L. Chen, H. Yin, C. Guo, and Y. Chu. User identity linkage across social networks via linked heterogeneous network embedding. World Wide Web, 22(6):2611 2632, 2019. [78] M. Waniek, T. P. Michalak, M. J. Wooldridge, and T. Rahwan. Hiding individuals and communities in a social network. Nature Human Behaviour, 2:139 147, 2018. [79] R. J. Williams. Simple statistical gradient-following algorithms for connectionist reinforcement learning. Machine Learning, 8:229 256, 1992. [80] H. Wu, C. Wang, Y. Tyshetskiy, A. Docherty, K. Lu, and L. Zhu. Adversarial examples for graph data: Deep insights into attack and defense. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI 2019, Macao, China, August 10-16, 2019, pages 4816 4823, 2019. [81] S. Wu, Y. Li, D. Zhang, Y. Zhou, and Z. Wu. Diverse and informative dialogue generation with context-specific commonsense knowledge awareness. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020, Online, pages 5811 5820, July 5-10 2020. [82] S. Wu, Y. Li, D. Zhang, Y. Zhou, and Z. Wu. Topicka: Generating commonsense knowledgeaware dialogue responses towards the recommended topic fact. In Proceedings of the Twenty Ninth International Joint Conference on Artificial Intelligence, IJCAI 2020, pages 3766 3772, Yokohama, Japan, January 2021. [83] W. Xie, X. Mu, R. K. Lee, F. Zhu, and E. Lim. Unsupervised user identity linkage via factoid embedding. In IEEE International Conference on Data Mining, ICDM 2018, Singapore, November 17-20, 2018, pages 1338 1343, 2018. [84] H. Xu, D. Luo, H. Zha, and L. Carin. Gromov-wasserstein learning for graph matching and node embedding. In Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 9-15 June 2019, Long Beach, California, USA, pages 6932 6941, 2019. [85] H. Xu, Y. Ma, H. Liu, D. Deb, H. Liu, J. Tang, and A. K. Jain. Adversarial attacks and defenses in images, graphs and text: A review. Co RR, abs/1909.08072, 2019. [86] K. Xu, H. Chen, S. Liu, P. Chen, T. Weng, M. Hong, and X. Lin. Topology attack and defense for graph neural networks: An optimization perspective. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI 2019, Macao, China, August 10-16, 2019, pages 3961 3967, 2019. [87] K. Xu, L. Wang, M. Yu, Y. Feng, Y. Song, Z. Wang, and D. Yu. Cross-lingual knowledge graph alignment via graph matching neural network. In Proceedings of the 57th Conference of the Association for Computational Linguistics, ACL 2019, Florence, Italy, July 28August 2, 2019, Volume 1: Long Papers, pages 3156 3161, 2019. [88] A. Yasar and Ü. V. Çatalyürek. An iterative global structure-assisted labeled network aligner. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, KDD 2018, London, UK, August 19-23, 2018, pages 2614 2623, 2018. [89] X. Zang, Y. Xie, J. Chen, and B. Yuan. Graph universal adversarial attacks: A few bad actors ruin graph learning models. Co RR, abs/2002.04784, 2020. [90] H. Zhang, T. Zheng, J. Gao, C. Miao, L. Su, Y. Li, and K. Ren. Data poisoning attack against knowledge graph embedding. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI 2019, Macao, China, August 10-16, 2019, pages 4853 4859, 2019. [91] J. Zhang and P. S. Yu. Integrated anchor and social link predictions across social networks. In Proceedings of the Twenty-Fourth International Joint Conference on Artificial Intelligence, IJCAI 2015, Buenos Aires, Argentina, July 25-31, 2015, pages 2125 2132, 2015. [92] J. Zhang, X. Kong, and P. S. Yu. Predicting social links for new users across aligned heterogeneous social networks. In 2013 IEEE 13th International Conference on Data Mining, ICDM 2013, Dallas, TX, USA, December 7-10, 2013, pages 1289 1294, 2013. [93] S. Zhang and H. Tong. FINAL: fast attributed network alignment. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD 2016, San Francisco, CA, USA, August 13-17, 2016, pages 1345 1354, 2016. [94] S. Zhang and H. Tong. Attributed network alignment: Problem definitions and fast solutions. IEEE Trans. Knowl. Data Eng., 31(9):1680 1692, 2019. [95] S. Zhang, H. Tong, J. Tang, J. Xu, and W. Fan. ineat: Incomplete network alignment. In 2017 IEEE International Conference on Data Mining, ICDM 2017, New Orleans, LA, USA, November 18-21, 2017, pages 1189 1194, 2017. [96] S. Zhang, H. Tong, R. Maciejewski, and T. Eliassi-Rad. Multilevel network alignment. In The World Wide Web Conference, WWW 2019, San Francisco, CA, USA, May 13-17, 2019, pages 2344 2354, 2019. [97] S. Zhang, H. Tong, J. Xu, Y. Hu, and R. Maciejewski. ORIGIN: non-rigid network alignment. In 2019 IEEE International Conference on Big Data (Big Data), Los Angeles, CA, USA, December 9-12, 2019, pages 998 1007, 2019. [98] Y. Zhang, J. Tang, Z. Yang, J. Pei, and P. S. Yu. COSNET: connecting heterogeneous social networks with local and global consistency. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Sydney, NSW, Australia, August 10-13, 2015, pages 1485 1494, 2015. [99] V. W. Zheng, M. Sha, Y. Li, H. Yang, Y. Fang, Z. Zhang, K. Tan, and K. C. Chang. Heterogeneous embedding propagation for large-scale e-commerce user alignment. In IEEE International Conference on Data Mining, ICDM 2018, Singapore, November 17-20, 2018, pages 1434 1439, 2018. [100] Z. Zhong, Y. Cao, M. Guo, and Z. Nie. Colink: An unsupervised framework for user identity linkage. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, (AAAI-18), the 30th innovative Applications of Artificial Intelligence (IAAI-18), and the 8th AAAI Symposium on Educational Advances in Artificial Intelligence (EAAI-18), New Orleans, Louisiana, USA, February 2-7, 2018, pages 5714 5721, 2018. [101] F. Zhou, L. Liu, K. Zhang, G. Trajcevski, J. Wu, and T. Zhong. Deeplink: A deep learning approach for user identity linkage. In 2018 IEEE Conference on Computer Communications, INFOCOM 2018, Honolulu, HI, USA, April 16-19, 2018, pages 1313 1321, 2018. [102] F. Zhou, Z. Wen, G. Trajcevski, K. Zhang, T. Zhong, and F. Liu. Disentangled network alignment with matching explainability. In 2019 IEEE Conference on Computer Communications, INFOCOM 2019, Paris, France, April 29 - May 2, 2019, pages 1360 1368, 2019. [103] J. Zhou and J. Fan. Translink: User identity linkage across heterogeneous social networks via translating embeddings. In 2019 IEEE Conference on Computer Communications, INFOCOM 2019, Paris, France, April 29 - May 2, 2019, pages 2116 2124, 2019. [104] K. Zhou, T. P. Michalak, M. Waniek, T. Rahwan, and Y. Vorobeychik. Attacking similaritybased link prediction in social networks. In Proceedings of the 18th International Conference on Autonomous Agents and Multi Agent Systems, AAMAS 19, Montreal, QC, Canada, May 13-17, 2019, pages 305 313, 2019. [105] X. Zhou, X. Liang, X. Du, and J. Zhao. Structure based user identification across social networks. IEEE Trans. Knowl. Data Eng., 30(6):1178 1191, 2018. [106] Y. Zhou and L. Liu. Clustering analysis in large graphs with rich attributes. In D. E. Holmes and L. C. Jain, editors, Data Mining: Foundations and Intelligent Paradigms: Volume 1: Clustering, Association and Classification. Springer, 2011. [107] Y. Zhou and L. Liu. Social influence based clustering of heterogeneous information networks. In Proceedings of the 19th ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD 13), pages 338 346, Chicago, IL, August 11-14 2013. [108] Y. Zhou and L. Liu. Activity-edge centric multi-label classification for mining heterogeneous information networks. In Proceedings of the 20th ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD 14), pages 1276 1285, New York, NY, August 24-27 2014. [109] Y. Zhou and L. Liu. Social influence based clustering and optimization over heterogeneous information networks. ACM Transactions on Knowledge Discovery from Data (TKDD), 10(1): 1 53, 2015. [110] Y. Zhou, H. Cheng, and J. X. Yu. Graph clustering based on structural/attribute similarities. Proceedings of the VLDB Endowment (PVLDB), 2(1):718 729, 2009. [111] Y. Zhou, H. Cheng, and J. X. Yu. Clustering large attributed graphs: An efficient incremental approach. In Proceedings of the 10th IEEE International Conference on Data Mining (ICDM 10), pages 689 698, Sydney, Australia, December 14-17 2010. [112] Y. Zhou, L. Liu, and D. Buttler. Integrating vertex-centric clustering with edge-centric clustering for meta path graph analysis. In Proceedings of the 21st ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD 15), pages 1563 1572, Sydney, Australia, August 10-13 2015. [113] Y. Zhou, L. Liu, K. Lee, C. Pu, and Q. Zhang. Fast iterative graph computation with resource aware graph parallel abstractions. In Proceedings of the 24th ACM Symposium on High Performance Parallel and Distributed Computing (HPDC 15), pages 179 190, Portland, OR, June 15-19 2015. [114] Y. Zhou, L. Liu, K. Lee, and Q. Zhang. Graphtwist: Fast iterative graph computation with two-tier optimizations. Proceedings of the VLDB Endowment (PVLDB), 8(11):1262 1273, 2015. [115] Y. Zhou, L. Liu, S. Seshadri, and L. Chiu. Analyzing enterprise storage workloads with graph modeling and clustering. IEEE Journal on Selected Areas in Communications (JSAC), 34(3): 551 574, 2016. [116] Y. Zhou, A. Amimeur, C. Jiang, D. Dou, R. Jin, and P. Wang. Density-aware local siamese autoencoder network embedding with autoencoder graph clustering. In Proceedings of the 2018 IEEE International Conference on Big Data (Big Data 18), pages 1162 1167, Seattle, WA, December 10-13 2018. [117] Y. Zhou, S. Wu, C. Jiang, Z. Zhang, D. Dou, R. Jin, and P. Wang. Density-adaptive local edge representation learning with generative adversarial network multi-label edge classification. In Proceedings of the 18th IEEE International Conference on Data Mining (ICDM 18), pages 1464 1469, Singapore, November 17-20 2018. [118] Y. Zhou, C. Jiang, Z. Zhang, D. Dou, R. Jin, and P. Wang. Integrating local vertex/edge embedding via deep matrix fusion and siamese multi-label classification. In Proceedings of the 2019 IEEE International Conference on Big Data (Big Data 19), pages 1018 1027, Los Angeles, CA, December 9-12 2019. [119] Y. Zhou, J. Ren, S. Wu, D. Dou, R. Jin, Z. Zhang, and P. Wang. Semi-supervised classificationbased local vertex ranking via dual generative adversarial nets. In Proceedings of the 2019 IEEE International Conference on Big Data (Big Data 19), pages 1267 1273, Los Angeles, CA, December 9-12 2019. [120] Y. Zhou, L. Liu, K. Lee, B. Palanisamy, and Q. Zhang. Improving collaborative filtering with social influence over heterogeneous information networks. To appear in ACM Transactions on Internet Technology (TOIT), 2020. [121] Y. Zhou, J. Ren, D. Dou, R. Jin, J. Zheng, and K. Lee. Robust meta network embedding against adversarial attacks. In Proceedings of the 20th IEEE International Conference on Data Mining (ICDM 20), Online, November 17-20 2020. [122] Y. Zhou, J. Ren, R. Jin, Z. Zhang, D. Dou, and D. Yan. Unsupervised multiple network alignment with multinominal gan and variational inference. In Proceedings of the 2020 IEEE International Conference on Big Data (Big Data 20), Online, December 10-13 2020. [123] D. Zhu, Z. Zhang, P. Cui, and W. Zhu. Robust graph convolutional networks against adversarial attacks. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, KDD 2019, Anchorage, AK, USA, August 4-8, 2019, pages 1399 1407, 2019. [124] Q. Zhu, X. Zhou, J. Wu, J. Tan, and L. Guo. Neighborhood-aware attentional representation for multilingual knowledge graphs. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI 2019, Macao, China, August 10-16, 2019, pages 1943 1949, 2019. [125] D. Zügner and S. Günnemann. Adversarial attacks on graph neural networks via meta learning. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019, 2019. [126] D. Zügner, A. Akbarnejad, and S. Günnemann. Adversarial attacks on neural networks for graph data. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, KDD 2018, London, UK, August 19-23, 2018, pages 2847 2856, 2018. [127] D. Zügner, O. Borchert, A. Akbarnejad, and S. Guennemann. Adversarial attacks on graph neural networks: Perturbations and their patterns. To appear in ACM Transactions on Knowledge Discovery from Data (TKDD), 2020.