# characterizing_the_evasion_attackability_of_multilabel_classifiers__18ad4cbf.pdf Characterizing the Evasion Attackability of Multi-label Classifiers Zhuo Yang,1 Yufei Han, 2 Xiangliang Zhang 1 1 King Abdullah University of Science and Technology, Thuwal, Saudi Arabia 2 Norton Research Group, Sophia Antipolis, France zhuo.yang@kaust.edu.sa, yfhan.hust@gmail.com, xiangliang.zhang@kaust.edu.sa Evasion attack in multi-label learning systems is an interesting, widely witnessed, yet rarely explored research topic. Characterizing the crucial factors determining the attackability of the multi-label adversarial threat is the key to interpret the origin of the adversarial vulnerability and to understand how to mitigate it. Our study is inspired by the theory of adversarial risk bound. We associate the attackability of a targeted multi-label classifier with the regularity of the classifier and the training data distribution. Beyond the theoretical attackability analysis, we further propose an efficient empirical attackability estimator via greedy label space exploration. It provides provably computational efficiency and approximation accuracy. Substantial experimental results on real-world datasets validate the unveiled attackability factors and the effectiveness of the proposed empirical attackability indicator. Introduction Evasion attack has been witnessed widely in real-world practices of multi-label learning (Song et al. 2018). For example, a creepware/stalkware usually has multiple malicious labels as it sniffs the victim s privacy via different mobile services. To avoid being flagged, the entities authoring these malwares (Roundy et al. 2020; Freed et al. 2018) tend to hide their key malicious labels, such as rending remotely recording phone calls or accessing private files, by slightly reprogramming app binary structures. Meanwhile, they preserve less harmful labels like GPS tracking to pretend to be benign parental control. Another example can be found in image recommendation systems. An adversary tends to embed spam/toxic advertisements (Gupta et al. 2013) into a recommended image with other harmless contents. These malicious contents are so well tuned that the sanitary check system is deceived by the camouflaged image, while recognizing correctly other harmless scenarios. Despite of the widely existence of multi-label adversarial threats, it has been a rarely investigated, yet important topic to evaluate the robustness of a multi-label classifier under evasion attack (a.k.a. attackability). Intuitively, assessing the attackability of a multi-label classifier h with an input instance is to explore the maximal perturbation on h s output that an input adversarial noise of bounded magnitudes Copyright c 2021, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. can ever cause. The problem of attackability assessment in a general setting can be defined as: given a magnitude bound of the adversarial perturbation and the distribution of legal input data instances, what is the worst-case missclassification risk of h under the attack? Classifier h is more attackable if it has a higher risk, while h is certified to be not attackable if its output cannot be changed by any adversarial noise within the magnitude bound. Via attackability assessment, we aim at answer the following questions: What are the factors determining the attackability level of a multi-label classifier? Can we derive an empirically computable attackability measurement for a multi-label classifier? Echoing the questions raises two challenges: First, analyzing the worst-case classification risk on adversarial instances with PAC-learning framework requires a fixed distribution of adversarial instances. However, it is a well received fact that such a distribution depends radically on the targeted classifier s property, thus it is not fixed and closely associated with the classifier s architecture. The celebrated works (Yin, Ramchandran, and Bartlett 2019; Khim and Loh 2018; Tu, Zhang, and Tao 2019) proposed to mitigate the gap via the lens of Rademacher complexity. Nevertheless, they all focused on single-label scenarios, thus can t be applied to answer the questions above. Second, evaluating the worstcase risk for an input instance needs to explore the maximal set of jointly attackable labels. Since labels are not mutually exclusive in multilabel tasks, the label exploration process is in nature an NP-hard mixed-integer programming problem. The adversarial noise generation method in (Song et al. 2018) applies only in the targeted attack scenario, where the attacked labels are given. Few effort has been dedicated to study the feasibility of label space exploration. To address the challenges above, we conduct both theoretical and empirical attackability analysis of a multi-label classifier. Our contributions can be summarized as follows: We measure the attackability of a multi-label classifier by evaluating the expected worst-case miss-classification loss over the distribution of adversarial examples. We instantiate the study to linear and deep neural networks, which are popularly deployed in multi-label applications. Our analysis unveils that the attack strength, the regularity of the targeted classifier and the empirical loss on un- The Thirty-Fifth AAAI Conference on Artificial Intelligence (AAAI-21) perturbed data are the external and intrinsic driving force jointly determining the attackability level. We further reveal the theoretical rationality of the low-rank regularization and adversarial training in hardening the classifier. We cast the problem of evaluating the empirical attackability level as a label space exploration process for each of the legal input instances. We further demonstrate the triviality of the label space exploration problem by formulating it as a submodular set function optimization problem. Thus it can be solved via greedy search with certified approximation accuracy. We propose a Greedy Attack Space Expansion (GASE) algorithm to address the computational bottleneck of the primitive greedy search for empirical attackability measurement. The proposed method provides a computationally economic estimator of the marginal gain obtained by adding a candidate label into the set of attacked labels. It selects the labels with the largest marginal gain to deliver an efficient exploration of attack targets. Related Works Robustness against adversaries. The emergence of evasion attack raises a severe challenge to practitioners trust on machine learning in performance-critic applications (Battista and Fabio 2018; Biggio et al. 2013; Carlini and Wagner 2017). Considerable efforts have been dedicated to detect adversarial samples, improve model designs and propose robust training methods (Cullina et al. 2019; Goodfellow, Shlens, and Szegedy 2015; Athalye, Carlini, and Wagner 2018; Fawzi, Moosavi-Dezfooli, and Frossard 2016; Szegedy et al. 2013; Xu, Evans, and Qi 2018; Madry et al. 2018; Ross and Doshi-Velez 2018; Jakubovitz and Giryes 2018; Hein and Andriushchenko 2017; Zugner and Gunnemann 2019; Bojchevski and G unnemann 2019; Raghunathan, Steinhardt, and Liang 2018; Florian et al. 2018; Papernot et al. 2016; Zugner and G unnemann 2020; Cohen, Rosenfeld, and Kolter 2019; Lee et al. 2019). Especially, (Wang et al. 2019; Shafahi et al. 2019; Gao et al. 2019) discussed convergence guarantee and high sample complexity of adversarial training. In contrast, few literature focuses on the essential problem of evaluating the vulnerability of a classifier under a given evasion attack setting and identifying key factors determining the feasibility of evasion attack against the targeted classifier. Pioneering works of this topic (Hein and Andriushchenko 2017; Wang, Jha, and Chaudhuri 2018; Fawzi, Moosavi-Dezfooli, and Frossard 2016; Gilmer et al. 2018) focused on identifying the upper bound of adversarial noise, which guarantees the stability of the targeted classifier s output, a.k.a adversarial sphere. Notably, (Fawzi, Moosavi-Dezfooli, and Frossard 2016) pointed out the association between adversarial robustness and the curvature of the classification boundary. Strengthened further by (Yin, Ramchandran, and Bartlett 2019; Khim and Loh 2018; Tu, Zhang, and Tao 2019), the expected classification risk under adversarial perturbation can be bounded by the Rademacher complexity of the targeted classifier. Moreover in (Qi et al. 2019; Wang et al. 2020), attackability of a recurrent neural net based classifier on discrete inputs was measured by checking the regularity of the targeted classifier. Apparently, the regularity of the targeted classifier play an equally important role as the attack strength in causing adversarial vulnerability. Inspiring as they are, these works focus on singlelabel learning tasks. Due to the label co-occurrence in multilabel learning, searching for the combinations of attacked labels causing the worst-case loss is NP-hard. It is thus an open issue to evaluate the adversarial risk of multi-label learners. Noise-tolerant multi-label learning. A relevant topic is to learn multi-label classifier with imperfect training instances. Miss-observations and noise corruptions of features and labels of training instances can introduce severe bias into the derived classifier. Most research efforts in this domain recognised that the key to success is to encode label correlation and the predicative relation between features and labels (Sun, Zhang, and Zhou 2010; Zhu, Yan, and Ma 2010; Liu et al. 2010; Lin et al. 2013; Wu, Jin, and Jain 2013; Zhao and Guo 2015; Yu et al. 2014; Bi and Kwok 2014; Goldberg et al. 2010; Cabral et al. 2015; Xu, Jin, and Zhou 2013; Chiang, Hsieh, and Dhillon 2015; Guo 2017; Zhu, Kwok, and Zhou 2018; Hsieh, Natarajan, and Dhillon 2015). They exploited not only low-rank structures of feature/label matrices for missing data imputation, but also gained stable performances by enforcing the low-rank regularization on the predictive model capturing the feature-label correlation. Especially (Xu et al. 2016) proposed to regularize the local Rademacher complexity of a linear multi-label classifier in the training process. The study indicated the link between the Rademacher complexity and the low-rank structure of the classifier s coefficients. The reported results showed that a lower-rank structured linear classifier can better recover missing labels. Nevertheless, all the previous works focus on adversary-free scenarios. Furthermore, the analysis over the role of low-rank structures was limited to linear multilabel classifiers. It is thus interesting to study whether the low rank driven regularization can help to mitigate the adversarial threat against both linear and DNN based multi-label classifiers. Attackability of Multi-label Classifiers Notations and Problem Definition. We assume Z = X Y as a measurable multi-label instance space, with X Rd and Y = { 1, 1}m, where d is the feature dimension and m is the number of labels. Given n i.i.d. training examples {(xi, yi)} drawn from P(Z), the classifier h H : X Y is learnt by minimizing the empirical loss Pn i=1 ℓ(xi, yi), with the loss function ℓ: X Y R. Eq.(1) defines a typical scenario of empirical attackability evaluation for a multi-label classifier h given an input xi, perturbed by r. The classification output sgn(h(xi + r )) has been flipped on as many as possible labels. C (xi) = max T, r µr j=1 I(yij =sgn(hj(xi + r ))), where r = arg min r r , s.t. yijhj(xi + r ) 0 (j T), yijhj(xi + r ) > 0 (j / T). (1) where hj(xi + r) denotes the classification score for the label j of the adversarial example, and sgn is the sign function outputting 1 based on the sign of hj(xi +r). The indicator function I( ) outputs 1 if the attack flips a label and 0 otherwise. T denotes the set of flipped labels. The magnitude of C indicates the attackability of the classifier given the attack strength limit µr and the input xi. Given the same input x and the bound of perturbation µr, one multi-label classifier h is more attackable than the other h , if C h > C h . Bound of Adversarial Attackability Beyond the attackability measurement given a local fixed input instance, we pursue a theoretical and empirical insight into the attackability of h in the space of adversarial samples, which are sampled from a new distribution P translated from P after injecting the adversarial perturbation. The distribution shift from P to P is the origin of adversarial threat, as it violates the i.i.d. assumption of the learning process. By assuming that P lies within a Wassernstein ball centered at P with a radius of ϵ, we have the following definition about classification risk under evasion attack. Definition 1. For a multi-label classifier h and legal input samples {xi, yi} P and its corresponding adversarial samples {x i, yi} P , the worst case expected and empirical risk under the evasion attack are: RP (h) = E(x,y) P[ max (x ,y) P ,W(P,P ) ϵ l(h(x ), y)], Remp P (h) = 1 i=1 [ max (x i,yi) P ,W(P,P ) ϵ l(h(x i), yi)], (2) where W(P , P) denotes the Wassernstein distance between P and P, and ϵ is the radius of the adversarial space. The W(P , P) can be bounded with the magnitude of the adversarial perturbation after (Tu, Zhang, and Tao 2019), which gives W(P , P) supx ,x x i xi 2 µr. Without loss of generality, we use the Euclidean distance x i xi 2 to constrain the attack budget hereafter. Consistent with the defined attackability evaluation scenario in Eq.1, RP (h) measures the attackability of h. A higher RP (h) indicates a more attackable h. And Remp P (h) is the empirical estimator of the attackability level. By definition, if we derive C by solving Eq.(1) for each instance (xi, yi) and adopt the binary 0-1 loss, an aggregation of the local worst-case loss Pn i=1 C (xi) gives Remp P (h). In the followings, we establish the upper bound of the attackability measurement with respect to linear and feedforwad neural network multi-label classifiers. It reveals the key factors determining the attackability level of a classifier. Given x Rd and y { 1, 1}m as the feature and label vector of a data instance, a linear multi-label classifier is h(x) = xw. The linear coefficient matrix w Rd m is defined with the spectral norm w δ Λ. Furthermore, we constrain the range of legal inputs and the adversary s strength as x 2 µx and r 2 µr respectively. Without loss of generality, a Least-Squared Error (LSE) loss is adopted to compute the classification risk, such as ℓ(x, y) = y xw 2. A distance metric for z = {x, y} is defined as d(zi, zj) = xi xj 2 + yi yj 2. Theorem 1. [Upper-bound of attackability for a linear multi-label classifier] The upper bound of RP (h) holds with at least probability of 1-σ: RP (h) Remp P (h) + 96 µxΛR(1 + µxΛ) + 12Ch π(m + 2µx) n + (m + Λµx) and the worst case empirical loss has the upper bound: Remp P (h) 1 i=1 ℓ(h(xi), yi) + Chµr, (4) where R denotes the rank of the coefficient matrix w and Ch = max{ w δ, 1}. The proof is presented in supplementary document. Remark 1. We have three observations from the derived analysis in Eq.(3-4). 1) The worst-case empirical loss Remp P can be used as a sensitive indicator of the worst-case expected adversarial risk RP . Thus it can be used as an empirical measure of attackability of the classifier over the adversarial data space. A lower Remp P implies a lower expected miss-classification risk in the adversarial space. 2) The spectrum of linear coefficient matrix w plays an important role in deciding the attackability level of h. Especially, h with lower rank w has lower expected missclassification risk. This is consistent with what was unveiled in previous research of multi-label classification: enforcing low-rank constraints over the linear classifier usually brings robustness improvement against noise corruption. 3) The empirical risk over unperturbed data, the magnitude of the adversarial perturbation and the spectrum of the classifier s coefficient matrix are the three main factors jointly determining the attackability of the multi-label classifier. On one hand, the risk upper bound depends on the external driving force of the adversarial threat, which is the magnitude of the adversarial perturbation µr . On the other hand, the internal factors on the riks upper bound are the regularity of the classifier (low-rank structure) and the profile of the training data distribution. Moreover, by dropping the terms with µr in Eq.(3), we can find that an adversary-free generalization bound of the linear multi-label classifier heavily depends on the low rank structure of the classifier. It is consistent with the results unveiled by previous works (Xu, Jin, and Zhou 2013; Yu et al. 2014; Zhu, Kwok, and Zhou 2018): low-rank structured classifiers are favorable in multi-label classification. Due to the page limit, we leave this discussion in the supplementary document. Inherited the setting of the attack scenario from Theorem.1, we consider a neural network based multi-label classifier hnn with L layers, where: The dimension of each layer is d1, d2, ..., d L, and d0 = d for taking input x and d L = m for outputting labels y . At each layer i, Ai Rdi 1 di denotes the linear coefficient matrix (connecting weights). The spectral norm of Ai is bounded as Ai δ Λi. Ri denotes the rank of Ai. The activation functions used in the same layer are Lipschitz continuous and share the same Lipschitz constant ρi. We use gi to denote the activation functions used at the layer i. The output of each layer i can be defined recursively as Hi = gi(Hi 1Ai). Theorem 2. [Upper-bound of attackability for neural nets based multi-label classifier] The upper bound of RP (hnn) holds with at least probability of 1 σ: RP (hnn) Remp P (hnn) + 2m 96 dmΛd PL i=1 Ri diΛi Ci n + 12Cnn(2µx + m) π n and the empirical loss Remp P has the upper bound: Remp P (hnn) 1 i=1 ℓ(hnn(xi), yi) + Cnnµr, (6) where C1 = ρd and Ci = Qd j=d i+1 ρj Qd j=d+2 i Aj δ with i 2, and Cnn = max{1, QL i=1 ρi Ai δ}. The proof is presented in supplementary document. Remark 2. Similar to the observations in Remark 1, the attackability of hnn depends heavily on the spectrum of linear coefficients at each layer of the neural nets, the empirical loss of hnn on legal input samples and the attack strength µr . More specifically, the linear coefficient matrices {Ai} with lower ranks and lower spectral norm can make hnn more robust. Indeed, enforcing regularization on the spectral norm of the linear coefficients can improve the generalization capability of DNN (Yoshida and Miyato 2017; Miyato et al. 2018). Our analysis not only provides the theoretical rationality behind the reported empirical observations, but also, unveils the impact of the low-rank constraint on {Ai} in controlling hnn s attackability. From Remark 1 and 2, we find that both reducing the worst-case empirical risk of the targeted classifier and enforcing low-rank constraints on its coefficients can help to reduce the attackability and mitigate the risk on evasion attack. The former can be achieved by conducting adversarial training with the crafted worst-case multi-label adversarial samples. The latter aims at controlling the Rademacher complexity of the targeted classifier, which improves the generalization capability of the targeted classifier. In (Xu and Mannor 2010), the close association between generalization and robustness was explored. Better generalization capability indicates more robustness against noise corruption. Empirical Attackability Evaluation by Greedy Exploration Problem Reformulation Solving Eq.(1) to compute the worst-case loss Remp P (h) over legal input instances is an NP-hard mixed-integer non-linear constraint problem (MINLP). Traditional solutions to this problem, such as Branch-and-Bound, has an exponential complexity in the worst case. To achieve an efficient evaluation, we propose to empirically approximate Remp P via greedy forward expansion of the set of the attacked labels. We re-formulate the label exploration problem in Eq.(1) as a bi-level set function optimization problem: S = arg max S ψ(S), where ψ(S) = max S,|S| k {|S| g(S)}, g(S) = min T S, r µr r 2 2, s.t. (1 2bj)yjhj(x + r) tj, j = 1, 2, ..., m, bj = 1 (for j T), bj = 0 (for j / T), where label yj = {+1, 1}, and tj is the minimum classification margin value enforced on label j. The core components of the constraints are the binary indicators {bj}. With bj = 1, label yj is flipped, while with bj = 0, the label remains unchanged. The set function g(S) returns the minimal magnitude of the perturbation r ever achieved via attacking the labels indicated by subsets of S. In this sense, the inner layer of Eq.(7) defines an evasion attack against the multi-label classifier targeting at the labels indicated by S. The optimization objective of the outer layer aims at expanding the set S as much as possible while minimizing as much as possible the required attack cost r 2. Notably, we set a lower bound k for |S| in Eq.(7) for the convenience of presentation. In a naive way, we can gradually increase the lower bound k until the attack cost valued by ψ(S) surpasses the budget limit. The volume |S| of the derived set gives an estimate of Remp P with the binary loss. Lemma 1. The outer layer of Eq.(7) defines a problem of non-monotone submodular function maximization. Let ψ( ˆS) and ψ(S ) denote respectively the objective function value obtained by randomized greedy forward search proposed in (Buchbinder et al. 2014) and the underlying global optimum following the cardinality lower bound constraint. The greedy search based solution has the following certified approximation accuracy: 4ψ(S ). (8) Fast Greedy Attack Space Exploration According to Lemma.1, the set S derived from the random greedy search produces an attack cost r 2 that is close to the one achieved by the global optimum solution. It guarantees the quality of the greedy search based solution. The primitive greedy forward expansion is thus designed as follows: We initialize an empty S (flipped labels), which assumes no labels are attacked at the beginning. In each round of the greedy expansion, for the current set S and current adversarial noise r(S), we choose each of the candidate labels j out of S and compute the marginal gain r(S j) 2 r(S) 2 by conducting targeted multilabel evasion attack. r(S j) 2 is the magnitude of the adversarial noise to flip all the labels in S j. We then select randomly one of the candidate labels j with the least marginal gains to update S = S j. We update r 2 by conducting an evasion attack targeting at the labels indicated by S. The expansion stops when r 2 µr. In each iteration, the primitive greedy forward expansion needs to perform evasion attack for each candidate label. It requires (m + 1)k k(k 1)/2 1 evasion attacks before including k labels in S. It is costly when the label dimension is high. To break the bottleneck, we propose a computationally economic estimator to the magnitude of the marginal gain = r(S j) 2 r(S) 2. Lemma 2. In each iteration of the greedy forward expansion, the magnitude of the marginal gain is proportional to |yjhj(x+r)| j(x+r) , where r is the current feasible adversarial perturbation. j(x + r) denotes the L2 norm of the gradient vector hj(ˆx) ˆx at the point ˆx = x + r. Therefore, instead of running evasion attack for each candidate label, we can simply choose the one with the smallest ratio |yjhj(x+r)| j(x+r) . Algorithm 1 presents the proposed Greedy Attack Space Expansion (GASE) algorithm. It only runs in total k 1 evasion attacks to reach |S| = k. In the proposed GASE algorithm, the step of greedy label expansion is equivalent to conducting the orthogonal matching pursuit guided greedy search (Elenberg et al. 2016). It enjoys fast computation, the optimal value of the objective function in Eq.(7) achieved by GASE has a guaranteed approximation accuracy to the underlying global optimum according to Theorem 1.3 in (Buchbinder et al. 2014). The step of greedy label expansion in Algorithm.1 benefits from label correlation in multi-label instances. A successful attack targeted at one label tends to bias the classification output of another highly correlated label simultaneously. The candidate label with the weakest classification margin while a large j(x+r) is thus likely to be flipped with minor update on the adversarial perturbation. Notably, the proposed GASE algorithm is independent of the choice of evasion attack methods in the step targeted evasion attack. Once the greedy search for each input instance x finishes, we use the average |S| computed over all x as the empirical attackability indicator. A larger average |S| indicates a higher attackability of the targeted multi-label classifier. Experiments In the experimental study, we aim at 1) validating the theoretical attackability analysis in Theorem 1 and 2; and 2) evaluating the empirical attackability indictor estimated by GASE for targeted classifiers. Datasets. We include 4 datasets collected from various real-world multi-label applications, cyber security practices (Creepware), biology research (Genbase) (Tsoumakas, Katakis, and Vlahavas 2010), object recognition (VOC2012) (Everingham et al. 2012) and environment research (Planet) (Kaggle 2017). The 4 datasets are summarized in Table.7. Targeted Classifiers. We instantiate the study empirically with linear Support Vector Machine (SVM) and Deep Neural Nets (DNN) based multi-label classifiers. Linear SVM is applied on Creepware and Genbase. DNN model Inception V3 is used on VOC2012 and Planet. On each data set, we Algorithm 1: Greedy Attack Space Expansion 1 Input: Instance example x, a trained multi-label classifier h, perturbation norm budget µr. 2 Output: The set of attacked labels S. 3 Initialize S as an empty set and r = 0. 4 while |S| < m and r 2 < µr do 5 Greedy label expansion: Calculate dj in Eq.(9) for each label j outside S, where hj(x + r) is the probabilistic classification output of label j, and tj is the threshold of label decision; dj = |yjhj(x + r)| j(x + r) . (9) Update S = S j, where label j(j / S) is selected randomly from the labels with the least values of Eq.(9). 6 Targeted evasion attack: Solve the targeted evasion attack problem with updated S and get the optimized perturbation r ; Update r = r . Dataset N m lavg Micro F1 Macro F1 Classifiertarget Creepware 966 16 2.07 0.76 0.66 SVM Genbase 662 27 1.25 0.99 0.73 SVM VOC2012 17,125 20 1.39 0.83 0.74 Inception-V3 Planet 40,479 17 2.87 0.82 0.36 Inception-V3 Table 1: Summary of the used real-world datasets. N is the number of instances. m is the total number of labels. lavg is the average number of labels per instance. The F1-scores of the targeted classifiers on different datasets are also reported. randomly choose 50%, 30% and 20% data instances for training, validation and testing to build the targeted multilabel classifier. In Table.7, we show Micro-F1 and Macro F1 scores derived on the unperturbed testing data. Note that feature engineering and model design of the classifiers for better classification is beyond the scope of this study. These classifiers are trained to achieve comparable classification accuracy w.r.t. the reported state-of-the-art methods on their corresponding datasets, so as to set up the test bed for the attackability analysis. Due to space limit, more experimental setting and results are provided in the supplementary file. Attack and Adversarial Training. We use adversarialrobustness-toolbox (Nicolae et al. 2018) to implement the step of targeted adversarial attack in Algorithm.1 and adversarial training. Specifically, projected gradient decent (PGD) (Madry et al. 2018) is employed to conduct the targeted attack in Algorithm.1. The decision threshold ti in Algorithm.1 is set to 0 without loss of generality. Performance Benchmark. We gradually increase the attack strength by varying the attack budget µr. Given a fixed value of µr, we calculate the average number of flipped labels on test data as an estimation of the empirical classifica- tion risk Remp P induced by the attack. This is the empirical attackability indicator, as defined in the end of the section of fast greedy attack space exploration. Validation of Empirical Attackability Indicator We assess here the empirical attackability indicator estimated by the proposed GASE algorithm, by comparing it with four baselines of label exploration strategies. PGS (Primitive Greedy Search): This is the costly primitive greedy search that requires (m+1)k k(k 1)/2 1 evasion attacks before including k labels in S. RS (Random Search): In each round of RS, one label is selected purely by random from the candidate set without evaluating the marginal gain and added to the current set S. OS (Oblivious Search): This method first computes the norm of the adversarial perturbation induced by flipping each candidate label while keeping the other labels unchanged. The labels causing the least perturbation magnitudes are selected to form the set S. LS (Loss-guided Search): In each iteration, LS updates the adversarial perturbation r along the direction where the multi-label classification loss increased the most. The set of the attacked labels are reported when r 2 surpasses the cost limit. This strategy aims at pushing the originally miss-classified instances even further from the decision plane, instead of flipping the labels of the originally correctly predicted instances. It misleads the search of the attackable labels by just maximizing the loss, and thus has bad performance as shown in Fig. 1. Fig. 1 shows the number of flipped labels obtained by the proposed GASE algorithm and the baselines on linear and DNN based multi-label classifiers. Since we limit the maximum iterations and perturbation norm bounds of attacks in our experiments, few cases of the involved label exploration methods can flip all of the labels in each dataset. Not surprisingly, the proposed GASE and PGS method achieve significantly more flipped labels than RS, OS and LS methods, especially when the constraint of attack budget is strict (with small perturbation norms). It confirms the reasonableness of greedy search stated in Lemma.1. Over all the datasets, GASE performs similarly or even better compared to PGS. It empirically demonstrates the merits of GASE: it is much less costly than PGS, while obtains attackability indicators with certified quality. Attackability Evaluation with Countermeasures for Evasion Attack Following in our Theorem 1 and 2, we study the impact of the countermeasures on multi-label classifiers attackability: controlling the model complexity by enforcing the low-rank nuclear norm constraint and conducting adversarial training. For the DNN based classifier, we enforce the nuclear norm constraint only on the linear coefficients of the final layer. We include 4 different settings on controlling the model complexity when training classifiers: With neither the low-rank nuclear-norm constraint nor adversarial training over the linear transformation coefficients, noted as lmd no and noadt, respectively. With both the nuclear norm constraint and adversarial training, noted as lmd λ and adt, where λ is the regularization parameter of the nuclear norm constraint. Without adversarial training while with the nuclear norm constraint, noted as lmd λ and noadt, respectively. With adversarial training while without the nuclear norm constraint, noted as lmd no and adt, respectively. The attackability indicators of all complexity-controlled classifiers are estimated by the proposed GASE. The results are shown in Fig. 2. The figure also shows robustness evaluation, as a low attackability indicates a high robustness. Consistently found in all datasets, the low-rank constraint has a significant stronger impact on the classifier s attackability compared to adversarial training, because the variation of λ caused larger change among curves in different colors. Classifiers trained with larger λ are more robust. Though adversarial training alone doesn t change drastically the robustness, combined with the low-rank constraint, they can make the classifiers more robust than using solely either one. The results confirmed our remarks from Theorem 1 and 2. Our experimental observations show that adversarial training doesn t change drastically classifiers performances on unperturbed test data. In contrast, there is indeed an obvious trade-off between improving a classifier s robustness by imposing the nuclear norm constraint and preserving its good utility on unperturbed test data. A strong nuclear norm constraint improves greatly the adversarial robustness. Nevertheless, it also causes accuracy loss to the classifiers. More evaluation results about the accuracy of classifiers under complexity control are in the supplementary document. Conclusion In this paper, we propose to assess the attackability of multilabel learning systems under adversarial evasion attack. We theoretically analyze the bound of the expected worst-case risk on adversarial data instances for linear and neural nets based multi-label classifiers. The resultant risk bound is used to evaluate the attackability of the targeted multi-label learning model. We unveil that the attckability depends heavily on 1) the empirical loss on the unperturbed data, 2) the rank of the targeted classifier s linear transformation coefficients and 3) the attack strength. The former two perspectives characterize the attacked multi-label learning task. The latter is decided purely by the adversary. They are the intrinsic cause and external driving force of the adversarial threat. Practically, we propose a greedy-expansion based label space exploration method to provide the empirical attackability measurement. Enjoying the submdoularity of the label space exploration problem, the empirical attackability evaluation has a certified approximation accuracy to the underlying true value. Our study intrigues the interpretability of adversarial threats of multi-label learning models. The future work will focus on proposing defensive methods for multi-learning systems with provably robustness. (a) attackability of SVM on Creepware (b) attackability of SVM on Genbase (c) attackability of Inception-V3 on VOC2012 (d) attackability of Inception-V3 on Planet Figure 1: The empirical attackability indicator estimated by different label exploration strategies. (a) attackability of controlled SVM on Creepware (b) attackability of controlled SVM on Genbase (c) attackability of controlled Inception-V3 on VOC2012 (d) attackability of controlled Inception-V3 on Planet Figure 2: The evaluation of classifiers attackability under different complexity controls. References Athalye, A.; Carlini, N.; and Wagner, D. 2018. Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples. In ICML, volume 80, 274 283. Battista, B.; and Fabio, R. 2018. Wild patterns: Ten years after the rise of adversarial machine learning. Pattern Recognition 84: 317 331. Bi, W.; and Kwok, J. T. 2014. Multilabel Classification with Label Correlations and Missing Labels. In AAAI, 1680 1686. Biggio, B.; Corona, I.; Maiorca, D.; Nelson, B.; ˇSrndi c, N.; Laskov, P.; Giacinto, G.; and Roli, F. 2013. Evasion Attacks against Machine Learning at Test Time. In ECML PKDD. Bojchevski, A.; and G unnemann, S. 2019. Certifiable Robustness to Graph Perturbations. In Neur IPS, 8319 8330. Buchbinder, N.; Feldman, M.; Naor, J.; and Schwartz, R. 2014. Submodular maximization with cardinality constraints. In SODA. Cabral, R.; la Torre, F. D.; Costeira, J. P.; and Bernardino, A. 2015. Matrix Completion for Weakly-Supervised Multi Label Image Classification. TPAMI 37(1): 121 135. Carlini, N.; and Wagner, D. 2017. Towards evaluating the robustness of neural networks. In IEEE S&P. Chiang, K.-Y.; Hsieh, C.-J.; and Dhillon, I. S. 2015. Matrix Completion with Noisy Side Information. In NIPS, 3447 3455. Cohen, J.; Rosenfeld, E.; and Kolter, Z. 2019. Certified Adversarial Robustness via Randomized Smoothing. In ICML, 1310 1320. Cullina, D.; Bhagoji, A.; Ramchandran; and Mittal, P. 2019. PAC-Learning in the presence of adversaries. In Neur IPS. Elenberg, E. R.; Khanna, R.; Dimakis, A. G.; and Negahban, S. 2016. Restricted Strong Convexity Implies Weak Submodularity. Annuals of Statistics . Everingham, M.; Gool, L. V.; Williams, C.; Winn, J.; and Zisserman, A. 2012. The PASCAL Visual Object Classes Challenge 2012 (VOC2012) Results. http://www.pascal-network.org/challenges/VOC/ voc2012/workshop/index.html . Fawzi, A.; Moosavi-Dezfooli, S.; and Frossard, P. 2016. Robustness of Classifiers: From Adversarial to Random Noise. In NIPS, 1632 1640. Florian, T.; Kurakin, A.; Papernot, N.; Boneh, D.; and Mc Daniel, P. 2018. Ensemble Adversarial Training: Attacks and Defenses. In ICLR. Freed, D.; Palmer, J.; Minchala, D.; Levy, K.; Ristenpart, T.; and Dell, N. 2018. A Stalker s Paradise : How Intimate Partner Abusers Exploit Technology. In CHI, 1 13. Gao, R.; Cai, T.; Li, H.; Hsieh, C. J.; Wang, L.; and Lee, J. D. 2019. Convergence of Adversarial Training in Overparametrized Neural Networks. In Neur IPS, 13029 13040. Gilmer, J.; Metz, L.; Faghri, F.; Schoenholz, S.; Raghu, M.; Wattenberg, M.; and Goodfellow, I. 2018. Adversarial Spheres. Co RR URL http://arxiv.org/abs/1801.02774. Goldberg, A. B.; Zhu, X.; Recht, B.; Xu, J.-M.; and Nowak, R. 2010. Transduction with Matrix Completion: Three Birds with One Stone. In NIPS, 757 765. Goodfellow, I.; Shlens, J.; and Szegedy, C. 2015. Explaining and Harnessing Adversarial Examples. In ICLR. Guo, Y. 2017. Convex Co-Embedding for Matrix Completion with Predictive Side Information. In AAAI, 1955 1961. Gupta, A.; Lamba, H.; Kumaraguru, P.; and Joshi, A. 2013. Faking Sandy: Characterizing and Identifying Fake Images on Twitter during Hurricane Sandy. In WWW, 729 736. Hein, M.; and Andriushchenko, M. 2017. Formal Guarantees on the Robustness of a Classifier against Adversarial Manipulation. In NIPS, 2266 2276. Hsieh, C.-J.; Natarajan, N.; and Dhillon, I. S. 2015. PU Learning for Matrix Completion. In ICML, 663 672. Jakubovitz, D.; and Giryes, R. 2018. Improving DNN Robustness to Adversarial Attacks Using Jacobian Regularization. In ECCV, 525 541. Springer International Publishing. Kaggle. 2017. Planet: Understanding the Amazon from Space. https://www.kaggle.com/c/planet-understandingthe-amazon-from-space/overview. Khim, J.; and Loh, P. 2018. Adversarial Risk Bounds for Binary Classification via Function Transformation. ar Xiv . Lee, G.; Yuan, Y.; Chang, S.; and Jaakkola, T. 2019. Tight Certificates of Adversarial Robustness for Randomly Smoothed Classifiers. In Neur IPS, 4910 4921. Lin, Z.; Ding, G.; Hu, M.; Wang, J.; and Ye, X. 2013. Image Tag Completion via Image-Specific and Tag-Specific Linear Sparse Reconstructions. In CVPR, 1618 1625. Liu, D.; Hua, X.-S.; Wang, M.; and Zhang, H.-J. 2010. Image Retagging. In ACM Multi Media, 491 500. Madry, A.; Makelov, A.; Schmidt, L.; Tsipras, D.; and Vladu, A. 2018. Towards Deep Learning Models Resistant to Adversarial Attacks. In ICLR. Miyato, T.; Kataoka, T.; Koyama, M.; and Yoshida, Y. 2018. Spectral Normalization for Generative Adversarial Networks. In ICLR. Nicolae, M.; Sinn, M.; Minh, T. N.; Rawat, A.; Wistuba, M.; Zantedeschi, V.; Molloy, I. M.; and Edwards, B. 2018. Adversarial Robustness Toolbox v0.2.2. Co RR URL http: //arxiv.org/abs/1807.01069. Papernot, N.; Mc Daniel, P.; Wu, X.; Jha, S.; and Swami, A. 2016. Distillation as a Defense to Adversarial Perturbations Against Deep Neural Networks. In IEEE S&P, 582 597. Qi, L.; Wu, L.; Chen, P.; Dimakis, A.; Dhillon, I.; and Witbrock, M. 2019. Discrete Attacks and Submodular Optimization with Applications to Text Classification. In Sys ML. Raghunathan, A.; Steinhardt, J.; and Liang, P. 2018. Semidefinite Relaxations for Certifying Robustness to Adversarial Examples. In Neur IPS, 10900 10910. Ross, A.; and Doshi-Velez, F. 2018. Improving the Adversarial Robustness and Interpretability of Deep Neural Networks by Regularizing their Input Gradients. In AAAI. Roundy, K. A.; Mendelberg, P.; Dell, N.; Mc Coy, D.; Nissani, D.; Ristenpart, T.; and Tamersoy, A. 2020. The Many Kinds of Creepware Used for Interpersonal Attacks. In IEEE S&P), 626 643. Shafahi, A.; Najibi, M.; Ghiasi, M. A.; Xu, Z.; Dickerson, J.; Studer, C.; Davis, L. S.; Taylor, G.; and Goldstein, T. 2019. Adversarial training for free! In Neur IPS, 3358 3369. Song, Q.; Jin, H.; Huang, X.; and Hu, X. 2018. Multi-label Adversarial Perturbations. In ICDM, 1242 1247. Sun, Y.-Y.; Zhang, Y.; and Zhou, Z.-H. 2010. Multi-Label Learning with Weak Label. In AAAI, 593 598. Szegedy, C.; Zaremba, W.; Sutskever, I.; Bruna, J.; Erhan, D.; Goodfellow, I. J.; and Fergus, R. 2013. Intriguing properties of neural networks. In ICLR. Tsoumakas, G.; Katakis, I.; and Vlahavas, I. 2010. Mining Multi-label Data, 667 685. Springer US. Tu, Z.; Zhang, J.; and Tao, D. 2019. Theoretical Analysis of Adversarial Learning: A Minimax Approach. In Neur IPS, 12259 12269. Wang, Y.; Han, Y.; Bao, H.; Shen, Y.; Ma, F.; Li, J.; and Zhang, X. 2020. Attackability Characterization of Adversarial Evasion Attack on Discrete Data. In KDD. Wang, Y.; Jha, S.; and Chaudhuri, K. 2018. Analyzing the Robustness of Nearest Neighbors to Adversarial Examples. In ICML. Wang, Y.; Ma, X.; Bailey, J.; Yi, J.; Zhou, B.; and Gu, Q. 2019. On the Convergence and Robustness of Adversarial Training. In ICML, 6586 6595. Wu, L.; Jin, R.; and Jain, A. K. 2013. Tag Completion for Image Retrieval. TPAMI 35(3): 716 727. Xu, C.; Liu, T.; Tao, D.; and Xu, C. 2016. Local Rademacher Complexity for Multi-Label Learning. IEEE Transactions on Image Processing 25(3): 1495 1507. Xu, H.; and Mannor, S. 2010. Robustness and Generalization. In COLT, 503 515. Xu, M.; Jin, R.; and Zhou, Z.-H. 2013. Speedup Matrix Completion with Side Information: Application to Multilabel Learning. In NIPS, 2301 2309. Xu, W.; Evans, D.; and Qi, Y. 2018. Feature Squeezing: Detecting Adversarial Examples in Deep Neural Networks. In NDSS. Yin, D.; Ramchandran, K.; and Bartlett, P. 2019. Rademacher Complexity for Adversarially Robust Generalization. In ICML. Yoshida, Y.; and Miyato, T. 2017. Spectral Norm Regularization for Improving the Generalizability of Deep Learning. Ar Xiv abs/1705.10941. Yu, H.-F.; Jain, P.; Kar, P.; and Dhillon, I. S. 2014. Largescale Multi-label Learning with Missing Labels. In ICML. Zhao, F.; and Guo, Y. 2015. Semi-supervised Multi-label Learning with Incomplete Labels. In IJCAI, 4062 4068. Zhu, G.; Yan, S.; and Ma, Y. 2010. Image Tag Refinement Towards Low-rank, Content-tag Prior and Error Sparsity. In ACM Multi Media, 461 470. Zhu, Y.; Kwok, J. T.; and Zhou, Z.-H. 2018. Multi-Label Learning with Global and Local Label Correlation. TKDE . Zugner, D.; and Gunnemann, S. 2019. Certifiable Robustness and Robust Training for Graph Convolutional Networks. In KDD, 246 256. Zugner, D.; and G unnemann, S. 2020. Certifiable Robustness of Graph Convolutional Networks under Structure Perturbations. In KDD, 1656 1665.