# adversarial_nonnegative_matrix_factorization__63346353.pdf Adversarial Nonnegative Matrix Factorization Lei Luo 1 2 Yanfu Zhang 2 Heng Huang 1 2 Nonnegative Matrix Factorization (NMF) has become an increasingly important research topic in machine learning. Despite all the practical success, most of existing NMF models are still vulnerable to adversarial attacks. To overcome this limitation, we propose a novel Adversarial NMF (ANMF) approach in which an adversary can exercise some control over the perturbed data generation process. Different from the traditional NMF models which focus on either the regular input or certain types of noise, our model considers potential test adversaries that are beyond the pre-defined constraints, which can cope with various noises (or perturbations). We formulate the proposed model as a bilevel optimization problem and use Alternating Direction Method of Multipliers (ADMM) to solve it with convergence analysis. Theoretically, the robustness analysis of ANMF is established under mild conditions dedicating asymptotically unbiased prediction. Extensive experiments verify that ANMF is robust to a broad categories of perturbations, and achieves state-of-the-art performances on distinct real-world benchmark datasets. 1. Introduction The nonnegative matrix factorization (NMF) has been a prevalent nonnegative dimensionality reduction method and successfully applied to many fields such as feature extraction (Zhi et al., 2011), video tracking (Bucak & Gunsel, 2007), image processing (Geng et al., 2012) and document clustering (Guan et al., 2012). Given a data matrix Y ℜm N with non-negative entries, the goal of NMF is to factorize Y into the product AX of two nonnegative matrices, with n columns in A, where n is generally small. 1JD Finance America Corporation, Mountain View, CA, USA 2Department of Electrical and Computer Engineering, University of Pittsburgh, PA, USA. Correspondence to: Heng Huang . Proceedings of the 37 th International Conference on Machine Learning, Online, Austria, PMLR 119, 2020. Copyright 2020 by the author(s). Since NMF was popularized by Lee and Seung (Lee & Seung, 1999), various NMF methods have been proposed. However, most of them independently deal with each element of X, regardless of the relationship between elements. To address this issue, some scholars considered the structural information of data or variate in modeling. Specifically, Cai et al. (Cai et al., 2010) exploited the intrinsic geometry of the data distribution and constructed a nearest neighbor graph to model the manifold structure term. Kim et al. (Kim et al., 2012) leveraged a mixed norm regularization to promote group sparsity in the factor matrices of NMF. Haeffele et al. (Haeffele et al., 2014) explored a matrix factorization technique suitable for large datasets that captures additional structure in the factors by using a projective tensor norm. It is noteworthy that the NMF models mentioned above are limited to the regular input, and they are more inclined to use the simple L2-norm to characterize the residual between matrices Y with AX. Such a strategy obviously contradicts to the practical observations where data often contains noise which may have heavy-tailed attribute. Thus, in recent research, Robust Nonnegative Matrix Factorization (RNMF) methods have gained increasing attentions. For instance, Kong et al. (Kong et al., 2011) presented a novel robust formulation of NMF by using L2,1-norm loss function which can accommodate outliers and noises in a better way than L2-norm. Subsequently, Huang et al. (Huang et al., 2014) generalized (Kong et al., 2011) by adopting manifold regularization. Based on the correntropy induced metric, Du et al. (Du et al., 2012) introduced a robust NMF method which can effectively cope with the practical noise. Guan et al. (Guan et al., 2017) proposed a Truncated Cauchy NMF loss that handles outliers by truncating large errors, while Gao et al. (Gao et al., 2015) employed capped norm to remove the effect of extreme data outliers in NMF. However, the RNMF methods are only suitable for some special types of noises, e.g., Laplacian or Cauchy noise, which cannot show the flexibility in facing the worst-case (i.e., adversarial) perturbations of data points. Compared with random noise, perturbations caused by specific features or the noise are more common and complex, but often contaminate the input data in practical applications. Sometimes, they are hardly perceptible to the human eye, yet sufficient to change the output of an algorithm. As a result, how to increase the robustness of models against the general Adversarial Nonnegative Matrix Factorization perturbations has become a very important task in NMF. To address this challenging problem, in this paper, we introduce a novel Adversarial Nonnegative Matrix Factorization (ANMF) model by emphasizing potential test adversaries that are beyond the pre-defined constraints. Specifically, we leverage adversarial perturbations of Y to learn the adversarial feature matrix A data and weight matrix X. Differing from the traditional NMF models which either focus on the regular data point or use the simple matrix norm to characterize error between matrices AX and Y, our model fully utilizes an adversary of input data Y to exercise some control over the data generation process to improve the stability of A. Thus, it does not rely on any noise assumption. The proposed model is formulated as a bilevel optimization problem and an efficient optimization algorithm is derived to solve it with the convergence analysis. In addition, we establish the complete theoretical guarantees for ANMF under mild conditions. These results soundly support the rationality of ANMF. The main contributions of this paper are summarized as follows: From learner and attacker perspectives, we propose a novel Adversarial Nonnegative Matrix Factorization (ANMF) model which can handle different types of noise or perturbations. The robustness analysis is provided under different conditions, which theoretically guarantees the soundness of our model. Alternating Direction Method of Multipliers is applied to solving ANMF, where we can achieve the closedform solution for each sub-problem. Empirical studies are performed on real-world benchmark data sets with various noise conditions. All results demonstrate that the proposed algorithm can consistently outperform other related methods. Notations: Throughout this paper, the bold capital and bold lowercase symbols are used to represent matrices and vectors, respectively. If all elements of a matrix A are greater than or equal to 0, we denote it by A 0. ||A||F and ||A||2,1 mean Frobenius norm and L2,1-norm of the matrix A, respectively. Finally, a p p-identity matrix is denoted by Ip and 0 denotes is a zero matrix. 2. Backgrounds Nonnegative matrix factorization. The standard NMF models factorize an observation matrix Y = [y1, y2, , y N] ℜm N as two nonnegative factors A = [a1, a2, , an] ℜm n and X = [x1, x2, , x N] ℜn N such that Y AX. A is often referred to as feature matrix and X referred as weights. It has been shown that the non-negativity constraint on the coefficients forcing features to combine, but not cancel out, can lead to much more interpretable features and improved downstream performance of the learned features. The standard NMF can be formulated as follows: min A,X Y AX 2 F , s.t., A 0, X 0. (1) Numerous algorithms have been developed in the literature for finding its high-quality solutions. The most representative algorithm is multiplicative update which alternates between solving certain surrogate functions for A and X, respectively (Gonzalez & Zhang, 2005). The Alternating Nonnegative Least Square (Lin, 2007) is another class of useful algorithms, which includes the projected gradient descent method (Hajinezhad et al., 2016). Since the L2-norm is sensitive to the practical noises, intensive efforts have been put to design robust NMF models. The common strategy of these methods is to adopt a special loss function to characterize the errors between Y and AX, e.g., L2,1-norm (Kong et al., 2011), Capped norm (Gao et al., 2015), Correntropy induced metric (Du et al., 2012) and Truncated Cauchy function (Guan et al., 2017). Adversarial Perturbations. State-of-the-art machine learning models have achieved high accuracy on a broad range of datasets, yet can be easily misled by small perturbations of their input. While such perturbations may be simple noise to a human or even imperceptible, they often cause modern models to misclassify their input with high confidence. To provide adversarial robustness against the adversarial attack, a standard technique, which is called adversarial training (Farnia et al., 2018), follows empirical risk minimization training over the adversarially-perturbed samples by solving i=1 l(fw(vi + w(vi)), bi), (2) where l( , ) is a loss function, fw( ) is the output function, w(vi) is the adversarial additive perturbation (or noise) for sample vi and each bi is a test sample or the label vector. 3. Adversarial Nonnegative Matrix Factorization Model (1) can be further written as: min A,x1, ,x N i=1 yi Axi 2 2, s.t., A 0, xi 0, i = 1, 2, , N. Although some robust NMF methods (Kong et al., 2011; Gao et al., 2015; Du et al., 2012) may be effective for handling practical noise, their aim is to search for an appropriate loss function to characterize the errors with special Adversarial Nonnegative Matrix Factorization properties such that outlying data have a relatively smaller influence in the process of learning variables. Thus, the given loss function is only suitable for some specific noise. For other types of noises, it may be ineffective. Additionally, these robust NMF methods highly rely on the independence assumption of noise pixels, which is unrealistic for some structured noises caused by occlusions or illumination. More recently, several efforts have focused on proactive approaches of modeling the learner and adversary as players in a game in which the learner chooses a classifier or a learning algorithm, and the attacker modifies either the training or test data (Li & Vorobeychik, 2014; Großhans et al., 2013; Tong et al., 2018). These methods have shown great potential when handling data with adversarial perturbations. In this paper, we consider the data perturbed by noises as the attacker and present an adversarial version of (3) to improve the performance of NMF, where our main task is to model the adversary s attack strategy and develop robust learning models to mitigate the attack. For the convenience, we assume that the learned feature data A and given data Y are drawn from an unknown distribution D at training time. By contrast, at application time, the test data can be generated either from D, the same distribution as the training data, or from D, a modification of D generated by an attacker. The action of the learner is to select parameters {x1, x2, , x N} of the Eq. (3). It is assumed that the attacker has an instance-specific target, and encourages that the prediction made by learner on the modified instance, yi = Axi (i = 1, , N), is close to this target. The cost functions of each learner (Cl) and the attacker (Ca) are estimated by: Cl(X, A, A) = αG( AX, Y) + βG(AX, Y) (4) and Ca(X, A, A) = G( AX, Z) + λG( A, A). (5) Here parameters α, β, λ > 0, G( , ) is a given metric and Z = Z(Y) is the instance-specific target for attacker. Ultimately, our model is expressed as: min X,A Cl(X, A (X), A) (6) s.t. A (X) = arg min A Ca(X, A, A), X 0, A 0. Our method provides a general framework for dealing with practical perturbations which may be caused by Laplace, Gaussian or structured noises. It does not depend on any assumption on perturbations (or noises). Thus, compared with those robust methods, model (6) is more adaptive for the practical NMF problems. Remark 1. In general, Z can be set using two approaches: (a) Z = Y + t1, where t is sampled from [0, 5σr] and σr is the standard deviation of test image samples [y1, y2, , y N], 1 is a matrix with all elements equal to one; (b) Z = Y + , where is the practical noises caused by outlier, illuminations or occlusions. Remark 2. We use manual perturbations in model (6). In fact, we can impose some priors on perturbations to automatically learn them. But here we omit such a process. Model (6) is factually a bilevel optimization problem. Specially, if the feature matrix A is fixed, it can be considered as the matrix version of the Stackelberg Equilibrium problem (Tong et al., 2018) (More details can be seen in the supplementary materials). In this paper, we set G( A, A) = A A 2 F . Then, for the lower level, we have the following closed-form solution: Theorem 1. Given X, the best response of the attacker is A (X) = (λA + ZXT )(λIn + XXT ) 1. (7) Since there is an inverse of complicated matrix in (7), it is difficult to solve problem (6) by directly substituting (7) into (6). To mitigate this limitation, we consider (7) as a constraint of (6), which leads to the following problem: min X,A, A Cl(X, A, A), (8) s.t. A(λI + XXT ) (λA + ZXT ) = 0, X 0, A 0. Let ϕ( A, X) = A(λI + XXT ) (λA + ZXT ). Problem (8) can be approximated as: min X,A, A Cl(X, A, A) + γ ϕ( A, X) 2 F , s.t., X 0, A 0. (9) Thus, we focus on model (9) in this paper. 4. Theoretical Analysis We define the empirical reconstruction error of NMF as follows: i=1 min x ||yi Ax||2 2. (10) Denote the expectation operator by E. Then, the expected reconstruction error of ANMF can be written as R(A) = Ey RN(A). (11) The Rademacher complexity is defined as: R(F) = Eσ,x sup f F i=1 σif(xi) (12) σ1, , σn are independent Rademacher variables, f( ) belongs to the [a, b]-value function class on X and x = Adversarial Nonnegative Matrix Factorization (x1, x2, , x N)T RN are independent and identically distributed examples. Based on the above theorems and definitions, we can obtain the generalization bound of ANMF. Theorem 2. For ANMF problem, assume that Y is upper bound by 1. For any learned normalized A and any δ > 0, with probability at least 1 δ, we have |R(A) RN(A)| min mnln(4(1 + n) mn N) ln δ It is easy to see that standard NMF problem (1) is a special case (α, γ = 0) of ANMF. Similar to (Liu et al., 2016), we can obtain a dimensionality independent generalization bound for standard NMF problem (1) when the feature matrx A is orthogonal, i.e., Theorem 3. For orthogonal NMF problem, assume that Y upper bounded by 1. For any learned normalized A with and any δ > 0, with probability at least 1 δ, we have |R(A) RN(A)| 6n r π Both Theorems 2 and 3 point out the asymptotically unbiased prediction of NMF methods. Specifically, Theorem 2 provides a slightly weaker but meaningful generalization bound for ANMF. Although Theorem 3 depends on the mild orthogonality assumption, its explicit representation can reflect more clearly the effectiveness of orthogonal ANMF than Theorem 2. All proofs are in supplementary materials. 5. Optimization Algorithm In this section, we apply the Alternating Direction Method of Multipliers (ADMM) optimization algorithm to solve problem (9). The ADMM or the Augmented Lagrange Multipliers (ALM) method was presented originally in (Gabay & Mercier, 1975) and (Fortin & Glowinski, 2000), which has been extensively studied in the theoretical frameworks of Lagrangian functions (Fortin & Glowinski, 2000). Recently, it has been shown that ADMM are efficient for many convex and nonconvex programming problems arising from various applications (Chartrand & Wohlberg, 2013; Wang et al., 2015; Mei et al., 2018). It should be noted that the iterative scheme of ADMM integrates the Gaussian-Seidel decomposition into iterations of the ALM in (Wang et al., 2015), which implies that the functions with regard to different variables are treated individually. Accordingly, the easier sub-problems could be Algorithm 1 Solving Eq. (9) via ADMM Input: Y Rm N and instance-specific target Z Output: feature matrix A Rm n and weight matrix X Rn N. Initialization: A and X using the traditional K-means method, A0 = AK, where AK is the clustering centroid obtained by K-means method, X0 = XK + 0.2, where XK is the K-means clustering result. U0 = X0T , B 0 = A 0 = A0, M0 = 0. repeat Update (H, B, J) by Hk+1 R1(2αIN + µIN + 2γUk Uk T ) 1, B k+1 max(0, A + 1 µMk 1), Jk+1 R2 2β + µ, Update (U, B) by Uk+1 max(0,(2γ(Hk+1 Z)T (Hk+1 Z) + µIN) 1R3), B k+1 max(0, Ak + 1 Update A by Ak+1 R4(2γλ2In + µIn + Xk Xk T ) 1, Update A by A k+1 R5(2γλ2In + µIn + Xk Xk T ) 1, Update X by Xk+1 (( A k+1)T A k+1+Ak+1T Ak+1+In) 1R6; Update M by Mk+1 Mk + µ(φ( A k+1, Ak+1)ψ(Xk+1) ν( Hk+1, Jk+1, B k+1, Bk+1, Uk+1)). until Converge generated. For the splitting, several auxiliary variables are introduced, and (9) is transformed into the following equivalent form: min H,J,U,X,A, A,B, B α H Y 2 F + β J Y 2 F + γ HU + λ A λA ZU 2 F , s.t. AX = H, AX = J, XT = U, A = B, A = B, B 0, B 0, U 0. Adversarial Nonnegative Matrix Factorization Let us write the augmented Lagrange function for the problem (15) as : Lµ(H, B, J, U, B, A, A, X, M) = α H Y 2 F + β J Y 2 F + γ HU + λ A λA ZU 2 F + µ 2 ( A B + 1 µM2 2 F + AX H + 1 µM4 2 F + XT U + 1 where µ > 0 is a tunable penalty parameter. Then, ADMM is applied to minimizing the augmented Lagrangian problem (16) with respect to H, J, U, X, A, A, B, B alternately. The iterative scheme of ADMM for problem (15) is summarized in Algorithm 1, where φ( A, A) = diag( A, A, A, A, IN), ψ(X) = diag(X, X, In, In, XT ), (17) ν( H, J, B, B, U) = diag(H, J, B, B, U), M = diag(M1, M2, M3, M4, M5), (18) R1 = 2αY + 2γ(λAk + ZUk λ A k)Uk T + µ( A k Xk + 1 R2 = 2βY + µ( A k Xk + 1 R3 = 2γ(Hk+1 Z)T (λAk λ A k) + µ(Xk T + 1 µMk 5), (20) R4 =2γλ(Hk+1Uk+1 + λ A k ZUk+1) µMk 2) + µ(Jk+1 1 µMk 4)Xk T , (21) R5 =2γλ(λAk+1 + ZUk+1 Hk+1Uk+1) + µ( B k+1 1 µMk 1) + µ(Hk+1 1 µMk 3)Xk T , R6 = ( A k+1)T (Hk+1 1 + (Ak+1)T (Jk+1 1 µMk 4) + Uk+1T 1 For detailed derivations of Algorithm 1, we refer the readers to the supplementary materials. Convergence Analysis. We provide a partial result on the convergence of the proposed Algorithm 1 by virtue of KKT conditions of problem (9). To simplify notations, let us define Ω= (H, B, J, U, B, A, A, X, M). (24) Theorem 4. Let {Ωk} k=1 be a sequence generated by Algorithm 1 that satisfies the condition lim k (Ωk+1 Ωk) = 0. (25) Then any accumulation point of {Ωk} k=1 is a KKT point of problem (15). Consequently, any accumulation point of (Ak, A k, Xk) k is a KKT point of problem (9). Corollary 1. Whenever {Ωk} k=1 converges, it converges to a KKT point. As a consequence, for Algorithm 1, it is enough that we only need to choose a proper termination parameter ϵ > 0, and use the termination condition: Y Ak Xk 2 F ϵ and Ak+1 Ak|2 F ϵ. (26) All proofs of above theorem and corollary can be found in supplementary materials 6. Experiments and Discussions Experiments were carried out on multiple real-world data sets. Throughout the experiments, we set ANMF parameters as α = 0.6, β = 10 5, γ = 10 3, λ = 10 3, and µ = 1. For detailed description of datasets and more comprehensive results, we refer the readers to supplementary materials. 6.1. Comparison with Baselines Some representative methods, including Standard Nonnegative Matrix Factorization (SNMF), L2,1-norm based NMF model (Ding et al., 2006), Orthogonal Nonnegative Matrix Factorization (ONMF) (Kong et al., 2011), and Capped norm Nonnegative Matrix Factorization (CNMF) (Gao et al., 2015), are compared with the ANMF. It should noted that the main novelty of this paper is to consider potential test adversaries in modeling, not the robust characterization for noise. Thus, it is unfair to compare our method with some robust methods such as Correntropy induced NMF model and Truncated Cauchy NMF model. The detailed results for clustering accuracy and normalized mutual information results are shown in Table 2 and Table 3 (The best results are marked in bold). It can be observed that the advantage of ANMF is quite evident. Although L2,1norm based NMF is a robust NMF method, the ignoring of test adversaries leads to the undesired performance. Compared to NMF, ONMF achieves the better results, which indicates that the orthogonal constraint w.r.t. weight matrix X is helpful for improving the performance. However, our Adversarial Nonnegative Matrix Factorization Table 1. Description of Benchmark Datasets Dataset Number of Instances Dimensions Classes Category MNIST 150 784 10 image Yale 165 1024 15 image ORL 400 644 40 image UMIST 575 644 20 image COIL-20 1440 1024 20 image USPS 9298 256 10 image BBCsports 737 4613 5 text BBCNews 2225 9635 5 text Web KB 4199 7770 4 text Reuters 9298 256 10 text RCV 9625 29992 4 text TDT2 9394 36771 30 text Table 2. ACC of noise-free Real Datasets. The best results are marked in bold. Dataset NMF ONMF L2, 1-norm NMF CNMF ANMF MNIST 0.7933( 0.0497) 0.7987( 0.0441) 0.8027( 0.0410) 0.7947( 0.0228) 0.8067( 0.0490) Yale 0.4388( 0.0233) 0.4145( 0.0360) 0.4424( 0.0235) 0.4036( 0.0380) 0.4509( 0.0164) ORL 0.7005( 0.0060) 0.6420( 0.0356) 0.6895( 0.0139) 0.5935( 0.0323) 0.7305( 0.0294) UMIST 0.4880( 0.0285) 0.4616( 0.0295) 0.4845( 0.0255) 0.4442( 0.0235) 0.4946( 0.0186) COIL-20 0.6692( 0.0215) 0.6626( 0.0264) 0.6578( 0.0130) 0.6601( 0.0300) 0.6833( 0.0162) USPS 0.7468( 0.0004) 0.7738( 0.0003) 0.7550( 0.0002) 0.7429( 0.0050) 0.7780( 0.0002) BBCSport 0.9493( 0.0007) 0.9460( 0.0024) 0.9468( 0.0006) 0.9327( 0.0064) 0.9531( 0.0031) BBC 0.9604( 0.0011) 0.9619( 0.0028) 0.9597( 0.0002) 0.9202( 0.0032) 0.9649( 0.0010) Web KB 0.6619( 0.0095) 0.6657( 0.0038) 0.6618( 0.0083) 0.6525( 0.0117) 0.6672( 0.0084) Reuters 0.7836( 0.0059) 0.7495( 0.0164) 0.7788( 0.0071) 0.7197( 0.0112) 0.8047( 0.0098) RCV 0.6458( 0.0194) 0.6493( 0.0054) 0.6420( 0.0183) 0.6280( 0.0021) 0.6516( 0.0137) TDT2 0.8546( 0.0067) 0.8246( 0.0119) 0.8448( 0.0046) 0.8062( 0.0150) 0.8638( 0.0176) Table 3. ACC of noisy Real Datasets. The best results are marked in bold. Noise Dataset NMF ONMF L2, 1-norm NMF CNMF ANMF MNIST 0.8067( 0.0464) 0.8093( 0.0379) 0.8080( 0.0477) 0.8067( 0.0254) 0.8160( 0.0421) Yale 0.3879( 0.0321) 0.3527( 0.0248) 0.3806( 0.0168) 0.3576( 0.0223) 0.4036( 0.0180) UMIST 0.4800( 0.0150) 0.4602( 0.0268) 0.4814( 0.0086) 0.4275( 0.0162) 0.5078( 0.0124) ORL 0.6155( 0.0252) 0.5475( 0.0083) 0.6225( 0.0275) 0.5350( 0.0173) 0.6670( 0.0248) COIL-20 0.6723( 0.0247) 0.6547( 0.0190) 0.6762( 0.0175) 0.6782( 0.0316) 0.6830( 0.0194) USPS 0.7542( 0.0004) 0.7716( 0.0003) 0.7592( 0.0003) 0.7505( 0.0058) 0.7793( 0.0002) MNIST 0.7880( 0.0417) 0.7733( 0.0194) 0.7800( 0.0481) 0.7453( 0.0202) 0.8027( 0.0293) Yale 0.3867( 0.0301) 0.3261( 0.0464) 0.3576( 0.0424) 0.3394( 0.0346) 0.4012( 0.0286) UMIST 0.4706( 0.0261) 0.4483( 0.0170) 0.4720( 0.0235) 0.4310( 0.0269) 0.4866( 0.0223) ORL 0.5370( 0.0141) 0.4850( 0.0157) 0.5145( 0.0192) 0.4650( 0.0190) 0.5600( 0.0366) COIL-20 0.6829( 0.0117) 0.6469( 0.0209) 0.6850( 0.0293) 0.6229( 0.0345) 0.6924( 0.0337) USPS 0.7520( 0.0002) 0.7638( 0.0009) 0.7527( 0.0007) 0.7269( 0.0003) 0.7654( 0.0006) MNIST 0.8107( 0.0494) 0.8040( 0.0376) 0.8093( 0.0543) 0.7920( 0.0311) 0.8160( 0.0423) Yale 0.3597( 0.0175) 0.3547( 0.0309) 0.3651( 0.0330) 0.3519( 0.0158) 0.3852( 0.0273) UMIST 0.4525( 0.0302) 0.4737( 0.0242) 0.4710( 0.0259) 0.4223( 0.0248) 0.4828( 0.0251) ORL 0.5465( 0.0243) 0.5145( 0.0288) 0.5520( 0.0198) 0.4695( 0.0368) 0.5680( 0.0207) COIL-20 0.5274( 0.0209) 0.5145( 0.0121) 0.5278( 0.0054) 0.5293( 0.0284) 0.5315( 0.0198) USPS 0.5210( 0.0005) 0.5296( 0.0074) 0.5195( 0.0018) 0.5306( 0.0078) 0.5327( 0.0065) MNIST 0.2493( 0.0037) 0.2440( 0.0060) 0.2427( 0.0060) 0.2480( 0.0056) 0.2497( 0.0163) Yale 0.5468( 0.0261) 0.4982( 0.0458) 0.5406( 0.0266) 0.4861( 0.0262) 0.5549( 0.0301) UMIST 0.2247( 0.0056) 0.2115( 0.0089) 0.2235( 0.0087) 0.2136( 0.0051) 0.2271( 0.0057) ORL 0.3250( 0.0127) 0.2880( 0.0132) 0.3230( 0.0110) 0.2780( 0.0082) 0.3485( 0.0146) COIL-20 0.6792( 0.0202) 0.6706( 0.0228) 0.6782( 0.0181) 0.6586( 0.0177) 0.6827( 0.0145) USPS 0.7388( 0.0002) 0.7602( 0.0001) 0.7455( 0.0001) 0.7320( 0.0001) 0.7559( 0.0003) Adversarial Nonnegative Matrix Factorization (c) regular (d) irregular (g) regular (h) irregular Figure 1. Illustrations of face datasets Yale (a)-(d) and ORL (e)-(h) with different types of noises (pixel, regular, irregular). From top row to bottom row: origin, noisy data, noisy Z, A, A -1000 -800 -600 -400 -200 0 200 400 600 800 1000 -600 (a) A on ORL -500 -400 -300 -200 -100 0 100 200 300 400 -400 (b) A on ORL -600 -500 -400 -300 -200 -100 0 100 200 300 -300 (c) A on Yale -500 0 500 1000 -800 (d) A on Yale Figure 2. Visualizing Feature Matrices A and A via T-SNE on ORL and Yale Datasets Table 4. ACC of Real Datasets with Salt & Pepper Noise. The best results are marked in bold. Dataset K-Means NMF ONMF ℓ2,1 NMF CNMF ANMF MNIST 0.7893( 0.0342) 0.8067( 0.0464) 0.8093( 0.0379) 0.8080( 0.0477) 0.8067( 0.0254) 0.8160( 0.0421) Yale 0.3503( 0.0259) 0.3879( 0.0321) 0.3527( 0.0248) 0.3806( 0.0168) 0.3576( 0.0223) 0.4036( 0.0180) UMIST 0.4734( 0.0157) 0.4800( 0.0150) 0.4602( 0.0268) 0.4814( 0.0086) 0.4275( 0.0162) 0.5078( 0.0124) ORL 0.5720( 0.0195) 0.6155( 0.0252) 0.5475( 0.0083) 0.6225( 0.0275) 0.5350( 0.0173) 0.6670( 0.0248) COIL-20 0.6678( 0.0123) 0.6723( 0.0247) 0.6547( 0.0190) 0.6762( 0.0175) 0.6782( 0.0316) 0.6830( 0.0194) USPS 0.7706( 0.0008) 0.7542( 0.0004) 0.7716( 0.0003) 0.7592( 0.0003) 0.7505( 0.0058) 0.7793( 0.0002) Table 5. NMI of Real Datasets with Salt & Pepper Noise. The best results are marked in bold. Dataset K-Means NMF ONMF ℓ2,1 NMF CNMF ANMF Yale 0.4008( 0.0268) 0.4393( 0.0261) 0.4070( 0.0277) 0.4255( 0.0229) 0.4144( 0.0203) 0.4509( 0.0168) UMIST 0.5917( 0.0123) 0.5941( 0.0168) 0.5821( 0.0198) 0.5959( 0.0098) 0.5467( 0.0157) 0.6123( 0.0083) ORL 0.7651( 0.0136) 0.7807( 0.0132) 0.7410( 0.0049) 0.7844( 0.0234) 0.7265( 0.0156) 0.8077( 0.0154) MNIST 0.7464( 0.0175) 0.7648( 0.0241) 0.7689( 0.0217) 0.7676( 0.0242) 0.7580( 0.0179) 0.7776( 0.0252) USPS 0.6565( 0.0011) 0.6349( 0.0003) 0.6581( 0.0002) 0.6478( 0.0001) 0.6356( 0.0053) 0.6594( 0.0006) COIL-20 0.7528( 0.0149) 0.7506( 0.0174) 0.7497( 0.0131) 0.7546( 0.0079) 0.7512( 0.0136) 0.7584( 0.0173) method is more competitive than others on the all databases. Therefore, considering the adversarial perturbations in modeling can increase the robustness of NMF. Figure 1 provides the visual decomposition results on Yale and ORL under different noise. For each figure, we present the noise-free images, Z, Z, A and A from top to bottom. The results indicate that our method can successfully avoid redundancy of features and learn the desired feature matrices A and A with good representation performance. Figure 2 illustrates distribution of A and A with dimensionality reduced using T-SNE (Maaten & Hinton, 2008). We can see that all data points are fully separated. To demonstrate the robustness of the proposed method, we also include the experimental results on image data corrupted by various noise. For each type we corrupt images successively to generate X, Z and Z (which includes three Adversarial Nonnegative Matrix Factorization Table 6. ACC of Real Datasets with Corrupt Pixel Noise. The best results are marked in bold. Dataset K-Means NMF ONMF ℓ2,1 NMF CNMF ANMF Yale 0.3188( 0.0355) 0.3867( 0.0301) 0.3261( 0.0464) 0.3576( 0.0424) 0.3394( 0.0346) 0.4012( 0.0286) UMIST 0.4557( 0.0134) 0.4706( 0.0261) 0.4483( 0.0170) 0.4720( 0.0235) 0.4310( 0.0269) 0.4866( 0.0223) ORL 0.4775( 0.0137) 0.5370( 0.0141) 0.4850( 0.0157) 0.5145( 0.0192) 0.4650( 0.0190) 0.5600( 0.0366) MNIST 0.7547( 0.0145) 0.7880( 0.0417) 0.7733( 0.0194) 0.7800( 0.0481) 0.7453( 0.0202) 0.8027( 0.0293) USPS 0.7619( 0.0007) 0.7520( 0.0002) 0.7638( 0.0009) 0.7527( 0.0007) 0.7269( 0.0003) 0.7654( 0.0006) COIL-20 0.6332( 0.0262) 0.6829( 0.0117) 0.6469( 0.0209) 0.6850( 0.0293) 0.6229( 0.0345) 0.6924( 0.0337) Table 7. NMI of Real Datasets with Corrupt Pixel Noise. The best results are marked in bold. Dataset K-Means NMF ONMF ℓ2,1 NMF CNMF ANMF Yale 0.3618( 0.0334) 0.4200( 0.0274) 0.3755( 0.0466) 0.3983( 0.0360) 0.3908( 0.0287) 0.4332( 0.0226) UMIST 0.5627( 0.0061) 0.5760( 0.0162) 0.5542( 0.0115) 0.5782( 0.0143) 0.5416( 0.0247) 0.5857( 0.0147) ORL 0.6868( 0.0120) 0.7195( 0.0148) 0.6830( 0.0132) 0.7120( 0.0134) 0.6722( 0.0161) 0.7295( 0.0240) MNIST 0.7170( 0.0278) 0.7576( 0.0306) 0.7540( 0.0202) 0.7534( 0.0364) 0.7192( 0.0180) 0.7642( 0.0144) USPS 0.6487( 0.0007) 0.6358( 0.0002) 0.6486( 0.0008) 0.6358( 0.0005) 0.6079( 0.0007) 0.6495( 0.0011) COIL-20 0.7548( 0.0139) 0.7670( 0.0100) 0.7529( 0.0151) 0.7659( 0.0174) 0.7369( 0.0181) 0.7685( 0.0093) Table 8. ACC of Real Datasets with regular patch noise. The best results are marked in bold. Dataset K-Means NMF ONMF ℓ2,1 NMF CNMF ANMF MNIST 0.7947( 0.0296) 0.8107( 0.0494) 0.8040( 0.0376) 0.8093( 0.0543) 0.7920( 0.0311) 0.8160( 0.0423) Yale 0.3464( 0.0308) 0.3597( 0.0175) 0.3547( 0.0309) 0.3651( 0.0330) 0.3519( 0.0158) 0.3852( 0.0273) UMIST 0.4525( 0.0302) 0.4737( 0.0242) 0.4449( 0.0272) 0.4710( 0.0259) 0.4223( 0.0248) 0.4828( 0.0251) ORL 0.5430( 0.0151) 0.5465( 0.0243) 0.5145( 0.0288) 0.5520( 0.0198) 0.4695( 0.0368) 0.5680( 0.0207) COIL-20 0.5188( 0.0095) 0.5274( 0.0209) 0.5145( 0.0121) 0.5278( 0.0054) 0.5293( 0.0284) 0.5315( 0.0198) USPS 0.5218( 0.0057) 0.5210( 0.0005) 0.5296( 0.0074) 0.5195( 0.0018) 0.5306( 0.0078) 0.5327( 0.0065) Table 9. NMI of Real Datasets with regular patch Noise. The best results are marked in bold. Dataset K-Means NMF ONMF ℓ2,1 NMF CNMF ANMF MNIST 0.7538( 0.0141) 0.7731( 0.0237) 0.7725( 0.0197) 0.7723( 0.0327) 0.7420( 0.0222) 0.7775( 0.0217) Yale 0.3805( 0.0467) 0.3967( 0.0184) 0.3860( 0.0175) 0.3989( 0.0536) 0.3905( 0.0267) 0.4166( 0.0319) UMIST 0.5311( 0.0247) 0.5458( 0.0167) 0.5204( 0.0260) 0.5424( 0.0207) 0.4942( 0.0210) 0.5511( 0.0209) ORL 0.7117( 0.0150) 0.7146( 0.0169) 0.6876( 0.0266) 0.7159( 0.0140) 0.6579( 0.0334) 0.7257( 0.0141) COIL-20 0.6104( 0.0147) 0.6073( 0.0153) 0.6138( 0.0119) 0.6027( 0.0064) 0.6157( 0.0218) 0.6157( 0.0155) USPS 0.3924( 0.0079) 0.3720( 0.0031) 0.3846( 0.0039) 0.3730( 0.0065) 0.3703( 0.0173) 0.4001( 0.0137) stages of noise). In detail, two types of noise are considered: salt & pepper noise, in which we corrupt 10% pixels in each stage and the results are summarized in Table 4 and Table 5; random corrupted pixels, in which we corrupt 20%, 5%, and 5% pixels in each stage and the results are summarized in Table 6 and Table 7; random corrupted regular patches (), in which we corrupt 50%, 25%, and 25% pixels in each stage (see Figure 1) and the results are summarized in Table 8 and Table 9. The robustness of the proposed method is clearly verified according to these tables. Due to the space limitation, experimental results for more types of noise, convergence curves of Algorithm 1 and the comparison with other methods can be found in supplementary materials. 7. Conclusion This paper focuses on nonnegative matrix factorization problem. To provide the robustness against real perturbations, we propose a new Adversarial Nonnegative Matrix Factorization model The adversarial perturbations of Y is used to learn the desire feature matrix A and weight matrix X. Different from the traditional NMF models which only focus on the regular data points, our models emphasizes potential test adversaries that are beyond the pre-defined constraints. Adversarial Nonnegative Matrix Factorization We formulate the proposed model as a bilevel optimization problem and utilize ADMM to solve it with convergence guarantee. Experimental results on real data sets validate the effectiveness and robustness of the proposed algorithm. Acknowledgements This work was partially supported by U.S. NSF IIS 1836945, IIS 1836938, IIS 1845666, IIS 1852606, IIS 1838627, IIS 1837956. Bucak, S. S. and Gunsel, B. Video content representation by incremental non-negative matrix factorization. In Image Processing, 2007. ICIP 2007. IEEE International Conference on, volume 2, pp. II 113. IEEE, 2007. Cai, D., He, X., Han, J., and Huang, T. S. Graph regularized nonnegative matrix factorization for data representation. IEEE transactions on pattern analysis and machine intelligence, 33(8):1548 1560, 2010. Chartrand, R. and Wohlberg, B. A nonconvex admm algorithm for group sparsity with sparse groups. In Acoustics, Speech and Signal Processing (ICASSP), 2013 IEEE International Conference on, pp. 6009 6013. IEEE, 2013. Ding, C., Li, T., Peng, W., and Park, H. Orthogonal nonnegative matrix t-factorizations for clustering. In Proceedings of the 12th ACM SIGKDD international conference on Knowledge discovery and data mining, pp. 126 135. ACM, 2006. Du, L., Li, X., and Shen, Y.-D. Robust nonnegative matrix factorization via half-quadratic minimization. In Data Mining (ICDM), 2012 IEEE 12th International Conference on, pp. 201 210. IEEE, 2012. Farnia, F., Zhang, J. M., and Tse, D. Generalizable adversarial training via spectral normalization. ar Xiv preprint ar Xiv:1811.07457, 2018. Fortin, M. and Glowinski, R. Augmented Lagrangian methods: applications to the numerical solution of boundaryvalue problems, volume 15. Elsevier, 2000. Gabay, D. and Mercier, B. A dual algorithm for the solution of non linear variational problems via finite element approximation. Institut de recherche d informatique et d automatique, 1975. Gao, H., Nie, F., Cai, W., and Huang, H. Robust capped norm nonnegative matrix factorization: Capped norm nmf. In Proceedings of the 24th ACM International on Conference on Information and Knowledge Management, pp. 871 880. ACM, 2015. Geng, B., Tao, D., Xu, C., Yang, L., and Hua, X.-S. Ensemble manifold regularization. IEEE Transactions on Pattern Analysis and Machine Intelligence, 34(6):1227 1233, 2012. Gonzalez, E. F. and Zhang, Y. Accelerating the lee-seung algorithm for nonnegative matrix factorization. Technical report, 2005. Großhans, M., Sawade, C., Br uckner, M., and Scheffer, T. Bayesian games for adversarial regression problems. In International Conference on Machine Learning, pp. 55 63, 2013. Guan, N., Tao, D., Luo, Z., and Yuan, B. Nenmf: An optimal gradient method for nonnegative matrix factorization. IEEE Transactions on Signal Processing, 60(6): 2882 2898, 2012. Guan, N., Liu, T., Zhang, Y., Tao, D., and Davis, L. S. Truncated cauchy non-negative matrix factorization for robust subspace learning. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2017. Haeffele, B., Young, E., and Vidal, R. Structured low-rank matrix factorization: Optimality, algorithm, and applications to image processing. In International conference on machine learning, pp. 2007 2015, 2014. Hajinezhad, D., Chang, T.-H., Wang, X., Shi, Q., and Hong, M. Nonnegative matrix factorization using admm: Algorithm and convergence analysis. In Acoustics, Speech and Signal Processing (ICASSP), 2016 IEEE International Conference on, pp. 4742 4746. IEEE, 2016. Huang, J., Nie, F., Huang, H., and Ding, C. Robust manifold nonnegative matrix factorization. ACM Transactions on Knowledge Discovery from Data (TKDD), 8(3):11, 2014. Kim, J., Monteiro, R. D., and Park, H. Group sparsity in nonnegative matrix factorization. In Proceedings of the 2012 SIAM International Conference on Data Mining, pp. 851 862. SIAM, 2012. Kong, D., Ding, C., and Huang, H. Robust nonnegative matrix factorization using l21-norm. In Proceedings of the 20th ACM international conference on Information and knowledge management, pp. 673 682. ACM, 2011. Lee, D. D. and Seung, H. S. Learning the parts of objects by non-negative matrix factorization. Nature, 401(6755): 788, 1999. Li, B. and Vorobeychik, Y. Feature cross-substitution in adversarial classification. In Advances in neural information processing systems, pp. 2087 2095, 2014. Adversarial Nonnegative Matrix Factorization Lin, C.-J. Projected gradient methods for nonnegative matrix factorization. Neural computation, 19(10):2756 2779, 2007. Liu, T., Gong, M., and Tao, D. Large-cone nonnegative matrix factorization. IEEE transactions on neural networks and learning systems, 28(9):2129 2142, 2016. Maaten, L. v. d. and Hinton, G. Visualizing data using t-sne. Journal of machine learning research, 9(Nov): 2579 2605, 2008. Mei, J.-J., Dong, Y., Huang, T.-Z., and Yin, W. Cauchy noise removal by nonconvex admm with convergence guarantees. Journal of Scientific Computing, 74(2):743 766, 2018. Tong, L., Yu, S., Alfeld, S., and Vorobeychik, Y. Adversarial regression with multiple learners. ar Xiv preprint ar Xiv:1806.02256, 2018. Wang, Y., Yin, W., and Zeng, J. Global convergence of admm in nonconvex nonsmooth optimization. Journal of Scientific Computing, pp. 1 35, 2015. Zhi, R., Flierl, M., Ruan, Q., and Kleijn, W. B. Graphpreserving sparse nonnegative matrix factorization with application to facial expression recognition. IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics), 41(1):38 52, 2011.