# nearly_optimal_robust_matrix_completion__2e877f55.pdf Nearly Optimal Robust Matrix Completion Yeshwanth Cherapanamjeri 1 Kartik Gupta 1 Prateek Jain 1 Abstract In this paper, we consider the problem of Robust Matrix Completion (RMC) where the goal is to recover a low-rank matrix by observing a small number of its entries out of which a few can be arbitrarily corrupted. We propose a simple projected gradient descent-based method to estimate the low-rank matrix that alternately performs a projected gradient descent step and cleans up a few of the corrupted entries using hard-thresholding. Our algorithm solves RMC using nearly optimal number of observations while tolerating a nearly optimal number of corruptions. Our result also implies significant improvement over the existing time complexity bounds for the low-rank matrix completion problem. Finally, an application of our result to the robust PCA problem (low-rank+sparse matrix separation) leads to nearly linear time (in matrix dimensions) algorithm for the same; existing state-of-the-art methods require quadratic time. Our empirical results corroborate our theoretical results and show that even for moderate sized problems, our method for robust PCA is an order of magnitude faster than the existing methods. 1. Introduction In this paper, we study the Robust Matrix Completion (RMC) problem where the goal is to recover an underlying low-rank matrix by observing a small number of sparsely corrupted entries. Formally, RMC: Find rank-r matrix L Rm n using Ωand PΩ(L ) + S , (1) where Ω [m] [n] is the set of observed entries (throughout the paper we assume that m n), S denotes the sparse corruptions of the observed entries, i.e., 1Microsoft Research India. Correspondence to: Prateek Jain . Proceedings of the 34 th International Conference on Machine Learning, Sydney, Australia, PMLR 70, 2017. Copyright 2017 by the author(s). Supp(S ) Ωand sampling operator PΩ: Rm n Rm n is defined as: (PΩ(A))ij = ( Aij, if (i, j) Ω 0, otherwise. (2) RMC is an important problem with several applications. It is used to model recommendation systems with outliers and to perform PCA under gross outliers as well as erasures (Jalali et al., 2011). In addition, it is a strict generalization of the following two fundamental problems in machine learning: Matrix Completion (MC): Matrix completion is the task of completing a low rank matrix given a subset of entries of the matrix. It is widely used in recommender systems and also finds applications in system identification and global positioning. Note that it is a special case of the RMC problem where the matrix S = 0. State-of-theart result for MC uses nuclear norm minimization and requires |Ω| µ2nr2 log2 n under standard µ-incoherence assumption (see Section 3). But it requires O(m2n) time in general. The best sample complexity result for a nonconvex iterative method (with at most logarithmic dependence on the condition number of L ) achieve exact recovery when |Ω| µ6nr5 log2 n and needs O(|Ω|r) computational steps. Robust PCA (RPCA): RPCA aims to decompose a sparsely corrupted low rank matrix into its low rank and sparse components. It corresponds to another special case of RMC where the whole matrix is observed. State-ofthe-art results for RPCA shows exact recovery of a rank-r, µ-incoherent L (see Assumption 1, Section 3) if at most ρ = O 1 µ2r fraction of the entries in each row/column of S are corrupted (Hsu et al., 2011; Netrapalli et al., 2014). Moreover, St-Nc RPCA algorithm (Netrapalli et al., 2014) solves the problem in time O(mnr2). Therefore, an efficient solution to the RMC problem implies efficient solutions to both the MC and RPCA problems. In this work, we attempt to answer the following open question (assuming m n): Can RMC be solved exactly by using |Ω| = O(rn log n) observations out of which O( 1 µ2r) fraction of the observed Nearly Optimal Robust Matrix Completion entries in each row/column are corrupted? Note that both |Ω| (for uniformly random Ω) and ρ values mentioned in the question above denote the information theoretic limits. Hence, the goal is to solve RMC for nearly-optimal number of samples and nearly-optimal fraction of corruptions. The existing state-of-the-art results for RMC with optimal ρ = 1 µ2r fraction of corrupted entries, either require at least a constant fraction of the entries of L to be observed (Chen et al., 2011; Cand es et al., 2011) or require restrictive assumptions like the support of corruptions being uniformly random (Li, 2013). (Klopp et al., 2014) also considers RMC problem but studies the noisy setting and does not provide exact recovery bounds. Moreover, most of the existing methods for RMC use convex relaxation for both low-rank and sparse components, and in general exhibit large time complexity (O(m2n)). Under standard assumptions on L , S , Ωand for n = O(m), we answer the above question in affirmative albeit with |Ω| which is O(r) (ignoring log factors) larger than the optimal sample complexity (see Theorem 1). In particular, we propose a simple projected gradient (PGD) style method for RMC that alternately cleans up corrupted entries by hard-thresholding and performs a projected gradient descent update onto the space of low rank matrices; our method s computational complexity is also nearly optimal (O(|Ω|r + (m + n)r2 + r3)). Our algorithm is based on projected gradient descent for estimating L and alternating projection on the set of sparse matrices for estimating S . Note that projection is onto non-convex sets of lowrank matrices (for L ) and sparse matrices (for S ), hence standard convex analysis techniques cannot be used for our algorithm. Several recent results (Jain & Netrapalli, 2015; Netrapalli et al., 2014; Jain et al., 2014; Hardt & Wootters, 2014; Blumensath, 2011) show that under certain assumptions, projection onto non-convex sets indeed lead to provable algorithms with fast convergence to the global optima. However, as explained in Section 3, RMC presents a unique set of challenges as we have to perform error analysis with the errors arising due to missing entries as well as sparse corruptions, both of which interact among themselves as well. In contrast, MC and RPCA only handle noise arising from one of the two sources. To overcome this challenge, we perform error analysis by segregating the effects due to random sampling and sparse corruptions and leverage their unique structure to obtain our result (See proof of Lemma 14). Another consequence of our careful error analysis is improved results for the MC as well as the RPCA problem. Our empirical results on synthetic data demonstrates effectiveness of our method. We also apply our method to the foreground background separation problem and find that our method is an order of magnitude faster than the stateof-the-art method (St-Nc RPCA) while achieving similar accuracy. In a concurrent and independent work, (Yi et al., 2016) also studied the RMC problem and obtained similar results. They study an alternating gradient descent style algorithm while our algorithm is based on low-rank projected gradient descent. Our sample complexity, corruption tolerance as well as time complexity differ along certain critical parameters: a) Sample complexity: Our sample complexity bound is dependent only logarithmically on κ, the condition number of the matrix L (see Table 1). On the other hand, result of (Yi et al., 2016) depends quadratically on κ, which can be significantly large. Another difference is that our sample complexity bound depends logarithmically on the final error ϵ (defined as ϵ = L L 2); which for typical finite precision computation only introduces an extra constant factor. b) Corruption tolerance: Our result allows the fraction of corrupted entries to be information theoretic optimal (up to a constant) O( 1 µ2r), while the result of (Yi et al., 2016) allows only O min 1 µ2r rκ, 1 µ2κ2r fraction of corrupted entries. c) Time Complexity: As a consequence of the sample complexity bounds, running time of the method by (Yi et al., 2016) depends quintically on κ whereas our algorithm only has a polylogarithmic dependence. In summary, this paper s main contributions are: (a) RMC: We propose a nearly linear time method that solves RMC with |Ω| = O(nr2 log2 n log2 M 2/ϵ) random entries and with optimal fraction of corruptions upto constant factors (ρ = O( 1 µ2r)). (b) Matrix Completion: Our result improves upon the existing linear time algorithm s sample complexity by an O(r3) factor, and time complexity by O(r4) factor, although with an extra O(log L /ϵ) factor in both time and sample complexity. (c) RPCA: We present a nearly linear time (O(nr3)) algorithm for RPCA under optimal fraction of corruptions, improving upon O(mnr2) time complexity of the existing methods. Notations: We assume that M = L + e S and PΩ(M) = PΩ(L ) + S , i.e., S = PΩ(e S ). v p denotes ℓp norm of a vector v; v denotes ℓ2 norm of v. A 2, A F , A denotes the operator, Frobenius, and nuclear norm of A, respectively; by default A = A 2. Operator PΩis given by (2), operators Pk(A) and HT ζ(A) are defined in Section 2. σi(A) denotes i-th singular value of A and σ i denotes the i-th singular value of L . Nearly Optimal Robust Matrix Completion Paper Organization: We present our main algorithm in Section 2 and our main results in Section 3. We also present an overview of the proof in Section 3. Section 4 presents our empirical result. Due to lack of space, we present most of the proofs and useful lemmas in the appendix. 2. Algorithm In this section we present our algorithm for solving the RMC (Robust Matrix Completion) problem: given Ωand PΩ(M) where M = L + e S Rm n, rank(L ) r, e S 0 s and S = PΩ(e S ), the goal is to recover L . To this end, we focus on solving the following non-convex optimization problem: (L , S ) = arg min L,S PΩ(M) PΩ(L) S 2 F s.t., rank(L) r, PΩ(S) = S, S 0 s. (3) For the above problem, we propose a simple iterative algorithm that combines projected gradient descent (for L) with alternating projections (for S). In particular, we maintain iterates L(t) and S(t), where L(t) is the current low-rank approximation of L and S(t) is the current sparse approximation of S . L(t+1) is computed using gradient descent step for objective (3) and then projecting back onto the set of rank k matrices. That is, L(t+1) = Pk p PΩ(M L(t) S(t)) , (4) where Pk(A) denotes projection of A onto the set of rankk matrices and can be computed efficiently using SVD of A, p = |Ω| mn. S(t+1) is computed by projecting the residual PΩ(M L(t+1)) onto set of sparse matrices using a hardthresholding operator, i.e., S(t+1) = HT ζ(M L(t+1)), (5) where HT ζ : Rm n Rm n is the hard thresholding operator defined as: (HT ζ(A))ij = Aij if |Aij| ζ and 0 otherwise. Intuitively, a better estimate of the sparse corruptions for each iteration will reduce the noise of the projected gradient descent step and a better estimate of the low rank matrix will enable better estimation of the sparse corruptions. Hence, under correct set of assumptions, the algorithm should recover L , S exactly. Unfortunately, just the above two simple iterations cannot handle problems where L has poor condition number, as the intermediate errors can be significantly larger than the smallest singular values of L , making recovery of the corresponding singular vectors challenging. To alleviate this issue, we propose an algorithm that proceeds in stages. In the q-th stage, we project L(t) onto set of rank-kq matrices. Rank kq is monotonic w.r.t. q. Under standard assumptions, we show that we can increase kq in a manner such that after each stage L(t) L decreases by at least a constant factor. Hence, the number of stages is only logarithmic in the condition number of L . See Algorithm 1 (PG-RMC ) for a pseudo-code of the algorithm. We require an upper bound of the first singular value for our algorithm to work. Specifically, we require σ = O (σ 1). Alternatively, we can also obtain an estimate of σ 1 by using the thresholding technique from (Yi et al., 2016) although this requires an estimate of the number of corruptions in each row and column. We also use a simplified version of Algorithm 5 from (Hardt & Wootters, 2014) to form independent sets of samples for each iteration which is required for our theoretical analysis. Our algorithm has an outer loop (see Line 6) which sets rank kq of iterates L(t) appropriately (see Line 7). We then update L(t) and S(t) in the inner loop using (4), (5). We set threshold for the hard-thresholding operator using singular values of current gradient descent update (see Line 12). Note that, we divide Ωuniformly into Q T sets, where Q is an upper bound on the number of outer iterations and T is the number of inner iterations. This division ensures independence across iterates that is critical to application of standard concentration bounds; such division is a standard technique in the matrix completion related literature (Jain & Netrapalli, 2015; Hardt & Wootters, 2014; Recht, 2011). Also, η is a tunable parameter which should be less than one and is smaller for easier problems. Note that updating S(t) requires O(|Ω| r + (m + n) r) computational steps. Computation of L(t+1) requires computing SVD for projection Pr, which can be computed in O(|Ω| r + (m + n) r2 + r3) time (ignoring log factors); see (Jain et al., 2010) for more details. Hence, the computational complexity of each step of the algorithm is linear in |Ω| r (assuming |Ω| r (m + n)). As we show in the next section, the algorithm exhibits geometric convergence rate under standard assumptions and hence the overall complexity is still nearly linear in |Ω| (assuming r is just a constant). Rank based Stagewise algorithm: We also provide a rank-based stagewise algorithm (R-RMC) where the outer loop increments kq by one at each stage, i.e., the rank is q in the q-th stage. Our analysis extends for this algorithm as well, however, its time and sample complexity trades off a factor of O(log(σ1/ϵ)) from the complexity of PG-RMC with a factor of r (rank of L ). We provide the detailed algorithm in Appendix 5.3 due to lack of space (see Algorithm 3). Nearly Optimal Robust Matrix Completion Algorithm 1 b L = PG-RMC (Ω, PΩ(M), ϵ, r, µ, η, σ) 1: Input: Observed entries Ω, Matrix PΩ(M) Rm n, convergence criterion ϵ, target rank r, incoherence parameter µ, thresholding parameter η, estimate of first singular value σ 2: T 10 log 20µ2nrσ ϵ , Q T 3: Partition Ωinto Q T + 1 subsets {Ω0} {Ωq,t : q [Q], t [T]} with p set to Ω QT +1 using algorithm 2 4: L(0) = 0, ζ(0) ησ 5: M (0) = mn |Ω0|PΩ0(M HT ζ(M)) 6: k0 0, q 0 7: while σkq+1(M (0)) > ϵ 2ηn do 8: q q + 1, 9: kq {i : σi(M (0)) σkq 1+1(M (0)) 10: for Iteration t = 0 to t = T do 11: S(t) = HT ζ(PΩq,t(M L(t))) 12: M (t) = L(t) mn |Ωq,t|PΩq,t(L(t) + S(t) M) 13: L(t+1) = Pkq(M (t)) 14: ζ(t+1) η σkq+1(M (t)) + 1 2 t 2 σkq(M (t)) 15: end for 16: S(0) = S(T ), L(0) = L(T +1), M (0) = M (T ), ζ(0) = ζ(T +1) 17: end while 18: Return: L(T ) Algorithm 2 {Ω1, . . . , ΩT } = Split Samples(Ω, p, T) 1: Input: Random samples with probability Tp Ω, Required Sampling Probability p, Number of Sets T 2: Output: T independent sets of entries {Ω1, . . . , ΩT } sampled with sampling probability p 3: p 1 (1 p)T 4: Ω be sampled from Ωwith each entry being included independently with probability p /p 5: for r = 1 to r = T do 6: qr ( T r)pr(1 p)T r p 7: end for 8: Initialize Ωt {} for t {1, . . . , T} 9: for Sample s Ω do 10: Draw r {1, . . . , T} with probability qr 11: Draw a random subset S of size r from {1, . . . , T} 12: Add s to Ωi for i S 13: end for 3. Analysis We now present our analysis for both of our algorithms PGRMC (Algorithm 1) and R-RMC (Algorithm 3). In general the problem of Robust PCA with Missing Entries (3) is harder than the standard Matrix Completion problem and hence is NP-hard (Hardt et al., 2014). Hence, we need to impose certain (by now standard) assumptions on L , e S , and Ωto ensure tractability of the problem: Assumption 1. Rank and incoherence of L : L Rm n is a rank-r incoherent matrix, i.e., e i U 2 µp r m, e j V 2 µp r n, i [m], j [n], where L = U Σ (V ) is the SVD of L . Assumption 2. Sampling (Ω): Ωis obtained by sampling each entry with probability p = |Ω| mn. Assumption 3. Sparsity of e S , S : We assume that at most ρ c µ2r fraction of the elements in each row and column of e S are non-zero for a small enough constant c. Moreover, we assume that Ωis independent of e S . Hence, S = PΩ(e S ) also has p ρ fraction of the entries in expectation. Assumptions 1, 2 are standard assumptions in the provable matrix completion literature (Cand es & Recht, 2009; Recht, 2011; Jain & Netrapalli, 2015), while Assumptions 1, 3 are standard assumptions in the robust PCA (lowrank+sparse matrix recovery) literature (Chandrasekaran et al., 2011; Cand es et al., 2011; Hsu et al., 2011). Hence, our setting is a generalization of both the standard and popular problems and as we show later in the section, our result can be used to meaningfully improve the state-of-the-art for both of these problems. We first present our main result for Algorithm 1 under the assumptions given above. Theorem 1. Let Assumptions 1, 2 and 3 on L , e S and Ωhold respectively. Let m n, n = O(m), and let the number of samples |Ω| satisfy: E[|Ω|] Cα2µ4r2n log2 (n) log2 µ2rσ1 where C is a global constant. Then, with probability at least 1 n log α 2 , Algorithm 1 with η = 4µ2r m , at most O(log( M 2/ϵ))) outer iterations and O(log( µ2r M 2 ϵ )) inner iterations, outputs a matrix ˆL such that: ˆL L F ϵ. Note that the number of samples matches information theoretic bound upto O(r log n log2 σ 1/ϵ) factor. Also, the number of allowed corruptions in e S also matches the known lower bounds (up to a constant factor) and cannot be improved upon information theoretically. We now present our result for the rank based stagewise algorithm (Algorithm 3). Theorem 2. Under Assumptions 1, 2 and 3 on L , e S and Ωrespectively and Ωsatisfying: E[|Ω|] Cα2µ4r3n log2 (n) log µ2rσ1 Nearly Optimal Robust Matrix Completion for a large enough constant C, then Algorithm 3 with η set to 4µ2r m outputs a matrix ˆL such that: ˆL L F ϵ, w.p. Notice that the sample complexity of Algorithm 3 has an additional multiplicative factor of O(r) when compared to that of Algorithm 1, but shaves off a factor of O(log(κ)). Similarly, computational complexity of Algorithm 3 also trades off a O(log κ) factor for O(r) factor from the computational complexity of Algorithm 1. Result for Matrix Completion: Note that for e S = 0, the RMC problem with Assumptions 1,2 is exactly the same as the standard matrix completion problem and hence, we get the following result as a corollary of Theorem 1: Corollary 1 (Matrix Completion). Suppose we observe Ω and PΩ(L ) where Assumptions 1,2 hold for L and Ω. Also, let E[|Ω|] Cα2µ4r2n log2 n log2 σ1/ϵ and m n. Then, w.p. 1 n log α 2 , Algorithm 1 outputs ˆL s.t. ˆL L 2 ϵ. Table 1 compares our sample and time complexity bounds for low-rank MC. Note that our sample complexity is nearly the same as that of nuclear-norm methods while the running time of our algorithm is significantly better than the existing results that have at most logarithmic dependence on the condition number of L . Result for Robust PCA: Consider the standard Robust PCA problem (RPCA), where the goal is to recover L from M = L + e S . For RPCA as well, we can randomly sample |Ω| entries from M, where Ωsatisfies the assumption required by Theorem 1. This leads us to the following corollary: Corollary 2 (Robust PCA). Suppose we observe M = L + e S , where Assumptions 1, 3 hold for L and e S . Generate Ω [m] [n] by sampling each entry independently with probability p, s.t., E[|Ω|] Cα2µ4r2n log2 n log2 σ1/ϵ. Let m n. Then, w.p. 1 n log α 2 , Algorithm 1 outputs ˆL s.t. ˆL L 2 ϵ. Hence, using Theorem 1, we will still be able to recover L but using only the sampled entries. Moreover, the running time of the algorithm is only O(µ4nr3 log2 n log2(σ1/ϵ)), i.e., we are able to solve RPCA problem in time almost linear in n. To the best of our knowledge, the existing stateof-the-art methods for RPCA require at least O(n2r) time to perform the same task (Netrapalli et al., 2014; Gu et al., 2016). Similarly, we don t need to load the entire data matrix in memory, but we can just sample the matrix and work with the obtained sparse matrix with at most linear number of entries. Hence, our method significantly reduces both run-time and space complexities, and as demonstrated empirically in Section 4 can help scale our algorithm to very large data sets without losing accuracy. 3.1. Proof Outline for Theorem 1 We now provide an outline of our proof for Theorem 1 and motivate some of our proof techniques; the proof of Theorem 2 follows similarly. Recall that we assume that M = L + e S and define S = PΩ(e S ). Similarly, we define e S(t) = HT ζ(M L(t)). Critically, S(t) = PΩ(e S(t)) (see Line 9 of Algorithm 1), i.e., e S(t) is the set of iterates that we could obtain if entire M was observed. Note that we cannot compute e S(t), it is introduced only to simplify our analysis. We first re-write the projected gradient descent step for L(t+1) as described in (4): L(t+1) = Pkq L + (e S e S(t)) | {z } E1 E2 z }| { (L(t) L ) +(e S(t) e S )) | {z } E3 That is, L(t+1) is obtained by rank-kq SVD of a perturbed version of L : L + E1 + E3. As we perform entrywise thresholding to reduce e S e S(t) , we need to bound L(t+1) L . To this end, we use techniques from (Jain & Netrapalli, 2015), (Netrapalli et al., 2014) that explicitly model singular vectors of L(t+1) and argue about the infinity norm error using a Taylor series expansion. However, in our case, such an error analysis requires analyzing the following key quantities (H = E1 + E3): 1 j, s.t., j even : Aj := max q [n] e q H H j Bj := max q [m] e q HH j 1 j, s.t., j odd : Cj := max q [n] e q H HH j Dj := max q [m] e q H H H j Note that E1 = 0 in the case of standard RPCA which was analyzed in (Netrapalli et al., 2014), while E3 = 0 in the case of standard MC which was considered in (Jain & Netrapalli, 2015). In contrast, in our case both E1 and E3 are non-zero. Moreover, E3 is dependent on random variable Ω. Hence, for j 2, we will get cross terms between E3 and E1 that will also have dependent random variables which precludes application of standard Bernstein-style tail bounds. To overcome this issue, we first carefully segregate the errors arising due to the randomness in the sampling process and the deterministic sparse corruptions in e S . We do this by introducing e S(t) which is the sparse iterate we Nearly Optimal Robust Matrix Completion Table 1: Comparison of PG-RMC and R-RMC with Other Matrix Completion Methods Sample Complexity Computational Complexity Nuclear norm (Recht, 2011) O µ2rn log2 n SVP (Jain & Netrapalli, 2015) O µ4r5n log3 n O µ4r7n log3 n log( 1 Alt. Min. (Hardt & Wootters, 2014) O nµ4r9 log3 (κ) log2 n O nµ4r13 log3 (κ) log2 n Alt. Grad. Desc. (Sun & Luo, 2015) O nrκ2 max{µ2 log n, µ4r6κ4} O n2r6κ4 log 1 R-RMC (This Paper) O µ4r3n log2 (n) log σ 1 ϵ O µ4r4n log2 (n) log σ 1 ϵ PG-RMC (This Paper) O µ4r2n log2 (n) log2 σ 1 ϵ O µ4r3n log2 (n) log2 σ 1 ϵ would have obtained had the whole matrix been observed. This allows us to decompose the error term into the sum of E1 and E3 where E1 represents the error due to the sparse corruptions and E3 represents the error arising from the randomness in the sampling procedure. We then incorporate this decomposition into a careful combinatorial-style argument similar to that of (Erdos et al., 2013; Jain & Netrapalli, 2015) to bound the above given quantity. That is, we can provide the following key lemma: Lemma 1. Let L , Ω, and e S satisfy Assumptions 1, 2 and 3 respectively. Let L = U Σ (V ) be the singular value decomposition of L . Furthermore, suppose that in the tth iteration of the qth stage, e S(t) defined as HTζ(M L(t)) satisfies Supp(e S(t)) Supp(e S ), then we have: max{Aa, Ba, Ca, Da} µ r r p ( E1 E2 ) log n a , c > 0, 0 j log n w.p. 1 n 2 log c 4 +4, where E1, E2 and E3 are defined in (6), Aa, Ba, Ca, Da are defined in (7). Remark: We would like to note that even for the standard MC setting, i.e., when E1 = 0, we obtain better bound than that of (Jain & Netrapalli, 2015) as we can bound maxi e T i (E3)q U 2 directly rather than the weaker r maxi e T i (E3)quj bound that (Jain & Netrapalli, 2015) uses. Now, using Lemmas 1 and 7 and by using a hardthresholding argument we can bound L(t+1) L 2µ2r m (σ kq+1 + 1 2 t σ kq) (see Lemma 9) in the q-th stage. Hence, after O(log(σ 1/ϵ)) inner iterations, we can guar- antee in the q-th stage: L(T ) L 4µ2r E1 + E2 20µ2r Moreover, by using sparsity of e S and the special structure of E3 (See Lemma 7), we have: E1 + E3 2 c σ kq+1, where c is a small constant. Now, the outer iteration sets the next stage s rank kq+1 as: kq+1 = |{i : σi(L + E1 + E3) 0.5 σkq+1(L + E1 + E3)}|. Using the bound on E1 + E3 2 and Weyl s eigenvalue perturbation bound (Lemma 2), we have: σ kq+1+1 0.6 σ kq+1. Hence, after Q = O(log(σ 1/ϵ)) outer iterations, Algorithm 1 converges to an ϵ-approximate solution to L . 4. Experiments In this section we discuss the performance of Algorithm 1 on synthetic data and its use in foreground background separation. The goal of the section is two-fold: a) to demonstrate practicality and effectiveness of Algorithm 1 for the RMC problem, b) to show that Algorithm 1 indeed solves RPCA problem in significantly smaller time than that required by the existing state-of-the-art algorithm (St-Nc RPCA (Netrapalli et al., 2014)). To this end, we use synthetic data as well as video datasets where the goal is to perform foreground-background separation (Cand es et al., 2011). We implemented our algorithm in MATLAB and the results for the synthetic data set were obtained by averaging over 20 runs. We obtained a matlab implementation of St Nc RPCA (Netrapalli et al., 2014) from the authors of (Netrapalli et al., 2014). Note that if the sampling probability is p = 1, then our method is similar to St-Nc RPCA; the key difference being how the rank is selected in each stage. Nearly Optimal Robust Matrix Completion We also implemented the Alternating Minimzation based algorithm from (Gu et al., 2016). However, we found it to be an order of magnitude slower than Algorithm 1 on the foreground-background separation task. For example, on the escalator video, the algorithm did not converge in less than 150 seconds despite discounting for the expensive sorting operation in the truncation step. On the other hand, our algorithm finds the foreground in about 8 seconds. Parameters. The algorithm has three main parameters: 1) threshold ζ, 2) incoherence µ and 3) sampling probability p (E[|Ω|] = p mn). In the experiments on synthetic data we observed that keeping ζ µ M S(t) 2 / n speeds up the recovery while for background extraction keeping ζ µ M S(t) 2 /n gives a better quality output. The value of µ for real world data sets was figured out using cross validation while for the synthetic data the same value was used as used in data generation. The sampling probability for the synthetic data could be kept as low as (θ =)2r log2(n)/n while for the real world data set we get good results for p = 0.05. We define effective sample size as the ratio between the sampling probability and θ. Also, rather than splitting samples, we use the entire set of observed entries to perform our updates (see Algorithm 1). Synthetic data. We generate M = L + e S of two sizes, where L = UV R2000 2000 (and R5000 5000) is a random rank-5 (and rank-10 respectively) matrix with incoherence 1. e S is generated by considering a uniformly random subset of size e S 0 from [m] [n] where every entry is i.i.d. from the uniform distribution in [ r 2 mn, r mn]. This is the same setup as used in (Cand es et al., 2011). Figure 1 (a) plots recovery error ( L L F ) vs computational time for our PG-RMC method (with different sampling probabilities) as well as the St-Nc RPCA algorithm. Note that even for very small values of sampling p, we can achieve the same recovery error as when using significantly small values. For example, our method with p = 0.1 achieve 0.01 error ( L L F ) in 2.5s while St-Nc RPCA method requires 10s to achieve the same accuracy. Note that we do not compare against the convex relaxation based methods like IALM from (Cand es et al., 2011), as (Netrapalli et al., 2014) shows that St-Nc RPCA is significantly faster than IALM and several other convex relaxation solvers. Figure 1 (b) plots time required to achieve different recovery errors ( L L F ) as the sampling probability p increases. As expected, we observe a linear increase in the run-time with p. Interestingly, for very small values of p, we observe an increase in running time. In this regime, Time(s) 5 10 15 log ||L ˆL||F 2 n = 5000, µ = 1, r = 10, ρ = 0.01 p = 0.05 p = 0.1 St-Nc RPCA Sampling probability 0.1 0.15 0.2 0.25 0.3 12 n = 2000, µ = 1, r = 5, ρ = 0.01 err = 1e-5 err = 1e-3 err = 1e-1 Rank 5 10 15 20 40 n = 2000, r = 5, ρ = 0.01, p = 0.5 err = 1e-1 err = 1e-3 err = 1e-5 µ 0.5 1 1.5 2 10 n = 2000, r = 5, ρ = 0.01, p = 0.1 err = 1e-5 err = 1e-3 err = 1e-1 effective sample size 0 0.5 1 success probability r = 5, µ = 1, ρ = 0.01 n = 1000 n = 2000 n = 3000 Figure 1: Performance of PG-RMC on synthetic data. 1a: time vs error for various sampling probabilities; time taken by St-Nc RPCA 1b: sampling probability vs time for constant error; time taken decreases with decreasing sampling probability upto an extent and then increases 1c: time vs rank for constant error 1d: incoherence vs time for constant error 1e: success probability vs effective sample size for various matrix sizes p becomes very large (as p doesn t satisfy the sampling requirements). Hence, the increase in the number of iterations (T log PΩ(M) 2 pϵ ) dominates the decrease in per iteration time complexity. Figure 1 (c), (d) plot computation time required by our method (PG-RMC , Algorithm 1) versus rank and incoherence, respectively. As expected, as these two problem parameters increase, our method requires more time. Note that our run-time dependence on rank seems to be linear, while our results require O(r3) time. This hints at the possibility of further improving the computational complexity analysis of our algorithm. Figure 1 (e) plots the effective sample size against success probability. We see that the probability of recovering the underlying low rank matrix undergoes a rapid phase Nearly Optimal Robust Matrix Completion Time(s) 0 100 200 300 log ||M ˆL ˆ S||F µ = 1, r = 5 p = 0.05 p = 0.1 St-Nc RPCA Time(s) 0 10 20 30 log ||M ˆL ˆ S||F µ = 1, r = 5 p = 0.01 p = 0.05 p = 0.1 St-Nc RPCA Figure 2: PG-RMC on Shopping video. 2a: a video frame 2c: an extracted background frame 2e: time vs error for different sampling probabilities; PG-RMC takes 38.7s while St-Nc PCA takes 204.4s. PG-RMC on Restaurant video. 2b: a video frame 2d: an extracted background frame 2f: time vs error for different sampling probabilities; PGRMC takes 7.9s while St-Nc PCA takes 27.8s transition from 0 to 1 when sampling probability crosses r log2 n/n. We also study phase transition for different values of sampling probability p. Figure 3 (a) in Appendix 5.5 show a phase transition phenomenon where beyond p > .06 the probability of recovery is almost 1 while below it, it is almost 0. Foreground-background separation. We also applied our technique to the problem of foreground-background separation. We use the usual method of stacking up the vectorized video frames to construct a matrix. The background, being static, will form the low rank component while the foreground is considered to be the noise. We applied our PG-RMC method (with varying p) to sev- eral videos. Figure 2 (a), (d) show one frame each from two such videos (a shopping center video, a restaurant video). Figure 2 (b), (d) show the extracted background from the two videos by using our method (PG-RMC , Algorithm 1) with probability of sampling p = 0.05. Figure 2 (c), (f) compare objective function value for different p values. Clearly, PG-RMC can recover the true background with p as small as 0.05. We also observe an order of magnitude speedup ( 5x) over St-Nc RPCA (Netrapalli et al., 2014). We present results on the video Escalator in Appendix 5.5. Conclusion. In this work, we studied the Robust Matrix Completion problem. For this problem, we provide exact recovery of the low-rank matrix L using nearly optimal number of observations as well as nearly optimal fraction of corruptions in the observed entries. Our RMC result is based on a simple and efficient PGD algorithm that has nearly linear time complexity as well. Our result improves state-of-the-art sample and run-time complexities for the related Matrix Completion as well as Robust PCA problem. For Robust PCA, we provide first nearly linear time algorithm under standard assumptions. Our sample complexity depends on ϵ, the desired accuracy in L . Removing this factor will be an interesting future work. Moreover, improving dependence of sample complexity on r (from r2 to r) also represents an important direction. Finally, similar to foreground background separation, we would like to explore more applications of RMC/RPCA. Bhatia, Rajendra. Matrix Analysis. Springer, 1997. Blumensath, Thomas. Sampling and reconstructing signals from a union of linear subspaces. IEEE Trans. Information Theory, 57(7):4660 4671, 2011. doi: 10.1109/ TIT.2011.2146550. URL http://dx.doi.org/ 10.1109/TIT.2011.2146550. Cand es, Emmanuel J. and Recht, Benjamin. Exact matrix completion via convex optimization. Foundations of Computational Mathematics, 9(6):717 772, December 2009. Cand es, Emmanuel J., Li, Xiaodong, Ma, Yi, and Wright, John. Robust principal component analysis? J. ACM, 58 (3):11, 2011. Chandrasekaran, Venkat, Sanghavi, Sujay, Parrilo, Pablo A., and Willsky, Alan S. Rank-sparsity incoherence for matrix decomposition. SIAM Journal on Optimization, 21(2):572 596, 2011. Nearly Optimal Robust Matrix Completion Chen, Yudong, Jalali, Ali, Sanghavi, Sujay, and Caramanis, Constantine. Low-rank matrix recovery from errors and erasures. In 2011 IEEE International Symposium on Information Theory Proceedings, ISIT 2011, St. Petersburg, Russia, July 31 - August 5, 2011, pp. 2313 2317, 2011. doi: 10.1109/ISIT.2011.6033975. URL http: //dx.doi.org/10.1109/ISIT.2011.6033975. Erdos, L aszl o, Knowles, Antti, Yau, Horng-Tzer, and Yin, Jun. Spectral statistics of Erdos R enyi graphs I: Local semicircle law. The Annals of Probability, 41(3B):2279 2375, 2013. Gu, Quanquan, Wang, Zhaoran Wang, and Liu, Han. Lowrank and sparse structure pursuit via alternating minimization. In Proceedings of the Nineteenth International Conference on Artificial Intelligence and Statistics, AISTATS 2016, C adiz, Spain, May 9-11, 2016, 2016. URL http://jmlr.org/proceedings/ papers/v51/gu16.html. Hardt, Moritz and Wootters, Mary. Fast matrix completion without the condition number. In Proceedings of The 27th Conference on Learning Theory, COLT 2014, Barcelona, Spain, June 13-15, 2014, pp. 638 678, 2014. URL http://jmlr.org/proceedings/ papers/v35/hardt14a.html. Hardt, Moritz, Meka, Raghu, Raghavendra, Prasad, and Weitz, Benjamin. Computational limits for matrix completion. In Proceedings of The 27th Conference on Learning Theory, COLT 2014, Barcelona, Spain, June 13-15, 2014, pp. 703 725, 2014. URL http://jmlr.org/proceedings/papers/ v35/hardt14b.html. Hsu, Daniel, Kakade, Sham M, and Zhang, Tong. Robust matrix decomposition with sparse corruptions. Information Theory, IEEE Transactions on, 57(11):7221 7234, 2011. Jain, Prateek and Netrapalli, Praneeth. Fast exact matrix completion with finite samples. In Proceedings of The 28th Conference on Learning Theory, COLT 2015, Paris, France, July 3-6, 2015, pp. 1007 1034, 2015. URL http://jmlr.org/proceedings/ papers/v40/Jain15.html. Jain, Prateek, Meka, Raghu, and Dhillon, Inderjit S. Guaranteed rank minimization via singular value projection. In NIPS, pp. 937 945, 2010. Jain, Prateek, Tewari, Ambuj, and Kar, Purushottam. On iterative hard thresholding methods for high-dimensional m-estimation. In Advances in Neural Information Processing Systems 27: Annual Conference on Neural Information Processing Systems 2014, December 8-13 2014, Montreal, Quebec, Canada, pp. 685 693, 2014. URL http://papers.nips.cc/paper/5293-oniterative-hard-thresholding-methodsfor-high-dimensional-m-estimation. Jalali, Ali, Ravikumar, Pradeep, Vasuki, Vishvas, and Sanghavi, Sujay. On learning discrete graphical models using group-sparse regularization. In Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics, AISTATS 2011, Fort Lauderdale, USA, April 11-13, 2011, pp. 378 387, 2011. URL http://www.jmlr.org/proceedings/ papers/v15/jalali11a/jalali11a.pdf. Klopp, Olga, Lounici, Karim, and Tsybakov, Alexandre B. Robust matrix completion. ar Xiv preprint ar Xiv:1412.8132, 2014. Li, Xiaodong. Compressed sensing and matrix completion with constant proportion of corruptions. Constructive Approximation, 37(1):73 99, 2013. ISSN 14320940. doi: 10.1007/s00365-012-9176-9. URL http: //dx.doi.org/10.1007/s00365-012-9176-9. Netrapalli, Praneeth, U N, Niranjan, Sanghavi, Sujay, Anandkumar, Animashree, and Jain, Prateek. Nonconvex robust pca. In Ghahramani, Z., Welling, M., Cortes, C., Lawrence, N. D., and Weinberger, K. Q. (eds.), Advances in Neural Information Processing Systems 27, pp. 1107 1115. Curran Associates, Inc., 2014. URL http://papers.nips.cc/paper/ 5430-non-convex-robust-pca.pdf. Recht, Benjamin. A simpler approach to matrix completion. Journal of Machine Learning Research, 12:3413 3430, 2011. URL http://dl.acm.org/ citation.cfm?id=2185803. Sun, Ruoyu and Luo, Zhi-Quan. Guaranteed matrix completion via nonconvex factorization. In IEEE 56th Annual Symposium on Foundations of Computer Science, FOCS 2015, Berkeley, CA, USA, 17-20 October, 2015, pp. 270 289, 2015. doi: 10.1109/FOCS.2015.25. URL http://dx.doi.org/10.1109/FOCS.2015.25. Yi, Xinyang, Park, Dohyung, Chen, Yudong, and Caramanis, Constantine. Fast algorithms for robust PCA via gradient descent. Co RR, abs/1605.07784, 2016. URL http://arxiv.org/abs/1605.07784.