# partial_matrix_completion__503acdfe.pdf Partial Matrix Completion Elad Hazan Princeton University Google Deep Mind Adam Tauman Kalai Microsoft Research Varun Kanade University of Oxford Clara Mohri Harvard University Y. Jennifer Sun Princeton University Google Deep Mind The matrix completion problem aims to reconstruct a low-rank matrix based on a revealed set of possibly noisy entries. Prior works consider completing the entire matrix with generalization error guarantees. However, the completion accuracy can be drastically different over different entries. This work establishes a new framework of partial matrix completion, where the goal is to identify a large subset of the entries that can be completed with high confidence. We propose an efficient algorithm with the following provable guarantees. Given access to samples from an unknown and arbitrary distribution, it guarantees: (a) high accuracy over completed entries, and (b) high coverage of the underlying distribution. We also consider an online learning variant of this problem, where we propose a low-regret algorithm based on iterative gradient updates. Preliminary empirical evaluations are included. 1 Introduction In the classical matrix completion problem, a subset of entries of a matrix are revealed and the goal is to reconstruct the full matrix. This is in general impossible, but if the matrix is assumed to be low rank and the distribution over revealed entries is uniformly random, then it can be shown that reconstruction is possible. A common application of matrix completion is in recommendation systems. For example, the rows of the matrix can correspond to users, the columns to movies, and an entry of the matrix is the preference score for the user over the corresponding movie. The completed entries can then be used to predict user preferences over unseen movies. The low rank assumption in this case is justified if the preference of users over movies is mostly determined by a small number of latent factors such as the genre, director, artistic style, and so forth. However, in many cases, it is both infeasible and unnecessary to complete the entire matrix. First consider the trivial example of movie recommendations with users and movies coming from two countries A and B, where each user rates random movies only from their country. Without any cross ratings of users from country A on movies from country B or vice versa, it is impossible to accurately complete these entries. A solution here is straightforward: partition the users and movies into their respective groups and then complete only the part of the matrix corresponding to user ratings of movies from their own country. In reality, many users have categories of movies with few or no ratings based on genre, language, or time period. For such users, it is difficult to accurately complete unrated movie categories which they do not like to watch from those which they have not even been exposed to. Thus it may be preferable to abstain from making rating predictions for such users on these unpredictable categories. This is Authors ordered alphabetically 37th Conference on Neural Information Processing Systems (Neur IPS 2023). further complicated by the fact that such categories are not crisply defined and relevant categories may vary across user demographics like country and age. Figure 1: A simple example where some users rate only comedies and some rate only horror movies. However, these two groups are unknown and the rows and columns are permuted according to = C R. The partial completions made by our algorithm, shown on simulated data, correspond to completing the ratings for the half of the matrix as one would like. Note that our algorithms can handle arbitrary revealed subsets, not only stochastic block models. The experimental setup and more preliminary experiments are described in Section 6. This motivates a more general question: can we identify a subset of the matrix that can be completed by the existing observations with high confidence? We call this problem partial matrix completion. Let M ? be a target matrix. A non-convex formulation would be to find a matrix ˆ M and a set of entries C [m] [n] which maximizes: max C [m] [n] |C| s.t. 1 |C| ij ˆ Mij)2 ". (1) This is non-convex because the set C is discrete and also because the choice of ˆ M interacts C. Fortunately, there are already several matrix completion algorithms which output ˆ M that guarantee low generalization error with respect to future examples from the same distribution of entries as those from the training distribution, though they are not accurate over the entire matrix. We will fix any one of these completions ˆ M. We thus relax the goal of partial matrix completion to identify a matrix C which gives a confidence score for each entry between 0 and 1. A value of 0 indicates that this entry should not be completed, whereas 1 indicates we have absolute confidence. We call this matrix the confidence matrix. Ideally, we would like that the confidence matrix to have a value of 1 on all entries that are well supported by the observation distribution. We formalize this problem through the following definitions. The precise definitions of some of the terms will be introduced in subsequent sections. Let M ? be a target matrix and ˆ M be a completion given by a matrix completion algorithm that guarantees low generalization error over future examples from the same entry distribution (formal definition given in Def. 8), and let " > 0. The coverage of C is defined to be k Ck1 Definition 1 (Partial matrix completion problem). Let M ? be a target matrix and ˆ M be a full completion (Def. 8), and " > 0 be an error tolerance. Find a confidence matrix C 2 [0, 1]m n with maximal coverage, i.e. arg max C2[0,1]m n k Ck1 s.t. 1 k Ck1 i2[m],j2[n] ij ˆ Mij)2 ". The matrix C that solves the above optimization problem has the following property: Cij indicates whether the (i, j)-th entry shall be completed, and a fractional value can be interpreted as either a fractional completion or a probability of completion. C provides the largest coverage among all that satisfy the property that the average error is at most " over the completed entries. Challenges of partial matrix completion. The above formulation is intuitive and convex, but notice that we do not know M ?, and thus cannot apply convex optimization directly. Key challenges are illustrated by two intuitive, but naïve approaches. First, consider completing just those matrix entries (i, j) where all rank-k completions have nearly the same value ˆ Mij. Even in a simple case where all revealed entries are 1, any single entry could be 1 for some rank 2 completion of the matrix. Based on this, one often will not be able to complete any entries whatsoever. Second, consider the generalization of our previous example where the sampling distribution is uniform over an arbitrary subset U of m n possible indices. For simplicity, think of |U| = mn/2 as covering half the matrix. It follows from previous generalization bounds on matrix completion that one could accurately complete all entries in U. However, in general U is unknown and arbitrarily complex. The second naïve approach would be to try to learn U, but this requires |U| = mn/2 m + n observations. 1.1 Our contribution We present an inefficient algorithm and its efficient relaxation to solve the partial matrix completion problem with provable guarantees over arbitrary sampling distributions. In particular, if µ is uniform over some arbitrary subset U [m] [n] of a rank-k matrix M ? 2 [ 1, 1]m n, the theoretical results obtained in this work have the following implications. The first, is an inefficient algorithm that solves the partial matrix completion problem for uniform distribution: Corollary 2 (Inefficient algorithm, uniform sampling distribution). For any ", δ > 0, with probability 1 δ over N c0 k(m+n)+log(1/δ) "2 revealed entries from M ? drawn from µ, for some constant c0, Algorithm 1 (Sec. 3) outputs ˆ M and C 2 [0, 1]m n, such that (1) : k Ck1 |U|, (2) : 1 k Ck1 i2[m],j2[n] ij ˆ Mij)2 ". Next, we describe the guarantees of an efficient algorithm 2. It has a worse sample complexity, up to polynomial factors in k, ". The advantage is that it runs in polynomial time. Otherwise, its guarantees are similar to that of the first algorithm. Corollary 3 (Efficient algorithm, uniform sampling distribution). For any ", δ > 0, with probability 1 δ over N c0 k2 "4 (k(m + n) + log(1/δ)) revealed entries from M ? drawn from µ, for some constant c0, Algorithm 2 (Sec. 4) outputs ˆ M and C 2 [0, 1]n m, such that (1) : k Ck1 |U|, (2) : 1 k Ck1 i2[m],j2[n] ij ˆ Mij)2 ". Corollary 2 is a simplification of Theorem 10, and Corollary 3 is a simplification of Theorem 11. The results can be extended to the setting where the observed entries are noisy with zero-mean bounded noise. A remarkable feature of our algorithms is that once the full completion ˆ M is obtained using existing procedures, they only rely on the locations of observed entries and not the values. Thus, the decisions regarding which entries to complete, i.e., to add to C is completely agnostic to the actual values of the revealed entries. Our framework is able to handle arbitrary sampling distributions, which captures scenarios such as overlapping groups or idiosyncratic rating habits. The online setting. We formulate the partial matrix completion problem in the online learning setting and propose an iterative gradient based online algorithm with provable guarantees. Corollary 4 is a special case of Theorem 22. A simple simulation of the online algorithm in this special case of uniform sampling over a subset of entries is demonstrated in Fig. 1. Corollary 4 (Online algorithm, uniform sampling distribution). Suppose M ? 2 [ 1, 1]n n is a bounded matrix with max-norm (Eq. 2) bounded by K. The sampling distribution µ is uniform over a fraction 0 < c 1 of the n2 entries. For any δ > 0, after T = O(δ 2K2n) iterations, the confidence matrix C 2 [0, 1]m n output by the online algorithm ODD satisfies with probability at least 1 c1 exp( c2δ2T) for some universal constants c1, c2 > 0, (1) : k Ck1 (c δ1/6)n2, (2) : 1 k Ck1 i2[m],j2[n] Cij(Mij M ? where V is the set of all matrices M 2 [ 1, 1]n n with max-norm bounded by K satisfying that Ei,j µ[(Mij M ? 1.2 Related work Matrix completion and recommendation systems. The common approach to collaborative filtering is that of matrix completion with trace norm minimization [Srebro, 2004]. It was shown that the trace norm regularization requires ((n + m)3/2) observations to complete the matrix under arbitrary distributions, even when the rank of the underlying matrix is constant, and this suffices even in the more challenging agnostic online learning setting [Hazan et al., 2012a, Shamir and Shalev Shwartz, 2014]. Srebro and Shraibman [2005] study matrix completion with the max-norm, which behaves better under arbitrary distributions, namely for low-rank matrices, O(n + m) observations suffice. Matrix completion and incoherence assumptions. A line of works in the matrix completion literature considered the goal of finding a completion of the ground truth matrix with low Frobenius norm error guarantee [Candès and Recht, 2009, Candès and Tao, 2010, Keshavan et al., 2010, Gross, 2011, Recht, 2011, Negahban and Wainwright, 2012, Jain et al., 2013, Chen et al., 2020, Abbe et al., 2020]. Such guarantee is strong but usually requires two restrictive assumptions on sampling distribution of observations and the ground truth matrix structure: (1) the sampling of observations is uniform across all entries, and (2) the ground truth matrix M ? satisfies some incoherence conditions. The incoherence condition is an assumption imposed on the singular vectors of M ? = U V >, which ensures the vectors ui, vj s to be sufficiently spread out on the unit sphere. The main reason the incoherence condition is necessary in establishing meaningful guarantees in low-rank matrix completion is that without such assumptions, there might exist multiple low-rank matrices that could explain the observed entries equally well but differ substantially on unobserved entries. When the incoherence condition is satisfied, it implies that the observed entries capture enough information about the low-rank structure, making it possible to recover the original matrix with sufficiently well. However, uniform sampling and incoherence assumptions do not hold in many realistic scenarios. Subsequently, to circumvent these restrictive assumptions, another line of work evaluates the generalization error as an alternative metric for completion performance [Srebro et al., 2004, Shamir and Shalev-Shwartz, 2014]. To formalize, consider the task of completing M ? with an arbitrary observation sampling distribution µ over [m] [n]. This can be conceptualized as predicting a hypothesis mapping from the domain [m] [n] to the codomain [ 1, 1]. The goal is to characterize the guarantees on the expected prediction error, quantified for a given completion ˆ M as E(i,j) µ[( ˆ Mij M ? ij)2]. When µ is the uniform distribution over [m] [n], the generalization error bound translates to a guarantee over the average error across all entries. However, if µ is arbitrary (e.g. supported over a fraction of the entries), then no guarantee could be established for entries that lie in the complement of the support of µ. This prohibits the use of ˆ M in settings where abstention is imperative. Our work takes an alternative approach to this fundamental limitation in matrix completion. The focus of our work is to identify entries where we can predict with high confidence to guarantee a low completion error weighted by evaluated confidence. We provide formulation of the problem, its convex relaxation, and an efficient gradient-based online algorithm due to a formulation of the problem as an online two-player game. Randomized rounding and semi-definite relaxations. For the efficient algorithm, the main analysis technique we deploy is based on randomized rounding solutions of semi-definite programming, due to the seminal work of Goemans and Williamson [1995]. These were originally developed in the context of approximation algorithms for MAX-CUT and other combinatorial problems. Here, we use this analysis technique for analyzing inner products of high dimensional vectors, but in a different context and to argue about a convex relaxation of a continuous optimization problem. Abstention in classification and regression. Abstaining from prediction has a long history in other learning problems such as binary classification Chow [1957]. There are several notable similarities. First, many of the algorithms work as ours does, by first fitting a complete classifier and then deciding afterwards where to abstain [e.g., Hellman, 1970, Goldwasser et al., 2020]. Second, numerous abstention models have been considered, including a fixed cost for abstaining [e.g., Chow, 1957, Kalai and Kanade, 2021], as well as fixing the error rate and maximizing the coverage [e.g., Geifman and El-Yaniv, 2017], as in our work, and fixing the coverage rate while minimizing the error [e.g., Geifman and El-Yaniv, 2019]. Third, there are algorithmic similarities, in particular the observation that worst-case performance guarantees do not depend on knowing the underlying distribution Kalai and Kanade [2021]. Online models of abstention have also been considered [e.g., Li et al., 2011]. For detailed survey on selective classification see [El-Yaniv and Wiener, 2010]. 2 Problem Setup and Preliminaries For a natural number n, let [n] = {1, 2, . . . , n}. Consider a fixed set of indices X = [m] [n] throughout. Thus, in this paper x 2 X is x = (i, j) where i 2 [m], j 2 [n]. The set of m n real-valued matrices is thus written as RX . We will thus view a matrix M over index set X as a function from X ! R, and for x 2 X, by slight abuse of notation denote by both Mx and M(x) the entry of M at index x. We focus on the squared loss in this paper, though it may be interesting to consider other loss functions in future work. Suppose is a distribution over X, for matrices, M, M 0, we denote by ( , M, M 0) def = Ex [(M(x) M 0(x))2]; we will also use the shorthand k M M 0k2 def = ( , M, M 0). When T 2 X N is a sequence of elements of X of length N, we denote by (T, M, M 0) x2T (M(x) M 0(x))2 and the shorthand k M M 0k2 def = (T, M, M 0). 2.1 Matrix Classes, Version Space and Generalization Matrix Norms. Given a matrix M, we denote by k Mk2,1 the maximum row norm of M. We denote by k Mkmax the max-norm of M defined as, def = min UV >=M k Uk2,1 k V k2,1. (2) The trace norm of a matrix M, denoted by k Mktr, is the sum of its singular values. The Frobenius norm of a matrix M, denoted by k Mkfr, is the square root of the sum of its squared entries. We will restrict attention to matrices with entries in [ 1, 1]. The following three classes of matrices with restrictions on respectively the rank, max-norm and trace-norm are of interest in this work. Formally, let K = {M 2 [ 1, 1]X }, and define, def = {rank(M) k} \ K, Mmax def = {k Mkmax K} \ K, Mtr def = {k Mktr } \ K. The standard assumption in matrix completion is that the target matrix is low-rank; however, most completion algorithms exploit convex optimization methods and optimize over max-norm or tracenorm bounded matrices. It is known that k Mkmax rank(M) if M 2 [ 1, 1]X [cf. Linial et al., 2007, Lemma 4.2] and it is easy to see that k Mktr rank(M)pmn for M 2 [ 1, 1]X . Thus, Mrk kpmn. Hence, we will work with the latter two classes which satisfy the following lemma which is proved in Appendix A. Lemma 5. The classes, Mmax , are closed under negations and are convex. Version Space. We define the notion of version spaces that are used in our key results. For a sequence T 2 X N, for any class of matrices M, a matrix M 2 M and β > 0, we define the version space around M of radius β based on T w.r.t. M as V(M, β, T; M) = {M 0 2 M | (T, M, M 0) β}. (3) Intuitively, version space is the set of matrices in a particular matrix class that are close to a given matrix with respect to T. Generalization Bounds. For a general class of matrices M, we define a notion of sample complexity that will guarantee the proximity of empirical and population measures of interest for all matrices in the class; we denote this notion by sc(", δ, M). Definition 6 (Sample Complexity). For ", δ > 0 and a class of matrices M, denote by sc(", δ, M), the sample complexity, to be the smallest natural number N0, such that for any distribution µ over X, for S µN0, and for any fixed ˆ M 2 M (possibly depending on S), with probability at least 1 δ, $$$ (S, M, ˆ M) (µ, M, ˆ M) If no such N0 exists, sc(", δ, F) Bounds on the sample complexity for matrix completion can be derived in terms of rank, max-norm and trace-norm using standard results in the literature, and the fact that the squared loss is 2-Lipschitz and bounded by 4 when both its arguments take values in [ 1, 1]. In the proposition below, the max-norm result follows from Theorem 5 in [Srebro and Shraibman, 2005] and the trace-norm result follows from Theorem 4 in [Shamir and Shalev-Shwartz, 2014]. Proposition 7. For the classes, Mmax , the following hold, 1. sc(", δ, Mmax K2(m + n) + log 1 2. sc(", δ, Mtr pm + n + log 1 It is worth comparing the two bounds in Proposition 7 above. Consider the matrix M ? consisting of all 1 s from a matrix completion point of view this is particularly easy as every user likes every movie . Note however that k M ?kmax = 1 and k M ?ktr = pmn. For this example, ignoring the dependence on ", δ, the sample complexity bound obtained using the trace-norm result would be O((m + n)pm + n), while that using max-norm would be O(m + n). In general, it is always the case that k Mktr/pmn k Mkmax, so it may seem that the bounds in terms of trace-norm are weaker. However, there are matrices for which this gap can be large and the sample complexity bound in terms of trace-norm is shown to be tight in general (see [Srebro and Shraibman, 2005, Shamir and Shalev-Shwartz, 2014] for further details). 2.2 Full Completion Problem (with Noise) We now define the matrix completion problem with zero-mean noise, in fact a matrix estimation problem in the noisy case. Let D be a distribution supported on X [ 1, 1]; the results in the paper can all be easily extended when D is supported on X [ B, B], increasing squared error by a B2 factor. Let µ be the marginal distribution of D over X. Let SXY = h(xt, yt)i N t=1 be an iid sample drawn from DN. Note here that xt = (it, jt) denotes the index of the matrix and yt the observed value. We let S = (x1, . . . , x N) denote the sequence of xt s from SXY (with repetitions allowed); note that S is distributed as µN. Let M ? 2 [ 1, 1]X be the matrix where M ? ij = ED[Y |X = (i, j)]. We say that ˆ M is an "-accurate completion of M ?, if (µ, M ?, ˆ M) ". Definition 8 (Full Completion Algorithm). We say that Full Comp(SXY , ", δ, M) is a full completion algorithm with sample complexity sc FC(", δ, M) for M, if provided SXY DN for some D over X [ 1, 1] with M ? 2 M, N sc FC(", δ, M), Full Comp(SXY , ", δ, M) outputs ˆ M 2 M that with probability at least 1 δ satisfies, (µ, M ?, ˆ M) ". The following result follows from [Srebro and Shraibman, 2005, Shamir and Shalev-Shwartz, 2014] (see also Prop. 7). Proposition 9. There exists polynomial time (in mn/") full completion algorithms for Mmax (e.g. ERM methods) with the following sample complexity bounds: 1. sc FC(", δ, Mmax K2(m + n) + log 1 2. sc Mtr (", δ, Mtr pm + n + log 1 2.3 The Partial Matrix Completion Problem Recall the definition of the (fractional) partial matrix completion in Eq. 1 and Def. 1. In this work, we will assume that a full completion matrix ˆ M is already obtained. The focus of the present work is finding a completion matrix C 2 {0, 1}X (or C 2 [0, 1]X for fractional coverage) with large coverage. The coverage in both cases are defined as: |C| x2X Cx, and a low loss, defined as: (C, M ?, ˆ M) def = 1 |C| Note that C 2 [0, 1]X can be viewed as a fractional set and we are overloading the notation to allow that. Such a fractional coverage can be randomly rounded to a set whose size is within 1 of |C|. In an ideal world, we would have at most " loss measured over the cells we complete as defined above, and large or even full coverage |C| = mn. Note that although the requirement in the optimization problems in Eq. (1) and Def. 1 is to guarantee (C, M ?, M) ", as we don t know M ?, the only guarantee we have by using a full completion algorithm (and generalization bounds) is that (with high probability) M ? is in some version space, say V, centered at ˆ M. So we actually show a stronger guarantee that, sup M2V (C, M, ˆ M) ". A second observation is that if M is a version space around ˆ M, then by convexity of the matrix classes, the matrix (M ˆ M)/2, must be in some version space, say V0, centered around the zero matrix, 0X . Thus, we will actually find a C that guarantees, (C, M, 0X ) " for every M 2 V0. This means that for maximizing coverage our algorithm has the remarkable property we only need to know the locations of the revealed entries as indicated by S. Section 3 presents a computationally inefficient but statistically superior algorithm in terms of sample complexity for the coverage problem and its consequences for partial matrix completion. Section 4 presents a computationally efficient algorithm at the cost of a slightly worse sample complexity when using the class Mmax 3 An Inefficient Algorithm The main novelty in the partial matrix completion problem is that of finding an optimal coverage, defined as the matrix C in the formulation in Def. 1. In this section we give an (inefficient) algorithm for finding the optimal coverage, and the generalization error bounds for Partial Matrix Completion that arise from it. In the next section we give an efficient approximation algorithm for doing the same, with slightly worse sample complexity bounds and more complex analysis. This differs substantially from prior work on abstention in classification and regression [Kalai and Kanade, 2021] where the optimal solution can be found in polynomial time. Let C 2 [0, 1]X be a target confidence matrix. For such a C, we will denote by C the probability distribution where C(x) = Cx/k Ck1 for x 2 X. Let S µN be a sample obtained from the target distribution µ. We output a C which is an optimal solution to the following optimization problem, inspired by a similar problem studied by Kalai and Kanade [2021] for classification with abstention. Parameters and inputs : γ, β, S, M maximize k Ck1 (4) subject to C 2 [0, 1]X , and 8M 2 V(0X , β, S; M), ( C, M, 0X ) γ Algorithm 1 1: Inputs: SXY = h(xt, yt)i N t=1 DN, ", δ, M, Full Comp (cf. Def. 8). 2: Obtain ˆ M 2 M using Full Comp(SXY , "/4, δ/3, M). 3: Obtain C using MP 4 with γ def = "/4, β def = "/8, S = (x1, x2, . . . , x N) from SXY . 4: return ( ˆ M, C). MP 4 defines a family of optimization problems; we will typically use M to be Mmax . The above optimization problem is in fact a linear program. The only unknowns are Cx for x 2 X. However, the set of constraints is infinite and it is unclear how a separation oracle for the constraint set may be designed. Such an oracle could be designed by solving the following optimization problem: max M2V(0X ,β,S;M) ( C, M, 0X ). (5) This optimization problem requires maximizing a quadratic function, and we prove it to be NP-hard with the additional PSD and symmetric constraints in Appendix A.2. The following result subsumes Corollary 2. We defer the proof of this theorem to Appendix A, as the proof is similar to that of our efficient algorithm in the next section. Theorem 10. Let M be either Mmax . Let D be distribution over X [ 1, 1], µ the marginal of D over X, µmax = maxi,j Pµ((i, j) is sampled). Suppose that M ?, defined as M ? ij = E(X,Y ) D[Y |X = (i, j)] satisfies that M ? 2 M. Furthermore, suppose that SXY DN and that Full Comp is a full completion algorithm as in Defn. 8. Then, provided N max{sc FC("/4, δ/3, M), sc("/8, δ/3, M)}, for ( ˆ M, C) output by Alg. 1, it holds that: 1. k Ck1 1/µmax, Cx( ˆ Mx M ? 4 An Efficient Algorithm In this section, we show how the result from Proposition 14 can be achieved using an efficient algorithm at a modest cost in terms of sample complexity when using the matrix class Mmax K . In particular, we consider the optimization problem defined in MP 6, where the constraint k Mk2 C γ for all M 2 V(0X , β, S; Mmax K ) from MP 4 is replaced by Ex C[Mx] γ. The efficiency comes from the fact that we can now implement a separation oracle for the constraint set by solving, for a given C, the problem, max M2V(0X ,β,S,Mmax K ) Ex C[Mx]. Mathematical program 6 is a convex optimization problem since the constraint set is convex and the objective function is linear. However, it is not immediately clear how this relaxed optimization problem relates to the original, which is the main technical contribution of this section. Parameters and inputs : γ, β, S, Mmax K maximize k Ck1 (6) subject to C 2 [0, 1]X , and 8M 2 V(0X , β, S; Mmax K ), Ex C[Mx] γ Algorithm 2 (Efficient) Offline Algorithm for Partial Matrix Completion 1: Inputs: SXY = h(xt, yt)i N t=1 DN, Full Comp (cf. Def. 8), ", δ. 2: Obtain ˆ M 2 Mmax K using Full Comp(SXY , "2/(4 2K2), δ/3, Mmax 3: Obtain C using MP 6 with γ def = "/(2 K), β def = "2/(8 2K2), S = (x1, x2, . . . , x N) from SXY . 4: return ( ˆ M, C). The following result subsumes Corollary 3. Theorem 11. Let D be a distribution over X [ 1, 1], µ the marginal of D over X, µmax = maxi,j Pµ((i, j) is sampled). Suppose that M ?, defined as M ? ij = E(X,Y ) D[Y |X = (i, j)] satisfies that M ? 2 Mmax K . Furthermore, suppose that SXY DN and that Full Comp is a full completion algorithm as in Defn. 8. Then, provided N max{sc FC("2/(4 2K2), δ/3, Mmax K ), sc("2/(8 2K2), δ/3, Mmax K )}, for ( ˆ M, C) output by Alg. 2, it holds that: 1. k Ck1 1/µmax, Cx( ˆ Mx M ? 5 Online Setting The previous section introduces the offline setting for the partial matrix completion problem. In Appendix B, we describe the online version of the problem, which is motivated by two important considerations. First, in many applications, the observation pattern is more general than a fixed distribution. It can be a changing distribution or be comprised of adversarial observations. Second, our online algorithm incrementally updates the solution via iterative gradient methods, which is more efficient than the offline methods. For space considerations, details are deferred to Appendix B which contains the setting, definitions, algorithm specification, and main results, and Appendix C, which details the proofs. In particular, the online algorithm, called Online Dual Descent (ODD), is described in Algorithm 3, and its regret guarantee in Theorem 18. The online regret guarantee implies the statistical learning guarantees of the previous sections when the support size is a constant fraction of the full matrix, and this implication is spelled out precisely in Corollary 22. 6 Experiments and Implementation The Movie Lens dataset ([Harper and Konstan, 2015]) consists of, among other data, a set of users with their rankings of a set of movies. It is a common benchmark for matrix completion because some of the rankings are missing, and one can make predictions on the rankings of user preferences. We used the dataset differently, aiming to test our online algorithm on generating a satisfying confidence matrix. The experimental procedure is outlined as follows: we used training data from 250 users and their ratings on 250 movies, giving us a total of 5189 completed samples from the incomplete matrix of size 250 250. We ran our algorithm, ODD (Algorithm 3), to get a confidence matrix C. In parallel, we used another standard matrix completion tool, fancyimpute ([Rubinsteyn and Feldman, 2016]), to fill in the missing entries of the matrix and obtain a completion ˆ Mf. After C and ˆ Mf are obtained, we reveal the true ratings at the missing entries using the validation set and computed the mean squared error of the predicted rating and true rating at each entry. The following plots show the distribution of the mean squared error with respect to the confidence score C assign at the particular entry. Figure 2: (a) ODD heuristic on 250 250 user-movie rating data from the Movie Lens Dataset. Squared error in entry-wise prediction and entry-wise confidence level estimated by the ODD heuristic. (b) The average squared error over various confidence levels suggests that a higher confidence level is correlated with a lower squared error. 7 Conclusion In this work we define the setting of partial matrix completion, with our high-level contributions outlined as the following: A new framework. We propose a new framework called partial matrix completion. In this problem, one is given possibly noisy observations from a low-rank matrix, and the goal is to identify entries that can be predicted with high confidence. The partial matrix completion problem answers two key questions: Based on possibly noisy samples from an arbitrary unknown sampling distribution µ, how many entries can be completed with " error? How can we efficiently identify the entries that can be accurately completed? When the underlying matrix has low rank k, we show that it is possible to complete a large fraction of the matrix using only O(k(m + n)) observations while simultaneously guaranteeing high accuracy over the completed set. We then study the complexity of identifying the optimal completion matrix. We show that a naïve mathematical programming formulation of the problem is hard. However, we propose a relaxation that gives rise to efficient algorithms, which results in a slightly worse dependence on k for the sample complexity. These guarantees are outlined both in Corollary 2 and Corollary 3, and Theorem 10 and Theorem 11 for their more general versions. Online game formulation. Furthermore, we consider the partial matrix completion problem in the online setting, where the revealed observations are not required to follow any particular fixed distribution. The goal henceforth is to minimize regret, the gap between the algorithm s performance and the single best decision in hindsight. This is a more general setting, as when imposed with distribution assumptions, regret guarantees in online learning algorithms naturally translate to statistical guarantees. Our proposed online partial matrix completion algorithm is derived from an online repeated game. The version space is the set of all valid completions with low generalization error. High confidence should be assigned to entries where all completions in the version space are similar. Therefore, we formulate the problem as a two-player online game. One player iteratively updates a confidence matrix, and the other learns the version space. We gave an iterative gradient-based method with provable regret guarantees and concluded with preliminary experimental evidence of the validity of our framework. 8 Acknowledgements Elad Hazan acknowledges funding from the Office of Naval Research grant N000142312156, the NSF award 2134040, and Open Philanthropy. This work was done in part when Clara Mohri was visiting and supported by Princeton University. Emmanuel Abbe, Jianqing Fan, Kaizheng Wang, and Yiqiao Zhong. Entrywise eigenvector analysis of random matrices with low expected rank. Annals of statistics, 48(3):1452, 2020. Emmanuel J Candès and Benjamin Recht. Exact matrix completion via convex optimization. Foun- dations of Computational mathematics, 9(6):717 772, 2009. Emmanuel J Candès and Terence Tao. The power of convex relaxation: Near-optimal matrix completion. IEEE Transactions on Information Theory, 56(5):2053 2080, 2010. Yuxin Chen, Yuejie Chi, Jianqing Fan, Cong Ma, and Yuling Yan. Noisy matrix completion: Understanding statistical guarantees for convex relaxation via nonconvex optimization. SIAM journal on optimization, 30(4):3098 3121, 2020. Chi-Keung Chow. An optimum character recognition system using decision functions. IRE Transac- tions on Electronic Computers, 4:247 254, 1957. Ran El-Yaniv and Yair Wiener. On the foundations of noise-free selective classification. Journal of Machine Learning Research, 11(53):1605 1641, 2010. URL http://jmlr.org/papers/ v11/el-yaniv10a.html. Yoav Freund and Robert E Schapire. Adaptive game playing using multiplicative weights. Games and Economic Behavior, 29(1-2):79 103, 1999. Yonatan Geifman and Ran El-Yaniv. Selective classification for deep neural networks. Advances in neural information processing systems, 30, 2017. Yonatan Geifman and Ran El-Yaniv. Selectivenet: A deep neural network with an integrated reject option. In Kamalika Chaudhuri and Ruslan Salakhutdinov, editors, Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 9-15 June 2019, Long Beach, California, USA, volume 97 of Proceedings of Machine Learning Research, pages 2151 2159. PMLR, 2019. Michel X Goemans and David P Williamson. Improved approximation algorithms for maximum cut and satisfiability problems using semidefinite programming. Journal of the ACM (JACM), 42(6): 1115 1145, 1995. ShafiGoldwasser, Adam Tauman Kalai, Yael Tauman Kalai, and Omar Montasser. Selective classification algorithms provably robust to arbitrary adversarial examples. In Advances in Neural Information Processing Systems, 2020. David Gross. Recovering low-rank matrices from few coefficients in any basis. IEEE Transactions on Information Theory, 57(3):1548 1566, 2011. Martin Grötschel, László Lovász, and Alexander Schrijver. The ellipsoid method and its consequences in combinatorial optimization. Combinatorica, 1(2):169 197, 1981. F. Maxwell Harper and Joseph A. Konstan. The movielens datasets: History and context. ACM Trans. Interact. Intell. Syst., 5(4), dec 2015. ISSN 2160-6455. doi: 10.1145/2827872. Johan Hastad. Clique is hard to approximate within n/sup 1-/spl epsiv. In Proceedings of 37th Conference on Foundations of Computer Science, pages 627 636. IEEE, 1996. Elad Hazan, Satyen Kale, and Shai Shalev-Shwartz. Near-optimal algorithms for online matrix prediction. In COLT, pages 38.1 38.13, 2012a. Elad Hazan, Satyen Kale, and Shai Shalev-Shwartz. Near-optimal algorithms for online matrix prediction. In Conference on Learning Theory, pages 38 1. JMLR Workshop and Conference Proceedings, 2012b. Elad Hazan et al. Introduction to online convex optimization. Foundations and Trends in Optimiza- tion, 2(3-4):157 325, 2016. Martin E. Hellman. The nearest neighbor classification rule with a reject option. IEEE Transactions on Systems Science and Cybernetics, 6(3):179 185, 1970. doi: 10.1109/TSSC.1970.300339. Prateek Jain, Praneeth Netrapalli, and Sujay Sanghavi. Low-rank matrix completion using alternating minimization. In Proceedings of the forty-fifth annual ACM symposium on Theory of computing, pages 665 674, 2013. Adam Kalai and Varun Kanade. Towards optimally abstaining from prediction with ood test exam- ples. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, volume 34, pages 12774 12785. Curran As- sociates, Inc., 2021. URL https://proceedings.neurips.cc/paper/2021/file/ 6a26c75d6a576c94654bfc4dda548c72-Paper.pdf. Raghunandan H Keshavan, Andrea Montanari, and Sewoong Oh. Matrix completion from a few entries. IEEE transactions on information theory, 56(6):2980 2998, 2010. Lihong Li, Michael L Littman, Thomas J Walsh, and Alexander L Strehl. Knows what it knows: a framework for self-aware learning. Machine learning, 82(3):399 443, 2011. Nati Linial, Shahar Mendelson, Gideon Schechtman, and Adi Shraibman. Complexity measures of sign matrices. Combinatorica, 27(4):439 463, 2007. Mehrdad Mahdavi, Rong Jin, and Tianbao Yang. Trading regret for efficiency: online convex optimization with long term constraints. The Journal of Machine Learning Research, 13(1): 2503 2528, 2012. Sahand Negahban and Martin J Wainwright. Restricted strong convexity and weighted matrix completion: Optimal bounds with noise. The Journal of Machine Learning Research, 13(1): 1665 1697, 2012. Benjamin Recht. A simpler approach to matrix completion. Journal of Machine Learning Research, 12(12), 2011. Alex Rubinsteyn and Sergey Feldman. fancyimpute: An imputation library for python, 2016. URL https://github.com/iskandr/fancyimpute. Ohad Shamir and Shai Shalev-Shwartz. Matrix completion with the trace norm: Learning, bounding, and transducing. Journal of Machine Learning Research, 15(1):3401 3423, 2014. Nathan Srebro. Learning with Matrix Factorizations. Ph D thesis, Massachusetts Institute of Technol- Nathan Srebro and Adi Shraibman. Rank, trace-norm and max-norm. In International Conference on Computational Learning Theory, pages 545 560. Springer, 2005. Nathan Srebro, Noga Alon, and Tommi Jaakkola. Generalization error bounds for collaborative prediction with low-rank matrices. Advances In Neural Information Processing Systems, 17, 2004. Manfred K Warmuth. The blessing and the curse of the multiplicative updates. In International Conference on Discovery Science, pages 382 382. Springer, 2010.