# implicit_regularization_for_optimal_sparse_recovery__ebf987fe.pdf Implicit Regularization for Optimal Sparse Recovery Tomas Vaškeviˇcius1, Varun Kanade2, Patrick Rebeschini1 1 Department of Statistics, 2 Department of Computer Science University of Oxford {tomas.vaskevicius, patrick.rebeschini}@stats.ox.ac.uk varunk@cs.ox.ac.uk We investigate implicit regularization schemes for gradient descent methods applied to unpenalized least squares regression to solve the problem of reconstructing a sparse signal from an underdetermined system of linear measurements under the restricted isometry assumption. For a given parametrization yielding a non-convex optimization problem, we show that prescribed choices of initialization, step size and stopping time yield a statistically and computationally optimal algorithm that achieves the minimax rate with the same cost required to read the data up to poly-logarithmic factors. Beyond minimax optimality, we show that our algorithm adapts to instance difficulty and yields a dimension-independent rate when the signal-to-noise ratio is high enough. Key to the computational efficiency of our method is an increasing step size scheme that adapts to refined estimates of the true solution. We validate our findings with numerical experiments and compare our algorithm against explicit 1 penalization. Going from hard instances to easy ones, our algorithm is seen to undergo a phase transition, eventually matching least squares with an oracle knowledge of the true support. 1 Introduction Many problems in machine learning, science and engineering involve high-dimensional datasets where the dimensionality of the data d is greater than the number of data points n. Linear regression with sparsity constraints is an archetypal problem in this setting. The goal is to estimate a d-dimensional vector w Rd with k non-zero components from n data points (xi, yi) Rd R, i {1, . . . , n}, linked by the linear relationship yi = xi, w + ξi, where ξi is a possible perturbation to the ith observation. In matrix-vector form the model reads y = Xw + ξ, where xi corresponds to the ith row of the n d design matrix X. Over the past couple of decades, sparse linear regression has been extensively investigated from the point of view of both statistics and optimization. In statistics, sparsity has been enforced by designing estimators with explicit regularization schemes based on the 1 norm, such as the lasso [46] and the closely related basis pursuit [15] and Dantzig selector [13]. In the noiseless setting (ξ = 0), exact recovery is possible if and only if the design matrix satisfies the restricted nullspace property [16, 17, 19]. In the noisy setting (ξ = 0), exact recovery is not feasible and a natural criterion involves designing estimators w that can recover the minimax-optimal rate kσ2 log(d/k)/n for the squared 2 error w w 2 2 in the case of i.i.d. sub-Gaussian noise with variance proxy σ2 when the design matrix satisfies restricted eigenvalue conditions [9, 48]. The lasso estimator, defined as any vector w that minimizes the objective Xw y 2 2 + λ w 1, achieves the minimax-optimal rate upon proper tuning of the regularization parameter λ. The restricted isometry property (RIP) [14] has been largely considered in the literature, as it implies both the restricted nullspace and eigenvalue conditions [16, 49], and as it is satisfied when the entries of X are i.i.d. sub-Gaussian and subexponential with sample size n = Ω(k log(d/k)) and n = Ω(k log2(d/k)) respectively [32, 1], or when the columns are unitary, e.g. [23, 24, 39, 41]. 33rd Conference on Neural Information Processing Systems (Neur IPS 2019), Vancouver, Canada. In optimization, computationally efficient iterative algorithms have been designed to solve convex problems based on 1 constraints and penalties, such as composite/proximal methods [4, 35]. Under restricted eigenvalue conditions, such as restricted strong convexity and restricted smoothness, various iterative methods have been shown to yield exponential convergence to the problem solution globally up to the statistical precision of the model [2], or locally once the iterates are close enough to the optimum and the support of the solution is identified [10, 28, 45]. In some regimes, for a prescribed choice of the regularization parameter, these algorithms are computationally efficient. They require O(1) iterations, where the notation O hides poly-logarithmic terms, and each iteration costs O(nd). Hence the total running cost is O(nd), which is the cost to store/read the data in/from memory. These results attest that there are regimes where optimal methods for sparse linear regression exist. However, these results reply upon tuning the hyperparameters for optimization, such as the step size, carefully, which in turn depends on identifying the correct hyperparameters, such as λ, for regularization. In practice, one has to resort to cross-validation techniques to tune the regularization parameter. Cross-validation adds an additional burden from a computational point of view, as the optimization algorithms need to be run for different choices of the regularization terms. In the context of linear regression with 2 penalty, a.k.a. ridge regression, potential computational savings have motivated research on the design of implicit regularization schemes where model complexity is directly controlled by tuning the hyper-parameters of solvers applied to unpenalized/unconstrained programs, such as choice of initialization, step-size, iteration/training time. There has been increasing interest in understanding the effects of implicit regularization (sometimes referred to as implicit bias) of machine learning algorithms. It is widely acknowledged that the choice of algorithm, parametrization, and parameter-tuning, all affect the learning performance of models derived from training data. While implicit regularization has been extensively investigated in connection to the 2 norm, there seem to be no results for sparse regression, which is surprising considering the importance of the problem. 1.1 Our Contributions In this work, we merge statistics with optimization, and propose the first statistically and computationally optimal algorithm based on implicit regularization (initialization/step-size tuning and early stopping) for sparse linear regression under the RIP. The algorithm that we propose is based on gradient descent applied to the unregularized, underdetermined objective function Xw y 2 2 where w is parametrized as w = u u v v, with u, v Rd and denotes the coordinate-wise multiplication operator for vectors. This parametrization yields a non-convex problem in u and v. We treat this optimization problem as a proxy to design a sequence of statistical estimators that correspond to the iterates of gradient descent applied to solve the sparse regression problem, and hence are cheap to compute iteratively. The matrix formulation of the same type of parametrization that we adopt has been recently considered in the setting of low-rank matrix recovery where it leads to exact recovery via implicit regularization in the noiseless setting under the RIP [25, 30]. In our case, this choice of parametrization yields an iterative algorithm that performs multiplicative updates on the coordinates of u and v, in contrast to the additive updates obtained when gradient descent is run directly on the parameter w, as in proximal methods. This feature allows us to reduce the convergence analysis to one-dimensional iterates and to differentiate the convergence on the support set S = {i {1, . . . , d} : w i = 0} from the convergence on its complement Sc = {1, . . . , d} \ S. We consider gradient descent initialized with u0 = v0 = α1, where 1 is the all-one vector. We show that with a sufficiently small initialization size α > 0 and early stopping, our method achieves exact reconstruction with precision controlled by α in the noiseless setting, and minimax-optimal rates in the noisy setting. To the best of our knowledge, our results are the first to establish non2 implicit regularization for a gradient descent method in a general noisy setting.1 These results rely on a constant choice of step size η that satisfies a bound related to the unknown parameter w max = w . We show how this choice of η can be derived from the data itself, i.e. only based on known quantities. If the noise vector ξ is made up of i.i.d. sub-Gaussian components with variance proxy σ2, this choice of η yields O((w max n)/(σ log d) log α 1) iteration complexity to achieve minimax rates. In order to achieve computational optimality, we design a preconditioned version of gradient descent (on 1However, see Remark 1 for work concurrent to ours that achieves similar goals. the parameters u and v) that uses increasing step-sizes and has running time O(nd). The iterationdependent preconditioner relates to the statistical nature of the problem. It is made by a sequence of diagonal matrices that implement a coordinate-wise increasing step-size scheme that allows different coordinates to accelerate convergence by taking larger steps based on refined estimates of the corresponding coordinates of w . This algorithm yields O(log((w max n)/(σ log d)) log α 1) iteration complexity to achieve minimax rates in the noisy setting. Since each iteration costs O(nd), the total computation complexity is, up to poly-logarithmic factors, the same as simply storing/reading the data. This algorithm is minimax-optimal and, up to logarithmic factors, computationally optimal. In contrast, we are not aware of any work on implicit 2 regularization that exploits an increasing step sizes scheme in order to attain computational optimality. To support our theoretical results we present a simulation study of our methods and comparisons with the lasso estimator and with the gold standard oracle least squares estimator, which performs least squares regression on S assuming oracle knowledge of it. We show that the number of iterations t in our method plays a role similar to the lasso regularization parameter λ. Despite both algorithms being minimax-optimal with the right choice of t and λ respectively, the gradient descent optimization path which is cheaper to compute as each iteration of gradient descent yields a new model exhibits qualitative and quantitative differences from the lasso regularization path which is more expensive to compute as each model requires solving a new lasso optimization program. In particular, the simulations emphasize how the multiplicative updates allow gradient descent to fit one coordinate of w at a time, as opposed to the lasso estimator that tends to fit all coordinates at once. Beyond minimax results, we prove that our methods adapt to instance difficulty: for easy problems where the signal is greater than the noise, i.e. w min XTξ /n with w min = mini S |w i |, our estimators achieve the statistical rate kσ2 log(k)/n, which does not depend on d. The experiments confirm this behavior and further attest that our estimators undergo a phase transition that is not observed for the lasso. Going from hard instances to easy ones, the learning capacity of implicitly-regularized gradient descent exhibits a qualitative transition and eventually matches the performance of oracle least squares. 1.2 Related Work Sparse Recovery. The statistical properties of explicit 1 penalization techniques are well studied [48, 13, 9, 31, 33]. Minimax rates for regression under sparsity constraints are derived in [37]. Computing the whole lasso regularization path can be done via the lars algorithm [18]. Another widely used approach is the glmnet which uses cyclic coordinate-descent with warm starts to compute regularization paths for generalized linear models with convex penalties on a pre-specified grid of regularization parameters [22]. [4] reviews various optimization techniques used in solving empirical risk minimization problems with sparsity inducing penalties. Using recent advances in mixed integer optimization, [8] shows that the best subset selection problem can be tackled for problems of moderate size. For such problem sizes, comparisons between the lasso and best subset selection problem ( 0 regularization) were recently made, suggesting that the best subset selection performs better in high signal-to-noise ratio regimes whereas the lasso performs better when the signal-to-noise ratio is low [29]. In this sense, our empirical study in Section 5 suggests that implicitly-regularized gradient descent is more similar to 0 regularization than 1 regularization. Several other techniques related to 1 regularization and extensions to the lasso exist. We refer the interested reader to the books [11, 47]. Implicit Regularization/Bias. Connections between 2 regularization and gradient descent optimization paths have been known for a long time and are well studied [12, 20, 52, 7, 38, 51, 34, 44, 3]. In contrast, the literature on implicit regularization inducing sparsity is scarce. Coordinate-descent optimization paths have been shown to be related to 1 regularization paths in some regimes [21, 18, 40, 54]. Understanding such connections can potentially allow transferring the now wellunderstood theory developed for penalized forms of regularization to early-stopping-based regularization which can result in lower computational complexity. Recently, [53] have shown that neural networks generalize well even without explicit regularization despite the capacity to fit unstructured noise. This suggests that some implicit regularization effect is limiting the capacity of the obtained models along the optimization path and thus explaining generalization on structured data. Understanding such effects has recently drawn a lot of attention in the machine learning community. In particular, it is now well understood that the optimization algorithm itself can be biased towards a particular set of solutions for underdetermined problems with many global minima where, in contrast to the work cited above, the bias of optimization algorithm is investigated at or near convergence, usually in a noiseless setting [43, 27, 26, 25, 30]. We compare our assumptions with the ones made in [30] in Appendix G. Remark 1 (Concurrent Work). After completing this work we became aware of independent concurrent work [56] which considers Hadamard product reparametrization wt = ut vt in order to implicitly induce sparsity for linear regression under the RIP assumption. Our work is significantly different in many aspects discussed in Appendix H. In particular, we obtain computational optimality and can properly handle the general noisy setting. 2 Model and Algorithms We consider the model defined in the introduction. We denote vectors with boldface letters and real numbers with normal font; thus, w denotes a vector and wi denotes the ith coordinate of w. For any index set A we let 1A denote a vector that has a 1 entry in all coordinates i A and a 0 entry elsewhere. We denote coordinate-wise inequalities by . With a slight abuse of notation we write w2 to mean the vector obtained by squaring each component of w. Finally, we denote inequalities up to multiplicative absolute constants, meaning that they do not depend on any parameters of the problem, by . A table of notation can be found in Appendix J. We now define the restricted isometry property which is the key assumption in our main theorems. Definition 1 (Restricted Isometry Property (RIP)). A n d matrix X/ n satisfies the (δ, k)-(RIP) if for any k-sparse vector w Rd we have (1 δ) w 2 2 Xw/ n 2 2 (1 + δ) w 2 2 . The RIP assumption was introduced in [14] and is standard in the compressed sensing literature. It requires that all n k sub-matrices of X/ n are approximately orthonormal where δ controls extent to which this approximation holds. Checking if a given matrix satisfies the RIP is NP-hard [5]. In compressed sensing applications the matrix X/ n corresponds to how we measure signals and it can be chosen by the designer of a sparse-measurement device. Random matrices are known to satisfy the RIP with high probability, with δ decreasing to 0 as n increases for a fixed k [6]. We consider the following problem setting. Let u, v Rd and define the mean squared loss as L(u, v) = 1 n X (u u v v) y 2 2 . Letting w = u u v v and performing gradient descent updates on w, we recover the original parametrization of mean squared error loss which does not implicitly induce sparsity. Instead, we perform gradient descent updates on (u, v) treating it as a vector in R2d and we show that the corresponding optimization path contains sparse solutions. Let η > 0 be the learning rate, (mt)t 0 be a sequence of vectors in Rd and diag(mt) be a d d diagonal matrix with mt on its diagonal. We consider the following general form of gradient descent: (ut+1, vt+1) = (ut, vt) η diag(mt, mt) L(ut, vt) (ut, vt) . (1) We analyze two different choices of sequences (mt)t 0 yielding two separate algorithms. Algorithm 1. Let α, η > 0 be two given parameters. Let u0 = v0 = α and for all t 0 we let mt = 1. Perform the updates given in (1). Algorithm 2. Let α, τ N and w max ˆz 2w max be three given parameters. Set η = 1 20ˆz and u0 = v0 = α. Perform the updates in (1) with m0 = 1 and mt adaptively defined as follows: 1. Set mt = mt 1. 2. If t = mτ log α 1 for some natural number m 2 then let mt,j = 2mt 1,j for all j such that u2 t,j v2 t,j 2 m 1ˆz. Algorithm 1 corresponds to gradient descent with a constant step size, whereas Algorithm 2 doubles the step-sizes for small enough coordinates after every τ log α 1 iterations. Before stating the main results we define some key quantities. First, our results are sensitive to the condition number κ = κ(w ) = w max/w min of the true parameter vector w . Since we are not able to recover coordinates below the maximum noise term XTξ /n, for a desired precision ε we can treat all coordinates of w below ε ( XTξ /n) as 0. This motivates the following definition of an effective condition number for given w , X, ξ and ε: κeff = κeff(w , X, ξ, ε) = w max/(w min ε ( XTξ /n)). We remark that κeff(w , X, ξ, ε) κ(w ). Second, we need to put restrictions on the RIP constant δ and initialization size α. These restrictions are given by the following: δ(k, w , X, ξ, ε) = 1/( k(1 log κeff(w ))), α(w , ε, d) := ε2 ε 1 (2d + 1)2 (w max)2 3 Main Results The following result is the backbone of our contributions. It establishes rates for Algorithm 1 in the norm as opposed to the typical rates for the lasso that are often only derived for the 2 norm. Theorem 1. Fix any ε > 0. Suppose that X/ n satisfies the (k + 1, δ)-RIP with δ δ (k, w , X, ξ, ε) and let the initialization α satisfy α α(w , ε, d). Then, Algorithm 1 with η 1/(20w max) and t = O((κeff(w ))/(ηw max) log α 1) iterations satisfies |wt,i w i | n XTξ ε if i S and w min 1 n XTξ 1S ε if i S and w min 1 n XTξ ε, α if i / S. This result shows how the parameters α, η and t affect the learning performance of gradient descent. The size of α controls the size of the coordinates outside the true support S at the stopping time. We discuss the role and also the necessity of small initialization size to achieve the desired statistical performance in Section 5. A different role is played by the step size η whose size affects the optimal stopping time t. In particular, (ηt)/ log α 1 can be seen as a regularization parameter closely related to λ 1 for the lasso. To see this, suppose that the noise ξ is σ2-sub-Gaussian with independent components. Then with high probability XTξ /n (σ log d)/ n). In such a setting an optimal choice of λ for the lasso is Θ((σ log d)/ n). On the other hand, letting t be the optimal stopping time given in Theorem 1, we have (ηt )/ log α 1 = O(1/w min(X, ξ, ε)) = O( n/(σ log d)). The condition η 1/(20w max) is also necessary up to constant factors in order to prevent explosion. If we can set 1/w max η 1/(20w max) then the iteration complexity of Theorem 1 reduces to O(κeff(w ) log α 1). The magnitude of w max is, however, an unknown quantity. Similarly, setting the proper initialization size α depends on w max, w min, d and the desired precision ε. The requirement that α w min/2 is an artifact of our proof technique and tighter analysis could replace this condition by simply α ε. Hence the only unknown quantity for selecting a proper initialization size is w max. The next theorem shows how w max can be estimated from the data up to a multiplicative factor 2 at the cost of one gradient descent iteration. Once this estimate is computed, we can properly set the initialization size and the learning rate η 1 w max which satisfies our theory and is tight up to constant multiplicative factors. We remark that η used in Theorem 2 can be set arbitrarily small (e.g., set η = 10 10) and is only used for one gradient descent step in order to estimate w max. Theorem 2 (Estimating w max). Set α = 1 and suppose that X/ n satisfies the (k + 1, δ)-RIP with δ 1/(20 k). Let the step size η be any number satisfying 0 < η 1/(5w max) and suppose that w max 5 XTξ /n. Perform one step of gradient descent and for each i {1, . . . , d} compute the update factors defined as f + i = (u1)i and f i = (v1)i. Let fmax = f + f . Then w max (fmax 1)/(3 η) < 2w max. We present three main corollaries of Theorem 1. The first one shows that in the noiseless setting exact recovery is possible and is controlled by the desired precision ε and hence by the initialization size α. Corollary 1 (Noiseless Recovery). Let ξ = 0. Under the assumptions of Theorem 1, the choice of η given by Theorem 2 and t = O(κeff(w ) log α 1), Algorithm 1 yields wt w 2 2 kε2. In the general noisy setting exact reconstruction of w is not possible. In fact, the bounds in Theorem 1 do not improve with ε chosen below the maximum noise term XTξ /n. In the following corollary we show that with a small enough ε if the design matrix X is fixed and the noise vector ξ is sub Gaussian, we recover minimax-optimal rates for 2 error. Our error bound is minimax-optimal in the setting of sub-linear sparsity, meaning that there exists a constant γ > 1 such that kγ d. Corollary 2 (Minimax Rates in the Noisy Setting). Let the noise vector ξ be made of independent σ2-sub-Gaussian entries. Let ε = 4 σ2 log(2d)/ n. Under the assumptions of Theorem 1, the choice of η given by Theorem 2 and t = O(κeff(w ) log α 1) = O((w max n)/(σ log d) log α 1), Algorithm 1 yields wt w 2 2 (kσ2 log d)/n with probability at least 1 1/(8d3). The next corollary states that gradient descent automatically adapts to the difficulty of the problem. The statement of Theorem 1 suggests that our bounds undergo a phase-transition when w min XTξ /n which is also supported by our empirical findings in Section 5. In the σ2-sub-Gaussian noise setting the transition occurs as soon as n (σ2 log d)/(w min)2. As a result, the statistical bounds achieved by our algorithm are independent of d in such a setting. To see that, note that while the term XTξ /n grows as O(log d), the term XTξ 1S /n grows only as O(log k). In contrast, performance of the lasso deteriorates with d regardless of the difficulty of the problem. We illustrate this graphically and give a theoretical explanation in Section 5. We remark that the following result does not contradict minimax optimality because we now treat the true parameter w as fixed. Corollary 3 (Instance Adaptivity). Let the noise vector ξ be made of independent σ2-sub-Gaussian entries. Let ε = 4 σ2 log(2k)/ n. Under the assumptions of Theorem 1, the choice of η given by Theorem 2 and t = O(κeff(w ) log α 1) = O((w max n)/(σ log k) log α 1), Algorithm 1 yields wt w 2 2 (kσ2 log k)/n. with probability at least 1 1/(8k3). The final theorem we present shows that the same statistical bounds achieved by Algorithm 1 are also attained by Algorithm 2. This algorithm is not only optimal in a statistical sense, but it is also optimal computationally up to poly-logarithmic factors. Theorem 3. Compute ˆz using Theorem 2. Under the setting of Theorem 1 there exists a large enough absolute constant τ so that Algorithm 2 parameterized with α, τ and ˆz satisfies the result of Theorem 1 and t = O(log κeff log α 1) iterations. Corollaries 1, 2 and 3 also hold for Algorithm 2 with stopping time equal to O(log κeff log α 1). We emphasize that both Theorem 1 and 3 use gradient-based updates to obtain a sequence of models with optimal statistical properties instead of optimizing the objective function L. In fact, if we let t for Algorithm 2 the iterates would explode. 4 Proof Sketch In this section we prove a simplified version of Theorem 1 under the assumption XTX/n = I. We further highlight the intricacies involved in the general setting and present the intuition behind the key ideas there. The gradient descent updates on ut and vt as given in (1) can be written as ut+1 = ut 1 (4η/n)XT(Xwt y) , vt+1 = vt 1 + (4η/n)XT(Xwt y) . (2) The updates can be succinctly represented as ut+1 = ut (1 r) and vt+1 = vt (1 + r), where by our choice of η, r 1. Thus, (1 r) (1 + r) 1 and we have ut vt u0 v0 = α21. Hence for any i, only one of |ut,i| and |vt,i| can be larger then the initialization size while the other is effectively equal to 0. Intuitively, ut,i is used if w i > 0, vt,i if w i < 0 and hence one of these terms can be merged into an error term bt,i as defined below. The details appear in Appendix B.4. To avoid getting lost in cumbersome notation, in this section we will assume that w 0 and w = u u. Theorem 4. Assume that w 0, 1 n XTX = I, and that there is no noise (ξ = 0). Parameterize w = u u with u0 = α1 for some 0 < α < w min. Letting η 1/(10w max) and t = O(log(w max/α2)/(ηw min)), Algorithm 1 yields wt w α2. Proof. As XTX/n = I, y = Xw , and vt = 0, the updates given in equation (2) reduce componentwise to updates on wt given by wt+1,i = wt,i (1 4η(wt,i w i ))2. For i such that w i = 0, wt,i is non-increasing and hence stays below α2. For i such that w i > α2, the update rule given above ensures that as long as wt,i < w i /2, wt,i increases at an exponential rate with base at least (1+2ηw i )2. As w0,i = α2, in O(log(w i /α2)/(ηw i )) steps, it holds that wt,i w i /2. Subsequently, the gap (w i wt,i) halves every O(1/ηw i ) steps; thus, in O(log(w max/α2)/(ηw min)) steps we have wt w α2. The exact details are an exercise in calculus, albeit a rather tedious one, and appear in Appendix B.1. The proof of Theorem 4 contains the key ideas of the proof of Theorem 1. However, the presence of noise (ξ = 0) and only having restricted isometry of XTX rather than isometry requires a subtle and involved analysis. We remark that we can prove tighter bounds in Theorem 4 than the ones in Theorem 1 because we are working in a simplified setting. Error Decompositions. We decompose wt into st := wt 1S and et := wt 1Sc so that wt = st + et. We define the following error sequences: bt = XTXet/n + XTξ/n, pt = XTX/n I (st w ), which allows us to write updates on st and et as st+1 = st (1 4η(st w + pt + bt))2, et+1 = et (1 4η(pt + bt))2. Error Sequence bt. Since our theorems require stopping before et exceeds α, the term XTXet/n can be controlled entirely by the initialization size. Hence bt XTξ/n and it represents an irreducible error arising due to the noise on the labels. For any i S at stopping time t we cannot expect the error on the ith coordinate |wt,i w i | to be smaller than |(XTξ)/n)i|. If we assume pt = 0 and ξ = 0 then in light of our simplified Theorem 4 we see that the terms in et grow exponentially with base at most (1 + 4η XTξ /n). We can fit all the terms in st such that |w i | XTξ /n which leads to minimax-optimal rates. Moreover, if w min XTξ /n then all the elements in st grow exponentially at a faster rate than all of the error terms. This corresponds to the easy setting where the resulting error depends only on XTξ/n 1S yielding dimension-independent error bounds. For more details see Appendix B.2. Error Sequence pt. Since st w is a k-sparse vector using the RIP we can upper-bound pt kδ st w . Note that for small t we have s0 α2 0 and hence, ignoring the logarithmic factor in the definition of δ in the worst case we have Cw max pt < w max for some absolute constant 0 < C < 1. If w max XTξ /n then the error terms can grow exponentially with base (1 + 4η Cw max) whereas the signal terms such that |w i | w max can shrink exponentially at rate (1 4η Cw max). On the other hand, in the light of Theorem 4 the signal elements converge exponentially fast to the true parameters w max and hence the error sequence pt should be exponentially decreasing. For small enough C and a careful choice of initialization size α we can ensure that elements of pt decrease before the error components in et get too large or the signal components in st get too small. For more details see Appendix A.2 and B.3. Tuning Learning Rates. The proof of Theorem 2 is given in Appendix D. If we choose 1/w max η 1/(10w max) in Theorem 4 then all coordinates converge in O(κ log(w max/α2)) iterations. The reason the factor κ appears is the need to ensure that the convergence of the component w i = w max is stable. However, this conservative setting of the learning rate unnecessarily slows down the convergence for components with w i w max. In Theorem 4, oracle knowledge of w would allow to set an individual step size for each coordinate i S equal to ηi = 1/(10w i ) yielding the total number of iterations equal to O(log(w max/α2)). In the setting where XTX/n = I this would not be possible even with the knowledge of w , since the error sequence pt can be initially too large which would result in explosion of the coordinates i with |w i | w max. Instead, we need to wait for pt to get small enough before we increase the step size for some of the coordinates as described in Algorithm 2. The analysis is considerably involved and the full proof can be found in Appendix E. We illustrate effects of increasing step sizes in Section 5. 5 Simulations Unless otherwise specified, the default simulation set up is as follows. We let w = γ1S for some constant γ. For each run the entries of X are sampled as i.i.d. Rademacher random variables and the noise vector ξ follows i.i.d. N(0, σ2) distribution. For 2 plots each simulation is repeated a total of 30 times and the median 2 error is depicted. The error bars in all the plots denote the 25th and 75th percentiles. Unless otherwise specified, the default values for simulation parameters are n = 500, d = 104, k = 25, α = 10 12, γ = 1, σ = 1 and for Algorithm 2 we set τ = 10. Effects of Initialization Size. As discussed in Section 4 each coordinate grows exponentially at a different rate. In Figure 1 we illustrate the necessity of small initialization for bringing out the exponential nature of coordinate paths allowing to effectively fit them one at a time. For more intuition, suppose that coordinates outside the true support grow at most as fast as (1 + ε)t while the coordinates on the true support grow at least as fast as (1 + 2ε)t. Since exponential function is very sensitive to its base, for large enough t we have (1 + ε)t (1 + 2ε)t. The role of the initialization size α is then finding a small enough α such that for large enough t we have α2(1 + ε)t 0 while α2(1 + 2ε)t is large enough to ensure convergence of the coordinates on the true support. 0 50 100 150 200 Number of iterations t log2 ||wt w ||2 Effect of initiliaztion size α 0 20 40 60 80 100 Number of iterations t Coordinates paths with α = 10 3 0 100 200 300 Number of iterations t Coordinates paths with α = 10 12 Figure 1: Effects of initialization size. We set k = 5, n = 100, η = 0.05, σ = 0.5 and run Algorithm 1. We remark that the X axes in the two figures on the right differ due to different choices of α. Exponential Convergence with Increasing Step Sizes. We illustrate the effects of Algorithm 2 on an ill-conditioned target with κ = 64. Algorithm 1 spends approximately twice the time to fit each coordinate that the previous one, which is expected, since the coordinate sizes decrease by half. On the other hand, as soon as we increase the corresponding step size, Algorithm 2 fits each coordinate at approximately the same number of iterations, resulting in O(log κ log α 1) total iterations. Figure 2 confirms this behavior in simulations. 0 10000 20000 30000 40000 Number of iterations t log2 ||wt w ||2 Constant step size 0 500 1000 1500 2000 2500 Number of iterations t log2 ||wt w ||2 2 Increasing step sizes Figure 2: Comparison of Algorithms 1 and 2. Let n = 250, k = 7 and the non-zero coordinates of w be {2i : i = 0, . . . , 6}. For both algorithms let η = 1/(20 26). The vertical lines in the figure on the right are equally spaced at τ log α 1 iterations. The scale of x axes differ by a factor of 16. The shaded region corresponds to 25th and 75th percentiles over 30 runs. Phase Transitions. As suggested by our main results, we present empirical evidence that when w min XTξ /n our algorithms undergo a phase transition with dimension-independent error bounds. We plot results for three different estimators. First we run Algorithm 2 for 2000 iterations and save every 10th model. Among the 200 obtained models we choose the one with the smallest error on a validation dataset of size n/4. We run the lasso for 200 choices of λ equally spaced on a logarithmic scale and for each run we select a model with the smallest 2 parameter estimation error using an oracle knowledge of w . Finally, we perform a least squares fit using an oracle knowledge of the true support S. Figure 3 illustrates, that with varying γ, σ and n we can satisfy the condition w min XTξ /n at which point our method approaches an oracle-like performance. Given exponential nature of the coordinate-wise convergence, all coordinates of the true support grow at a strictly larger exponential rate than all of the coordinates on Sc as soon as w min XTξ /n > XTξ /n. An approximate solution of this equation is shown in Figure 3 using vertical red lines. 0.0 0.2 0.4 0.6 0.8 1.0 γ || w w ||2 2 Comparing 2 errors gradient descent lasso least squares oracle 0.0 2.5 5.0 7.5 10.0 12.5 σ Comparing 2 errors gradient descent lasso least squares oracle 8 9 10 11 12 log2 n log2 || w w ||2 Comparing 2 errors gradient descent lasso least squares oracle Figure 3: Phase transitions. The figure on the right uses γ = 1/4. The red vertical lines show solutions of the equation w min = γ = 2 E XTξ /n 2 σ 2 log(2d)/ n. Dimension Free Bounds in the Easy Setting. Figure 4 shows that when w min XTξ /n our algorithm matches the performance of oracle least squares which is independent of d. In contrast, the performance of the lasso deteriorates as d increases. To see why this is the case, in the setting where XTX/n = I, the lasso solution with parameter λ has a closed form solution wλ i = sign(w LS i )(|w LS i | λ)+, where w LS is the least squares solution. In the sub-Gaussian noise setting, the minimax rates are achieved by the choice λ = Θ( σ2 log(d)/n) introducing a bias which depends on log d. Such a bias is illustrated in Figure 4 and is not present at the optimal stopping time of our algorithm. 0 10 20 30 ||wt||1 Gradient descent 0 10 20 30 ||wλ||1 7 8 9 log d Comparing 2 errors gradient descent lasso least squares oracle Figure 4: Dimension dependent bias for the lasso. In contrast, in a high signal-to-noise ratio setting gradient descent is able to recover coordinates on S without a visible bias. 6 Summary and Further Improvements In this paper, we have provided the first statistical and computational guarantees for two algorithms based on implicit regularization applied to a sparse recovery problem under the RIP assumption in the general noisy setting. We show that Algorithms 1 and 2 yield optimal statistical rates and, in contrast to 1-penalization, adapt to the problem difficulty. In particular, given enough data both algorithms yield dimension-independent rates. While our algorithms are parametrized by step-size, initializationsize and the number of iterations, we show that a suitable choice of step-size and initialization-size can be computed from the data at the cost of one gradient-descent iteration. Consequently, the only tuneable hyper-parameter in practice is the number of iterations, which can be done using cross-validation as we have done in our experiments. With the provided choice of hyper-parameters, we show that Algorithm 1 attains statistical optimality after O( n) gradient-descent iterations, yielding a computationally sub-optimal algorithm. To circumvent this issue we propose a novel increasing step-sizes scheme (Algorithm 2) under which only O(1) iterations are required, resulting in a computationally optimal algorithm up to poly-logarithmic factors. Our results can be improved in two different aspects. First, our constraints on the RIP parameter δ result in sub-optimal sample complexity with respect to the sparsity parameter k. Second, the RIP condition could potentially be replaced by the restricted eigenvalue (RE) condition which allows correlated designs. We expand on both of these points in Appendix I and provide empirical evidence suggesting that both inefficiencies are artifacts of our analysis and not inherent limitations of our algorithms. Acknowledgments Tomas Vaškeviˇcius is supported by the EPSRC and MRC through the Ox Wa SP CDT programme (EP/L016710/1). Varun Kanade and Patrick Rebeschini are supported in part by the Alan Turing Institute under the EPSRC grant EP/N510129/1. [1] Radoslaw Adamczak, Alexander E Litvak, Alain Pajor, and Nicole Tomczak-Jaegermann. Restricted isometry property of matrices with independent columns and neighborly polytopes by random sampling. Constructive Approximation, 34(1):61 88, 2011. [2] Alekh Agarwal, Sahand Negahban, and Martin J Wainwright. Fast global convergence rates of gradient methods for high-dimensional statistical recovery. In Advances in Neural Information Processing Systems, pages 37 45, 2010. [3] Alnur Ali, J Zico Kolter, and Ryan J Tibshirani. A continuous-time view of early stopping for least squares regression. ar Xiv preprint ar Xiv:1810.10082, 2018. [4] Francis Bach, Rodolphe Jenatton, Julien Mairal, and Guillaume Obozinski. Optimization with sparsity-inducing penalties. Foundations and Trends R in Machine Learning, 4(1):1 106, 2012. [5] Afonso S Bandeira, Edgar Dobriban, Dustin G Mixon, and William F Sawin. Certifying the restricted isometry property is hard. IEEE transactions on information theory, 59(6):3448 3450, 2013. [6] Richard Baraniuk, Mark Davenport, Ronald De Vore, and Michael Wakin. A simple proof of the restricted isometry property for random matrices. Constructive Approximation, 28(3):253 263, 2008. [7] Frank Bauer, Sergei Pereverzev, and Lorenzo Rosasco. On regularization algorithms in learning theory. Journal of complexity, 23(1):52 72, 2007. [8] Dimitris Bertsimas, Angela King, and Rahul Mazumder. Best subset selection via a modern optimization lens. The Annals of Statistics, 44(2):813 852, 2016. [9] Peter J Bickel, Ya acov Ritov, and Alexandre B Tsybakov. Simultaneous analysis of Lasso and Dantzig selector. The Annals of Statistics, 37(4):1705 1732, 2009. [10] Kristian Bredies and Dirk A. Lorenz. Linear convergence of iterative soft-thresholding. Journal of Fourier Analysis and Applications, 14(5):813 837, Dec 2008. ISSN 1531-5851. doi: 10.1007/s00041-008-9041-1. URL https://doi.org/10.1007/s00041-008-9041-1. [11] Peter Bühlmann and Sara Van De Geer. Statistics for high-dimensional data: methods, theory and applications. Springer Science & Business Media, 2011. [12] Peter Bühlmann and Bin Yu. Boosting with the 2 loss: Regression and classification. Journal of the American Statistical Association, 98(462):324 339, 2003. [13] Emmanuel Candes and Terence Tao. The Dantzig selector: Statistical estimation when p is much larger than n. The Annals of Statistics, 35(6):2313 2351, 2007. [14] Emmanuel J Candes and Terence Tao. Decoding by linear programming. IEEE Transactions on Information Theory, 51(12):4203 4215, 2005. [15] Scott Shaobing Chen, David L Donoho, and Michael A Saunders. Atomic decomposition by basis pursuit. SIAM Journal on Scientific Computing, 20(1):33, 1998. [16] Albert Cohen, Wolfgang Dahmen, and Ronald De Vore. Compressed sensing and best k-term approximation. Journal of the American mathematical society, 22(1):211 231, 2009. [17] David L Donoho and Xiaoming Huo. Uncertainty principles and ideal atomic decomposition. IEEE transactions on information theory, 47(7):2845 2862, 2001. [18] Bradley Efron, Trevor Hastie, Iain Johnstone, and Robert Tibshirani. Least angle regression. The Annals of Statistics, 32(2):407 499, 2004. [19] Arie Feuer and Arkadi Nemirovski. On sparse representation in pairs of bases. IEEE Transactions on Information Theory, 49(6):1579 1581, 2003. [20] Jerome Friedman and Bogdan E Popescu. Gradient directed regularization. Technical report, 2004. [21] Jerome Friedman, Trevor Hastie, and Robert Tibshirani. The elements of statistical learning, volume 1. Springer series in statistics New York, 2001. [22] Jerome Friedman, Trevor Hastie, and Rob Tibshirani. Regularization paths for generalized linear models via coordinate descent. Journal of Statistical Software, 33(1):1, 2010. [23] Olivier Guédon, Shahar Mendelson, Alain Pajor, and Nicole Tomczak-Jaegermann. Subspaces and orthogonal decompositions generated by bounded orthogonal systems. Positivity, 11(2): 269 283, 2007. [24] Olivier Guédon, Shahar Mendelson, Alain Pajor, and Nicole Tomczak-Jaegermann. Majorizing measures and proportional subsets of bounded orthonormal systems. Revista matemática iberoamericana, 24(3):1075 1095, 2008. [25] Suriya Gunasekar, Blake E Woodworth, Srinadh Bhojanapalli, Behnam Neyshabur, and Nati Srebro. Implicit regularization in matrix factorization. In Advances in Neural Information Processing Systems, pages 6151 6159, 2017. [26] Suriya Gunasekar, Jason Lee, Daniel Soudry, and Nathan Srebro. Characterizing implicit bias in terms of optimization geometry. In International Conference on Machine Learning, pages 1827 1836, 2018. [27] Suriya Gunasekar, Jason D Lee, Daniel Soudry, and Nati Srebro. Implicit bias of gradient descent on linear convolutional networks. In Advances in Neural Information Processing Systems, pages 9461 9471, 2018. [28] Elaine T Hale, Wotao Yin, and Yin Zhang. Fixed-point continuation for _1-minimization: Methodology and convergence. SIAM Journal on Optimization, 19(3):1107 1130, 2008. [29] Trevor Hastie, Robert Tibshirani, and Ryan J Tibshirani. Extended comparisons of best subset selection, forward stepwise selection, and the lasso. ar Xiv preprint ar Xiv:1707.08692, 2017. [30] Yuanzhi Li, Tengyu Ma, and Hongyang Zhang. Algorithmic regularization in over-parameterized matrix sensing and neural networks with quadratic activations. In Conference On Learning Theory, pages 2 47, 2018. [31] Nicolai Meinshausen and Bin Yu. Lasso-type recovery of sparse representations for highdimensional data. The Annals of Statistics, 37(1):246 270, 2009. [32] Shahar Mendelson, Alain Pajor, and Nicole Tomczak-Jaegermann. Uniform uncertainty principle for bernoulli and subgaussian ensembles. Constructive Approximation, 28(3):277 289, 2008. [33] Sahand N Negahban, Pradeep Ravikumar, Martin J Wainwright, and Bin Yu. A unified framework for high-dimensional analysis of M-estimators with decomposable regularizers. Statistical Science, 27(4):538 557, 2012. [34] Gergely Neu and Lorenzo Rosasco. Iterate averaging as regularization for stochastic gradient descent. Proceedings of Machine Learning Research vol, 75:1 21, 2018. [35] Neal Parikh and Stephen Boyd. Proximal algorithms. Foundations and Trends R in Optimization, 1(3):127 239, 2014. [36] Garvesh Raskutti, Martin J Wainwright, and Bin Yu. Restricted eigenvalue properties for correlated gaussian designs. Journal of Machine Learning Research, 11(Aug):2241 2259, 2010. [37] Garvesh Raskutti, Martin J Wainwright, and Bin Yu. Minimax rates of estimation for highdimensional linear regression over q-balls. IEEE transactions on information theory, 57(10): 6976 6994, 2011. [38] Garvesh Raskutti, Martin J Wainwright, and Bin Yu. Early stopping and non-parametric regression: an optimal data-dependent stopping rule. The Journal of Machine Learning Research, 15(1):335 366, 2014. [39] Justin Romberg. Compressive sensing by random convolution. SIAM Journal on Imaging Sciences, 2(4):1098 1128, 2009. [40] Saharon Rosset, Ji Zhu, and Trevor Hastie. Boosting as a regularized path to a maximum margin classifier. Journal of Machine Learning Research, 5(Aug):941 973, 2004. [41] Mark Rudelson and Roman Vershynin. On sparse reconstruction from fourier and gaussian measurements. Communications on Pure and Applied Mathematics: A Journal Issued by the Courant Institute of Mathematical Sciences, 61(8):1025 1045, 2008. [42] Mark Rudelson and Shuheng Zhou. Reconstruction from anisotropic random measurements. In Conference on Learning Theory, pages 10 1, 2012. [43] Daniel Soudry, Elad Hoffer, Mor Shpigel Nacson, Suriya Gunasekar, and Nathan Srebro. The implicit bias of gradient descent on separable data. The Journal of Machine Learning Research, 19(1):2822 2878, 2018. [44] Arun Suggala, Adarsh Prasad, and Pradeep K Ravikumar. Connecting optimization and regularization paths. In Advances in Neural Information Processing Systems, pages 10608 10619, 2018. [45] Shaozhe Tao, Daniel Boley, and Shuzhong Zhang. Local linear convergence of ISTA and FISTA on the LASSO problem. SIAM Journal on Optimization, 26(1):313 336, 2016. [46] Robert Tibshirani. Regression shrinkage and selection via the Lasso. Journal of the Royal Statistical Society: Series B (Methodological), 58(1):267 288, 1996. [47] Robert Tibshirani, Martin Wainwright, and Trevor Hastie. Statistical learning with sparsity: the lasso and generalizations. Chapman and Hall/CRC, 2015. [48] Sara van de Geer. The deterministic Lasso. Research Report, 140, 2007. [49] Sara A Van De Geer and Peter Bühlmann. On the conditions used to prove oracle results for the lasso. Electronic Journal of Statistics, 3:1360 1392, 2009. [50] Martin J Wainwright. High-dimensional statistics: A non-asymptotic viewpoint, volume 48. Cambridge University Press, 2019. [51] Yuting Wei, Fanny Yang, and Martin J Wainwright. Early stopping for kernel boosting algorithms: A general analysis with localized complexities. In Advances in Neural Information Processing Systems, pages 6065 6075, 2017. [52] Yuan Yao, Lorenzo Rosasco, and Andrea Caponnetto. On early stopping in gradient descent learning. Constructive Approximation, 26(2):289 315, 2007. [53] Chiyuan Zhang, Samy Bengio, Moritz Hardt, Benjamin Recht, and Oriol Vinyals. Understanding deep learning requires rethinking generalization. ar Xiv preprint ar Xiv:1611.03530, 2016. [54] Tong Zhang and Bin Yu. Boosting with early stopping: Convergence and consistency. The Annals of Statistics, 33(4):1538 1579, 2005. [55] Yuchen Zhang, Martin J Wainwright, and Michael I Jordan. Lower bounds on the performance of polynomial-time algorithms for sparse linear regression. In Conference on Learning Theory, pages 921 948, 2014. [56] Peng Zhao, Yun Yang, and Qiao-Chu He. Implicit regularization via hadamard product overparametrization in high-dimensional linear regression. ar Xiv preprint ar Xiv:1903.09367, 2019.