# efficient_dictionary_learning_with_gradient_descent__5be9aee4.pdf Efficient Dictionary Learning with Gradient Descent Dar Gilboa 1 2 Sam Buchanan 3 2 John Wright 3 2 Randomly initialized first-order optimization algorithms are the method of choice for solving many high-dimensional nonconvex problems in machine learning, yet general theoretical guarantees cannot rule out convergence to critical points of poor objective value. For some highly structured nonconvex problems however, the success of gradient descent can be understood by studying the geometry of the objective. We study one such problem complete orthogonal dictionary learning, and provide converge guarantees for randomly initialized gradient descent to the neighborhood of a global optimum. The resulting rates scale as low order polynomials in the dimension even though the objective possesses an exponential number of saddle points. This efficient convergence can be viewed as a consequence of negative curvature normal to the stable manifolds associated with saddle points, and we provide evidence that this feature is shared by other nonconvex problems of importance as well. 1. Introduction Many central problems in machine learning and signal processing are most naturally formulated as optimization problems. These problems are often both nonconvex and highdimensional. High dimensionality makes the evaluation of second-order information prohibitively expensive, and thus randomly initialized first-order methods are usually employed instead. This has prompted great interest in recent years in understanding the behavior of gradient descent on nonconvex objectives (Hardt et al., 2015; Ge et al., 2015; Hardt et al., 2016; Dauphin et al., 2014). General analysis of firstand second-order methods on such problems can provide guarantees for convergence to critical points but 1Department of Neuroscience, Columbia University 2Data Science Institute, Columbia University 3Department of Electrical Engineering, Columbia University. Correspondence to: Dar Gilboa . Proceedings of the 36 th International Conference on Machine Learning, Long Beach, California, PMLR 97, 2019. Copyright 2019 by the author(s). these may be highly suboptimal, since nonconvex optimization is in general an NP-hard probem (Bertsekas, 1999). Outside of a convex setting (Nesterov, 2013) one must assume additional structure in order to make statements about convergence to optimal or high quality solutions. It is a curious fact that for certain classes of problems such as ones that involve sparsification (Lee et al., 2013; Bronstein et al., 2005) or matrix/tensor recovery (Keshavan et al., 2010; Jain et al., 2013; Anandkumar et al., 2014) first-order methods can be used effectively. Even for some highly nonconvex problems where there is no ground truth available such as the training of neural networks first-order methods converge to high-quality solutions (Zhang et al., 2016). Dictionary learning is a problem of inferring a sparse representation of data that was originally developed in the neuroscience literature (Olshausen & Field, 1996), and has since seen a number of important applications including image denoising, compressive signal acquisition and signal classification (Elad & Aharon, 2006; Mairal et al., 2014). In this work we study a formulation of the dictionary learning problem that can be solved efficiently using randomly initialized gradient descent despite possessing a number of saddle points exponential in the dimension. A feature that appears to enable efficient optimization is the existence of sufficient negative curvature in the directions normal to the stable manifolds of all critical points that are not global minima 1. This property ensures that the regions of the space that feed into small gradient regions under gradient flow do not dominate the parameter space. Figure 1 illustrates the value of this property: negative curvature prevents measure from concentrating about the stable manifold. As a consequence randomly initialized gradient methods avoid the slow region of around the saddle point. The main results of this work is a convergence rate for randomly initialized gradient descent for complete orthogonal dictionary learning to the neighborhood of a global minimum of the objective. Our results are probabilistic since they rely on initialization in certain regions of the parameter space, yet they allow one to flexibly trade off between the maximal number of iterations in the bound and the probabil- 1As well as a lack of spurious local minimizers, and the existence of large gradients or strong convexity in the remaining parts of the space Efficient Dictionary Learning with Gradient Descent Figure 1. Negative curvature helps gradient descent. Red: slow region of small gradient around a saddle point. Green: stable manifold associated with the saddle point. Black: points that flow to the slow region. Left: global negative curvature normal to the stable manifold. Right: positive curvature normal to the stable manifold randomly initialized gradient descent is more likely to encounter the slow region. ity of the bound holding. The sample complexity required for the concentration results in the paper to hold with high probability is p O(n4) up to polylog(n) factors, where p is the number of samples and n is the dimensionality of the space. While our focus is on dictionary learning, it has been recently shown that for other important nonconvex problems such as phase retrieval (Chen et al., 2018) performance guarantees for randomly initialized gradient descent can be obtained as well. In fact, in Appendix C we show that negative curvature normal to the stable manifolds of saddle points (illustrated in Figure 1) is also a feature of the population objective of generalized phase retrieval, and can be used to obtain an efficient convergence rate. 2. Related Work Easy nonconvex problems. There are two basic impediments to solving nonconvex problems globally: (i) spurious local minimizers, and (ii) flat saddle points, which can cause methods to stagnate in the vicinity of critical points that are not minimizers. The latter difficulty has motivated the study of strict saddle functions (Sun et al., 2015b; Ge et al., 2015), which have the property that at every point in the domain of optimization, there is a large gradient, a direction of strict negative curvature, or the function is strongly convex. By leveraging this curvature information, it is possible to escape saddle points and obtain a local minimizer in polynomial time.2 Perhaps more surprisingly, many known 2This statement is nontrivial: finding a local minimum of a smooth function is NP-hard. strict saddle functions also have the property that every local minimizer is global; for these problems, this implies that efficient methods find global solutions. Examples of problems with this property include variants of sparse dictionary learning (Sun et al., 2017), phase retrieval (Sun et al., 2016), tensor decomposition (Ge et al., 2015), community detection (Bandeira et al., 2016) and phase synchronization (Boumal, 2016). Minimizing strict saddle functions. Strict saddle functions have the property that at every saddle point there is a direction of strict negative curvature. A natural approach to escape such saddle points is to use second order methods (e.g., trust region (Conn et al., 2000) or curvilinear search (Goldfarb, 1980)) that explicitly leverage curvature information. Alternatively, one can attempt to escape saddle points using first order information only. However, some care is needed: canonical first order methods such as gradient descent will not obtain minimizers if initialized at a saddle point (or at a point that flows to one) at any critical point, gradient descent simply stops. A natural remedy is to randomly perturb the iterate whenever needed. A line of recent works shows that noisy gradient methods of this form efficiently optimize strict saddle functions (Lee et al., 2016; Du et al., 2017; Jin et al., 2017). For example, (Jin et al., 2017) obtains rates on strict saddle functions that match the optimal rates for smooth convex programs up to a polylogarithmic dependence on dimension.3 Randomly initialized gradient descent? The aforementioned results are broad, and nearly optimal. Nevertheless, important questions about the behavior of first order methods for nonconvex optimization remain unanswered. For example: in every one of the aforemented benign nonconvex optimization problems, randomly initialized gradient descent rapidly obtains a minimizer. This may seem unsurprising: general considerations indicate that the stable manifolds associated with non-minimizing critical points have measure zero (Nicolaescu, 2011), this implies that a variety of small-stepping first order methods converge to minimizers in the large-time limit (Lee et al., 2017). However, it is not difficult to construct strict saddle problems that are not amenable to efficient optimization by randomly initialized gradient descent see (Du et al., 2017) for an example. This contrast between the excellent empirical performance of randomly initialized first order methods and worst case examples suggests that there are important geometric and/or topological properties of easy nonconvex problems that are not captured by the strict saddle hypothesis. Hence, the motivation of this paper is twofold: (i) to provide theoretical corroboration (in certain specific situa- 3This work also proves convergence to a second-order stationary point under more general smoothness assumptions. Efficient Dictionary Learning with Gradient Descent tions) for what is arguably the simplest, most natural, and most widely used first order method, and (ii) to contribute to the ongoing effort to identify conditions which make nonconvex problems amenable to efficient optimization. 3. Dictionary Learning over the Sphere Suppose we are given data matrix Y = y1, . . . yp Rn p. The dictionary learning problem asks us to find a concise representation of the data (Elad & Aharon, 2006), of the form Y AX, where X is a sparse matrix. In the complete, orthogonal dictionary learning problem, we restrict the matrix A to have orthonormal columns (A O(n)). This variation of dictionary learning is useful for finding concise representations of small datasets (e.g., patches from a single image, in MRI (Ravishankar & Bresler, 2011)). To analyze the behavior of dictionary learning algorithms theoretically, it useful to posit that Y = A0X0 for some true dictionary A0 O(n) and sparse coefficient matrix X0 Rn p, and ask whether a given algorithm recovers the pair (A0, X0).4 In this work, we further assume that the sparse matrix X0 is random, with entries i.i.d. Bernoulli Gaussian5. For simplicity, we will let A0 = I; our arguments extend directly to general A0 via the simple change of variables q 7 A 0q. (Spielman et al., 2012) showed that under mild conditions, the complete dictionary recovery problem can be reduced to the geometric problem of finding a sparse vector in a linear subspace (Qu et al., 2014). Notice that because A0 is orthogonal, row(Y ) = row(X0). Because X0 is a sparse random matrix, the rows of X0 are sparse vectors. Under mild conditions (Spielman et al., 2012), they are the sparsest vectors in the row space of Y , and hence can be recovered by solving the conceptual optimization problem min q Y 0 s.t. q Y = 0. This is not a well-structured optimization problem: the objective is discontinuous, and the constraint set is open. A natural remedy is to replace the ℓ0 norm with a smooth sparsity surrogate, and to break the scale ambiguity by constraining q to the sphere, giving min f DL(q) 1 k=1 hµ(q yk) s.t. q Sn 1. (1) Here, we choose hµ(t) = µ log(cosh(t/µ)) as a smooth sparsity surrogate. This choice is motivated by convenience of analysis and analogous performance guarantees should 4This problem exhibits a sign permutation symmetry: A0X0 = (A0Γ)(Γ X0) for any signed permutation matrix Γ. Hence, we only ask for recovery up to a signed permutation. 5[X0]ij = V ijΩij, with V ij N(0, 1), Ωij Bern(θ) independent. Figure 2. Left: The separable objective for n = 3. Note the similarity to the dictionary learning objective. Right: The objective for complete orthogonal dictionary learning (discussed in section 6) for n = 3. be obtainable for other choices. This objective was analyzed in (Sun et al., 2015a), which showed that (i) although this optimization problem is nonconvex, when the data are sufficiently large, with high probability every local optimizer is near a signed column of the true dictionary A0, (ii) every other critical point has a direction of strict negative curvature, and (iii) as a consequence, a second-order Riemannian trust region method efficiently recovers a column of A0.6 The Riemannian trust region method is of mostly theoretical interest: it solves complicated (albeit polynomial time) subproblems that involve the Hessian of f DL. In practice, simple iterative methods, including randomly initialized gradient descent are also observed to rapidly obtain high-quality solutions. In the sequel, we will give a geometric explanation for this phenomenon, and bound the rate of convergence of randomly initialized gradient descent to the neighborhood of a column of A0. Our analysis of f DL is probabilistic in nature: it argues that with high probability in the sparse matrix X0, randomly initialized gradient descent rapidly produces a minimizer. To isolate more clearly the key intuitions behind this analysis, we first analyze the simpler separable objective min f Sep(q) i=1 hµ(qi) s.t. q Sn 1. (2) Figure 2 plots both f Sep and f DL as functions over the sphere. Notice that many of the key geometric features in f DL are present in f Sep; indeed, f Sep can be seen as an ultrasparse version of f DL in which the columns of the true sparse matrix X0 are taken to have only one nonzero entry. A virtue of this model function is that its critical points and their stable manifolds have simple closed form expressions (see Lemma 1). 6Combining with a deflation strategy, one can then efficiently recover the entire dictionary A0. Efficient Dictionary Learning with Gradient Descent 4. Outline of Important Geometric Features Our problems of interest have the form min f(q) s.t. q Sn 1, where f : Rn R is a smooth function. We let f(q) and 2f(q) denote the Euclidean gradient and hessian (over Rn), and let grad [f] (q) and Hess [f] (q) denote their Riemannian counterparts (over Sn 1). The projection operator onto Sn 1 is denoted by PSn 1.We will obtain results for Riemannian gradient descent defined by the update q expq( η grad[f](q)) for some step size η > 0, where expq : Tq Sn 1 Sn 1 is the exponential map (Absil et al., 2009). The Riemannian gradient on the sphere is given by grad[f](q) = (I qq ) f(q). We let A denote the set of critical points of f over Sn 1 these are the points q s.t. grad [f] ( q) = 0. We let A denote the set of local minimizers, and A its complement. Both f Sep and f DL are Morse functions on Sn 1,7 we can assign an index α to every q A, which is the number of negative eigenvalues of Hess [f] ( q). Our goal is to understand when gradient descent efficiently converges to a local minimizer. In the small-step limit, gradient descent follows gradient flow lines γ : R M, which are solution curves of the ordinary differential equation γ(t) = grad [f] (γ(t)) To each critical point α A of index λ, there is an associated stable manifold of dimension dim(M) λ, which is roughly speaking, the set of points that flow to α under gradient flow: lim t γ(t) = α γ a gradient flow line s.t. γ(0) = q Our analysis uses the following convenient coordinate chart ϕ(w) = w, q 1 w 2 q(w) (3) where w B1(0). We also define two useful sets: C {q Sn 1|qn w } Cζ q Sn 1 qn w 1 + ζ . (4) 7Strictly speaking, f DL is Morse with high probability, due to results of (Sun et al., 2017). Figure 3. Negative curvature and efficient gradient descent. The union of the light blue, orange and yellow sets is the set C. In the light blue region, there is negative curvature normal to C, while in the orange region the gradient norm is large, as illustrated by the arrows. There is a single global minimizer in the yellow region. For the separable objective, the stable manifolds of the saddles and maximizers all lie on C (the black circles denote the critical points, which are either maximizers " ", saddles " ", or minimizers " "). The red dots denote Cζ with ζ = 0.2. The set C is simply the subset of Sn 1 where the n-th coordinate is largest in magnitude and positive. Since the problems considered here are symmetric with respect to a signed permutation of the coordinates we can consider a certain C and the results will hold for the other symmetric sections as well. We will show that at every point in C aside from a neighborhood of a global minimizer for the separable objective (or a solution to the dictionary problem that may only be a local minimizer), there is either a large gradient component in the direction of the minimizer or negative curvature in a direction normal to C. For the case of the separable objective, one can show that the stable manifolds of the saddles lie on this boundary, and hence this curvature is normal to the stable manifolds of the saddles and allows rapid progress away from small gradient regions and towards a global minimizer 8. These regions are depicted in Figure 3. In the sequel, we will make the above ideas precise for the two specific nonconvex optimization problems discussed in Section 3 and use this to obtain a convergence rate to a neighborhood of a global minimizer. Our analysis are specific to these problems. However, as we will describe in more detail later, they hinge on important geometric characteristics of these problems which make them amenable to 8The direction of this negative curvature is important here, and it is this feature that distinguishes these problems from other problems in the strict-saddle class where this direction may be arbitrary Efficient Dictionary Learning with Gradient Descent efficient optimization, which may obtain in much broader classes of problems. 5. Separable Function Convergence Rate In this section, we study the behavior of randomly initialized gradient descent on the separable function f Sep. We begin by characterizing the critical points: Lemma 1 (Critical points of f Sep). The critical points of the separable problem (2) are A = PSn 1[a] a { 1, 0, 1} n, a > 0 . (5) For every α A and corresponding a(α), for µ < c n log n the stable manifold of α takes the form PSn 1 [ a(α) + b ] | supp(a(α)) supp(b) = , b < 1 where c > 0 is a numerical constant. Proof. Please see Appendix A By inspecting the dimension of the stable manifolds, it is easy to verify that that there are 2n global minimizers at the 1-sparse vectors on the sphere bei, 2n maximizers at the least sparse vectors and an exponential number of saddle points of intermediate sparsity. This is because the dimension of W s(α) is simply the dimension of b in 6, and it follows directly from the stable manifold theorem that only minimizers will have a stable manifold of dimension n 1. The objective thus possesses no spurious local minimizers. When referring to critical points and stable manifolds from now on we refer only to those that are contained in C or on its boundary. It is evident from Lemma 1 that the critical points in A all lie on C and that S α A W s(α) = C , and there is a minimizer at its center given by q(0) = ben. 5.1. The effect of negative curvature on the gradient We now turn to making precise the notion that negative curvature normal to stable manifolds of saddle points enables gradient descent to rapidly exit small gradient regions. We do this by defining vector fields u(i)(q), i [n 1] such that each field is normal to a continuous piece of Cζ and points outwards relative to Cζ defined in 4. By showing that the Riemannian gradient projected in this direction is positive and proportional to ζ, we are then able to show that gradient descent acts to increase ζ(q(w)) = qn w 1 geometrically. This corresponds to the behavior illustrated in the light blue region in Figure 3. Lemma 2 (Separable objective gradient projection). For any w such that q(w) Cζ, i [n 1], we define a vector u(i) Tq(w)Sn 1 by 0 j / {i, n}, sign(wi) j = i, |wi| qn j = n. (7) If µ log 1 µ wi and µ < 1 16, then u(i) grad[f Sep](q(w)) c w ζ, where c > 0 is a numerical constant. Proof. Please see Appendix A. Since we will use this property of the gradient in Cζ to derive a convergence rate, we will be interested in bounding the probability that gradient descent initialized randomly with respect to a uniform measure on the sphere is initialized in Cζ. This will require bounding the volume of this set, which is done in the following lemma: Lemma 3 (Volume of Cζ). For Cζ defined as in (4) we have Vol(Cζ) Vol(Sn 1) 1 Proof. Please see Appendix D.3. 5.2. Convergence rate Using the results above, one can obtain the following convergence rate: Theorem 1 (Gradient descent convergence rate for separable function). For any 0 < ζ0 < 1, r > µ log 1 µ , Rie- mannian gradient descent with step size η < min c1 on the separable objective (2) with µ < c2 n log n, enters an L ball of radius r around a global minimizer in iterations with probability P 1 2 log(n)ζ0, where ci, C > 0 are numerical constants. Proof. Please see Appendix A. Efficient Dictionary Learning with Gradient Descent We have thus obtained a convergence rate for gradient descent that relies on the negative curvature around the stable manifolds of the saddles to rapidly move from these regions of the space towards the vicinity of a global minimizer. This is evinced by the logarithmic dependence of the rate on ζ. As was shown for orthogonal dictionary learning in (Sun et al., 2017), we also expect a linear convergence rate due to strong convexity in the neighborhood of a minimizer, but do not take this into account in the current analysis. 6. Dictionary Learning Convergence Rate The proofs in this section will be along the same lines as those of Section 5. While we will not describe the positions of the critical points explicitly, the similarity between this objective and the separable function motivates a similar argument. It will be shown that initialization in some Cζ will guarantee that Riemannian gradient descent makes uniform progress in function value until reaching the neighborhood of a global minimizer. We will first consider the population objective which corresponds to the infinite data limit f pop DL (q) E X0f DL(q) = Ex i.i.d.BG(θ) hµ(q x) . (8) and then bounding the finite sample size fluctuations of the relevant quantities. We begin with a lemma analogous to Lemma 2: Lemma 4 (Dictionary learning population gradient). For any w such that q(w) Cζ, r < |wi|, µ < c1r5/2 ζ the dictionary learning population objective 8 obeys u(i) grad[f pop DL ](q(w)) cθr3ζ where cθ depends only on θ, c1 is a positive numerical constant and u(i) is defined in 7. Proof. Please see Appendix B Using this result, we obtain the desired convergence rate for the population objective, presented in Lemma 11 in Appendix B. After accounting for finite sample size fluctuations in the gradient, one obtains a rate of convergence to the neighborhood of a solution (which is some signed basis vector due to our choice A0 = I) Theorem 2 (Gradient descent convergence rate for dictionary learning). For any 1 > ζ0 > 0, s > µ 4 2, Riemannian gradient descent with step size η < c5θs n log np on the dic- tionary learning objective 1 with µ < c6 ζ0 n5/4 , θ (0, 1 2), enters a ball of radius c3s from a target solution in s + n log 1 iterations with probability P 1 2 log(n)ζ0 Py c8p 6 where Py is given in Lemma 10 with y = c7θ(1 θ)ζ0 n3/2 and ci, Ci are positive constants. Additionally, Py exp c(θ,ζ0)p n3 + npolylog(n, 1 ζ0 , θ) + log n for some c(θ, ζ0) > 0. Proof. Please see Appendix B The two terms in the rate correspond to an initial geometric increase in the distance from the set containing the small gradient regions around saddle points, followed by convergence to the vicinity of a minimizer in a region where the gradient norm is large. The latter is based on results on the geometry of this objective provided in (Sun et al., 2017). 7. Discussion The above analysis suggests that second-order properties - namely negative curvature normal to the stable manifolds of saddle points - play an important role in the success of randomly initialized gradient descent in the solution of complete orthogonal dictionary learning. This was done by furnishing a convergence rate guarantee that holds when the random initialization is not in regions that feed into small gradient regions around saddle points, and bounding the probability of such an initialization. In Appendix C we provide an additional example of a nonconvex problem for which an efficient rate can be obtained based on an analysis that relies on negative curvature normal to stable manifolds of saddles - generalized phase retrieval. An interesting direction of further work is to more precisely characterize the class of functions that share this feature. The effect of curvature can be seen in the dependence of the maximal number of iterations T on the parameter ζ0. This parameter controlled the volume of regions where initialization would lead to slow progress and the failure probability of the bound 1 P was linear in ζ0, while T depended logarithmically on ζ0. This logarithmic dependence is due to a geometric increase in the distance from the stable manifolds of the saddles during gradient descent, which is a consequence of negative curvature. Note that the choice of ζ0 allows one to flexibly trade off between T and 1 P. By decreasing ζ0, the bound holds with higher probability, at the price of an increase in T. This is because the volume of acceptable initializations now contains regions of smaller minimal gradient norm. In a sense, the result is an extrapolation of works such as (Lee et al., 2017) that analyze the ζ0 = 0 case to finite ζ0. Our analysis uses precise knowledge of the location of the stable manifolds of saddle points. For less symmetric Efficient Dictionary Learning with Gradient Descent problems, including variants of sparse blind deconvolution (Zhang et al., 2017) and overcomplete tensor decomposition, there is no closed form expression for the stable manifolds. However, it is still possible to coarsely localize them in regions containing negative curvature. Understanding the implications of this geometric structure for randomly initialized first-order methods is an important direction for future work. One may also hope that studying simple model problems and identifying structures (here, negative curvature orthogonal to the stable manifold) that enable efficient optimization will inspire approaches to broader classes of problems. Acknowledgements The authors gratefully acknowledge support from NSF 1740833, 1733857, 1527809, and an NDSEG fellowship. Absil, P.-A., Mahony, R., and Sepulchre, R. Optimization algorithms on matrix manifolds. Princeton University Press, 2009. Anandkumar, A., Ge, R., and Janzamin, M. Guaranteed nonorthogonal tensor decomposition via alternating rank-1 updates. ar Xiv preprint ar Xiv:1402.5180, 2014. Balan, R., Casazza, P., and Edidin, D. On signal reconstruction without phase. Applied and Computational Harmonic Analysis, 20(3):345 356, 2006. Bandeira, A. S., Boumal, N., and Voroninski, V. On the low-rank approach for semidefinite programs arising in synchronization and community detection. In Conference on Learning Theory, pp. 361 382, 2016. Bertsekas, D. P. Nonlinear programming. Athena scientific Belmont, 1999. Boumal, N. Nonconvex phase synchronization. SIAM Journal on Optimization, 26(4):2355 2377, 2016. Bronstein, M. M., Bronstein, A. M., Zibulevsky, M., and Zeevi, Y. Y. Blind deconvolution of images using optimal sparse representations. IEEE Transactions on Image Processing, 14(6):726 736, 2005. Candes, E. J., Li, X., and Soltanolkotabi, M. Phase retrieval via wirtinger flow: Theory and algorithms. IEEE Transactions on Information Theory, 61(4):1985 2007, 2015. Chen, Y., Chi, Y., Fan, J., and Ma, C. Gradient descent with random initialization: Fast global convergence for nonconvex phase retrieval. ar Xiv preprint ar Xiv:1803.07726, 2018. Conn, A. R., Gould, N. I., and Toint, P. L. Trust region methods, volume 1. Siam, 2000. Corbett, J. V. The pauli problem, state reconstruction and quantum-real numbers. Reports on Mathematical Physics, 57:53 68, 2006. Dauphin, Y. N., Pascanu, R., Gulcehre, C., Cho, K., Ganguli, S., and Bengio, Y. Identifying and attacking the saddle point problem in high-dimensional non-convex optimization. In Advances in neural information processing systems, pp. 2933 2941, 2014. Du, S. S., Jin, C., Lee, J. D., Jordan, M. I., Poczos, B., and Singh, A. Gradient descent can take exponential time to escape saddle points. ar Xiv preprint ar Xiv:1705.10412, 2017. Elad, M. and Aharon, M. Image denoising via sparse and redundant representations over learned dictionaries. IEEE Transactions on Image processing, 15(12):3736 3745, 2006. Ge, R., Huang, F., Jin, C., and Yuan, Y. Escaping from saddle points?online stochastic gradient for tensor decomposition. In Conference on Learning Theory, pp. 797 842, 2015. Goldfarb, D. Curvilinear path steplength algorithms for minimization which use directions of negative curvature. Mathematical programming, 18(1):31 40, 1980. Hardt, M., Recht, B., and Singer, Y. Train faster, generalize better: Stability of stochastic gradient descent. ar Xiv preprint ar Xiv:1509.01240, 2015. Hardt, M., Ma, T., and Recht, B. Gradient descent learns linear dynamical systems. ar Xiv preprint ar Xiv:1609.05191, 2016. Jain, P., Netrapalli, P., and Sanghavi, S. Low-rank matrix completion using alternating minimization. In Proceedings of the forty-fifth annual ACM symposium on Theory of computing, pp. 665 674. ACM, 2013. Jin, C., Ge, R., Netrapalli, P., Kakade, S. M., and Jordan, M. I. How to escape saddle points efficiently. ar Xiv preprint ar Xiv:1703.00887, 2017. Keshavan, R. H., Montanari, A., and Oh, S. Matrix completion from a few entries. IEEE Transactions on Information Theory, 56(6):2980 2998, 2010. Kreutz-Delgado, K. The complex gradient operator and the cr-calculus. ar Xiv preprint ar Xiv:0906.4835, 2009. Lee, J. D., Simchowitz, M., Jordan, M. I., and Recht, B. Gradient descent only converges to minimizers. In Conference on Learning Theory, pp. 1246 1257, 2016. Efficient Dictionary Learning with Gradient Descent Lee, J. D., Panageas, I., Piliouras, G., Simchowitz, M., Jordan, M. I., and Recht, B. First-order methods almost always avoid saddle points. ar Xiv preprint ar Xiv:1710.07406, 2017. Lee, K., Wu, Y., and Bresler, Y. Near optimal compressed sensing of sparse rank-one matrices via sparse power factorization. ar Xiv preprint, 2013. Mairal, J., Bach, F., Ponce, J., et al. Sparse modeling for image and vision processing. Foundations and Trends R in Computer Graphics and Vision, 8(2-3):85 283, 2014. Miao, J., Ishikawa, T., Johnson, B., Anderson, E. H., Lai, B., and Hodgson, K. O. High resolution 3d x-ray diffraction microscopy. Physical review letters, 89(8):088303, 2002. Nesterov, Y. Introductory lectures on convex optimization: A basic course, volume 87. Springer Science & Business Media, 2013. Nicolaescu, L. An invitation to Morse theory. Springer Science & Business Media, 2011. Olshausen, B. A. and Field, D. J. Emergence of simple-cell receptive field properties by learning a sparse code for natural images. Nature, 381(6583):607, 1996. Qu, Q., Sun, J., and Wright, J. Finding a sparse vector in a subspace: Linear sparsity using alternating directions. In Advances in Neural Information Processing Systems, pp. 3401 3409, 2014. Ravishankar, S. and Bresler, Y. Mr image reconstruction from highly undersampled k-space data by dictionary learning. IEEE transactions on medical imaging, 30(5): 1028 1041, 2011. Shechtman, Y., Eldar, Y. C., Cohen, O., Chapman, H. N., Miao, J., and Segev, M. Phase retrieval with application to optical imaging: a contemporary overview. IEEE signal processing magazine, 32(3):87 109, 2015. Spielman, D. A., Wang, H., and Wright, J. Exact recovery of sparsely-used dictionaries. In Conference on Learning Theory, pp. 37 1, 2012. Sun, J., Qu, Q., and Wright, J. Complete dictionary recovery over the sphere. In Sampling Theory and Applications (Samp TA), 2015 International Conference on, pp. 407 410. IEEE, 2015a. Sun, J., Qu, Q., and Wright, J. When are nonconvex problems not scary? ar Xiv preprint ar Xiv:1510.06096, 2015b. Sun, J., Qu, Q., and Wright, J. A geometric analysis of phase retrieval. In Information Theory (ISIT), 2016 IEEE International Symposium on, pp. 2379 2383. IEEE, 2016. Sun, J., Qu, Q., and Wright, J. Complete dictionary recovery over the sphere i: Overview and the geometric picture. IEEE Transactions on Information Theory, 63(2):853 884, 2017. Zhang, C., Bengio, S., Hardt, M., Recht, B., and Vinyals, O. Understanding deep learning requires rethinking generalization. ar Xiv preprint ar Xiv:1611.03530, 2016. Zhang, Y., Lau, Y., Kuo, H.-w., Cheung, S., Pasupathy, A., and Wright, J. On the global geometry of sphereconstrained sparse blind deconvolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4894 4902, 2017.