# sharpness_restart_and_acceleration__90c0a0df.pdf Sharpness, Restart and Acceleration Vincent Roulet INRIA, ENS Paris France vincent.roulet@inria.fr Alexandre d Aspremont CNRS, ENS Paris France aspremon@ens.fr The Łojasiewicz inequality shows that sharpness bounds on the minimum of convex optimization problems hold almost generically. Sharpness directly controls the performance of restart schemes, as observed by Nemirovskii and Nesterov [1985]. The constants quantifying error bounds are of course unobservable, but we show that optimal restart strategies are robust, and searching for the best scheme only increases the complexity by a logarithmic factor compared to the optimal bound. Overall then, restart schemes generically accelerate accelerated methods. Introduction We study convex optimization problems of the form minimize f(x) (P) where f is a convex function defined on Rn. The complexity of these problems using first order methods is generically controlled by smoothness assumptions on f such as Lipschitz continuity of its gradient. Additional assumptions such as strong convexity or uniform convexity provide respectively linear [Nesterov, 2013b] and faster polynomial [Juditski and Nesterov, 2014] rates of convergence. However, these assumptions are often too restrictive to be applied. Here, we make a much weaker and generic assumption that describes the sharpness of the function around its minimizers by constants µ 0 and r 1 such that µ r d(x, X )r f(x) f , for every x K, (Sharp) where f is the minimum of f, K Rn is a compact set, d(x, X ) = miny X x y is the distance from x to the set X K of minimizers of f 1 for the Euclidean norm . This defines a lower bound on the function around its minimizers: for r = 1, f shows a kink around its minimizers and the larger is r the flatter is the function around its minimizers. We tackle this property by restart schemes of classical convex optimization algorithms. Sharpness assumption (Sharp) is better known as a Hölderian error bound on the distance to the set of minimizers. Hoffman [Hoffman, 1952] first introduced error bounds to study system of linear inequalities. Natural extensions were then developed for convex optimization [Robinson, 1975; Mangasarian, 1985; Auslender and Crouzeix, 1988], notably through the concept of sharp minima [Polyak, 1979; Burke and Ferris, 1993; Burke and Deng, 2002]. But the most striking discovery was made by Łojasiewicz [Łojasiewicz, 1963, 1993] who proved inequality (Sharp) for real analytic and subanalytic functions. It has then been extended to non-smooth subanalytic convex functions by Bolte et al. [2007]. Overall, since (Sharp) essentially measures the sharpness of minimizers, it holds somewhat generically. On the other hand, this inequality is purely descriptive as we have no hope of ever observing either r or µ, and deriving adaptive schemes is crucial to ensure practical relevance. 1We assume the problem feasible, i.e. X = . 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA. Łojasiewicz inequalities either in the form of (Sharp) or as gradient dominated properties [Polyak, 1979] led to new simple convergence results [Karimi et al., 2016], in particular for alternating and splitting methods [Attouch et al., 2010; Frankel et al., 2015], even in the non-convex case [Bolte et al., 2014]. Here we focus on Hölderian error bounds as they offer simple explanation of accelerated rates of restart schemes. Restart schemes were already studied for strongly or uniformly convex functions [Nemirovskii and Nesterov, 1985; Nesterov, 2013a; Juditski and Nesterov, 2014; Lin and Xiao, 2014]. In particular, Nemirovskii and Nesterov [1985] link a strict minimum condition akin to (Sharp) with faster convergence rates using restart schemes which form the basis of our results, but do not study the cost of adaptation and do not tackle the non-smooth case. In a similar spirit, weaker versions of this strict minimum condition were used more recently to study the performance of restart schemes in [Renegar, 2014; Freund and Lu, 2015; Roulet et al., 2015]. The fundamental question of a restart scheme is naturally to know when must an algorithm be stopped and relaunched. Several heuristics [O Donoghue and Candes, 2015; Su et al., 2014; Giselsson and Boyd, 2014] studied adaptive restart schemes to speed up convergence of optimal methods. The robustness of restart schemes was then theoretically studied by Fercoq and Qu [2016] for quadratic error bounds, i.e. (Sharp) with r = 2, that LASSO problem satisfies for example. Fercoq and Qu [2017] extended recently their work to produce adaptive restarts with theoretical guarantees of optimal performance, still for quadratic error bounds. Previous references focus on smooth problems, but error bounds appear also for non-smooth ones, Gilpin et al. [2012] prove for example linear converge of restart schemes in bilinear matrix games where the minimum is sharp, i.e. (Sharp) with r = 1. Our contribution here is to derive optimal scheduled restart schemes for general convex optimization problems for smooth, non-smooth or Hölder smooth functions satisfying the sharpness assumption. We then show that for smooth functions these schemes can be made adaptive with nearly optimal complexity (up to a squared log term) for a wide array of sharpness assumptions. We also analyze restart criterion based on a sufficient decrease of the gap to the minimum value of the problem, when this latter is known in advance. In that case, restart schemes are shown ot be optimal without requiring any additional information on the function. 1 Problem assumptions 1.1 Smoothness Convex optimization problems (P) are generally divided in two classes: smooth problems, for which f has Lipschitz continuous gradients, and non-smooth problems for which f is not differentiable. Nesterov [2015] proposed to unify point of views by assuming generally that there exist constants 1 s 2 and L > 0 such that f(x) f(y) L x y s 1, for all x, y Rn (Smooth) where f(x) is any sub-gradient of f at x if s = 1 (otherwise this implies differentiability of f). For s = 2, we retrieve the classical definition of smoothness [Nesterov, 2013b]. For s = 1 we get a classical assumption made in non-smooth convex optimization, i.e., that sub-gradients of the function are bounded. For 1 < s < 2, this assumes gradient of f to be Hölder Lipschitz continuous. In a first step, we will analyze restart schemes for smooth convex optimization problems, then generalize to general smoothness assumption (Smooth) using appropriate accelerated algorithms developed by Nesterov [2015]. 1.2 Error bounds In general, an error bound is an inequality of the form d(x, X ) ω(f(x) f ), where ω is an increasing function at 0, called the residual function, and x may evolve either in the whole space or in a bounded set, see Bolte et al. [2015] for more details. We focus on Hölderian Error Bounds (Sharp) as they are the most common in practice. They are notably satisfied by a analytic and subanalytic functions but the proof (see e.g. Bierstone and Milman [1988, Theorem 6.4]) is shown using topological arguments that are far from constructive. Hence, outside of some particular cases (e.g. strong convexity), we cannot assume that the constants in (Sharp) are known, even approximately. Error bounds can generically be linked to Łojasiewicz inequality that upper bounds magnitude of the gradient by values of the function [Bolte et al., 2015]. Such property paved the way to many recent results in optimization [Attouch et al., 2010; Frankel et al., 2015; Bolte et al., 2014]. Here we will see that (Sharp) is sufficient to acceleration of convex optimization algorithms by their restart. Note finally that in most cases, error bounds are local properties hence the convergence results that follow will generally be local. 1.3 Sharpness and smoothness Let f be a convex function on Rn satisfying (Smooth) with parameters (s, L). This property ensures that, f(x) f + L s x y s, for given x Rn and y X . Setting y to be the projection of x onto X , this yields the following upper bound on suboptimality s d(x, X )s. (1) Now, assume that f satisfies the error bound (Sharp) on a set K with parameters (r, µ). Combining (1) and (Sharp) this leads for every x K, sµ r L d(x, X )s r. This means that necessarily s r by taking x X . Moreover if s < r, this last inequality can only be valid on a bounded set, i.e. either smoothness or error bound or both are valid only on a bounded set. In the following, we write κ L 2 s /µ 2 r and τ 1 s respectively a generalized condition number for the function f and a condition number based on the ratio of powers in inequalities (Smooth) and (Sharp). If r = s = 2, κ matches the classical condition number of the function. 2 Scheduled restarts for smooth convex problems In this section, we seek to solve (P) assuming that the function f is smooth, i.e. satisfies (Smooth) with s = 2 and L > 0. Without further assumptions on f, an optimal algorithm to solve the smooth convex optimization problem (P) is Nesterov s accelerated gradient method [Nesterov, 1983]. Given an initial point x0, this algorithm outputs, after t iterations, a point x = A(x0, t) such that t2 d(x0, X )2, (3) where c > 0 denotes a universal constant (whose value will be allowed to vary in what follows, with c = 4 here). We assume without loss of generality that f(x) f(x0). More details about Nesterov s algorithm are given in Supplementary Material. In what follows, we will also assume that f satisfies (Sharp) with parameters (r, µ) on a set K X , which means µ r d(x, X )r f(x) f , for every x K. (Sharp) As mentioned before if r > s = 2, this property is necessarily local, i.e. K is bounded. We assume then that given a starting point x0 Rn, sharpness is satisfied on the sublevel set {x| f(x) f(x0)}. Remark that if this property is valid on an open set K X , it will also be valid on any compact set K K with the same exponent r but a potentially lower constant µ. The scheduled restart schemes we present here rely on a global sharpness hypothesis on the sublevel set defined by the initial point and are not adaptive to constant µ on smaller sublevel sets. On the other hand, restarts on criterion that we present in Section 4, assuming that f is known, adapt to the value of µ. We now describe a restart scheme exploiting this extra regularity assumption to improve the computational complexity of solving problem (P) using accelerated methods. 2.1 Scheduled restarts Here, we schedule the number of iterations tk made by Nesterov s algorithm between restarts, with tk the number of (inner) iterations at the kth algorithm run (outer iteration). Our scheme is described in Algorithm 1 below. Algorithm 1 Scheduled restarts for smooth convex minimization Inputs : x0 Rn and a sequence tk for k = 1, . . . , R. for k = 1, . . . , R do xk := A(xk 1, tk) end for Output : ˆx := x R The analysis of this scheme and the following ones relies on two steps. We first choose schedules that ensure linear convergence in the iterates xk at a given rate. We then adjust this linear rate to minimize the complexity in terms of the total number of iterations. We begin with a technical lemma which assumes linear convergence holds, and connects the growth of tk, the precision reached and the total number of inner iterations N. Lemma 2.1. Let xk be a sequence whose kth iterate is generated from the previous one by an algorithm that runs tk iterations and write N = PR k=1 tk the total number of iterations to output a point x R. Suppose setting tk = Ceαk, k = 1, . . . , R for some C > 0 and α 0 ensures that outer iterations satisfy f(xk) f νe γk, (4) for all k 0 with ν 0 and γ 0. Then precision at the output is given by, f(x R) f ν exp( γN/C), when α = 0, and f(x R) f ν (αe αC 1N + 1) γ α , when α > 0. Proof. When α = 0, N = RC, and inserting this in (4) at the last point x R yields the desired result. On the other hand, when α > 0, we have N = PR k=1 tk = Ceα eαR 1 eα 1 , which gives R = log eα 1 eαC N + 1 /α. Inserting this in (4) at the last point, we get f(x R) f ν exp γ eαC N + 1 ν (αe αC 1N+1) γ α , where we used ex 1 x. This yields the second part of the result. The last approximation in the case α > 0 simplifies the analysis that follows without significantly affecting the bounds. We also show in Supplementary Material that using tk = tk does not significantly affect the bounds above. Remark that convergence bounds are generally linear or polynomial such that we can extract a subsequence that converges linearly. Therefore our approach does not restrict the analysis of our scheme. It simplifies it and can be used for other algorithms like the gradient descent as detailed in Supplementary Material. We now analyze restart schedules tk that ensure linear convergence. Our choice of tk will heavily depend on the ratio between r and s (with s = 2 for smooth functions here), incorporated in the parameter τ = 1 s/r defined in (2). Below, we show that if τ = 0, a constant schedule is sufficient to ensure linear convergence. When τ > 0, we need a geometrically increasing number of iterations for each cycle. Proposition 2.2. Let f be a smooth convex function satisfying (Smooth) with parameters (2, L) and (Sharp) with parameters (r, µ) on a set K. Assume that we are given x0 Rn such that {x| f(x) f(x0)} K. Run Algorithm 1 from x0 with iteration schedule tk = C κ,τeτk, for k = 1, . . . , R, where C κ,τ e1 τ(cκ) 1 2 (f(x0) f ) τ with κ and τ defined in (2) and c = 4e2/e here. The precision reached at the last point ˆx is given by, f(ˆx) f exp 2e 1(cκ) 1 2 N (f(x0) f ) = O exp( κ 1 2 N) , when τ = 0, (6) f(ˆx) f f(x0) f τe 1(f(x0) f ) τ 2 (cκ) 1 τ , when τ > 0, (7) where N = PR k=1 tk is the total number of iterations. Proof. Our strategy is to choose tk such that the objective is linearly decreasing, i.e. f (xk) f e γk(f(x0) f ), (8) for some γ 0 depending on the choice of tk. This directly holds for k = 0 and any γ 0. Combining (Sharp) with the complexity bound in (3), we get f (xk) f cκ t2 k (f (xk 1) f ) 2 r , where c = 4e2/e using that r2/r e2/e. Assuming recursively that (8) is satisfied at iteration k 1 for a given γ, we have f (xk) f cκe γ 2 t2 k (f(x0) f ) 2 r , and to ensure (8) at iteration k, we impose t2 k (f(x0) f ) 2 r e γk(f(x0) f ). Rearranging terms in this last inequality, using τ defined in (2), we get 2 (cκ) 1 2 (f(x0) f ) τ For a given γ 0, we can set tk = Ceαk where 2 (cκ) 1 2 (f(x0) f ) τ 2 and α = τγ/2, (10) and Lemma 2.1 then yields, f(ˆx) f exp γe γ 2 N (f(x0) f ), when τ = 0, while f(ˆx) f (f(x0) f ) τ 2 γe γ 2 (f(x0) f ) τ 2 N+1 2 when τ > 0. These bounds are minimal for γ = 2, which yields the desired result. When τ = 0, bound (6) matches the classical complexity bound for smooth strongly convex functions [Nesterov, 2013b]. When τ > 0 on the other hand, bound (7) highlights a much faster convergence rate than accelerated gradient methods. The sharper the function (i.e. the smaller r), the faster the convergence. This matches the lower bounds for optimizing smooth and sharp functions functions [Arjevani and Shamir, 2016; Nemirovskii and Nesterov, 1985, Page 6] up to constant factors. Also, setting tk = C κ,τeτk yields continuous bounds on precision, i.e. when τ 0, bound (7) converges to bound (6), which also shows that for τ near zero, constant restart schemes are almost optimal. 2.2 Adaptive scheduled restart The previous restart schedules depend on the sharpness parameters (r, µ) in (Sharp). In general of course, these values are neither observed nor known a priori. Making our restart scheme adaptive is thus crucial to its practical performance. Fortunately, we show below that a simple logarithmic grid search strategy on these parameters is enough to guarantee nearly optimal performance. We run several schemes with a fixed number of inner iterations N to perform a log-scale grid search on τ and κ. We define these schemes as follows. Si,0 : Algorithm 1 with tk = Ci, Si,j : Algorithm 1 with tk = Cieτjk, (11) where Ci = 2i and τj = 2 j. We stop these schemes when the total number of inner algorithm iterations has exceed N, i.e. at the smallest R such that PR k=1 tk N. The size of the grid search in Ci is naturally bounded as we cannot restart the algorithm after more than N total inner iterations, so i [1, . . . , log2 N ]. We will also show that when τ is smaller than 1/N, a constant schedule performs as well as the optimal geometrically increasing schedule, which crucially means we can also choose j [1, . . . , log2 N ] and limits the cost of grid search. The following result details the convergence of this method, its notations are the same as in Proposition 2.2 and its technical proof can be found in Supplementary Material. Proposition 2.3. Let f be a smooth convex function satisfying (Smooth) with parameters (2, L) and (Sharp) with parameters (r, µ) on a set K. Assume that we are given x0 Rn such that {x| f(x) f(x0)} K and denote N a given number of iterations. Run schemes Si,j defined in (11) to solve (P) for i [1, . . . , log2 N ] and j [0, . . . , log2 N ], stopping each time after N total inner algorithm iterations i.e. for R such that PR k=1 tk N. Assume N is large enough, so N 2C κ,τ, and if 1 N > τ > 0, C κ,τ > 1. If τ = 0, there exists i [1, . . . , log2 N ] such that scheme Si,0 achieves a precision given by f(ˆx) f exp e 1(cκ) 1 2 N (f(x0) f ). If τ > 0, there exist i [1, . . . , log2 N ] and j [1, . . . , log2 N ] such that scheme Si,j achieves a precision given by f(ˆx) f f(x0) f τe 1(cκ) 1 2 (f(x0) f ) τ 2 (N 1)/4+1 2 Overall, running the logarithmic grid search has a complexity (log2 N)2 times higher than running N iterations using the optimal (oracle) scheme. As showed in Supplementary Material, scheduled restart schemes are theoretically efficient only if the algorithm itself makes a sufficient number of iterations to decrease the objective value. Therefore we need N large enough to ensure the efficiency of the adaptive method. If τ = 0, we naturally have C κ,0 1, therefore if 1 N > τ > 0 and N is large, assuming C κ,τ C κ,0, we get C κ,τ 1. This adaptive bound is similar to the one of Nesterov [2013b] to optimize smooth strongly convex functions in the sense that we lose approximately a log factor of the condition number of the function. However our assumptions are weaker and we are able to tackle all regimes of the sharpness property, i.e. any exponent r [2, + ], not just the strongly convex case. In the supplementary material we also analyze the simple gradient descent method under the sharpness (Sharp) assumption. It shows that simple gradient descent achieves a O(ϵ τ) complexity for a given accuracy ϵ. Therefore restarting accelerated gradient methods reduces complexity to O(ϵ τ/2) compared to simple gradient descent. This result is similar to the acceleration of gradient descent. We extend now this restart scheme to solve non-smooth or Hölder smooth convex optimization problem under the sharpness assumption. 3 Universal scheduled restarts for convex problems In this section, we use the framework introduced by Nesterov [2015] to describe smoothness of a convex function f, namely, we assume that there exist s [1, 2] and L > 0 on a set J Rn, i.e. f(x) f(y) L x y s 1, for every x, y J. Without further assumptions on f, the optimal rate of convergence for this class of functions is bounded as O(1/N ρ), where N is the total number of iterations and ρ = 3s/2 1, (12) which gives ρ = 2 for smooth functions and ρ = 1/2 for non-smooth functions. The universal fast gradient method [Nesterov, 2015] achieves this rate by requiring only a target accuracy ϵ and a starting point x0. It outputs after t iterations a point x U(x0, ϵ, t), such that 2 + c L 2 s d(x0, X )2 s ϵ 2, (13) where c is a constant (c = 2 4s 2 s ). More details about the universal fast gradient method are given in Supplementary Material. We will again assume that f is sharp with parameters (r, µ) on a set K X , i.e. µ r d(x, X )r f(x) f , for every x K. (Sharp) As mentioned in Section 1.2, if r > s, smoothness or sharpness are local properties, i.e. either J or K or both are bounded, our analysis is therefore local. In the following we assume for simplicity, given an initial point x0, that smoothness and sharpness are satisfied simultaneously on the sublevel set {x| f(x) f(x0)}. The key difference with the smooth case described in the previous section is that here we schedule both the target accuracy ϵk used by the algorithm and the number of iterations tk made at the kth run of the algorithm. Our scheme is described in Algorithm 2. Algorithm 2 Universal scheduled restarts for convex minimization Inputs : x0 Rn, ϵ0 f(x0) f , γ 0 and a sequence tk for k = 1, . . . , R. for k = 1, . . . , R do ϵk := e γϵk 1, xk := U(xk 1, ϵk, tk) end for Output : ˆx := x R Our strategy is to choose a sequence tk that ensures f(xk) f ϵk, for the geometrically decreasing sequence ϵk. The overall complexity of our method will then depend on the growth of tk as described in Lemma 2.1. The proof is similar to the smooth case and can be found in Supplementary Material. Proposition 3.1. Let f be a convex function satisfying (Smooth) with parameters (s, L) on a set J and (Sharp) with parameters (r, µ) on a set K. Given x0 Rn assume that {x|f(x) f(x0)} J K. Run Algorithm 2 from x0 for a given ϵ0 f(x0) f with γ = ρ, tk = C κ,τ,ρeτk, where C κ,τ,ρ e1 τ(cκ) s 2ρ ϵ τ ρ 0 where ρ is defined in (12), κ and τ are defined in (2) and c = 8e2/e here. The precision reached at the last point ˆx is given by, f(ˆx) f exp ρe 1(cκ) s 2ρ N ϵ0 = O exp( κ s 2ρ N) , when τ = 0, f(ˆx) f ϵ0 τe 1(cκ) s τ ρ 0 N + 1 ρ τ = O κ s 2τ N ρ τ , when τ > 0, where N = PR k=1 tk is total number of iterations. This bound matches the lower bounds for optimizing smooth and sharp functions [Nemirovskii and Nesterov, 1985, Page 6] up to constant factors. Notice that, compared to Nemirovskii and Nesterov [1985], we can tackle non-smooth convex optimization by using the universal fast gradient algorithm of Nesterov [2015]. The rate of convergence in Proposition 3.1 is controlled by the ratio between τ and ρ. If these are unknown, a log-scale grid search won t be able to reach the optimal rate, even if ρ is known since we will miss the optimal rate by a constant factor. If both are known, in the case of non-smooth strongly convex functions for example, a grid-search on C recovers nearly the optimal bound. Now we will see that if f is known, restart produces adaptive optimal rates. 4 Restart with termination criterion Here, we assume that we know the optimum f of (P), or have an exact termination criterion. This is the case for example in zero-sum matrix games problems or non-degenerate least-squares without regularization. We assume again that f satisfies (Smooth) with parameters (s, L) on a set J and (Sharp) with parameters (r, µ) on a set K. Given an initial point x0 we assume that smoothness and sharpness are satisfied simultaneously on the sublevel set {x| f(x) f(x0)}. We use again the universal gradient method U. Here however, we can stop the algorithm when it reaches the target accuracy as we know the optimum f , i.e. we stop after tϵ inner iterations such that x = U(x0, ϵ, tϵ) satisfies f(x) f ϵ, and write x C(x0, ϵ) the output of this method. Here we simply restart this method and decrease the target accuracy by a constant factor after each restart. Our scheme is described in Algorithm 3. Algorithm 3 Restart on criterion Inputs : x0 Rn, f , γ 0, ϵ0 = f(x0) f for k = 1, . . . , R do ϵk := e γϵk 1, xk := C(xk 1, ϵk) end for Output : ˆx := x R The following result describes the convergence of this method. It relies on the idea that it cannot do more iterations than the best scheduled restart to achieve the target accuracy at each iteration. Its proof can be found in Supplementary Material. Proposition 4.1. Let f be a convex function satisfying (Smooth) with parameters (s, L) on a set J and (Sharp) with parameters (r, µ) on a set K. Given x0 Rn assume that {x, f(x) f(x0)} J K. Run Algorithm 3 from x0 with parameter γ = ρ. The precision reached at the last point ˆx is given by, f(ˆx) f exp ρe 1(cκ) s 2ρ N (f(x0) f ) = O exp( κ s 2ρ N) , when τ = 0, f(ˆx) f f(x0) f τe 1(cκ) s 2ρ (f(x0) f ) τ ρ N + 1 ρ τ = O κ s 2τ N ρ τ , when τ > 0, where N is the total number of iterations, ρ is defined in (12), κ and τ are defined in (2) and c = 8e2/e here. Therefore if f is known, this method is adaptive, contrary to the general case in Proposition 3.1. It can even adapt to the local values of L or µ as we use a criterion instead of a preset schedule. Here, stopping using f(xk) f implicitly yields optimal choices of C and τ. A closer look at the proof shows that the dependency in γ of this restart scheme is a factor h(γ) = γe γ/ρ of the number of iterations. Taking γ = 1, leads then to a suboptimal constant factor of at most h(ρ)/h(1) e/2 1.3 for ρ [1/2, 2], so running this scheme with γ = 1 makes it parameter-free while getting nearly optimal bounds. 5 Numerical Results We illustrate our results by testing our adaptive restart methods, denoted Adap and Crit, introduced respectively in Sections 2.2 and 4 on several problems and compare them against simple gradient descent (Grad), accelerated gradient methods (Acc), and the restart heuristic enforcing monotonicity (Mono in [O Donoghue and Candes, 2015]). For Adap we plot the convergence of the best method found by grid search to compare with the restart heuristic. This implicitly assumes that the grid search is run in parallel with enough servers. For Crit we use the optimal f found by another solver. This gives an overview of its performance in order to potentially approximate it along the iterations in a future work as done with Polyak steps [Polyak, 1987]. All restart schemes were done using the accelerated gradient with backtracking line search detailed in the Supplementary Material, with large dots representing restart iterations. The results focused on unconstrained problems but our approach can directly be extended to composite problems by using the proximal variant of the gradient, accelerated gradient and universal fast gradient methods [Nesterov, 2015] as detailed in the Supplementary Material. This includes constrained optimization as a particular case by adding the indicator function of the constraint set to the objective (as in the SVM example below). In Figure 1, we solve classification problems with various losses on the UCI Sonar data set [Asuncion and Newman, 2007]. For least square loss on sonar data set, we observe much faster convergence of the restart schemes compared to the accelerated method. These results were already observed by O Donoghue and Candes [2015]. For logistic loss, we observe that restart does not provide much improvement. The backtracking line search on the Lipschitz constant may be sufficient to capture the geometry of the problem. For hinge loss, we regularized by a squared norm and optimize the dual, which means solving a quadratic problem with box constraints. We observe here that the scheduled restart scheme convergences much faster, while restart heuristics may be activated too late. We observe similar results for the LASSO problem. In general Crit ensures the theoretical accelerated rate but Adap exhibits more consistent behavior. This highlights the benefits of a sharpness assumption for these last two problems. Precisely quantifying sharpness from data/problem structure is a key open problem. 0 200 400 600 800 Iterations Grad Acc Mono Adap Crit 0 500 1000 Iterations Grad Acc Mono Adap Crit 0 500 1000 Iterations Grad Acc Mono Adap Crit 0 500 1000 Iterations Grad Acc Mono Adap Crit Figure 1: From left to right: least square loss, logistic loss, dual SVM problem and LASSO. We use adaptive restarts (Adap), gradient descent (Grad), accelerated gradient (Acc) and restart heuristic enforcing monotonicity (Mono). Large dots represent the restart iterations. Regularization parameters for dual SVM and LASSO were set to one. Acknowledgments The authors would like to acknowledge support from the chaire Économie des nouvelles données with the data science joint research initiative with the fonds AXA pour la recherche, a gift from Société Générale Cross Asset Quantitative Research and an AMX fellowship. The authors are affiliated to PSL Research University, Paris, France. Arjevani, Y. and Shamir, O. [2016], On the iteration complexity of oblivious first-order optimization algorithms, in International Conference on Machine Learning , pp. 908 916. Asuncion, A. and Newman, D. [2007], Uci machine learning repository . Attouch, H., Bolte, J., Redont, P. and Soubeyran, A. [2010], Proximal alternating minimization and projection methods for nonconvex problems: An approach based on the kurdyka-łojasiewicz inequality , Mathematics of Operations Research 35(2), 438 457. Auslender, A. and Crouzeix, J.-P. [1988], Global regularity theorems , Mathematics of Operations Research 13(2), 243 253. Bierstone, E. and Milman, P. D. [1988], Semianalytic and subanalytic sets , Publications Mathématiques de l IHÉS 67, 5 42. Bolte, J., Daniilidis, A. and Lewis, A. [2007], The łojasiewicz inequality for nonsmooth subanalytic functions with applications to subgradient dynamical systems , SIAM Journal on Optimization 17(4), 1205 1223. Bolte, J., Nguyen, T. P., Peypouquet, J. and Suter, B. W. [2015], From error bounds to the complexity of first-order descent methods for convex functions , Mathematical Programming pp. 1 37. Bolte, J., Sabach, S. and Teboulle, M. [2014], Proximal alternating linearized minimization for nonconvex and nonsmooth problems , Mathematical Programming 146(1-2), 459 494. Burke, J. and Deng, S. [2002], Weak sharp minima revisited part i: basic theory , Control and Cybernetics 31, 439 469. Burke, J. and Ferris, M. C. [1993], Weak sharp minima in mathematical programming , SIAM Journal on Control and Optimization 31(5), 1340 1359. Fercoq, O. and Qu, Z. [2016], Restarting accelerated gradient methods with a rough strong convexity estimate , ar Xiv preprint ar Xiv:1609.07358 . Fercoq, O. and Qu, Z. [2017], Adaptive restart of accelerated gradient methods under local quadratic growth condition , ar Xiv preprint ar Xiv:1709.02300 . Frankel, P., Garrigos, G. and Peypouquet, J. [2015], Splitting methods with variable metric for kurdyka łojasiewicz functions and general convergence rates , Journal of Optimization Theory and Applications 165(3), 874 900. Freund, R. M. and Lu, H. [2015], New computational guarantees for solving convex optimization problems with first order methods, via a function growth condition measure , ar Xiv preprint ar Xiv:1511.02974 . Gilpin, A., Pena, J. and Sandholm, T. [2012], First-order algorithm with O(log 1/ϵ) convergence for ϵ-equilibrium in two-person zero-sum games , Mathematical programming 133(1-2), 279 298. Giselsson, P. and Boyd, S. [2014], Monotonicity and restart in fast gradient methods, in 53rd IEEE Conference on Decision and Control , IEEE, pp. 5058 5063. Hoffman, A. J. [1952], On approximate solutions of systems of linear inequalities , Journal of Research of the National Bureau of Standards 49(4). Juditski, A. and Nesterov, Y. [2014], Primal-dual subgradient methods for minimizing uniformly convex functions , ar Xiv preprint ar Xiv:1401.1792 . Karimi, H., Nutini, J. and Schmidt, M. [2016], Linear convergence of gradient and proximal-gradient methods under the polyak-łojasiewicz condition, in Joint European Conference on Machine Learning and Knowledge Discovery in Databases , Springer, pp. 795 811. Lin, Q. and Xiao, L. [2014], An adaptive accelerated proximal gradient method and its homotopy continuation for sparse optimization., in ICML , pp. 73 81. Łojasiewicz, S. [1963], Une propriété topologique des sous-ensembles analytiques réels , Les équations aux dérivées partielles pp. 87 89. Łojasiewicz, S. [1993], Sur la géométrie semi-et sous-analytique , Annales de l institut Fourier 43(5), 1575 1595. Mangasarian, O. L. [1985], A condition number for differentiable convex inequalities , Mathematics of Operations Research 10(2), 175 179. Nemirovskii, A. and Nesterov, Y. [1985], Optimal methods of smooth convex minimization , USSR Computational Mathematics and Mathematical Physics 25(2), 21 30. Nesterov, Y. [1983], A method of solving a convex programming problem with convergence rate O(1/k2) , Soviet Mathematics Doklady 27(2), 372 376. Nesterov, Y. [2013a], Gradient methods for minimizing composite functions , Mathematical Programming 140(1), 125 161. Nesterov, Y. [2013b], Introductory lectures on convex optimization: A basic course, Vol. 87, Springer Science & Business Media. Nesterov, Y. [2015], Universal gradient methods for convex optimization problems , Mathematical Programming 152(1-2), 381 404. O Donoghue, B. and Candes, E. [2015], Adaptive restart for accelerated gradient schemes , Foundations of computational mathematics 15(3), 715 732. Polyak, B. [1979], Sharp minima institute of control sciences lecture notes, moscow, ussr, 1979, in IIASA workshop on generalized Lagrangians and their applications, IIASA, Laxenburg, Austria . Polyak, B. [1987], Introduction to optimization, Optimization Software. Renegar, J. [2014], Efficient first-order methods for linear programming and semidefinite programming , ar Xiv preprint ar Xiv:1409.5832 . Robinson, S. M. [1975], An application of error bounds for convex programming in a linear space , SIAM Journal on Control 13(2), 271 273. Roulet, V., Boumal, N. and d Aspremont, A. [2015], Renegar s condition number, shaprness and compressed sensing performance , ar Xiv preprint ar Xiv:1506.03295 . Su, W., Boyd, S. and Candes, E. [2014], A differential equation for modeling nesterov s accelerated gradient method: Theory and insights, in Advances in Neural Information Processing Systems , pp. 2510 2518.