# a_variational_perspective_on_highresolution_odes__54d44b41.pdf A Variational Perspective on High-Resolution ODEs Hoomaan Maskan Umeå University hoomaan.maskan@umu.se Konstantinos C. Zygalakis University of Edinburgh k.zygalakis@ed.ac.uk Alp Yurtsever Umeå University alp.yurtsever@umu.se We consider unconstrained minimization of smooth convex functions. We propose a novel variational perspective using forced Euler-Lagrange equation that allows for studying high-resolution ODEs. Through this, we obtain a faster convergence rate for gradient norm minimization using Nesterov s accelerated gradient method. Additionally, we show that Nesterov s method can be interpreted as a ratematching discretization of an appropriately chosen high-resolution ODE. Finally, using the results from the new variational perspective, we propose a stochastic method for noisy gradients. Several numerical experiments compare and illustrate our stochastic algorithm with state of the art methods. 1 Introduction Smooth convex minimization is a fundamental class in optimization with broad applications and a rich theoretical foundation that enables the development of powerful methods. In fact, the theory and methods developed for other problem classes, such as non-smooth convex or smooth non-convex optimization, often build upon the work done in smooth convex optimization. As a result, the study of smooth convex minimization has a significant impact on the broader field of optimization. Over the last two decades, first-order methods those relying solely on gradient information, in contrast to methods like Newton s that require Hessian information have seen a lot of interest both in theory and applications due to their efficiency and adaptability for large-scale data-driven applications. Among these, gradient descent stands as one of the simplest and oldest. When equipped with an appropriate step-size, gradient descent guarantees a suboptimality gap (objective residual) of order O(1/k) after k iterations. In his seminal work, Nesterov [1983] has shown that the gradient method can achieve faster rates by incorporating momentum deviations. Nesterov s accelerated gradient algorithm (NAG) ensures a convergence rate of O(1/k2) which is an order of magnitude faster than the gradient descent. Remarkably, this rate matches information-theoretical lower bounds for first-order oracle complexity, meaning that NAG is optimal and no other first-order method can guarantee an essentially faster convergence rate [Nesterov, 2003]. The original proof of Nesterov [1983], known as the estimate sequence technique, is a highly algebraic and complex procedure, difficult to interpret, and provides arguably limited insight into why momentum deviations help with the convergence rates [Hu and Lessard, 2017]. Therefore, many researchers have tried to provide a better understanding of momentum-based acceleration through different perspectives. For example, Su et al. [2016], Shi et al. [2021], and Sanz Serna and Zygalakis [2021] consider a continuous time perspective; Lessard et al. [2016] and Fazlyab et al. [2018] use integral quadratic constraints and control systems with non-linear feedbacks; Muehlebach and Jordan [2019, 2022, 2023] present a dynamical perspective; Attouch et al. [2020, 2021] utilize inertial dynamic involving both viscous damping and Hessian-driven damping; Zhu and Orecchia [2014] view acceleration as a linear coupling of gradient descent and mirror descent updates; and Ahn and Sra [2022] provide an understanding of the NAG algorithm through an approximation of the proximal point method. Our work is focused on the continuous-time perspective. 37th Conference on Neural Information Processing Systems (Neur IPS 2023). In [Su et al., 2016], the authors derive a second-order ordinary differential equation (ODE) that exhibits trajectories similar to those of the NAG algorithm in the limit of an infinitesimal step-size s 0. This result has inspired researchers to analyze various ODEs and discretization schemes to gain a better understanding of the acceleration phenomenon. Notably, Wibisono et al. [2016] demonstrate that the ODE presented in [Su et al., 2016] is a special case of a broader family of ODEs that extend beyond Euclidean space. They achieve this by minimizing the action on a Lagrangian that captures the properties of the problem template, an approach known as the variational perspective, and they discretize their general ODE using the rate-matching technique. In a related vein, Shi et al. [2021] proposed substituting the low-resolution ODEs (LR-ODEs) introduced in [Su et al., 2016] with high-resolution ODEs (HR-ODEs) which can capture trajectories of NAG more precisely. Our main contribution in this paper is a novel extension of the variational perspective for HRODEs. A direct combination of these two frameworks is challenging, as it remains unclear how the Lagrangian should be modified to recover HR-ODEs. To address this problem, we propose an alternative approach that preserves the Lagrangian but extends the variational perspective. More specifically, instead of relying on the conventional Euler-Lagrange equation, we leverage the forced Euler-Lagrange equation that incorporates external forces acting on the system. By representing the damped time derivative of the potential function gradients as an external force, our proposed variational perspective allows us to reconstruct various HR-ODEs through specific damping parameters. More details are provided in Section 2. Other contributions of our paper are as follows: In Section 3, we show that our proposed variational analysis yields a special representation of NAG leading to superior convergence rates than [Shi et al., 2019] in terms of gradient norm minimization. In Section 4, we propose an HR-ODE based on the rate-matching technique. We demonstrate that NAG can be interpreted as an approximation of the rate-matching technique applied to a specific ODE. Then, in Section 5, we extend our analysis to a stochastic setting with noisy gradients, introducing a stochastic method that guarantees convergence rates of O(1/k1/2) for the expected objective residual and O(1/k3/4) for the expected squared norm of the gradient. Finally, in Section 6, we present numerical experiments to demonstrate the empirical performance of our proposed method and to validate our theoretical findings. Problem template and notation. We consider a generic unconstrained smooth convex minimization template: min x Rn f(x) (1) where f : Rn R is convex and L-smooth, meaning that its gradient is Lipschitz continuous: f(x) f(y) L x y , x, y Rn, (2) with denoting the Euclidean norm. We denote the class of L-smooth convex functions by FL. Throughout, we assume that the solution set for Problem (1) is non-empty, and we denote an arbitrary solution by x , hence f := f(x ) f(x) for all x Rn. The NAG Algorithm. Given an initial state x0 = y0 Rn and a step-size parameter s > 0, the NAG algorithm updates the variables xk and yk iteratively as follows: yk+1 = xk s f(xk), xk+1 = yk+1 + k k + 3(yk+1 yk). (NAG) Low-Resolution and High-Resolution ODEs. Throughout, by LR-ODE, we refer to the secondorder ODE introduced and studied in [Su et al., 2016], which exhibits a behavior reminiscent of momentum-based accelerated methods like Heavy Ball (HB) and NAG in the limit of step-size s 0. Unfortunately, this generic ODE cannot distinguish the differences between HB and NAG. It is important to note, however, that the guarantees of HB and NAG differ significantly in discrete time. Specifically, HB has guarantees for only a specific subset of Problem (1), whereas NAG guarantees a solution for all instances of (1) with appropriate step sizes. Therefore, it s worth noting that the LR-ODE may not fully capture certain important aspects of the NAG trajectory. Conversely, by HR-ODEs, we refer to the ODEs proposed in [Shi et al., 2021], which extend the LR-ODE by incorporating gradient correction terms. Through this extension, HR-ODEs provide a more precise representation of the trajectory for these distinct algorithms. 2 External Forces and High-Resolution ODEs Consider the Lagrangian L(Xt, Xt, t) = eαt+γt 1 2 e αt Xt 2 eβtf(Xt) . (3) Here, Xt Rd is the first time-derivative of X(t), and αt, βt, γt : T R are continuously differentiable functions of time that correspond to the weighting of velocity, the potential function f, and the overall damping, respectively. Using variational calculus, we define the action for the curves {Xt : t R} as the functional A(X) = R R L(Xt, Xt, t)dt. In the absence of external forces, a curve is a stationary point for the problem of minimizing the action A(X) if and only if it satisfies the Euler Lagrange equation d Xt (Xt, Xt, t)} = L Xt (Xt, Xt, t). This was used in [Wibisono et al., 2016, Wilson et al., 2021] to calculate the LR-ODEs for convex and strongly convex functions1. Note that the Euler-Lagrange equation as written, does not account for an external force, F (which is non-conservative). In this case, the Euler-Lagrange equation should be modified to the forced Euler-Lagrange Equation Xt (Xt, Xt, t) L Xt (Xt, Xt, t) = F, (4) which itself is the result of integration by parts of Lagrange d Alembert principle [Campos et al., 2021]. Using the Lagrangian (3) we have L Xt (Xt, Xt, t) = eγt(e αt Xt), L Xt (Xt, Xt, t) = eγt+αt+βt( f(Xt)). (5) Substituting (5) in (4) gives Xt + ( γt αt) Xt + e2αt+βt f(Xt) = eαt γt F. (6) In what follows, we will present various choices of the external force F, including two for convex functions and one for strongly convex functions. 2.1 Convex Functions First choice for convex functions. Let us first consider the following external force: dt[e αt f(X)]. (7) In this case, (6) becomes Xt + ( γt αt) Xt + e2αt+βt f(Xt) = seαt d dt e αt f(Xt) . (8) It is possible to show the convergence of Xt to x and establish a convergence rate for this as shown in the following theorem (proof in Appendix A.1). The proof of this theorem (and the subsequent theorems in this section) is based on the construction of a suitable Lyapunov function for the corresponding ODE (e.g. see [Siegel, 2019, Shi et al., 2019, Attouch et al., 2020, 2021]). This non-negative function attains zero only at the stationary solution of the corresponding ODE and decreases along the trajectory of the ODE [Khalil, 2002]. For this theorem (and the subsequent theorems), we will define a proper Lyapunov function and prove sufficient decrease of the function f along the corresponding ODE trajectory. Theorem 2.1. Under the ideal scaling conditions βt eαt and γt = eαt, Xt in (8) will satisfy f(Xt) f(x ) O(e βt) 1Wilson et al. [2021] use different Lagrangian for strongly convex functions, but the methodology is the same. Now, choosing parameters as αt = log(n(t)), βt = log(q(t)/n(t)), γt = eαt = n(t), (9) in (8) gives ( Xt + (n(t) n(t) n(t) + s 2f(Xt)) Xt + (n(t)q(t) s n(t) n(t)) f(Xt) = 0, F = seγt d dt[e αt f(X)], (10) which reduces to t + s 2f(Xt) Xt + Cp2tp 2 + s f(Xt) = 0, (11) by taking n(t) = p t , q(t) = Cptp 1. Remark 2.1.1. For p = 2, C = 1/4, equation (11) corresponds to the (H-ODE) in [Laborde and Oberman, 2020]. Second choice for convex functions. Now, we consider an external force given by F = seγt βt d h e (αt βt) f(Xt) i (12) In this case, replacing F in (6) gives Xt + ( γt αt) Xt + e2αt+βt f = seαt βt d dt[e (αt βt) f(Xt)]. (13) We establish the following convergence result, and the proof can be found in Appendix A.2. Theorem 2.2. Under the modified ideal scaling conditions βt eαt, γt = eαt, βt eαt βt +2 αt βt, Xt in (13) will satisfy f(Xt) f(x ) O 1 eβt + se 2αt βt Taking the same parameters as in (9) gives ( Xt + (n(t) n(t) n(t) + s 2f(Xt)) Xt + (n(t)q(t) s( n(t) n(t) q(t)n(t) n(t)q(t) n(t)q(t) )) f(Xt) = 0, F = seγt βt d dt e (αt βt) f(Xt) . (14) which reduces to t + s 2f(Xt) Xt + Cp2tp 2 + s(p + 1) f(Xt) = 0, (15) for n(t) = p/t, q(t) = Cptp 1. Remark 2.2.1. Note that setting C = 1/4, p = 2 will lead to the ODE t + s 2f(Xt) Xt + 1 + 3 s f(Xt) = 0. (16) This ODE was discretized using the Semi-Implicit Euler (SIE) and the Implicit Euler (IE) discretization schemes in [Shi et al., 2019]. The corresponding optimization algorithms were shown to accelerate. In addition, note that the convergence rate proved in Theorem 2.2 is faster than its counterpart in Theorem 2.1. 2.2 Strongly Convex Functions Our analysis is applicable to strongly convex functions as well. Consider the Lagrangian proposed in [Wilson et al., 2021] for strongly convex functions L(Xt, Xt, t) = eαt+βt+γt µ 2 e αt Xt 2 f(Xt) . (17) Then, the forced Euler-Lagrange equation (4) becomes X + ( αt + γt + βt) X + 1 µe2αt f(X) = F µe αt+γt+βt . (18) Taking F = seαt+γt d dt(eβt f(Xt)) in (18) gives X + ( αt + γt + βt) X + 1 µe2αt f(X) = se2αt βt d dt(eβt f(Xt)) We can establish the following convergence result for Xt in (19) to the unique minimizer x . The proof of this result is deferred to Appendix A.3. Theorem 2.3. Under the modified ideal scaling conditions αt = α, βt eαt, γt = eαt, and βt 0 Xt in (19) satisfies f(Xt) f(x ) O(e βt) (20) for µ-strongly convex function f. Remark 2.3.1. Taking α = log( µ) and γt = βt = µt in (19) gives the NAG s corresponding HR-ODE Xt + (2 µ + s 2f(Xt)) Xt + (1 + µs) f(Xt) = 0, (21) for µ-strongly convex function f as in [Shi et al., 2021]. 3 Gradient Norm Minimization of NAG One of the implications of our variational study on HR-ODEs in Section 2 was the ODE (14). Reformulating this ODE gives ( Xt = n(t)(Vt Xt) s f(Xt) Vt = q(t) f(Xt) s q(t)n(t) n(t)q(t) n2(t)q(t) f(Xt). (22) Applying the SIE on (22) for X(t) X(tk), V (t) V (tk), n(tk) = p/tk, q(tk) = Cptp 1 k , p = 2, tk = k s and C = 1/4 gives xk+1 = xk + 2 k(vk xk+1) s f(xk), vk+1 = vk 1 2(ks) f(xk+1) s f(xk+1), (23) which is exactly the NAG algorithm. The interpretation of the NAG method as the SIE discretization of (22) has not been discussed before in the literature (see [Ahn and Sra, 2022] for the four most studied representations). It is precisely this connection with the ODE (22) though that inspires our choice of the Lyapunov function which in turn gives rise to a faster convergence rate. The following theorem formulates this result. The proof is in Appendix A.4 and it is based on the discrete Lyapunov analysis of (23). Similar convergence rate was very recently found by [Chen et al., 2022] through implicit velocity perspective on HR-ODEs which uses a different Lyapunov analysis than this work. Theorem 3.1. Consider the update (23). Then, if f FL we have min 0 i k 1 f(xi) 2 12 k3s2 x0 x 2, and f(xk) f(x ) 2 sk(k + 2) x0 x 2 for 0 s 1/L, v0 = x0, and any x0 Rn. Remark 3.1.1 (Comparison with state of the art). The rate in Theorem 3.1 is improved compared to the previous rate found in [Shi et al., 2021], which is min 0 i k f(xi) 2 8568 (k + 1)3s2 x0 x 2, for 0 < s 1/(3L) and k 0. 4 Rate-Matching Approximates the NAG Algorithm The ODE (11) when p = 2, C = 1/4 is equivalent to Xt = 2 t (Zt Xt) s f(Xt), Zt = t 2 f(Xt). (24) which can be viewed as a perturbation of the LR-ODE Xt = 2 t (Zt Xt), Zt = t 2 f(Xt). (25) We now show that when the rate-matching technique in [Wibisono et al., 2016] is applied to (25), the final algorithm reveals similar behavior as (24). This result is then used to approximately recover the NAG method using rate-matching discretization. Applying the rate-matching discretization on the ODE (25) gives xk+1 = 2 k+2zk + k k+2yk, yk = xk s f(xk), zk = zk 1 1 2sk f(yk). (26) which has a convergence rate of O(1/(sk2)) [Wibisono et al., 2016]. In the following proposition, we study the behavior of (26) in limit of s 0. The proof is given in Appendix A.5. Proposition 4.1. The continuous-time behavior of (26) is approximately t + s 2f(Xt) Xt + 1 + s f(Xt) = 0, (27) which is the the high-resolution ODE (11). The ODE (27) is the same as (24). In this sense, rate-matching implicitly perturbs the LR-ODE. The question that naturally arises is that when do we recover the HR-ODE (16) (which corresponds to the NAG algortihm through the SIE discretization) from the rate-matching technique? To answer, we will first perturb the LR-ODE (25) in the second line. Then, the rate-matching discretization is applied. Perturbing (25) gives Xt = 2 t (Zt Xt), Zt = t 2 f(Xt) s f(Xt). (28) Discretizing (28) using the rate-matching method with tk = k s gives xk+1 = 2 k+2zk + k k+2yk, yk = xk s f(xk), zk = zk 1 s 2(k + 2) f(yk), (29) which is extremely close to the NAG algorithm. Indeed, replacing f(yk) with f(xk) in the third line of (29) gives exactly the NAG method. Typically, xk and yk are very close. This is due to xk and yk having a difference of order s. Since in continuous time X(tk) Y (tk) (due to s 0), the HR-ODE of (29) is (16). This means that the corresponding HR-ODE of (29) is Xt = 2 t (Zt Xt) s f(Xt), Zt = t 2 f(Xt) s f(Xt). (30) which is the perturbed version of (28) and the HR-ODE associated with the NAG algorithm. 5 Stochastic Extensions In this section, we propose a stochastic variation of (23). We model noisy gradients by adding i.i.d noise ek with variance σ2 to the gradients. Consider the update ( xk+1 = xk + 2sk tk (vk xk+1) βsk L ( f(xk) + ek), vk+1 = vk 1 2(tksk + 2skβ L )( f(xk+1) + ek+1) (31) with β 2. This update reduces to (23) when ek = 0, sk = s = β/ L, tk = k s. We will refer to (31) as the Noisy NAG (NNAG) algorithm. NNAG is interesting due to its capability of dealing with perturbed gradients. This is the case in practical methods e.g. SGD [Bottou, 2010], SAG [Schmidt et al., 2017], SAGA [Defazio et al., 2014], SVRG [Johnson and Zhang, 2013], and etc. The following convergence result holds for NNAG, and its proof is in Appendix A.6. Theorem 5.1. Suppose f FL and consider the NNAG method detailed in (31) with the following parameter choices: β 2, sk = c kα , and tk = i=1 si for some c 1 4 α < 1. (32) We define the critical iteration k0 as the smallest positive integer that satisfies L 8 (Pk0 i=1 1 iα )2 Then, the following bounds hold for all k k0 : E[f(xk)] f(x ) E[ε(k0)]+ σ2c4 (1 α)2 [k3 4α 0 k3 4α]+ σ2c3β 2 L(1 α)(3α 2)[k2 3α 0 k2 3α]+ β2c2σ2 2L(2α 1)[k1 2α 0 k1 2α] 4(1 α)2 ((k1 α 1)2)+ cβ 2 L(1 α)(k(1 α) 1) if α > 3/4, and E[f(xk)] f(x ) E[ε(k0)]+2σ2c4h log( k k0 ) i + 8σ2c3β h k 1/4 0 k 1/4i + β2c2σ2 L h k 1/2 0 k 1/2i 4c2((k1/4 1)2)+ 2cβ L (k1/4 1) (34) if α = 3/4 with ε(k) = ( t2 k 4 + tkβ L)(f(xk) f(x )) + 1 Next, we show that slight modifications to the NNAG method gives rise to another stochastic method with a similar convergence rate as the NNAG algorithm, but more transparent proof (see Appendix A.7). This proof results in a convergence rate for E min0 i k 1 f(xi) 2 with a rate of O(log(k)/k(3/4)). It remains a future work to show similar result for the NNAG update. Theorem 5.2. Suppose f FL and consider the following modification of the NNAG method xk+1 = xk + 2sk tk (vk xk+1) sk L( f(xk) + ek), vk+1 = vk 1 2((tk)sk)( f(xk+1) + ek+1) s2 k( f(xk+1) + ek+1). (35) with the same parameter choices as in (32). Then, the following convergence bounds hold: E[f(xk)] f(x ) E[ε(0)]+ c4σ2 8 [16(1+log(k))+32+6] 2c2 h 2(k 1 4 1)2+k 3 4 (k 1 4 1) i α = 3 E[ε(0)]+ c4σ2 8 h (4α 2) (1 α)2(4α 3) + 4(4α 1) (1 α)(4α 2) + 4(4α) c2 2(1 α) h (k1 α 1)2 2(1 α) +k α(k(1 α) 1) i 1 > α > 3 with E[ε(0)] = 1 2 v0 x 2. In addition, for α = 3/4 we have E min 0 i k 1 f(xi) 2 2 LE[ε(0)] + (2c4σ2 L)(2 log(k) + 6 + 3 16c3 k3/4 1 2 + k1/2 . (37) Remark 5.2.1. The algorithm in (35) reduces to (23) when ek = 0, sk = s = 1/ L, tk = k s. Remark 5.2.2 (Connection to NAG). When there is no noise (σ = 0) we can also allow parameter α to be zero. This is because we do not need to alleviate the effect of the noise with a decreasing stepsize. Therefore, we recover the convergence rate of O(1/k2) for the NAG method when c = 1/ L. Remark 5.2.3 (Comparison with [Laborde and Oberman, 2020]). Laborde et al., proposed a stochastic method with noisy gradients. Their method uses another presentation of the NAG algorithm (the presentation from EE discretization). Our rate (36) has the same order of convergence as [Laborde and Oberman, 2020]. However, their analysis did not achieve the bound (37) (see [Laborde and Oberman, 2020] Appendix C.4). Remark 5.2.4 (Comparison between Theorems 5.1 and 5.2). The rate in (34) is asymptotically similar to (36). However, the transient behavior of (34) is faster than both (36) and the rate in [Laborde and Oberman, 2020] when L is large (see Figure 1 top row). This is due to the tuning parameter β which is usually set to L or higher. This scenario (Large L) often happens in practice, e.g. in training a two-layer Convolutional Neural Network (CNN) [Shi et al., 2022]. Remark 5.2.5 (Limitations and future directions). One limitation of our theoretical analysis is that our convergence result for NNAG holds only for a large enough number of iterations k k0. However, in our numerical experiments we observed that the same bounds hold also for k k0, so we believe our result can be improved. Additionally, our proposed forces are currently defined only in Euclidean space, and we see potential for extending the framework to non-Euclidean spaces. 6 Numerical Results In this section, we present our empirical results, divided into three parts: Theoretical upper bounds, binary classification with logistic regression, and classification with neural networks. Upper Bounds. First, we compare the bounds in (34) and (36) with the Proposition 4.5 in [Laborde and Oberman, 2020]. The results are depicted in Figure 1. In practical scenarios where L is large [Shi et al., 2022] the bound (34) is lower than the other two for large enough iterations. This observation has motivated us to analyze the behavior of NNAG in practical scenarios such as binary classification and CNN training tasks. 100 101 102 103 104 105 10 1 Bound value for L = 103 100 101 102 103 104 105 Bound value for L = 102 Upper-bound (34) Upper-bound (36), α = 3/4 Upper-bound [Laborde et al., [2020]] Figure 1: Comparison of our upper bounds with the state-of-the-art. Binary Classification. For this task we considered d = 1000 randomly generated samples of dimension n = 10 and labels. The problem can be written as min x Rn 1 d i=1 log(1 + e yi xi,x ), (38) where yi and xi denote the ith label and sample. For comparison, we considered the perturbed gradient descent method with Gaussian noise and decreasing step-size sk = 1/( Lk2/3) (Proposition 3.4 in [Laborde and Oberman, 2020]) together with accelerated noisy gradient descent (Per-FE-C) in [Laborde and Oberman, 2020]. For NNAG we used sk = 1/( Lk3/4) and β = L/10. All the perturbation was done using i.i.d. Gaussian noise with unit variance, and we conducted 100 Monte Carlo runs. The results are presented in Figure 2 right panel. As shown, NNAG outperforms all the other methods in this case. In a related experiment, we combined NNAG with SGD and SVRG. For the SGD-mixing, we replaced the noisy gradients with SGD-like gradients, while for the SVRG-mixing we evaluated all the gradients at the beginning of each epoch (essentially like taking a snapshot in SVRG) and set tk = 0. The step-sizes for SVRG (in accordance with the original implementation [Johnson and Zhang, 2013]), SGD, NNAG+SVRG, and NNAG+SGD were set as 1/(10L), 1/L, c = 1/ L, β = L/10, 20 40 60 80 100 10 9 Logistic loss 20 40 60 80 100 Logistic loss NNAG Perturbed Gradient Descent Per-FE-C SGD SVRG NNAG+SVRG NNAG+SGD Figure 2: Comparison of the performance of various methods in binary classification problem. L, β = L, respectively. We conducted 100 Monte-Carlo runs, and the results are displayed in Figure 2 left panel. Notably, when mixed with either SGD or SVRG, NNAG outperforms the original methods. This highlights NNAG s flexibility in terms of gradient noise and demonstrates its potential to accelerate various methods. Classification on CIFAR10. Finally, we tackled the non-convex optimization problem of training a CNN with CIFAR10 dataset [Krizhevsky et al., 2009] using the SGD, SVRG, NNAG, and NNAG+SVRG methods. The network consisted of two convolutional layers each followed by max pooling and 3 fully connected linear layers each followed by Re LU activation function. The stepsizes for SGD and SVRG were set as 0.01, and for the NNAG and NNAG+SVRG algorithms we had c = 0.05, β = 1502 and c = 0.001, β = 1002/10, respectively. The division by 10 is due to stepsize division by 10 in the SVRG method. The results for 20 Monte-Carlo simulations are depicted in Figure 3. Notably, SVRG+NNAG outperforms the other methods in terms of minimizing training error. Additionally, NNAG exhibits slightly better validation accuracy, hinting at its convergence toward a different local solution. It is worth noting that faster convergence rates of NNAG+SVRG do not pose problems of overfitting. If necessary, one can terminate the algorithm earlier to achieve optimal performance. In Figure 3, for instance, NNAG+SVRG reaches its peak "validation accuracy" after approximately 20 epochs, with a validation accuracy of roughly 0.6 and a training error of around 0.66. After this point, a slow overfitting phase begins. Similarly, for SGD and SVRG, their peak "validation accuracy" is achieved after about 50 epochs, with a validation accuracy of approximately 0.6 and a training error of about 0.66, followed by a slow overfitting phase. Finally, NNAG achieves comparable results after approximately 100 epochs. 0 20 40 60 80 100 0 Training Error 0 20 40 60 80 100 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 Validation Accuracy SGD SVRG NNAG NNAG+SVRG Figure 3: Training error and validation accuracy of NNAG, SGD, SVRG, and NNAG+SVRG when used for training a simple CNN on CIFAR10 dataset. Lower and upper confidence bounds with significance level of 0.68 are drawn with similar color to their corresponding line. 7 Related Work Polyak s Heavy-Ball (HB) method was one of the first momentum-based methods which could accelerate relative to the gradient descent method [Polyak, 1963]. However, this was not the case for every smooth function [Lessard et al., 2016]. Nesterov modified the HB and introduced the NAG method. This method achieved global convergence with a rate of O(1/k2) for smooth convex functions [Nesterov, 1983]. Nesterov used estimate sequence technique to show the convergence of the NAG method. This technique does not provide immediate insights toward the success of the NAG algorithm in acceleration. Thus, many have tried to understand the essence of acceleration. On a similar line of work, Wibisono et al. [2016] introduced a variational perspective on accelerated methods, leading to a general (non-Euclidean) ODE containing the ODE found by Su et al. [2016] as a special case. Their approach is based on the choice of a Lagrangian and its corresponding parameters. Since the choice of Lagrangian is not unique, Wilson et al. [2021] provided a variational perspective on different accelerated first-order methods using another Lagrangian. Fazlyab et al. [2017] developed a family of accelerated dual algorithms for constrained convex minimization through a similar variational approach. In a more recent development, Zhang et al. [2021] showed that the second-variation also plays an important role in optimality of the ODE found by Su et al. [2016]. Specifically, they showed that if the time duration is long enough, then the mentioned ODE for the NAG algorithm is the saddle point to the problem of minimizing the action functional. The dynamical system perspective on NAG was studied in [Muehlebach and Jordan, 2019]. They showed that the NAG algorithm is recovered from the SIE discretization of an ODE. The mentioned ODE was not the result of a vanishing step-size argument. They found that a curvature-dependent damping term accounts for the acceleration phenomenon. Interestingly, [Chen et al., 2022] also used similar ODE without the SIE discretization. They showed that implicit-velocity is the reason of the acceleration. In a recent analysis, [Muehlebach and Jordan, 2023] explores the connections between non-smooth dynamical systems and first-order methods for constrained optimization. 8 Conclusion In this work, we tackled the problem of unconstrained smooth convex minimization in Euclidean space. Through a variational analysis of HR-ODEs, we achieved improved convergence rates for NAG in terms of gradient norm. In addition, we showed that NAG can be viewed as an approximation of the rate-matching technique when applied on a specific ODE. Our analysis was then extended to stochastic scenarios. In particular, we proposed a method with both constant and varying stepsizes which performed comparable and sometimes better than state of the art methods. This work entails multiple future directions. Nesterov s oracle complexity lower bound on gradient norm minimization is O(k 4) [Nesterov, 2003]. It remains an open question to see if the NAG method can achieve this rate of convergence for gradient norm minimization. In this work, we noticed that the HR-ODEs follow the same external force structure. In the smooth-strongly convex case, Triple Momentum (TM) method is the fastest known globally convergent method [Van Scoy et al., 2018]. However, the HR-ODE associated with the TM method is not shown to achieve the similar convergence rate as the TM method [Sun et al., 2020]. One could use the external force structure proposed here to find a better convergence rate for the HR-ODE associated with the TM algorithm. In addition, our analysis was confined to the Euclidean space. We believe it is possible to explore non Euclidean forces using a Bregman Lagrangian as in [Wibisono et al., 2016]. Finally, we blended our noisy stochastic scheme with other known stochastic methods (e.g. SGD and SVRG). This technique improved the performance of those methods. As a future work, one can apply the same technique to other practical methods like ADAM, RMSprop, etc, and study the behavior of the final algorithm. Acknowledgements Alp Yurtsever and Hoomaan Maskan received support from the Wallenberg AI, Autonomous Systems and Software Program (WASP) funded by the Knut and Alice Wallenberg Foundation. Konstantinos C. Zygalakis acknowledges support from a Leverhulme Research Fellowship (RF/2020310), and the EPSRC grant EP/V006177/1. We gratefully acknowledge the support of NVIDIA Corporation with the donation of 2 Quadro RTX 6000 GPUs used for this research. K. Ahn and S. Sra. Understanding nesterov s acceleration via proximal point method. In Symposium on Simplicity in Algorithms (SOSA), pages 117 130. SIAM, 2022. H. Attouch, Z. Chbani, J. Fadili, and H. Riahi. First-order optimization algorithms via inertial systems with hessian driven damping. Mathematical Programming, pages 1 43, 2020. H. Attouch, Z. Chbani, J. Fadili, and H. Riahi. Convergence of iterates for first-order optimization algorithms with inertia and hessian driven damping. Optimization, pages 1 40, 2021. L. Bottou. Large-scale machine learning with stochastic gradient descent. In Proceedings of COMPSTAT 2010: 19th International Conference on Computational Statistics Paris France, August 2227, 2010 Keynote, Invited and Contributed Papers, pages 177 186. Springer, 2010. C. M. Campos, A. Mahillo, and D. M. de Diego. A discrete variational derivation of accelerated methods in optimization. ar Xiv preprint ar Xiv:2106.02700, 2021. S. Chen, B. Shi, and Y.-x. Yuan. Gradient norm minimization of nesterov acceleration: o(1/k3). ar Xiv preprint ar Xiv:2209.08862, 2022. A. Defazio, F. Bach, and S. Lacoste-Julien. Saga: A fast incremental gradient method with support for non-strongly convex composite objectives. Advances in neural information processing systems, 27, 2014. M. Fazlyab, A. Koppel, V. M. Preciado, and A. Ribeiro. A variational approach to dual methods for constrained convex optimization. In 2017 American Control Conference (ACC), pages 5269 5275, 2017. doi: 10.23919/ACC.2017.7963773. M. Fazlyab, A. Ribeiro, M. Morari, and V. M. Preciado. Analysis of optimization algorithms via integral quadratic constraints: Nonstrongly convex problems. SIAM Journal on Optimization, 28(3):2654 2689, 2018. doi: 10.1137/17M1136845. URL https://doi.org/10.1137/17M1136845. B. Hu and L. Lessard. Dissipativity theory for nesterov s accelerated method. In International Conference on Machine Learning, pages 1549 1557. PMLR, 2017. R. Johnson and T. Zhang. Accelerating stochastic gradient descent using predictive variance reduction. Advances in neural information processing systems, 26, 2013. H. K. Khalil. Nonlinear systems third edition. Patience Hall, 115, 2002. A. Krizhevsky, G. Hinton, et al. Learning multiple layers of features from tiny images. Toronto, ON, Canada, 2009. M. Laborde and A. Oberman. A lyapunov analysis for accelerated gradient methods: from deterministic to stochastic case. In S. Chiappa and R. Calandra, editors, Proceedings of the Twenty Third International Conference on Artificial Intelligence and Statistics, volume 108 of Proceedings of Machine Learning Research, pages 602 612. PMLR, 26 28 Aug 2020. URL https://proceedings.mlr.press/v108/laborde20a.html. L. Lessard, B. Recht, and A. Packard. Analysis and design of optimization algorithms via integral quadratic constraints. SIAM J. Optim., 26:57 95, 2016. M. Muehlebach and M. Jordan. A dynamical systems perspective on nesterov acceleration. In International Conference on Machine Learning, pages 4656 4662. PMLR, 2019. M. Muehlebach and M. I. Jordan. On constraints in first-order optimization: A view from nonsmooth dynamical systems. Journal of Machine Learning Research, 23(256):1 47, 2022. M. Muehlebach and M. I. Jordan. Accelerated first-order optimization under nonlinear constraints. ar Xiv preprint ar Xiv:2302.00316, 2023. Y. Nesterov. A method for solving the convex programming problem with convergence rate o(1/k2). Proceedings of the USSR Academy of Sciences, 269:543 547, 1983. Y. Nesterov. Introductory lectures on convex optimization: A basic course, volume 87. Springer Science & Business Media, 2003. B. Polyak. Gradient methods for the minimisation of functionals. Ussr Computational Mathematics and Mathematical Physics, 3:864 878, 1963. J. M. Sanz Serna and K. C. Zygalakis. The connections between lyapunov functions for some optimization algorithms and differential equations. SIAM Journal on Numerical Analysis, 59(3): 1542 1565, 2021. M. Schmidt, N. Le Roux, and F. Bach. Minimizing finite sums with the stochastic average gradient. Mathematical Programming, 162:83 112, 2017. B. Shi, S. S. Du, W. J. Su, and M. I. Jordan. Acceleration via symplectic discretization of highresolution differential equations. ar Xiv preprint ar Xiv:1902.03694, 2019. B. Shi, S. S. Du, M. I. Jordan, and W. J. Su. Understanding the acceleration phenomenon via high-resolution differential equations. Ar Xiv, abs/1810.08907, 2021. Z. Shi, Y. Wang, H. Zhang, J. Z. Kolter, and C.-J. Hsieh. Efficiently computing local lipschitz constants of neural networks via bound propagation. Advances in Neural Information Processing Systems, 35:2350 2364, 2022. J. W. Siegel. Accelerated first-order methods: Differential equations and lyapunov functions. ar Xiv preprint ar Xiv:1903.05671, 2019. W. Su, S. Boyd, and E. J. Candès. A differential equation for modeling nesterov s accelerated gradient method: Theory and insights. Journal of Machine Learning Research, 17(153):1 43, 2016. URL http://jmlr.org/papers/v17/15-084.html. B. Sun, J. George, and S. S. Kia. High-resolution modeling of the fastest firstorder optimization method for strongly convex functions. In 59th IEEE Conference on Decision and Control, CDC 2020, Jeju Island, South Korea, December 14-18, 2020, pages 4237 4242. IEEE, 2020. doi: 10.1109/CDC42340.2020.9304444. URL https://doi.org/10.1109/CDC42340.2020.9304444. B. Van Scoy, R. A. Freeman, and K. M. Lynch. The fastest known globally convergent first-order method for minimizing strongly convex functions. IEEE Control Systems Letters, 2(1):49 54, 2018. doi: 10.1109/LCSYS.2017.2722406. A. Wibisono, A. C. Wilson, and M. I. Jordan. A variational perspective on accelerated methods in optimization. Proceedings of the National Academy of Sciences, 113 (47):E7351 E7358, 2016. ISSN 0027-8424. doi: 10.1073/pnas.1614734113. URL https://www.pnas.org/content/113/47/E7351. A. C. Wilson, B. Recht, and M. I. Jordan. A lyapunov analysis of accelerated methods in optimization. J. Mach. Learn. Res., 22:113 1, 2021. P. Zhang, A. Orvieto, and H. Daneshmand. Rethinking the variational interpretation of accelerated optimization methods. Advances in Neural Information Processing Systems, 34:14396 14406, 2021. Z. A. Zhu and L. Orecchia. Linear coupling: An ultimate unification of gradient and mirror descent. In Information Technology Convergence and Services, 2014.