# stable_nonconvexnonconcave_training_via_linear_interpolation__18cc3845.pdf Stable Nonconvex-Nonconcave Training via Linear Interpolation Thomas Pethick EPFL (LIONS) thomas.pethick@epfl.ch Wanyun Xie EPFL (LIONS) wanyun.xie@epfl.ch Volkan Cevher EPFL (LIONS) volkan.cevher@epfl.ch This paper presents a theoretical analysis of linear interpolation as a principled method for stabilizing (large-scale) neural network training. We argue that instabilities in the optimization process are often caused by the nonmonotonicity of the loss landscape and show how linear interpolation can help by leveraging the theory of nonexpansive operators. We construct a new optimization scheme called relaxed approximate proximal point (RAPP), which is the first 1-SCLI method to achieve last iterate convergence rates for ρ-comonotone problems while only requiring ρ > 1 2L. The construction extends to constrained and regularized settings. By replacing the inner optimizer in RAPP we rediscover the family of Lookahead algorithms for which we establish convergence in cohypomonotone problems even when the base optimizer is taken to be gradient descent ascent. The range of cohypomonotone problems in which Lookahead converges is further expanded by exploiting that Lookahead inherits the properties of the base optimizer. We corroborate the results with experiments on generative adversarial networks which demonstrates the benefits of the linear interpolation present in both RAPP and Lookahead. 1 Introduction Stability is a major concern when training large scale models. In particular, generative adversarial networks (GANs) are known to be notoriously difficult to train. To stabilize training, the Lookahead algorithm of Zhang et al. (2019) was recently proposed for GANs Chavdarova et al. (2020) which linearly interpolates with a slow moving iterate. The mechanism has enjoyed superior empirical performance in both minimization and minimax problems, but it largely remains a heuristic with little theoretical motivation. One major obstacle for providing a theoretical treatment, is in capturing the (fuzzy) notion of stability. Loosely speaking, a training dynamics is referred to as unstable in practice when the iterates either cycle indefinitely or (eventually) diverge as has been observed for the Adam optimizer (see e.g. Gidel et al. (2018, Fig. 12) and Chavdarova et al. (2020, Fig. 6) respectively). Conversely, a stable dynamics has some bias towards stationary points. The notion of stability so far (e.g. in Chavdarova et al. (2020, Thm. 2-3)) is based on the spectral radius and thus inherently local. In this work, we are interested in establishing global convergence properties, in which case some structural assumptions are needed. One (nonmonotone) structure that lends itself well to the study of stability is that of cohypomonotonicity studied in Combettes & Pennanen (2004); Diakonikolas et al. (2021), since even the extragradient method has been shown to cycle and diverge in this problem class (see Pethick et al. (2022, Fig. 1) and Pethick et al. (2023, Fig. 2) respectively). We provide a geometric intuition behind these difficulties in Figure 1. Biasing the optimization schemes towards stationary points becomes a central concern and we demonstrate in Figure 2 that Lookahead can indeed converge for such nonmonotone problems. 37th Conference on Neural Information Processing Systems (Neur IPS 2023). Table 1: Overview of last iterate results with our contribution highlighted in blue . Prior to this work there existed no rates for 1-SCLI schemes handling ρ-comonotone problems with ρ ( 1/2L, ) and no global convergence guarantees for Lookahead beyond bilinear games. Method Setting ρ Handles constraints ρ-independent rates Reference PP Comonotone ( 1/2L, ) (Gorbunov et al., 2022b, Thm. 3.1) Relaxed PP Comonotone ( 1/2L, ) Theorem 6.2 Extrapolate EG Comonotone & Lips. ( 1/8L, ) (Gorbunov et al., 2022b, Thm. 4.1) EG+ Comonotone & Lips. Unknown rates RAPP Comonotone & Lips. ( 1/2L, ) Corollary 6.4 Local - - (Chavdarova et al., 2020, Thm. 2) Bilinear - - (Ha & Kim, 2022, Cor. 7) Comonotone & Lips. ( 1/3 3L, ) - Theorem 7.1 LA-EG Bilinear - - (Ha & Kim, 2022, Cor. 8) Monotone & Lips. - - Theorem F.1 LA-CEG+ Comonotone & Lips. ( 1/2L, ) - Corollary 7.7 A principled approach to cohypomonotone problems is the extragradient+ algorithm (EG+) proposed by Diakonikolas et al. (2021). However, the only known rates are on the best iterate, which can be problematic to pick in practice. It is unclear whether last iterate rates for EG+ are possible even in the monotone case (see discussion prior to Thm. 3.3 in Gorbunov et al. (2022a)). For this reason, the community has instead resorted to showing last iterate of extragradient (EG) method of Korpelevich (1977), despite originally being developed for the monotone case. Maybe not surprisingly, EG only enjoys a last iterate guarantee under mild form of cohypomonotonicity and have so far only been studied in the unconstrained case (Luo & Tran-Dinh; Gorbunov et al., 2022b). Recently, last iterate rate were established for the same (tight) range of cohypomonotone problems for which EG+ has best iterate guarantees. However, the analyzed scheme is implicit and the complexity blows up with increasing cohypomonotonicity (Gorbunov et al., 2022b). This leaves the questions: Can an explicit scheme enjoy last iterate rates for the same range of cohypomonotone problems? Can the rate be agnostic to the degree of cohypomonotonicity? We answer both in the affirmative. This work focuses on 1-SCLI schemes (Arjevani et al., 2015; Golowich et al., 2020), whose update rule only depends on the previous iterate in a time-invariant fashion. Another approach to establishing last iterate is Halpern-type methods with an explicit scheme developed in Lee & Kim (2021) for cohypomonotone problems and later extended to the constrained case in Cai et al. (2022) (c.f. Appendix A). As will become clear, a principled mechanism behind convergence in this nonmonotone class is the linear interpolation also used in Lookahead. This iterative interpolation is more broadly referred to as the Krasnosel ski ı-Mann (KM) iteration in the theory of nonexpansive operators. We show that the extragradient+ algorithm (EG+) of Diakonikolas et al. (2021), our proposed relaxed approximate proximal point method (RAPP), and Lookahead based algorithms are all instances of the (inexact) KM iteration and provide simple proofs of these schemes in the cohypomonotone case. More concretely we make the following contributions: 1. We prove global convergence rates for the last iterate of our proposed algorithm RAPP which additionally handles constrained and regularized settings. This makes RAPP the first 1-SCLI scheme to have non-asymptotic guarantees for ρ-comonotone problems while only requiring ρ > 1/2L. As a byproduct we obtain a last iterate convergence rate for an implicit scheme that is independent of the degree of cohypomonotonicity. The last iterate rates are established by showing monotonic decrease of the operator norm something which is not possible for EG+. This contrast is maybe surprising, since RAPP can be viewed as an extension of EG+, which simply takes multiple extrapolation steps. 2. By replacing the inner optimization routine in RAPP with gradient descent ascent (GDA) and extragradient (EG) we rediscover the Lookahead algorithms considered in Chavdarova et al. (2020). We obtain guarantees for the Lookahead variants by deriving nonexpansive properties of the base optimizers. By casting Lookahead as a KM iteration we find that the optimal interpolation constant is λ = 0.5. This choice corresponds to the default value used in practice for both minimization and minimax thus providing theoretical motivation for the parameter value. 3. For τ = 2 inner iterations we observe that LA-GDA reduces to a linear interpolation between GDA and EG+ which allows us to obtain global convergence in ρ-comonotone problems when ρ > 1/3 3L. However, for τ large, we provide a counterexample showing that LA-GDA cannot be guaranteed to converge. This leads us to instead propose LA-CEG+ which corrects the inner optimization to guarantee global convergence for ρ-comonotone problems when ρ > 1/2L. 4. We test the methods on a suite of synthetic examples and GAN training where we confirm the stabilizing effect. Interestingly, RAPP seems to provide a similar benefit as Lookahead, which suggest that linear interpolation could play a key role also experimentally. An overview of the theoretical results is provided in Table 1 and Figure 5 B. 2 Related work Lookahead The Lookahead algorithm was first introduced for minimization in Zhang et al. (2019). In the context of Federated Averaging in federated learning (Mc Mahan et al., 2017) and the Reptile algorithm in meta-learning (Nichol et al., 2018), the method can be seen as a single worker and single task instance respectively. Analysis for Lookahead was carried out for nonconvex minimization (Wang et al., 2020; Zhou et al., 2021) and a nested variant proposed in (Pushkin & Barba, 2021). Chavdarova et al. (2020) popularized the Lookahead algorithm for minimax training by showing state-of-the-art performance on image generation tasks. Apart from the original local convergence analysis in Chavdarova et al. (2020) and the bilinear case treated in Ha & Kim (2022) we are not aware of any convergence analysis for Lookahead for minimax problems and beyond. Cohypomonotone Cohypomontone problems were first studied in Iusem et al. (2003); Combettes & Pennanen (2004) for proximal point methods and later expanded on in greater detail in Bauschke et al. (2021). The condition was relaxed to the star-variant referred to as the weak Minty variational inequality (MVI) in Diakonikolas et al. (2021) and the extragradient+ algorithm (EG+) was analyzed. The analysis of EG+ was later tightened and extended to the constrained case in Pethick et al. (2022). Proximal point The proximal point method (PP) has a long history. For maximally monotone operators (and thus convex-concave minimax problems) convergence of PP follows from Opial (1967). The first convergence analysis of inexact PP dates back to Rockafellar (1976); Brézis & Lions (1978). It was later shown that convergence also holds for the relaxed inexact PP as defined in (8) (Eckstein & Bertsekas, 1992). In recent times, PP has gained renewed interest due to its success for certain nonmonotone structures. Inexact PP was studied for cohypomontone problems in Iusem et al. (2003). Asymptotic convergence was established of the relaxed inexact PP for a sum of cohypomonotone operators in Combettes & Pennanen (2004), and later considered in Grimmer et al. (2022) without inexactness. Last iterate rates were established for PP in ρ-comonotone problems (with a dependency on ρ) (Gorbunov et al., 2022b). Explicit approximations of PP through a contractive map was used for convex-concave minimax problems in Cevher et al. (2023) and was the original motivation for Mirror Prox of Nemirovski (2004). See Appendix A for additional references in the stochastic setting. Figure 1: Consider minx X maxy Y ϕ(z) with z = (x, y). As opposed to convexconcave minimax problems, the cohypomonotone condition allows the gradients Fz = ( xϕ(z), yϕ(z)) to point away from the solutions (see Appendix B.1 for the relationship between cohypomonotonicity and the weak MVI). This can lead to instability issues for standard algorithms such as the Adam optimizer. We are interested in finding a zero of an operator S : Rd Rd which decomposes into a Lipschitz continuous (but possibly nonmonotone) operator F and a maximally monotone operator A, i.e. find z Rd such that, 0 Sz := Az + Fz. (1) Most relevant in the context of GAN training is that (1) includes constrained minimax problems. Example 3.1. Consider the following minimax problem min x X max y Y ϕ(x, y). (2) The problem can be recast as the inclusion problem (1) by defining the joint iterates z = (x, y), the stacked gradients Fz = ( xϕ(x, y), yϕ(x, y)), and A = (NX , NY) where N denotes the normal cone. As will become clear (cf. Algorithm 1), A will only be accessed through the resolvent JγA := (id +γA) 1 which reduces to the proximal operator. More specifically JγA(z) = (proj X (x), proj Y(y)). We will rely on the following assumptions (see Appendix B for any missing definitions). Assumption 3.2. In problem (1), (i) The operator A : Rd Rd is maximally monotone. (ii) The operator F : Rd Rd is L-Lipschitz, i.e. for some L [0, ), Fz Fz L z z z, z Rd. (iii) The operator S := F + A is ρ-comonotone for some ρ ( 1/2L, ), i.e. v v , z z ρ v v 2 (v, z), (v , z ) grph S. Remark 3.3. Assumption 3.2(iii) is also known as |ρ|-cohypomonotonicity when ρ < 0, which allows for increasing nonmonotonicity as |ρ| grows. See Appendix B.1 for the relationship with weak MVI. When only stochastic feedback ˆFσ( , ξ) is available we make the following classical assumptions. Assumption 3.4. For the operator ˆFσ( , ξ) : Rd Rd the following holds. (i) Unbiased: Eξ ˆFσ(z, ξ) = Fz z Rd. (ii) Bounded variance: Eξ ˆFσ(z, ξ) Fz 2 σ2 z, z Rd. 4 Inexact Krasnosel ski ı-Mann iterations The main work horse we will rely on is the inexact Krasnosel ski ı-Mann (IKM) iteration from monotone operators (also known as the averaged iteration), which acts on an operator T : Rd Rd with inexact feedback, zk+1 = (1 λ)zk + λ(Tzk + ek), (IKM) where λ (0, 1) and ek is a random variable with dependency on all variables up until (and including) k. The operator e Tk : z 7 Tz + ek can crucially be an iterative optimization scheme in itself. This is important, since we can obtain RAPP, LA-GDA and LA-CEG+ by plugging in different optimization routines. In fact, RAPP is derived by taking e Tk to be a (contractive) fixed point iteration in itself, which approximates the resolvent. We note that also the extragradient+ (EG+) method of Diakonikolas et al. (2021), which converges for cohypomonotone and Lipschitz problems, can be seen as a Krasnosel ski ı-Mann iteration on an extragradient step EG(z) = z γF(z γFz) zk+1 = (1 λ)zk + λ EG(zk) (EG+) where λ (0, 1). We provide a proof of EG+ in Theorem G.1 which extends to the constrained case using the construction from Pethick et al. (2022) but through a simpler argument under fixed stepsize. Essentially, the IKM iteration leads to a conservative update that stabilizes the update using the previous iterate. This is the key mechanism behind showing convergence in the nonmonotone setting known as cohypomonotonicity. Very generally, it is possible to provide convergence guarantees for IKM when the following holds (Theorem C.1 is deferred to the appendix due to space limitations). Definition 4.1. An operator T : Rn Rd is said to be quasi-nonexpansive if Tz z z z z Rd, z fix T. (3) Remark 4.2. This notion is crucial to us since the resolvent JB := (id +B) 1 is (quasi)-nonexpansive if B is 1/2-cohypomonotone (Bauschke et al., 2021, Prop. 3.9(iii)). 5 Approximating the resolvent As apparent from Remark 4.2, the IKM iteration would provide convergence to a zero of the cohypomonotone operator S from Assumption 3.2 by using its resolvent T = JγS. However, the update is implicit, so we will instead approximate JγS. Given z Rd we seek z Rd such that z = JγS(z) = (id +γS) 1z = (id +γA) 1(z γFz ) This can be approximated with a fixed point iteration of Qz : w 7 (id +γA) 1(z γFw) (4) which is a contraction for small enough γ since F is Lipschitz continuous. It follows from Banach s fixed-point theorem Banach (1922) that the sequence converges linearly. We formalize this in the following theorem, which additionally applies when only stochastic feedback is available. wt+1 = (id +γA) 1(z γ ˆFσ(wt, ξt)) ξt P (5) Lemma 5.1. Suppose Assumptions 3.2(i), 3.2(ii) and 3.4. Given z Rd, the iterates generated by (5) with γ (0, 1/L) converges to a neighborhood linearly, i.e., E wτ JγS(z) 2 (γL)2τ w0 w 2 + γ2 (1 γL)2 σ2. (6) The resulting update in (5) is identical to GDA but crucially always steps from z. We use this as a subroutine in RAPP to get convergence under a cohypomonotone operator while only suffering a logarithmic factor in the rate. Interpretation In the special case of the constrained minimax problem in (2), the application of the resolvent JγS(z) is equivalent to solving the following optimization problem min x X max y Y n ϕµ(x , y ) := ϕ(x , y ) + 1 2µ y y 2 o . (7) for appropriately chosen µ (0, ). (5) can thus be interpreted as solving a particular regularized subproblem. Later we will drop this regularization to arrive at the Lookahead algorithm. 6 Last iterate under cohypomonotonicity As stated in Section 5, we can obtain convergence using the approximate resolvent through Theorem C.1. The convergence is provided in terms of the average, so additional work is needed for a last iterate result. IKM iteration on the approximate resolvent (i.e. e Tk(z) = JγS(z) + ek) becomes, zk = zk vk with vk γS( zk) (8a) zk+1 = (1 λ)zk + λ( zk + ek) (8b) with λ (0, 1) and γ > 0 and error ek Rd. Without error, (8) reduces to relaxed proximal point zk+1 = (1 λ)zk + λJγS(zk) (Relaxed PP) For a last iterate result it remains to argue that the residual JγS(zk) zk is monotonically decreasing (up to an error we can control). Showing monotonic decrease is fairly straightforward if λ = 1 (see Lemma E.1 and the associated proof). However, we face additional complication due to the averaging, which is apparent both from the proof and the slightly more complicated error term in the following lemma. Lemma 6.1. If S is ρ-comonotone with ρ > γ 2 then (8) satisfies for all z zer S, JγS(zk) zk 2 JγS(zk 1) zk 1 2 + δk(z ) where δk(z) := 4 ek ( zk+1 z + zk z ). The above lemma allows us to obtain last iterate convergence for IKM on the inexact resolvent by combing the lemma with Theorem C.1. Algorithm 1 Relaxed approximate proximal point method (RAPP) Require: z0 Rn λ (0, 1), γ ( 2ρ +, 1/L) Repeat for k = 0, 1, . . . until convergence 1: w0 k = zk 2: for all t = 0, 1, . . . , τ 1 do 3: ξk,t P 4: wt+1 k = (id +γA) 1(zk γ ˆFσk(wt k, ξk,t)) 5: zk+1 = (1 λ)zk + λwτ k Return zk+1 Theorem 6.2 (Last iterate of inexact resolvent). Suppose Assumptions 3.2 and 3.4 with σk. Consider the sequence (zk)k N generated by (8) with λ (0, 1) and ρ > γ 2 . Then, for all z zer S, E[ JγS(z K) z K 2] z0 z 2 + PK 1 k=0 εk(z ) λ(1 λ)K + 1 j=k δj(z ), where εk(z) := 2λE[ ek zk z ] + λ2E[ ek 2] and δk(z) := 4E[ ek ( zk+1 z + zk z )]. Remark 6.3. Notice that the rate in Theorem 6.2 has no dependency on ρ. Specifically, it gets rid of the factor γ/(γ + 2ρ) which Gorbunov et al. (2022b, Thm. 3.2) shows is unimprovable for PP. Theorem 6.2 requires that the iterates stays bounded. In Corollary 6.4 we will assume bounded diameter for simplicity, but it is relatively straightforward to show that the iterates can be guaranteed to be bounded by controlling the inexactness (see Lemma E.2). All that remains to get convergence of the explicit scheme in RAPP, is to expand and simplify the errors εk(z) and δk(z) using the approximation of the resolvent analyzed in Lemma 5.1. Corollary 6.4 (Explicit inexact resolvent). Suppose Assumption 3.2 holds. Consider the sequence (zk)k N generated by RAPP with deterministic feedback and ρ > γ 2 . Then, for all z zer S with D := supj N zj z < , (i) with τ = log K log(1/γL): 1 K PK 1 i=0 JγS(zk) zk 2 = O z0 z 2 λ(1 λ)K + D2 (1 λ)K . (ii) with τ = log K2 log(1/γL): JγS(z K) z K 2 = O z0 z 2 λ(1 λ)K + D2 K + D2 (1 λ)K2 , . Remark 6.5. Corollary 6.4(ii) implies an oracle complexity of O log(ε 2)ε 1 for ensuring that the last iterate satisfies JγS(z K) z K 2 ε. A stochastic extension is provided in Corollary E.3 by taking the batch size increasing. Notice that RAPP, for τ = 2 inner steps, reduces to EG+ in the unconstrained case where A 0. 7 Analysis of Lookahead The update in RAPP leads to a fairly conservative update in the inner loop, since it corresponds to optimizing a highly regularized subproblem as noted in Section 5. Could we instead replace the optimization procedure with gradient descent ascent (GDA)? If we replace the inner optimization routine we recover what is known as the Lookahead (LA) algorithm wt+1 k = wt k γFwt k t = 0, ..., τ 1 zk+1 = (1 λ)zk + λwτ k We empirically demonstrate that this scheme can converge for nonmonotone problems for certain choices of parameters (see Figure 2). However, what global guarantees can we provide theoretically? It turns out that for LA-GDA with two inner steps (τ = 2) we have an affirmative answer. After some algebraic manipulation it is not difficult to see that the update can be simplified as follows 2(zk 2λγFzk) + 1 2(zk 2λγF(zk γFzk)). (9) Figure 2: LA-GDA and RAPP can converge for Hsieh et al. (2021, Ex. 5.2). Interestingly, we can set the stepsize γ larger than 1/L while RAPP remains stable. Approximate proximal point (APP) with the same stepsize diverges (the iterates of APP are deferred to Figure 6). In this example, it is apparent from the rates, that there is a benefit in replacing the conservative inner update in RAPP with GDA in LA-GDA as explored in Section 7. This is the average of GDA and EG+ (when λ (0, 1/2)). This observation allows us to show convergence under cohypomonotonicity. This positive result for nonmonotone problems partially explains the stabilizing effect of LA-GDA. Theorem 7.1. Suppose Assumption 3.2 holds. Consider the sequence (zk)k N generated by LA-GDA with τ = 2, γ 1/L and λ (0, 1/2). Furthermore, suppose that 2ρ > (1 2λ)γ and 2ρ 2λγ (1 γ2L2)γ. (10) Then, for all z zer F, k=0 F zk 2 z0 z 2 λγ (1 2λ)γ + 2ρ K . (11) Remark 7.2. For λ 0 and γ = c/L where c (0, ), sufficient condition reduces to ρ γ(1 γ2L2)/2 = c(1 c2)/2L, of which the minimum is attained with c = 1/ 3, leading to the requirement ρ 1/3 3L. A similar statement is possible for zk. Thus, (LA-GDA) improves on the range of ρ compared with EG (see Table 1). For larger τ, LA-GDA does not necessarily converge (see Figure 3 for a counterexample). We next ask what we would require of the base optimizer to guarantee convergence for any τ. To this end, we replace the inner iteration with some abstract algorithm Alg : Rd Rd, i.e. wt+1 k = Alg(wt k) t = 0, ..., τ 1 zk+1 = (1 λ)zk + λwτ k Convergence follows from quasi-nonexpansiveness. Theorem 7.3. Suppose Alg : Rd Rd is quasi-nonexpansive. Then (zk)k N generated by (LA) converges to some z fix Alg. Remark 7.4. Even though the base optimizer Alg might not converge (since nonexpansiveness is not sufficient), Theorem 7.3 shows that the outer loop converges. Interestingly, this aligns with the benefit observed in practice of using the outer iteration of Lookahead (see Figure 4). Cocoercive From Theorem 7.3 we almost immediately get converge of LA-GDA for coercive problems since V = id γF is nonexpansive iff γF is 1/2-cocoercive. Corollary 7.5. Suppose F is 1/L-cocoercive. Then (zk)k N generated by LA-GDA with γ 2/L converges to some z zer F. Remark 7.6. Corollary 7.5 can trivially be extended to the constrained case by observing that also V = (id +γA) 1(id γF) is nonexpansive when A is maximally monotone. As a special case this captures constrained convex and gradient Lipschitz minimization problems. Figure 3: We test the Lookahead variants on Pethick et al. (2022, Ex. 3(iii)) where ρ ( 1/8L, 1/10L) (left) and Pethick et al. (2022, Ex. 5) with ρ = 1/3 (right). For the left example LA-GDA (provably) converges for τ = 2, but may be nonconvergent for larger τ as illustrate. Both variants of LA-GDA diverges in the more difficult example on the right, while LA-CEG+ in contrast provably converges. It seems that LA-CEG+ trades off a constant slowdown in the rate for convergence in a larger class. Monotone When only monotonicity and Lipschitz holds we may instead consider the following extragradient based version of Lookahead (first empirically investigated in Chavdarova et al. (2020)) wt+1 k = EG(wk t ) t = 0, ..., τ 1 zk+1 = (1 λ)zk + λwτ k where EG(z) = z γF(z γFz). We show in Theorem F.1 that the EG-operator of the inner loop is quasi-nonexpansive, which implies convergence of LA-EG through Theorem 7.3. Theorem F.1 extends even to cases where A 0 by utilizing the forward-backward-forward construction of Tseng (1991). This providing the first global convergence guarantee for Lookahead beyond bilinear games. Cohypomonotone For cohypomonotone problems large τ may prevent LA-GDA from converging (see Figure 3 for a counterexample). Therefore we propose replacing the inner optimization loop in LA-GDA with the method proposed in (Pethick et al., 2022, Alg. 1). Let H = id γF. We can write one step of the inner update with α (0, 1) as CEG+(w) = w + 2α(H w Hw) with w = (id + γA) 1 Hw. (12) The usefulness of the operator CEG+ : Rd Rd comes from the fact that it is quasi-nonexpansive under Assumption 3.2 (see Theorem G.1). Thus, Theorem 7.3 applies even when F is only cohypomonotone if we make the following modification to LA-GDA wt+1 k = CEG+(wt k) t = 0, ..., τ 1 zk+1 = (1 λ)zk + λwτ k In the unconstrained case (A 0) this reduces to using the EG+ algorithm of Diakonikolas et al. (2021) for the inner loop. We have the following convergence guarantee. Corollary 7.7. Suppose Assumption 3.2 holds. Then (zk)k N generated by LA-CEG+ with λ (0, 1), γ ( 2ρ +, 1/L) and α (0, 1 + 2ρ γ ) converges to some z zer S. 8 Experiments This section demonstrates that linear interpolation can lead to an improvement over common baselines. Synthetic examples Figures 2 and 3 demonstrate RAPP, LA-GDA and LA-CEG+ on a host of nonmonotone problems (Hsieh et al. (2021, Ex. 5.2), Pethick et al. (2022, Ex. 3(iii)), Pethick et al. (2022, Ex. 5)). See Appendix H.2 for definitions and further details. Table 2: Adam-based. The combination of Lookahead and extragradient-like methods performs the best. Adam 21.04 2.20 7.61 0.15 Extra Adam 18.23 1.13 7.79 0.08 Extra Adam+ 22.94 1.93 7.65 0.13 LA-Adam 17.63 0.65 7.86 0.07 LA-Extra Adam 15.88 0.67 7.97 0.12 LA-Extra Adam+ 17.86 1.03 8.08 0.15 Table 3: GDA-based. Both RAPP and Lookahead increases the scores substantially. GDA 19.36 0.08 7.84 0.07 EG 18.94 0.60 7.84 0.02 EG+ 19.35 4.28 7.74 0.44 LA-GDA 16.87 0.18 8.01 0.08 LA-EG 16.91 0.66 7.97 0.12 LA-EG+ 17.20 0.44 7.94 0.11 RAPP 17.76 0.82 7.98 0.08 Image generation We replicate the experimental setup of Chavdarova et al. (2020); Miyato et al. (2018), which uses hinge version of the non-saturated loss and a Res Net with spectral normalization for the discriminator (see Appendix H.2 for details). To evaluate the performance we rely on the commonly used Inception score (ISC) (Salimans et al., 2016) and Fréchet inception distance (FID) (Heusel et al., 2017) and report the best iterate. We demonstrate the methods on the CIFAR10 dataset (Krizhevsky et al., 2009). The aim is not to beat the state-of-the-art, but rather to complement the already exhaustive numerical evidence provided in Chavdarova et al. (2020). For a fair computational comparison we count the number of gradient computations instead of iterations k as in Chavdarova et al. (2020). Maybe surprisingly, we find that the extrapolation methods such as EG and RAPP still outperform the baseline, despite having fewer effective iterations. RAPP improves over EG, which suggest that it can be worthwhile to spend more computation on refining the updates at the cost of making fewer updates. Figure 4: Adam eventually diverges on CIFAR10 while Lookahead is stable with the outer iterate enjoying superior performance. The first experiment we conduct matches the setting of Chavdarova et al. (2020) by relying on the Adam optimizer and using and update ratio of 5 : 1 between the discriminator and generator. We find in Table 2 that LA-Extra Adam+ has the highest ISC (8.08) while LA-Extra Adam has the lowest FID (15.88). In contrast, we confirm that Adam is unstable while Lookahead prevents divergence as apparent from Figure 4, which is in agreement with Chavdarova et al. (2020). In addition, the outer loop of Lookahead achieves better empirical performance, which corroborate the theoretical result (cf. Remark 7.4). Notice that Extra Adam+ has slow convergence (without Lookahead), which is possibly due to the 1/2-smaller stepsize. We additionally simplify the setup by using GDA-based optimizers with an update ratio of 1 : 1, which avoids the complexity of diagonal adaptation, gradient history and multiple steps of the discriminator as in the Adam-based experiments. The results are found in Table 3. The learning rates are tuned for GDA and we use those parameters fixed across all other methods. Despite being tuned on GDA, we find that extragradient methods, Lookahead-based methods and RAPP all still outperform GDA in terms of FID. The biggest improvement comes from the linear interpolation based methods Lookahead and RAPP (see Figure 8 for further discussion on EG+). Interesting, the Lookahead-based methods are roughly comparable with their Adam variants (Table 2) while GDA even performs better than Adam. 9 Conclusion & limitations We have precisely characterized the stabilizing effect of linear interpolation by analyzing it under cohypomonotonicity. We proved last iterate convergence rates for our proposed method RAPP. The algorithm is double-looped, which introduces a log factor in the rate as mentioned in Remark E.4. It thus remains open whether last iterate is possible using only τ = 2 inner iterations (for which RAPP reduces to EG+ in the unconstrained case). By replacing the inner solver we subsequently rediscovered and analyzed Lookahead using nonexpansive operators. In that regard, we have only dealt with compositions of operators. It would be interesting to further extend the idea to understanding and developing both Federated Averaging and the meta-learning algorithm Reptile (of which Lookahead can be seen as a single client and single task instance respectively), which we leave for future work. 10 Acknowledgements Thanks to Leello Dadi and Stratis Skoulakis for helpful discussion. This work was supported by Google. This work was supported by Hasler Foundation Program: Hasler Responsible AI (project number 21043). This work was supported by the Swiss National Science Foundation (SNSF) under grant number 200021_205011. This project has received funding from the European Research Council (ERC) under the European Union s Horizon 2020 research and innovation programme (grant agreement n 725594 - time-data). Arjevani, Y., Shalev-Shwartz, S., and Shamir, O. On lower and upper bounds for smooth and strongly convex optimization problems. ar Xiv preprint ar Xiv:1503.06833, 2015. Banach, S. Sur les opérations dans les ensembles abstraits et leur application aux équations intégrales. Fund. math, 3(1):133 181, 1922. Bauschke, H. H. and Combettes, P. L. Convex analysis and monotone operator theory in Hilbert spaces. CMS Books in Mathematics. Springer, 2017. ISBN 978-3-319-48310-8. Bauschke, H. H., Moursi, W. M., and Wang, X. Generalized monotone operators and their averaged resolvents. Mathematical Programming, 189(1):55 74, 2021. Bertsekas, D. P. Incremental proximal methods for large scale convex optimization. Mathematical programming, 129(2):163 195, 2011. Bianchi, P. On the convergence of a stochastic proximal point algorithm. In CAMSAP, 2015. Bravo, M. and Cominetti, R. Stochastic fixed-point iterations for nonexpansive maps: Convergence and error bounds. ar Xiv preprint ar Xiv:2208.04193, 2022. Brézis, H. and Lions, P. L. Produits infinis de résolvantes. Israel Journal of Mathematics, 29(4): 329 345, 1978. Cai, Y., Oikonomou, A., and Zheng, W. Accelerated algorithms for monotone inclusions and constrained nonconvex-nonconcave min-max optimization. ar Xiv preprint ar Xiv:2206.05248, 2022. Cevher, V., Piliouras, G., Sim, R., and Skoulakis, S. Min-max optimization made simple: Approximating the proximal point method via contraction maps, 2023. URL https://arxiv.org/abs/ 2301.03931. Chavdarova, T., Pagliardini, M., Stich, S. U., Fleuret, F., and Jaggi, M. Taming GANs with Lookahead-Minmax. ar Xiv preprint ar Xiv:2006.14567, 2020. Combettes, P. L. Quasi-Fejérian analysis of some optimization algorithms. In Studies in Computational Mathematics, volume 8, pp. 115 152. Elsevier, 2001. Combettes, P. L. and Pennanen, T. Proximal methods for cohypomonotone operators. SIAM journal on control and optimization, 43(2):731 742, 2004. Diakonikolas, J. Halpern iteration for near-optimal and parameter-free monotone inclusion and strong solutions to variational inequalities. In Conference on Learning Theory, pp. 1428 1451. PMLR, 2020. Diakonikolas, J., Daskalakis, C., and Jordan, M. I. Efficient methods for structured nonconvexnonconcave min-max optimization. In International Conference on Artificial Intelligence and Statistics, pp. 2746 2754. PMLR, 2021. Eckstein, J. and Bertsekas, D. P. On the Douglas Rachford splitting method and the proximal point algorithm for maximal monotone operators. Mathematical Programming, 55(1):293 318, 1992. Gidel, G., Berard, H., Vignoud, G., Vincent, P., and Lacoste-Julien, S. A variational inequality perspective on generative adversarial networks. ar Xiv preprint ar Xiv:1802.10551, 2018. Golowich, N., Pattathil, S., Daskalakis, C., and Ozdaglar, A. Last iterate is slower than averaged iterate in smooth convex-concave saddle point problems. In Conference on Learning Theory, pp. 1758 1784. PMLR, 2020. Gorbunov, E., Loizou, N., and Gidel, G. Extragradient method: O (1/k) last-iterate convergence for monotone variational inequalities and connections with cocoercivity. In International Conference on Artificial Intelligence and Statistics, pp. 366 402. PMLR, 2022a. Gorbunov, E., Taylor, A., Horváth, S., and Gidel, G. Convergence of proximal point and extragradientbased methods beyond monotonicity: the case of negative comonotonicity. ar Xiv preprint ar Xiv:2210.13831, 2022b. Grimmer, B., Lu, H., Worah, P., and Mirrokni, V. The landscape of the proximal point method for nonconvex nonconcave minimax optimization. Mathematical Programming, pp. 1 35, 2022. Ha, J. and Kim, G. On convergence of Lookahead in smooth games. In International Conference on Artificial Intelligence and Statistics, pp. 4659 4684. PMLR, 2022. Halpern, B. Fixed points of nonexpanding maps. 1967. Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., and Hochreiter, S. GANs trained by a two time-scale update rule converge to a local Nash equilibrium. Advances in neural information processing systems, 30, 2017. Hsieh, Y.-P., Mertikopoulos, P., and Cevher, V. The limits of min-max optimization algorithms: Convergence to spurious non-critical sets. In International Conference on Machine Learning, pp. 4337 4348. PMLR, 2021. Iusem, A. N., Pennanen, T., and Svaiter, B. F. Inexact variants of the proximal point algorithm without monotonicity. SIAM Journal on Optimization, 13(4):1080 1097, 2003. Korpelevich, G. Extragradient method for finding saddle points and other problems. Matekon, 13(4): 35 49, 1977. Krizhevsky, A., Hinton, G., et al. Learning multiple layers of features from tiny images. 2009. Lee, S. and Kim, D. Fast extra gradient methods for smooth structured nonconvex-nonconcave minimax problems. Advances in Neural Information Processing Systems, 34:22588 22600, 2021. Lieder, F. On the convergence rate of the halpern-iteration. Optimization letters, 15(2):405 418, 2021. Luo, Y. and Tran-Dinh, Q. Extragradient-type methods for co-monotone root-finding problems. Mc Mahan, B., Moore, E., Ramage, D., Hampson, S., and y Arcas, B. A. Communication-efficient learning of deep networks from decentralized data. In Artificial intelligence and statistics, pp. 1273 1282. PMLR, 2017. Miyato, T., Kataoka, T., Koyama, M., and Yoshida, Y. Spectral normalization for generative adversarial networks. ar Xiv preprint ar Xiv:1802.05957, 2018. Nemirovski, A. Prox-method with rate of convergence o (1/t) for variational inequalities with Lipschitz continuous monotone operators and smooth convex-concave saddle point problems. SIAM Journal on Optimization, 15(1):229 251, 2004. Nichol, A., Achiam, J., and Schulman, J. On first-order meta-learning algorithms. ar Xiv preprint ar Xiv:1803.02999, 2018. Obukhov, A., Seitzer, M., Wu, P.-W., Zhydenko, S., Kyl, J., and Lin, E. Y.-J. High-fidelity performance metrics for generative models in Py Torch, 2020. URL https://github.com/toshas/ torch-fidelity. Version: 0.3.0, DOI: 10.5281/zenodo.4957738. Opial, Z. Weak convergence of the sequence of successive approximations for nonexpansive mappings. Bulletin of the American Mathematical Society, 73(4):591 597, 1967. Patrascu, A. and Irofti, P. Stochastic proximal splitting algorithm for composite minimization. Optimization Letters, 15(6):2255 2273, 2021. Patrascu, A. and Necoara, I. Nonasymptotic convergence of stochastic proximal point methods for constrained convex optimization. The Journal of Machine Learning Research, 18(1):7204 7245, 2017. Pethick, T., Latafat, P., Patrinos, P., Fercoq, O., and Cevher, V. Escaping limit cycles: Global convergence for constrained nonconvex-nonconcave minimax problems. In International Conference on Learning Representations, 2022. Pethick, T., Fercoq, O., Latafat, P., Patrinos, P., and Cevher, V. Solving stochastic weak Minty variational inequalities without increasing batch size. In The Eleventh International Conference on Learning Representations, 2023. Pushkin, D. and Barba, L. Multilayer Lookahead: a nested version of Lookahead. ar Xiv preprint ar Xiv:2110.14254, 2021. Rockafellar, R. T. Monotone operators and the proximal point algorithm. SIAM journal on control and optimization, 14(5):877 898, 1976. Salimans, T., Goodfellow, I., Zaremba, W., Cheung, V., Radford, A., and Chen, X. Improved techniques for training GANs. Advances in neural information processing systems, 29, 2016. Toulis, P., Horel, T., and Airoldi, E. M. The proximal Robbins-Monro method. ar Xiv preprint ar Xiv:1510.00967, 2015. Toulis, P., Tran, D., and Airoldi, E. Towards stability and optimality in stochastic gradient descent. In Artificial Intelligence and Statistics, pp. 1290 1298. PMLR, 2016. Tseng, P. Applications of a splitting algorithm to decomposition in convex programming and variational inequalities. SIAM Journal on Control and Optimization, 29(1):119 138, 1991. Wang, J., Tantia, V., Ballas, N., and Rabbat, M. Lookahead converges to stationary points of smooth non-convex functions. In ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 8604 8608. IEEE, 2020. Yoon, T. and Ryu, E. K. Accelerated algorithms for smooth convex-concave minimax problems with o (1/kˆ 2) rate on squared gradient norm. In International Conference on Machine Learning, pp. 12098 12109. PMLR, 2021. Zhang, M., Lucas, J., Ba, J., and Hinton, G. E. Lookahead optimizer: k steps forward, 1 step back. Advances in neural information processing systems, 32, 2019. Zhou, P., Yan, H., Yuan, X., Feng, J., and Yan, S. Towards understanding why lookahead generalizes better than SGD and beyond. Advances in Neural Information Processing Systems, 34:27290 27304, 2021.