# blackbox_generalization_stability_of_zerothorder_learning__4ffba9c8.pdf Black-Box Generalization: Stability of Zeroth-Order Learning Konstantinos E. Nikolakakis Yale University konstantinos.nikolakakis@yale.edu Farzin Haddadpour Yale University farzin.haddadpour@yale.edu Dionysios S. Kalogerias Yale University dionysis.kalogerias@yale.edu Amin Karbasi Yale University & Google Research amin.karbasi@yale.edu We provide the first generalization error analysis for black-box learning through derivative-free optimization. Under the assumption of a Lipschitz and smooth unknown loss, we consider the Zeroth-order Stochastic Search (Zo SS) algorithm, that updates a d-dimensional model by replacing stochastic gradient directions with stochastic differences of K + 1 perturbed loss evaluations per dataset (example) query. For both unbounded and bounded possibly nonconvex losses, we present the first generalization bounds for the Zo SS algorithm. These bounds coincide with those for SGD, and they are independent of d, K and the batch size m, under appropriate choices of a slightly decreased learning rate. For bounded nonconvex losses and a batch size m = 1, we additionally show that both generalization error and learning rate are independent of d and K, and remain essentially the same as for the SGD, even for two function evaluations. Our results extensively extend and consistently recover established results for SGD in prior work, on both generalization bounds and corresponding learning rates. If additionally m = n, where n is the dataset size, we recover generalization guarantees for full-batch GD as well. 1 Introduction Learning methods often rely on empirical risk minimization objectives that highly depend on a limited training data-set. Known gradient-based approaches such as SGD train and generalize effectively in reasonable time [1]. In contrast, emerging applications such as convex bandits [2 4], black-box learning [5], federated learning [6], reinforcement learning [7, 8], learning linear quadratic regulators [9,10], and hyper-parameter tuning [11] stand in need of gradient-free learning algorithms [11 14] due to an unknown loss/model or impossible gradient evaluation. Given two or more function evaluations, zeroth-order algorithms (see, e.g., [14,15]) aim to estimate the true gradient for evaluating and updating model parameters (say, of dimension d). In particular, Zeroth-order Stochastic Search (Zo SS) [13, Corollary 2], [16, Algorithm 1] uses K + 1 function evaluations (K 1), while deterministic zeroth-order approaches [5, Section 3.3] require at least K d + 1 queries. The optimization error of the Zo SS algorithm is optimal as shown in prior work for convex problems [13], and suffers at most a factor of p d/K in the convergence rate as compared with SGD. In addition to the optimization error, the importance of generalization error raises the question of how well zeroth-order algorithms generalize to unseen examples. In this paper, we show that the generalization error of Zo SS essentially coincides with that of SGD, under the choice of a slightly decreased learning rate. Assuming a Lipschitz and smooth loss function, we 36th Conference on Neural Information Processing Systems (Neur IPS 2022). establish generalization guarantees for Zo SS, by extending stability-based analysis for SGD [1], to the gradient-free setting. In particular, we rely on the celebrated result that uniform algorithmic stability implies generalization [1,17,18]. Early works [17,19 22] first introduced the notion of stability, and the connection between (uniform) stability and generalization. Recently, alternative notions of stability and generalization gain attention such as locally elastic stability [23], VC-dimension/flatness measures [24], distributional stability [25 27], information theoretic bounds [16,28 33] mainly based on assuming a sub-Gaussian loss, as well as connections between differential privacy and generalization [34 37]. In close relation to our paper, Hardt et al. [1] first showed uniform stability final-iterate bounds for vanilla SGD. More recent works develop alternative generalization error bounds based on high probability guarantees [38 41] and data-dependent variants [42], or under different assumptions than those of prior works such as as strongly quasi-convex [43], non-smooth convex [44 47], and pairwise losses [48,49]. In the nonconvex case, [50] provide bounds that involve on-average variance of the stochastic gradients. Generalization performance of other algorithmic variants lately gain further attention, including SGD with early momentum [51], randomized coordinate descent [52], look-ahead approaches [53], noise injection methods [54], and stochastic gradient Langevin dynamics [55 62]. Recently, stability and generalization of full-bath GD has also been studied; see, e.g., [63 67]. In particular, Charles and Papailiopoulos. [64] showed instability of GD for nonconvex losses. Still, such instability does not imply a lower bound on the generalization error of GD (in expectation). In fact, Hoffer et al. [63] showed empirically that the generalization of GD is not affected by the batch-size, and for large enough number of iterations GD generalizes comparably to SGD. Our analysis agree with the empirical results of Hoffer et al. [63], as we show that (for smooth losses) the generalization of Zo SS (and thus of SGD) is independent of the batch size. Notation. We denote the training data-set S of size n as {zi}n i=1, where zi are i.i.d. observations of a random variable Z with unknown distribution D. The parameters of the model are vectors of dimension d, denoted by W Rd, and Wt is the output at time t of a (randomized) algorithm AS. The (combined) loss function f( , z) : Rd R+ is uniformly Lipschitz and smooth for all z Z. We denote the Lipschitz constant as L and the smoothness parameter by β. The number of function (i.e., loss) evaluations (required at each iteration of the Zo SS algorithm) is represented by K + 1 N. We denote by f the smoothed approximation of the loss gradient, associated with parameter µ. The parameter Γd K p (3d 1)/K + 1 prominently appears in our results. We denote the gradient of the loss function with respect to model parameters W, by f(W, z) wf(w, z)|w=W . We denote the mini batch at t by Jt, and m |Jt|. 1.1 Contributions Under the assumption of Lipschitz and smooth loss functions, we provide generalization guarantees for black-box learning, extending the analysis of prior work by Hardt et al. [1] to the gradient free setting. In particular, we establish uniform stability and generalization error bounds for the final iterate of the Zo SS algorithm; see Table 1 for a summary of the results. In more detail, the contributions of this work are as follows: For unbounded and bounded losses, we show generalization error bounds identical to SGD, with a slightly decreased learning rate. Specifically, the generalization error bounds are independent of the dimension d, the number of evaluations K and the batch-size m. Further, a large enough number of evaluations (K) provide fast generalization even in the high dimensional regime. For bounded nonconvex losses and single (example) query updates (m = 1), we show that both the Zo SS generalization error and learning rate are independent of d and K, similar to that of SGD [1, Theorem 3.8]. This property guarantees efficient generalization even with two function evaluations. In the full information regime (i.e., when the number of function evaluations K grow to ), the Zo SS generalization bounds also provide guarantees for SGD by recovering the results in prior work [1]. Further, we derive novel SGD bounds for unbounded nonconvex losses, as well as mini-batch SGD for any batch size. Our results subsume generalization guarantees for full-batch Zo SS and GD algorithms. Generalization Error Bounds: Zo SS vs SGD Algorithm Bound NC UB MB Zo SS (this work) αt C/(tΓd K) n (2 + c)CL2 1 Cβ+1 (e T) SGD, αt C/t Hardt et al. [1] n 2CL2 1 Cβ+1 (e T) Zo SS (this work) αt C/t 3e 1 + (Cβ) 1 2 2n 1 + (2 + c)CL2 T (independent of both d and K) Zo SS (this work) Γd K (Γd K 1) SGD, αt C/T Hardt et al. [1] Zo SS (this work) αt C/(TΓd K) (2 + c)L2(e Cβ 1) Zo SS (this work) αt log(1+Cβ) n (proper choice of C in previous bound) Zo SS (this work) αt C/(tΓd K) (2 + c)L2(e T)Cβ n min{C + β 1, C log(T)} Table 1: A list of the generalization error bounds developed herein for Zo SS (Eq. 6) in comparison with SGD, with µ c LΓd K/nβ(3 + d)3/2, for c > 0. In the table, NC and UB stand for nonconvex and unbounded , respectively. MB corresponds to the mini-batch algorithm and for any batch size. Also, αt denotes the stepsize of Zo SS/SGD, and T the total number of iterations. 2 Problem Statement Given a data S {zi}n i=1 of i.i.d samples zi from an unknown distribution D, our goal is to find the parameters w of a learning model such that w arg minw R(w), where R(w) EZ D[f(w, Z)]. Since the distribution D is not known, we consider the empirical risk i=1 f(w, zi), (1) and the corresponding empirical risk minimization (ERM) problem to find w s arg minw RS(w). For a (randomized) algorithm AS with input S and output W = A(S), the excess risk ϵexcess is bounded by the sum of the generalization error ϵgen and the optimization error ϵopt, ϵexcess ES,A[R(W)] R(w ) = ES,A[R(W) RS(W)] | {z } ϵgen +(ES,A[RS(W)] R(w ) | {z } ϵopt To analyze and control ϵgen, we prove uniform stability bounds which imply generalization [1, Theorem 2.2]. Specifically, if for all i.i.d. sequences S, S Zn that differ in one entry, we have supz EA[f(A(S), z) f(A(S ), z)] ϵstab, for some ϵstab > 0, then ϵgen ϵstab. Because the loss is L-Lipschitz, ϵstab may then be chosen as L sup S,S EA A(S) A(S ) . Our primary goal in this work is to develop uniform stability bounds for a gradient-free algorithm AS of the form wt+1 = wt αt fwt,z, where fwt,z only depends on loss function evaluations. To achieve this without introducing unnecessary assumptions, we consider a novel algorithmic stability error decomposition approach. In fact, the stability error introduced at time t by AS breaks down into the stability error of SGD and an approximation error due to missing gradient information. Let Gt( ) and G t( ) be the following SGD update rules Gt(w) w αt f(w, zit), G t(w) w αt f(w, z it), (3) under inputs S, S respectively, and let it {1, 2, . . . , n} be a random index chosen uniformly and independently by the random selection rule of the algorithm, for all t T. Similarly we use the notation G( ) and G ( ) to denote the iteration mappings of AS, i.e., Gt(w) w αt f w,zit, G t(w) w αt f w,z it. (4) Then, as we also discuss later on (Lemma 1), the iterate stability error Gt(w) G t(w ) of AS, for any w, w Rd and for all at t T, may be decomposed as Gt(w) G t(w ) Gt(w) G t(w ) | {z } ϵGBstab + f(w, zit) f w,zit + f(w , z it) f w ,z it | {z } ϵest where ϵGBstab denotes the gradient-based stability error (associated with SGD), and ϵest denotes the gradient approximation error. We now proceed by formally introducing Zo SS. 3 Zeroth-Order Stochastic Search (Zo SS) As a gradient-free alternative of the classical SGD algorithm, we consider the Zo SS scheme, with iterates generated according to the following (single-example update) rule Wt+1 = Wt αt 1 K f(Wt + µU t k, zit) f(Wt, zit) µ U t k, U t k N(0, Id), µ R+, (6) where αt 0 is the corresponding learning rate (for the mini-batch update rule we refer the reader to Section 5). At every iteration t, Zo SS generates K i.i.d. standard normal random vectors U t k, k = 1, . . . , K, and obtains K + 1 loss evaluations on perturbed model inputs. Then Zo SS evaluates a smoothed approximation of the gradient for some µ > 0. In light of the discussion in Section 2, we define the Zo SS smoothed gradient step at time t as f K,µ w,zit f K,µ,Ut w,zit 1 f(w + µU t k, zit) f(w, zit) µ U t k. (7) 3.1 Zo SS Stability Error Decomposition To show stability bounds for Zo SS, we decompose its error into two parts through the stability error decomposition discussed in Section 2. Under the Zo SS update rule, Eq. (5) holds by considering the directions f w,zit and f w ,z it according to Zo SS smoothed approximations (7). Then for any w, w Rd, the iterate stability error Gt(w) G t(w ) of Zo SS at t, breaks down into the gradient based error ϵGBstab and approximation error ϵest. The error term ϵGBstab expresses the stability error of the gradient based mappings [1, Lemma 2.4] and inherits properties related to the SGD update rule. The error ϵest captures the approximation error of the Zo SS smoothed approximation and depends on K and µ. The consistency of the smoothed approximation with respect to SGD follows from lim K ,µ 0 f K,µ w,z = f(w, z) for all w R and z Z. Further, the stability error is also consistent since lim K ,µ 0 |ϵest| = 0. Later on, we use the Zo SS error decomposition in Eq. (5) together with a variance reduction lemma (Lemma 10), to derive exact expressions on the iterate stability error Gt(w) G t(w ) for fixed K and µ > 0 (see Lemma 1). Although in this paper we derive stability bounds and bounds on the ϵgen, the excess risk ϵexcess depends on both errors ϵgen and ϵopt. In the following section, we briefly discuss known results on the ϵopt of zeroth-order methods, including convex and nonconvex losses. 3.2 Optimization Error in Zeroth-Order Stochastic Approximation Convergence rates of the Zo SS optimization error and related zeroth-order variants have been extensively studied in prior works; see e.g., [14,15,68]. For the convex loss setting, when K + 1 function evaluations are available and no other information regarding the loss is given, the Zo SS algorithm achieves optimal rates with respect to the optimization error ϵopt. Specifically, under the assumption of a closed and convex loss, Duchi et al. [13] provided a lower bound for the minimax convergence rate and showed that ϵopt = Ω( p d/K), for any algorithm that approximates the gradient given K + 1 evaluations. In the nonconvex setting Ghadimi et al. [69, 70] established sample complexity guarantees for the zeroth-order approach to reach an approximate stationary point. 4 Main Results For our analysis, we introduce the same assumptions on the loss function (Lipschitz and smooth) as appears in prior work [1]. Additionally, we exploit the η-expansive and σ-bounded properties of the SGD mappings Gt( ) and G t( ) in Eq. (3).1 The mappings Gt( ) and G t( ) are introduced for analysis purposes due to the stability error decomposition given in Eq. (5) and no further assumptions or properties are required for the zeroth-order update rules Gt( ) and G t( ) given in Eq. (4). The η-expansivity of Gt( ) holds for η = 1 + βαt if the loss is nonconvex, and η = 1 if the loss is convex and αt 2/β [1, Lemma 3.6]. Note that Gt( ) is always σ-bounded (σ = Lαt) [1, Lemma 3.3.]. 4.1 Stability Analysis We derive generalization error bounds through uniform stability. To study the stability of Zo SS, we apply a variance reduction lemma that we provide in Appendix A. Exploiting the variance reduction lemma, we show a growth recursion lemma for the iterates of the Zo SS. Lemma 1 (Zo SS Growth Recursion) Consider the sequences of updates { Gt}T t=1 and { G t}T t=1. Let w0 = w 0 be the starting point, wt+1 = Gt(wt) and w t+1 = G t(w t) for any t {1, . . . , T}. Then for any wt, w t Rd and t 0 the following recursion holds E[ Gt(wt) G t(w t) ] K β wt w t + µβαt(3 + d)3/2, if Gt( ) = G t( ), wt w t + 2αt LΓd K + µβαt(3 + d)3/2, if Gt( ) = G t( ). The growth recursion of Zo SS characterizes the stability error that it is introduced by the Zo SS update and according to the outcome of the randomized selection rule at each iteration. Lemma 1 extends growth recursion results for SGD in prior work [1, Lemma 2.5] to the setting of the Zo SS algorithm. If K and µ 0 (while the rest of the parameters are fixed), then Γd K 1, and the statement recovers that of the SGD [1, Lemma 2.5]. Proof of Lemma 1. Let S and S be two samples of size n differing in only a single example, and let Gt( ), G t( ) be the update rules of the Zo SS for each of the sequences S, S respectively. First under the event Et { Gt( ) G t( )} (see Eq. (4)), by applying the Taylor expansion there exist vectors W k,t and W k,t with jth coordinates in the intervals w(j) t , w(j) t + µU (j) k,t w(j) t + µU (j) k,t , w(j) t and w (j) t , w (j) t + µU (j) k,t w (j) t + µU (j) k,t , w (j) t , respectively, such we find that for any wt, w t Rd it is true that Gt(wt) G t(w t) = Gt(wt) Gt(w t) = wt w t αt k=1 f(wt, zit) f(w t, zit), U t k U t k (8) 2 U T k 2 wf(w, zit)|w=W k,t U t k U t k + αt 2 U T k 2 wf(w, zit)|w=W k,t U t k U t k = wt αt f(wt, zit) | {z } G(wt) (w t αt f(w t, zit)) | {z } G (w t) G(w t) 1 [1, Definition 2.3]: An update rule G( ) is η-expansive if G(w) G(w ) η w w for all w, w Rd. If w G(w) σ then it is σ-bounded. 2 U T k 2 wf(w, zit)|w=W k,t U t k U t k + αt 2 U T k 2 wf(w, zit)|w=W k,t U t k U t k k=1 f(wt, zit) f(w t, zit), U t k U t k ( f(wt, zit) f(w t, zit)) . (9) We find (9) by adding and subtracting αt f(wt, zit) and αt f(w t, zit) in Eq. (8). Recall that U t k are independent for all k K, t T and that the mappings G( ) and G ( ) defined in Eq. (9), are η-expansive. The last display and the triangle inequality give E[ Gt(wt) Gt(w t) ] G(wt) G(w t) + 2αt 2 E U t k 3 +αt K E[ f(wt, zit) f(w t, zit) ] (10) η wt w t + 2αt 2 E U t k 3 + αt K β wt w t (11) wt w t + µβαt(3 + d)3/2, (12) where (10) follows from (9) and Lemma 10, and for (11) we applied the η-expansive property of G( ) (see [1, Lemma 2.4 and Lemma 3.6]) and the β-smoothness of the loss function.2 Finally (12) holds since the random vectors U t k N(0, Id) are identically distributed for all k {1, 2, . . . , K} and E U t k 3 (3 + d)3/2. Eq. (12) gives the first part of the recursion. Similar to (9), under the event Ec t { Gt( ) = G t( )}, we find Gt(wt) G t(w t) = wt αt f(wt, zit) | {z } G(wt) w t αt f(w t, z it) | {z } G (w t) 2 U T k 2 wf(w, zit)|w= W k,t U t k U t k + αt 2 U T k 2 wf(w, z it)|w= W k,t U t k U t k k=1 f(wt, zit) f(w t, z it), U t k U t k ( f(wt, zit) f(w t, z it)) . (13) By using the last display, triangle inequality, Lemma 10 and β-smoothness, we find E[ Gt(wt) Gt(w t) ] G(wt) G (w t) + 2αt 2 E[ U t k 3] + αt K E[ f(wt, zit) f(w t, z it) ] min{η, 1}δt + 2σt + 2αt 2 E[ U t k 3] + 2Lαt δt + 2αt LΓd K + µβαt(3 + d)3/2, (15) where (14) follows from the triangle inequality and L Lipschitz condition, while the upper bound on G(wt) G (w t) comes from [1, Lemma 2.4]. Finally, (15) holds since η 1 for both convex and nonconvex losses, σt = Lαt and E U t k 3 (3 + d)3/2 for all k {1, . . . , K}. This shows the second part of recursion. For sake of brevity, let I be an adapted stopping time that corresponds to the first iteration index that the single distinct instance of the two data-sets S, S is sampled by Zo SS. For any t0 {0, 1, . . . , n} we define the event Eδt0 {I > t0} {δt0 = 0}. The next result provides the stability bound. 2For all z Z and W Rd it is true that 2 wf(w, z)|w=W β. Lemma 2 (Zo SS Stability | Nonconvex Loss) Assume that the loss function f( , z) is L-Lipschitz and β-smooth for all z Z. Consider the Zo SS algorithm (6) with final-iterate estimates WT and W T , corresponding to the data-sets S, S , respectively (that differ in exactly one entry). Then the discrepancy δT WT W T , under the event Eδt0, satisfies the inequality E[δT |Eδt0] 2L n Γd K + µβ(3 + d)3/2 T X 1 + βαjΓd K The corresponding bound of Lemma 2 for convex losses is slightly tighter than the bound in (16). Since the two bounds differ only by a constant, the consequent results of Lemma 2 are essentially identical for convex losses as well. We provide the equivalent version of Lemma 2 for convex losses in Appendix B. Proof of Lemma 2. Consider the events Et { Gt( ) G t( )} and Ec t { Gt( ) = G t( )} (see Eq. (4)). Recall that P(Et) = 1 1/n and P(Ec t ) = 1/n for all t T. For any t0 0, a direct application of Lemma 1 gives E[δt+1|Eδt0] = P(Et)E[δt+1|Et, Eδt0] + P(Ec t )E[δt+1|Ec t , Eδt0] E[δt+1|Et, Eδt0] + 1 n E[δt+1|Ec t , Eδt0] n Γd K + µβαt(3 + d)3/2. (17) With Rt (η + αtβ(Γd K 1) + (1 η αtβ(Γd K 1))/n) solving the recursion in (17) gives E[δT |Eδt0] 2L n Γd K + µβ(3 + d)3/2 T X j=t+1 Rj. (18) We consider the last inequality for nonconvex loss functions with η = 1 + βαt and convex loss functions with η = 1 to derive Lemma 2 and Lemma 11 respectively (Appendix B). 4.2 Generalization Error Bounds For the first generalization error bound, we evaluate the right part of the inequality (16) for decreasing step size and bounded nonconvex loss. Then the Lipschitz condition provides a uniform stability condition for the loss and yields the next theorem. Theorem 3 (Nonconvex Bounded Loss | Decreasing Stepsize) Assume that the loss f( , z) [0, 1] is L-Lipschitz and β-smooth for all z Z. Consider the Zo SS update rule (6) with T the total number of iterates, αt C/tΓd K for some (fixed) C > 0 and for all t T, and fixed µ c LΓd K/nβ(3 + d)3/2 for some c > 0. Then the generalization error of Zo SS is bounded by (2 + c)CL2 1 Cβ+1 (e T) 1, 1 + (Cβ) 1 eβC 1 + (Cβ) 1 (2 + c)CL2 1 Cβ+1 Cβ Cβ+1 . (20) Inequality (19), as a tighter version of (20), provides a meaningful bound in marginal cases, i.e., lim β 0 E [|f(WT , z) f(W T , z)|] (2 + c)CL2 n max log e T (2 + c)CL2 By neglecting the negative term in (19) we find (20), that is the Zo SS equivalent of SGD [1, Theorem 3.8]. When K and c 0, then Γd K 1, and the inequalities (19), (20) reduce to a generalization bound for SGD. Inequality (20) matches that of [1, Theorem 3.8], and (19) provides a tighter generalization bound for SGD as well. We show Theorem 3 in Appendix A. Next, we provide a bound on the generalization error for nonconvex losses that comes directly from Theorem 3. In contrast to Theorem 3, the next result provides learning rate and a generalization error bounds, both of which are independent of the dimension and the number of function evaluations. Corollary 4 Assume that the loss function f( , z) [0, 1] is L-Lipschitz and β-smooth for all z Z. Consider the Zo SS update rule (6) with µ c LΓd K/(nβ(3 + d)3/2), T the total number of iterates, and αt C/t for some (fixed) C > 0 and for all t T. Then the generalization error of Zo SS is bounded by |ϵgen| 1 + (βC) 1 2 1 + (2 + c)CL2 3Te As a consequence, even in the high dimensional regime d , two function evaluations (i.e., K = 1) are sufficient for the Zo SS to achieve ϵgen = O(T/n), with the learning rate being no smaller than that of SGD. We continue by providing the proof of Theorem 3. For the proof of Corollary 4, see Appendix A.3. In light of Theorem 3 and Corollary 4, we observe that the over-fitting phenomenon occurs in the gradient-free approach similarly to gradient-based algorithms. For general nonconvex (and convex) losses under standard step-size choices, the generalization error increases with respect to T. Further, the effect of β affects both the stability (similarly to SGD in prior work) of the algorithm and the error approximation of the Zo SS. If β is large, then the expected approximation error (due to limited function evaluations) is also large [15] and the dependence on smoothness is unavoidable in blackbox learning. In our results, this is expressed through the Growth Recursion of Zo SS (Lemma 2), that involves both the stability and approximation error per iteration. However, a smaller step-size (αt = 1/2tβΓd K) mitigates the effect of β on the bound. We refer the reader to Appendix E for a unified analysis of the excess risk, that captures the over-fitting and under-fitting trade-off. Additionally, the number of iterations T is considered to be fixed and known (as in prior works including on average and high probability results on generalization). This is reasonable and quite standard because given the theoretical results, we know beforehand the appropriate choices of T that provide a good trade-off between generalization and optimization. A classical setting is that of a fixed step-size αt = 1/T with T = n, which provides the well known generalization error bound for SGD with order O(1/ n), as appears in very recent and timely prior works [32, Section 3.1], [45]. In the unbounded loss case, we apply Lemma 2 by setting t0 = 0 (recall that t0 is a free parameter, while the algorithm depends on the random variable I). The next result provides a generalization error bound for the Zo SS algorithm with constant step size. In the first case of the theorem, we also consider the convex loss as a representative result, as we show the same bound holds for an appropriate choice of greater learning rate than the learning rate of the nonconvex case. The convex case for the rest of the results of this work can be similarly derived. Theorem 5 (Unbounded Loss | Constant Step Size) Assume that the loss f( , z) is L-Lipschitz, βsmooth for all z Z. Consider the Zo SS update rule (6) with µ c LΓd K/(nβ(3 + d)3/2) for some c > 0. Let T be the total number of iterates and for any t T, if f( , z) is convex for all z Z and αt min{log 1 + Cβ(1 1/Γd K) /Tβ(Γd K 1), 2/β}, or if f( , z) is nonconvex and αt log 1 + Cβ /TβΓd K, for C > 0 then |ϵgen| (2 + c)CL2 if f( , z) is nonconvex and αt C/TΓd K, for some C > 0, then |ϵgen| L2 (2 + c) (e Cβ 1) For the proof of Theorem 5 see Appendix A.4. In the following, we present the generalization error of Zo SS for an unbounded loss with a decreasing step size. Recall that the results for unbounded nonconvex loss also hold for the case of a convex loss with similar bounds on the generalization error and learning rate (see the first case of Theorem 5). Theorem 6 (Unbounded Loss | Decreasing Step Size) Assume that the loss f( , z) is L-Lipschitz, β-smooth for all z Z. Consider Zo SS with update rule (6), T the total number of iterates, αt C/tΓd K for all t T and for some C > 0, and µ c LΓd K/(nβ(3 + d)3/2) for some c > 0. Then the generalization error of Zo SS is bounded by |ϵgen| (2 + c)L2(e T)Cβ n min C + β 1, C log(e T) . (25) For the proof of Theorem 6 see Appendix A.5. Note that the constant C is free and controls the learning rate. Furthermore, it quantifies the trade-off between the speed of training and the generalization of the algorithm. In the next section, we consider the Zo SS algorithm with a minibatch of size m for which we provide generalization error bounds. These results hold under the assumption of unbounded loss and for any batch size m including the case m = 1. 5 Generalization of Mini-Batch Zo SS For the mini-batch version of Zo SS, at each iteration t, the randomized selection rule (uniformly) samples a batch Jt of size m and evaluates the new direction of the update by averaging the smoothed approximation f K,µ w,z over the samples z Jt as f K,µ w,Jt f K,µ,Ut f(w + µU t k,i, z Jt,i) f(w, z Jt,i) µ U t k,i, (26) where U t k,i N(0, Id) are i.i.d. (standard normal), and µ R+. The update rule of the minibatch Zo SS is Wt+1 = Wt αt f K,µ Wt,Jt for all t T, and we define GJt(w) w αt f K,µ w,Jt, G J t(w) w αt f K,µ w,J t for Jt S and J t S respectively. Due to space limitation, we refer the reader to Appendix C for the detailed stability analysis of Zo SS with mini-batch. Specifically, we prove a growth recursion lemma for the mini-batch Zo SS updates (see Appendix C.1 for proof). Lemma 7 (Mini-Batch Zo SS Growth Recursion) Consider the sequences of updates { GJt}T t=1 and { G Jt}T t=1 and µ c LΓd K/(nβ(3 + d)3/2). Let w0 = w 0 be the starting point, wt+1 = GJt(wt) and w t+1 = G Jt(w t) for any t {1, . . . , T}. Then for any wt, w t Rd and t 0 the following recursion holds E[ GJt(wt) G Jt(w t) ] 1 + βαtΓd K δt + c Lαt n Γd K if GJt( ) = G Jt( ) 1 + m 1 m βαtΓd K δt + 2Lαt m Γd K + c Lαt n Γd K if GJt( ) = G Jt( ). Although the iterate stability error (at time t) in the growth recursion depends on the batch size m under the event { GJt( ) = G Jt( )}, the stability bound on the final iterates is independent of m, and coincides with the single example updates (m = 1, Lemma 2). Herein, we provide an informal statement of the result. Lemma 8 (Mini-Batch Zo SS Stability | Nonconvex Loss) Consider the mini-batch Zo SS with any batch size m n, and iterates Wt+1 = Wt αt f K,µ Wt,Jt, W t+1 = W t αt f K,µ W t,J t, for all t T, with respect to the sequences S, S . Then the stability error δT satisfies the inequality of Lemma 2. We refer the reader to Appendix Section C.1, Theorem 14 for the formal statement of the result.3 Through the Lipschitz condition of the loss and Lemma 8, we show that the mini-batch Zo SS enjoys the same generalization error bounds as in the case of single-query Zo SS (m = 1). As a consequence, the batch size does not affect the generalization error. Theorem 9 (Mini-batch Zo SS | Generalization Error) Let the loss function f( , z) be L-Lipschitz and β-smooth (possibly nonconvex, possibly unbounded) for all z Z. Then the bounds of Theorem 5 and Theorem 6 hold for the mini-batch Zo SS with iterate Wt+1 = Wt αt f K,µ Wt,Jt, for all t T and any batch size m n. 3As in the single-query (m = 1) Zo SS, under the assumption of convex loss, the stability error of mini-batch Zo SS satisfies the inequality (46), Appendix B, Lemma 11. By letting K and c 0, the generalization error bounds of mini-batch Zo SS reduce to those of mini-batch SGD, extending results of the single-query (m = 1) SGD that appeared in prior work [1]. Additionally, once K , c 0 and m = n we obtain generalization guarantees for full-batch GD. For the sake of clarity and completeness we provide dedicated stability and generalization analysis of full-batch GD in Appendix D, Corollary 15. 6 Discussion: Black-box Adversarial Attack Design and Future Work A standard, well-cited example of Zo SS application is adversarial learning as considered in [5], when the gradient is not known for the adversary (for additional applications for instance federated/reinforcement learning, linear quadratic regulators; see also Section 1 for additional references). Notice that the algorithm in [5] is restrictive in the high dimensional regime since it requires 2d function evaluations per iteration. In contrast, Zo SS can be considered with any K 2 functions evaluations (the trade-off is between accuracy and resource allocation, which is also controlled through K). If K = d + 1 evaluations are available we recover guarantees for the deterministic zeroth-order approaches (similar to [5]). Retrieving a large number of function evaluations often is not possible in practice. When a limited amount of function evaluations is available, the adversary obtains the solution (optimal attack) with an optimization error that scales by a factor of p d/K, and the generalization error of the attack is of the order T/n under appropriate choices of the step-size, the smoothing parameter µ and K. Fine tuning of the these parameters might be useful in practice, but in general K should be chosen as large as possible. In contrast, µ should be small and satisfy the inequality µ c LΓd K/nβ(3 + d)3/2 (Theorem 6). For instance, in practice µ is often chosen between 10 10 and 10 8 (or even lower) and the Zo SS algorithm remains (numerically) stable. For neural networks with smooth activation functions [71 73], the Zo SS algorithm does not require the smoothness parameter β to be necessarily known, however if β is large then the guarantees of the estimated model would be pessimistic. To ensure that the learning procedure is successful, the adversary can approximate β (since the loss is not known) by estimating the (largest eigenvalue of the) Hessian through the available function evaluations [74, Section 4.1]. Although the non-smooth (convex) loss setting lies out of the scope of this work, it is expected to inherit properties and rates of the SGD for non-smooth losses (at least for sufficiently small smoothing parameter µ). In fact, [45, page 3, Table 1] developed upper and lower bounds for the SGD in the non-smooth case, and they showed that standard step-size choices provide vacuous stability bound. Due to these inherent issues of non-smooth (and often convex only cases), the generalization error analysis of Zo SS for non-smooth losses remains open. Finally, information-theoretic generalization error bounds of Zo SS can potentially provide further insight into the problem, due to the noisy updates of the algorithm, and consist part of future work. 7 Conclusion In this paper, we characterized the generalization ability of black-box learning models. Specifically, we considered the Zeroth-order Stochastic Search (Zo SS) algorithm, which evaluates smoothed approximations of the unknown gradient of the loss by only relying on K + 1 loss evaluations. Under the assumptions of a Lipschitz and smooth (unknown) loss, we showed that the Zo SS algorithm achieves the same generalization error bounds as that of SGD, while the learning rate is slightly decreased compared to that of SGD. The efficient generalization ability of Zo SS, together with strong optimality results related to the optimization error by Duchi et al. [13], makes it a robust and powerful algorithm for a variety of black-box learning applications and problems. [1] Moritz Hardt, Ben Recht, and Yoram Singer. Train faster, generalize better: Stability of stochastic gradient descent. In Maria Florina Balcan and Kilian Q. Weinberger, editors, Proceedings of The 33rd International Conference on Machine Learning, volume 48 of Proceedings of Machine Learning Research, pages 1225 1234, New York, New York, USA, 20 22 Jun 2016. PMLR. URL: https://proceedings.mlr.press/v48/hardt16.html. [2] Alekh Agarwal, Dean P Foster, Daniel J Hsu, Sham M Kakade, and Alexander Rakhlin. Stochastic convex optimization with bandit feedback. In J. Shawe-Taylor, R. Zemel, P. Bartlett, F. Pereira, and K. Q. Weinberger, editors, Advances in Neural Information Processing Systems, volume 24. Curran Associates, Inc., 2011. URL: https://proceedings.neurips.cc/ paper/2011/file/67e103b0761e60683e83c559be18d40c-Paper.pdf. [3] Ohad Shamir. An optimal algorithm for bandit and zero-order convex optimization with twopoint feedback. The Journal of Machine Learning Research, 18(1):1703 1713, 2017. URL: https://www.jmlr.org/papers/volume18/16-632/16-632.pdf. [4] Arya Akhavan, Massimiliano Pontil, and Alexandre Tsybakov. Exploiting higher order smoothness in derivative-free optimization and continuous bandits. In H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 9017 9027. Curran Associates, Inc., 2020. URL: https://proceedings.neurips.cc/paper/2020/file/ 6646b06b90bd13dabc11ddba01270d23-Paper.pdf. [5] Pin-Yu Chen, Huan Zhang, Yash Sharma, Jinfeng Yi, and Cho-Jui Hsieh. Zoo: Zeroth order optimization based black-box attacks to deep neural networks without training substitute models. In Proceedings of the 10th ACM workshop on artificial intelligence and security, pages 15 26, 2017. URL: https://dl.acm.org/doi/10.1145/3128572.3140448. [6] Zan Li and Li Chen. Communication-efficient decentralized zeroth-order method on heterogeneous data. In 2021 13th International Conference on Wireless Communications and Signal Processing (WCSP), pages 1 6, 2021. doi:10.1109/WCSP52459.2021.9613620. [7] Anirudh Vemula, Wen Sun, and J. Bagnell. Contrasting exploration in parameter and action space: A zeroth-order optimization perspective. In Kamalika Chaudhuri and Masashi Sugiyama, editors, Proceedings of the Twenty-Second International Conference on Artificial Intelligence and Statistics, volume 89 of Proceedings of Machine Learning Research, pages 2926 2935. PMLR, 16 18 Apr 2019. URL: https://proceedings.mlr.press/v89/vemula19a.html. [8] Harshat Kumar, Dionysios S. Kalogerias, George J. Pappas, and Alejandro Ribeiro. Actoronly deterministic policy gradient via zeroth-order gradient oracles in action space. In 2021 IEEE International Symposium on Information Theory (ISIT), pages 1676 1681, 2021. doi: 10.1109/ISIT45174.2021.9518023. [9] Dhruv Malik, Ashwin Pananjady, Kush Bhatia, Koulik Khamaru, Peter Bartlett, and Martin Wainwright. Derivative-free methods for policy optimization: Guarantees for linear quadratic systems. In Kamalika Chaudhuri and Masashi Sugiyama, editors, Proceedings of the Twenty Second International Conference on Artificial Intelligence and Statistics, volume 89 of Proceedings of Machine Learning Research, pages 2916 2925. PMLR, 16 18 Apr 2019. URL: https://proceedings.mlr.press/v89/malik19a.html. [10] Hesameddin Mohammadi, Mahdi Soltanolkotabi, and Mihailo R. Jovanovi c. On the linear convergence of random search for discrete-time LQR. IEEE Control Systems Letters, 5(3):989 994, 2021. doi:10.1109/LCSYS.2020.3006256. [11] Jeremy Rapin and Olivier Teytaud. Nevergrad - A gradient-free optimization platform. https: //Git Hub.com/Facebook Research/Nevergrad, 2018. [12] Luis Miguel Rios and Nikolaos V Sahinidis. Derivative-free optimization: a review of algorithms and comparison of software implementations. Journal of Global Optimization, 56(3):1247 1293, 2013. URL: https://link.springer.com/article/10.1007/s10898-012-9951-y. [13] John C. Duchi, Michael I. Jordan, Martin J. Wainwright, and Andre Wibisono. Optimal rates for zero-order convex optimization: The power of two function evaluations. IEEE Transactions on Information Theory, 61(5):2788 2806, 2015. doi:10.1109/TIT.2015.2409256. [14] Yurii Nesterov and Vladimir Spokoiny. Random gradient-free minimization of convex functions. Foundations of Computational Mathematics, 17(2):527 566, 2017. URL: https: //link.springer.com/article/10.1007/s10208-015-9296-2. [15] Krishnakumar Balasubramanian and Saeed Ghadimi. Zeroth-order nonconvex stochastic optimization: Handling constraints, high dimensionality, and saddle points. Foundations of Computational Mathematics, pages 1 42, 2021. URL: https://link.springer.com/article/ 10.1007/s10208-021-09499-8. [16] Mahdi Haghifam, Gintare Karolina Dziugaite, Shay Moran, and Dan Roy. Towards a unified information-theoretic framework for generalization. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, volume 34, pages 26370 26381. Curran Associates, Inc., 2021. URL: https://proceedings.neurips.cc/paper/2021/file/ ddbc86dc4b2fbfd8a62e12096227e068-Paper.pdf. [17] Olivier Bousquet and André Elisseeff. Stability and generalization. The Journal of Machine Learning Research, 2:499 526, 2002. URL: https://www.jmlr.org/papers/v2/ bousquet02a.html. [18] Shai Shalev-Shwartz, Ohad Shamir, Nathan Srebro, and Karthik Sridharan. Learnability, stability and uniform convergence. The Journal of Machine Learning Research, 11:2635 2670, 2010. URL: https://www.jmlr.org/papers/volume11/shalev-shwartz10a/shalevshwartz10a.pdf. [19] Luc P. Devroye and et al. Distribution-free performance bounds for potential function rules, 1979. URL: https://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.128.8772. [20] Michael Kearns and Dana Ron. Algorithmic stability and sanity-check bounds for leaveone-out cross-validation. Neural Computation, 11(6):1427 1453, 1999. doi:10.1162/ 089976699300016304. [21] Sayan Mukherjee, Partha Niyogi, Tomaso Poggio, and Ryan Rifkin. Learning theory: stability is sufficient for generalization and necessary and sufficient for consistency of empirical risk minimization. Advances in Computational Mathematics, 25(1):161 193, 2006. URL: https: //link.springer.com/article/10.1007/s10444-004-7634-z. [22] Olivier Bousquet and André Elisseeff. Algorithmic stability and generalization performance. In T. Leen, T. Dietterich, and V. Tresp, editors, Advances in Neural Information Processing Systems, volume 13. MIT Press, 2001. URL: https://proceedings.neurips.cc/paper/ 2000/file/49ad23d1ec9fa4bd8d77d02681df5cfa-Paper.pdf. [23] Zhun Deng, Hangfeng He, and Weijie Su. Toward better generalization bounds with locally elastic stability. In Marina Meila and Tong Zhang, editors, Proceedings of the 38th International Conference on Machine Learning, volume 139 of Proceedings of Machine Learning Research, pages 2590 2600. PMLR, 18 24 Jul 2021. URL: https://proceedings.mlr.press/v139/ deng21b.html. [24] Yiding Jiang*, Behnam Neyshabur*, Hossein Mobahi, Dilip Krishnan, and Samy Bengio. Fantastic generalization measures and where to find them. In International Conference on Learning Representations, 2020. URL: https://openreview.net/forum?id=SJg IPJBFv H. [25] Raef Bassily, Kobbi Nissim, Adam Smith, Thomas Steinke, Uri Stemmer, and Jonathan Ullman. Algorithmic stability for adaptive data analysis. In Proceedings of the Forty-Eighth Annual ACM Symposium on Theory of Computing, STOC 16, page 1046 1059, New York, NY, USA, 2016. Association for Computing Machinery. doi:10.1145/2897518.2897566. [26] Jonathan Ullman, Adam Smith, Kobbi Nissim, Uri Stemmer, and Thomas Steinke. The limits of post-selection generalization. In S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa Bianchi, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 31. Curran Associates, Inc., 2018. URL: https://proceedings.neurips.cc/paper/2018/ file/77ee3bc58ce560b86c2b59363281e914-Paper.pdf. [27] Christopher Jung, Katrina Ligett, Seth Neel, Aaron Roth, Saeed Sharifi-Malvajerdi, and Moshe Shenfeld. A New Analysis of Differential Privacy s Generalization Guarantees (Invited Paper), page 9. Association for Computing Machinery, New York, NY, USA, 2021. URL: https: //doi.org/10.1145/3406325.3465358. [28] Thomas Steinke and Lydia Zakynthinou. Reasoning About Generalization via Conditional Mutual Information. In Jacob Abernethy and Shivani Agarwal, editors, Proceedings of Thirty Third Conference on Learning Theory, volume 125 of Proceedings of Machine Learning Research, pages 3437 3452. PMLR, 09 12 Jul 2020. URL: https://proceedings.mlr.press/v125/ steinke20a.html. [29] Ibrahim Alabdulmohsin. An Information-Theoretic Route from Generalization in Expectation to Generalization in Probability. In Aarti Singh and Jerry Zhu, editors, Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, volume 54 of Proceedings of Machine Learning Research, pages 92 100. PMLR, 20 22 Apr 2017. URL: https:// proceedings.mlr.press/v54/alabdulmohsin17a.html. [30] Aolin Xu and Maxim Raginsky. Information-theoretic analysis of generalization capability of learning algorithms. In I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc., 2017. URL: https://proceedings.neurips.cc/ paper/2017/file/ad71c82b22f4f65b9398f76d8be4c615-Paper.pdf. [31] Yuheng Bu, Shaofeng Zou, and Venugopal V. Veeravalli. Tightening mutual information based bounds on generalization error. In 2019 IEEE International Symposium on Information Theory (ISIT), pages 587 591, 2019. doi:10.1109/ISIT.2019.8849590. [32] Gergely Neu, Gintare Karolina Dziugaite, Mahdi Haghifam, and Daniel M. Roy. Informationtheoretic generalization bounds for stochastic gradient descent. In Mikhail Belkin and Samory Kpotufe, editors, Proceedings of Thirty Fourth Conference on Learning Theory, volume 134 of Proceedings of Machine Learning Research, pages 3526 3545. PMLR, 15 19 Aug 2021. URL: https://proceedings.mlr.press/v134/neu21a.html. [33] Hrayr Harutyunyan, Maxim Raginsky, Greg Ver Steeg, and Aram Galstyan. Informationtheoretic generalization bounds for black-box learning algorithms. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, volume 34, pages 24670 24682. Curran Associates, Inc., 2021. URL: https://proceedings.neurips.cc/paper/2021/file/ cf0d02ec99e61a64137b8a2c3b03e030-Paper.pdf. [34] Fengxiang He, Shaopeng Fu, Bohan Wang, and Dacheng Tao. Robustness, privacy, and generalization of adversarial training. ar Xiv preprint ar Xiv:2012.13573, 2020. URL: https: //arxiv.org/abs/2012.13573. [35] Fengxiang He, Bohan Wang, and Dacheng Tao. Tighter generalization bounds for iterative differentially private learning algorithms. In Cassio de Campos and Marloes H. Maathuis, editors, Proceedings of the Thirty-Seventh Conference on Uncertainty in Artificial Intelligence, volume 161 of Proceedings of Machine Learning Research, pages 802 812. PMLR, 27 30 Jul 2021. URL: https://proceedings.mlr.press/v161/he21a.html. [36] Zhenhuan Yang, Yunwen Lei, Siwei Lyu, and Yiming Ying. Stability and differential privacy of stochastic gradient descent for pairwise learning with non-smooth loss. In Arindam Banerjee and Kenji Fukumizu, editors, Proceedings of The 24th International Conference on Artificial Intelligence and Statistics, volume 130 of Proceedings of Machine Learning Research, pages 2026 2034. PMLR, 13 15 Apr 2021. URL: https://proceedings.mlr.press/v130/ yang21c.html. [37] Puyu Wang, Zhenhuan Yang, Yunwen Lei, Yiming Ying, and Hai Zhang. Differentially private empirical risk minimization for auc maximization. Neurocomputing, 461:419 437, 2021. [38] Vitaly Feldman and Jan Vondrak. Generalization bounds for uniformly stable algorithms. In S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 31. Curran Associates, Inc., 2018. URL: https://proceedings.neurips.cc/paper/2018/file/ 05a624166c8eb8273b8464e8d9cb5bd9-Paper.pdf. [39] Vitaly Feldman and Jan Vondrak. High probability generalization bounds for uniformly stable algorithms with nearly optimal rate. In Alina Beygelzimer and Daniel Hsu, editors, Proceedings of the Thirty-Second Conference on Learning Theory, volume 99 of Proceedings of Machine Learning Research, pages 1270 1279. PMLR, 25 28 Jun 2019. URL: https://proceedings.mlr.press/v99/feldman19a.html. [40] Liam Madden, Emiliano Dall Anese, and Stephen Becker. High probability convergence and uniform stability bounds for nonconvex stochastic gradient descent. ar Xiv e-prints, pages ar Xiv 2006, 2020. URL: https://arxiv.org/abs/2006.05610. [41] Yegor Klochkov and Nikita Zhivotovskiy. Stability and deviation optimal risk bounds with convergence rate o(1/n). In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, volume 34, pages 5065 5076. Curran Associates, Inc., 2021. URL: https://proceedings.neurips.cc/paper/ 2021/file/286674e3082feb7e5afb92777e48821f-Paper.pdf. [42] Ilja Kuzborskij and Christoph Lampert. Data-dependent stability of stochastic gradient descent. In Jennifer Dy and Andreas Krause, editors, Proceedings of the 35th International Conference on Machine Learning, volume 80 of Proceedings of Machine Learning Research, pages 2815 2824. PMLR, 10 15 Jul 2018. URL: https://proceedings.mlr.press/v80/ kuzborskij18a.html. [43] Robert Mansel Gower, Nicolas Loizou, Xun Qian, Alibek Sailanbayev, Egor Shulgin, and Peter Richtárik. SGD: General analysis and improved rates. In Kamalika Chaudhuri and Ruslan Salakhutdinov, editors, Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research, pages 5200 5209. PMLR, 09 15 Jun 2019. URL: https://proceedings.mlr.press/v97/qian19b.html. [44] Vitaly Feldman. Generalization of erm in stochastic convex optimization: The dimension strikes back. In D. Lee, M. Sugiyama, U. Luxburg, I. Guyon, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 29. Curran Associates, Inc., 2016. URL: https://proceedings.neurips.cc/paper/2016/file/ 8c01a75941549a705cf7275e41b21f0d-Paper.pdf. [45] Raef Bassily, Vitaly Feldman, Cristóbal Guzmán, and Kunal Talwar. Stability of stochastic gradient descent on nonsmooth convex losses. In H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 4381 4391. Curran Associates, Inc., 2020. URL: https://proceedings.neurips.cc/ paper/2020/file/2e2c4bf7ceaa4712a72dd5ee136dc9a8-Paper.pdf. [46] Yunwen Lei and Yiming Ying. Fine-grained analysis of stability and generalization for stochastic gradient descent. In Hal Daumé III and Aarti Singh, editors, Proceedings of the 37th International Conference on Machine Learning, volume 119 of Proceedings of Machine Learning Research, pages 5809 5819. PMLR, 13 18 Jul 2020. URL: https: //proceedings.mlr.press/v119/lei20c.html. [47] Yunwen Lei, Ting Hu, and Ke Tang. Generalization performance of multi-pass stochastic gradient descent with convex loss functions. The Journal of Machine Learning Research, 22(25):1 41, 2021. URL: http://jmlr.org/papers/v22/19-716.html. [48] Yunwen Lei, Mingrui Liu, and Yiming Ying. Generalization guarantee of SGD for pairwise learning. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, volume 34, pages 21216 21228. Curran Associates, Inc., 2021. URL: https://proceedings.neurips.cc/paper/2021/ file/b1301141feffabac455e1f90a7de2054-Paper.pdf. [49] Yunwen Lei, Antoine Ledent, and Marius Kloft. Sharper generalization bounds for pairwise learning. In H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 21236 21246. Curran Associates, Inc., 2020. URL: https://proceedings.neurips.cc/paper/2020/file/ f3173935ed8ac4bf073c1bcd63171f8a-Paper.pdf. [50] Yi Zhou, Yingbin Liang, and Huishuai Zhang. Understanding generalization error of SGD in nonconvex optimization. Machine Learning, pages 1 31, 2021. URL: https: //link.springer.com/article/10.1007/s10994-021-06056-w. [51] Ali Ramezani-Kebrya, Ashish Khisti, and Ben Liang. On the generalization of stochastic gradient descent with momentum. ar Xiv preprint ar Xiv:2102.13653, 2021. [52] Puyu Wang, Liang Wu, and Yunwen Lei. Stability and generalization for randomized coordinate descent. ar Xiv preprint ar Xiv:2108.07414, 2021. URL: https://arxiv.org/abs/ 2108.07414. [53] Pan Zhou, Hanshu Yan, Xiaotong Yuan, Jiashi Feng, and Shuicheng Yan. Towards understanding why lookahead generalizes better than sgd and beyond. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, volume 34, pages 27290 27304. Curran Associates, Inc., 2021. URL: https://proceedings.neurips.cc/paper/2021/file/ e53a0a2978c28872a4505bdb51db06dc-Paper.pdf. [54] Yue Xing, Qifan Song, and Guang Cheng. On the algorithmic stability of adversarial training. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, volume 34, pages 26523 26535. Curran Associates, Inc., 2021. URL: https://proceedings.neurips.cc/paper/2021/file/ df1f1d20ee86704251795841e6a9405a-Paper.pdf. [55] Ankit Pensia, Varun Jog, and Po-Ling Loh. Generalization error bounds for noisy, iterative algorithms. In 2018 IEEE International Symposium on Information Theory (ISIT), pages 546 550, 2018. doi:10.1109/ISIT.2018.8437571. [56] Wenlong Mou, Liwei Wang, Xiyu Zhai, and Kai Zheng. Generalization bounds of SGLD for non-convex learning: Two theoretical viewpoints. In Sébastien Bubeck, Vianney Perchet, and Philippe Rigollet, editors, Proceedings of the 31st Conference On Learning Theory, volume 75 of Proceedings of Machine Learning Research, pages 605 638. PMLR, 06 09 Jul 2018. URL: https://proceedings.mlr.press/v75/mou18a.html. [57] Jian Li, Xuanyuan Luo, and Mingda Qiao. On generalization error bounds of noisy gradient methods for non-convex learning. In International Conference on Learning Representations, 2020. URL: https://openreview.net/forum?id=Skxxtg HKPS. [58] Jeffrey Negrea, Mahdi Haghifam, Gintare Karolina Dziugaite, Ashish Khisti, and Daniel M. Roy. Information-theoretic generalization bounds for SGLD via data-dependent estimates. In H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 32. Curran Associates, Inc., 2019. URL: https://proceedings.neurips.cc/paper/2019/file/ 05ae14d7ae387b93370d142d82220f1b-Paper.pdf. [59] Yikai Zhang, Wenjia Zhang, Sammy Bald, Vamsi Pritham Pingali, Chao Chen, and Mayank Goswami. Stability of SGD: Tightness analysis and improved bounds. In The 38th Conference on Uncertainty in Artificial Intelligence, 2022. URL: https://openreview.net/forum?id= Sl-zm O8j5lq. [60] Tyler Farghly and Patrick Rebeschini. Time-independent generalization bounds for SGLD in non-convex settings. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, volume 34, pages 19836 19846. Curran Associates, Inc., 2021. URL: https://proceedings.neurips.cc/paper/ 2021/file/a4ee59dd868ba016ed2de90d330acb6a-Paper.pdf. [61] Bohan Wang, Huishuai Zhang, Jieyu Zhang, Qi Meng, Wei Chen, and Tie-Yan Liu. Optimizing information-theoretical generalization bound via anisotropic noise of SGLD. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, volume 34, pages 26080 26090. Curran Associates, Inc., 2021. URL: https://proceedings.neurips.cc/paper/2021/file/ db2b4182156b2f1f817860ac9f409ad7-Paper.pdf. [62] Hao Wang, Yizhe Huang, Rui Gao, and Flavio Calmon. Analyzing the generalization capability of SGLD using properties of Gaussian channels. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, volume 34, pages 24222 24234. Curran Associates, Inc., 2021. URL: https://proceedings.neurips.cc/paper/2021/file/ cb77649f5d53798edfa0ff40dae46322-Paper.pdf. [63] Elad Hoffer, Itay Hubara, and Daniel Soudry. Train longer, generalize better: closing the generalization gap in large batch training of neural networks. In I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc., 2017. URL: https://proceedings.neurips.cc/paper/2017/file/ a5e0ff62be0b08456fc7f1e88812af3d-Paper.pdf. [64] Zachary Charles and Dimitris Papailiopoulos. Stability and generalization of learning algorithms that converge to global optima. In Jennifer Dy and Andreas Krause, editors, Proceedings of the 35th International Conference on Machine Learning, volume 80 of Proceedings of Machine Learning Research, pages 745 754. PMLR, 10 15 Jul 2018. URL: https://proceedings.mlr.press/v80/charles18a.html. [65] Idan Amir, Yair Carmon, Tomer Koren, and Roi Livni. Never go full batch (in stochastic convex optimization). In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, volume 34, pages 25033 25043. Curran Associates, Inc., 2021. URL: https://proceedings.neurips.cc/paper/ 2021/file/d27b95cac4c27feb850aaa4070cc4675-Paper.pdf. [66] Idan Amir, Tomer Koren, and Roi Livni. SGD generalizes better than GD (and regularization doesn t help). In Mikhail Belkin and Samory Kpotufe, editors, Proceedings of Thirty Fourth Conference on Learning Theory, volume 134 of Proceedings of Machine Learning Research, pages 63 92. PMLR, 15 19 Aug 2021. URL: https://proceedings.mlr.press/v134/ amir21a.html. [67] Dominic Richards and Ilja Kuzborskij. Stability & generalisation of gradient descent for shallow neural networks without the neural tangent kernel. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, volume 34, pages 8609 8621. Curran Associates, Inc., 2021. URL: https://proceedings.neurips.cc/paper/2021/file/ 483101a6bc4e6c46a86222eb65fbcb6a-Paper.pdf. [68] Yurii Nesterov. Random gradient-free minimization of convex functions. core discussion papers 2011001, université catholique de louvain. Center for Operations Research and Econometrics (CORE), 2011. URL: https://econpapers.repec.org/paper/corlouvco/2011001.htm. [69] Saeed Ghadimi and Guanghui Lan. Stochastic first-and zeroth-order methods for nonconvex stochastic programming. Found Comput Math, 23:35 76, 2022. URL: https://doi.org/ 10.1007/s10208-021-09499-8. [70] Saeed Ghadimi, Guanghui Lan, and Hongchao Zhang. Mini-batch stochastic approximation methods for nonconvex stochastic composite optimization. Mathematical Programming, 155(12):267 305, 2016. URL: https://link.springer.com/content/pdf/10.1007/s10107014-0846-1.pdf. [71] Dan Hendrycks and Kevin Gimpel. Gaussian error linear units (gelus). ar Xiv preprint ar Xiv:1606.08415, 2016. URL: https://arxiv.org/abs/1606.08415. [72] Prajit Ramachandran, Barret Zoph, and Quoc V Le. Searching for activation functions. ar Xiv preprint ar Xiv:1710.05941, 2017. URL: https://arxiv.org/abs/1710.05941. [73] Hao Zheng, Zhanlei Yang, Wenju Liu, Jizhong Liang, and Yanpeng Li. Improving deep neural networks using softplus units. In 2015 International Joint Conference on Neural Networks (IJCNN), pages 1 4, 2015. doi:10.1109/IJCNN.2015.7280459. [74] Krishnakumar Balasubramanian and Saeed Ghadimi. Zeroth-order nonconvex stochastic optimization: Handling constraints, high dimensionality, and saddle points. Foundations of Computational Mathematics, 22(1):35 76, 2022. URL: https://doi.org/10.1137/120880811. [75] Moritz Hardt, Benjamin Recht, and Yoram Singer. Train faster, generalize better: Stability of stochastic gradient descent. ar Xiv preprint ar Xiv:1509.01240, 2015. URL: https:// arxiv.org/abs/1509.01240. [76] Robert M Corless, Gaston H Gonnet, David EG Hare, David J Jeffrey, and Donald E Knuth. On the Lambert W function. Advances in Computational mathematics, 5(1):329 359, 1996. URL: https://link.springer.com/content/pdf/10.1007/BF02124750.pdf. Acknowledgments and Disclosure of Funding We would like to thank the four anonymous reviewers for providing valuable comments and suggestions, which have improved the presentation of the results and the overall quality of our paper. Amin Karbasi acknowledges funding in direct support of this work from NSF (IIS-1845032), ONR (N0001419-1-2406), and the AI Institute for Learning-Enabled Optimization at Scale (TILOS). 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] See abstract and introduction. (b) Did you describe the limitations of your work? [Yes] See Section 4 for discussions on the assumptions adopted, which are standard and common in the related literature. (c) Did you discuss any potential negative societal impacts of your work? [N/A] This paper presents theoretical contributions of generic nature and broad applicability. (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [Yes] See Table 1, as well as all statement of our results. (b) Did you include complete proofs of all theoretical results? [Yes] All results include detailed proofs, either in the paper or in the supplement. 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [N/A] (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [N/A] (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [N/A] (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [N/A] 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [N/A] (b) Did you mention the license of the assets? [N/A] (c) Did you include any new assets either in the supplemental material or as a URL? [N/A] (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [N/A] (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [N/A] 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]