# directional_pruning_of_deep_neural_networks__09ec7755.pdf Directional Pruning of Deep Neural Networks Shih Kang Chao Department of Statistics University of Missouri Columbia, MO 65211 chaosh@missouri.edu Zhanyu Wang Department of Statistics Purdue University West Lafayette, IN 47907 wang4094@purdue.edu Yue Xing Department of Statistics Purdue University West Lafayette, IN 47907 xing49@purdue.edu Guang Cheng Department of Statistics Purdue University West Lafayette, IN 47907 chengg@purdue.edu In the light of the fact that the stochastic gradient descent (SGD) often finds a flat minimum valley in the training loss, we propose a novel directional pruning method which searches for a sparse minimizer in or close to that flat region. The proposed pruning method does not require retraining or the expert knowledge on the sparsity level. To overcome the computational formidability of estimating the flat directions, we propose to use a carefully tuned 1 proximal gradient algorithm which can provably achieve the directional pruning with a small learning rate after sufficient training. The empirical results demonstrate the promising results of our solution in highly sparse regime (92% sparsity) among many existing pruning methods on the Res Net50 with the Image Net, while using only a slightly higher wall time and memory footprint than the SGD. Using the VGG16 and the wide Res Net 28x10 on the CIFAR-10 and CIFAR-100, we demonstrate that our solution reaches the same minima valley as the SGD, and the minima found by our solution and the SGD do not deviate in directions that impact the training loss. The code that reproduces the results of this paper is available at https://github.com/ donlan2710/g RDA-Optimizer/tree/master/directional_pruning. 1 Introduction Deep neural networks (DNNs), after properly trained, provide the state-of-the-art performance in various domains. Overparameterization is a common practice in modern deep learning, which facilitates better expressive power and faster convergence. On the other hand, overparameterization makes DNN exceedingly large, especially for large-scale tasks. For example, the Image Net [10, 52] may need billions of parameters [4] to become sufficiently overparameterized. As the number of parameters in DNN is growing fast, the cost to deploy and process large DNNs can be prohibitive on devices with low memory/processing resources or with strict latency requirements, such as mobile phones, augmented reality devices and autonomous cars. Many achievements have been made in shrinking the DNN while maintaining accuracy, and the MIT Technological Review lists the tiny AI as one of the breakthroughs in 2020 [1]. Among many methods for shrinking DNN, sparse DNN has attracted much attention. Here, sparsity refers to the situation that most model parameters are zero in a DNN. Sparse DNN not only requires Corresponding author. 34th Conference on Neural Information Processing Systems (Neur IPS 2020), Vancouver, Canada. less memory and storage capacity, but also reduces inference time [9]. One of the popular ways to get sparse DNNs is magnitude pruning [27, 26, 43, 62, 40, 17, 18, 19]. Magnitude pruning first learns the model parameters with an optimizer, e.g. stochastic gradient descent (SGD), and then prunes based on the learned magnitude of parameters with an a priori threshold. However, determining a threshold requires some expert knowledge and trial-and-error, as a principle for setting the threshold is not available. In addition, naïvely masking parameters usually worsens the training loss and testing accuracy. Hence, retraining is needed for the pruned network to regain a similar performance as the dense network [27]. Unfortunately, retraining as an additional step requires some care [17] and additional computation. 1.1 Directional pruning In this paper, we try to answer when a coefficient can be pruned without paying the price of increasing the training loss, and how we can prune based on this. These answers rely on the local geometry of the DNN loss function (w), where w denotes the parameters. Suppose that w SGD 2 Rd, the parameter trained by the SGD, has reached a valley of minima. Hence, r (w SGD) 0. The Hessian r2 (w SGD) has multiple nearly zero eigenvalues [53, 54, 21, 48], and the directions associated with these eigenvalues are the flat directions on the loss landscape. Perturbation in these directions causes little change in the training loss by the second order Taylor expansion of (w) around w SGD. We denote the subspace generated by these directions as P0. Following [37, 29], pruning w SGD can be viewed as a perturbation of w SGD: w SGD A sign(w SGD). (1) Here, sign(w SGD) 2 { 1, 1}d is the sign vector of w SGD and A is a diagonal matrix with 0 Ajj |w SGD j | for j = 1, . . . , d. The jth coefficient is pruned if Ajj = |w SGD j |. For example, in a 2D illustration in the left panel of Figure 1, (1) is a vector starting from the origin to a point in the orange rectangle. Retraining is needed if A sign(w SGD) 62 P0. Some empirical studies even suggest P0 is nearly orthogonal to the w SGD [25, 21], so generally A sign(w SGD) 62 P0. Therefore, we instead consider w SGD λ where the perturbation direction 2 P0 and λ > 0. We maximize the number of j such that sign( j) = sign(w SGD j ) for j = 1, . . . , d, in order to decay as many coefficients in w SGD as possible. Specifically, we select as !!u sign(w SGD) i.e. = 0{sign(w SGD)}, where 0 denotes the projection on the subspace P0. The vector does not always decrease the magnitude of w SGD, and it does whenever sign(w SGD j ) j > 0, or sj := sign(w SGD 0{sign(w SGD)} Decreasing the magnitude of the coefficients with sj > 0 in w SGD would cause little changes in the training loss, as long as we simultaneously increase the magnitude of coefficients j0 6= j with sj0 < 0 proportional to |sj0|. As illustrated in the left panel of Figure 1, the adverse effect due to decreasing the magnitude of w2 (s2 > 0) can be compensated by increasing the magnitude of w1, so that the net change is the red vector in P0. Note that this argument has a similar spirit as the optimal brain surgeon [29], and it is the key to remove the need of retraining. The sj can thus be understood as a score to indicate whether pruning the jth coefficient causes a (ir)redeemable training loss change. We propose the novel directional pruning using the score sj in (2). Definition 1.1 (Directional pruning based on SGD). Suppose (w) is the training loss, and r (w SGD) = 0 where w SGD is the minimizer found by SGD. Suppose none of the coefficients in w SGD is zero. With λ > 0 and sj defined in (2), the directional pruning solves 1 2kw SGD wk2 sj|wj|. (3) In (3), the coefficients with sj > 0 are pruned with sufficiently large λ by the absolute value penalization, but the magnitude of wj0 with sj0 0 is un-penalized, and are even encouraged to 0sign(w SGD) sign(w SGD) Figure 1: Left: a 2D graphical illustration of the directional pruning. The orange region contains all possible locations of the vector w SGD A sign(w SGD). The directional pruning with different λ takes solutions on the red dashed line. Right: training loss contour of the wide Res Net28 10 (WRN28x10 [60]) on the CIFAR-100 around the minimal loss path (the white curve) between minimizers found by the SGD and (g RDA) [8] (the algorithm we propose to use) using [20]. While no coefficient of the SGD minimizer is zero, our solution has only 9.7% active parameters. Testing accuracy is 76.6% for the SGD and 76.81% for our solution. increase. For a 2D illustration, the solution path for different λ > 0 is the dashed red curve in the left panel of Figure 1. If λ is too large, the coefficients j with sj < 0 may overshoot, illustrated as the flat part on the dashed red line extended to the right of the red point. Remark 1.2 (Solution of (3)). The objective function in (3) is separable for each coefficient. The part with sj > 0 is solved by the 1 proximal operator. The part with sj < 0 is non-convex, but it still has the unique global minimizer if w SGD j 6= 0. The solution of (3) is bwj = sign(w SGD where [a]+ = max{0, a}. See Proposition A.1 in the appendix for a proof. Implementing the directional pruning is very challenging due to high dimensionality. Specifically, the matrix r2 of modern deep neural network is often very large so that estimating P0 is computationally formidable. Perhaps surprisingly, we will show that there is a very simple algorithm (g RDA) presented in Section 2, that can asymptotically solve (3) without explicitly estimating the Hessian. The right panel of Figure 1 shows that if λ is selected appropriately, our method achieves a similar training loss as the dense network with w SGD, while being highly sparse with a test accuracy comparable to the SGD. More detailed empirical analysis is in Section 4.2. Remark 1.3 (Major differences to the optimal brain surgeon ). It is worth noting that (3) is different from the optimization problem in [29, 28]. While an analytic map between directional pruning and optimal brain surgeon is interesting for future study, the two are generally nonequivalent. Particularly, directional pruning perturbs from w SGD continuously in λ like a restricted 1 weight decay on P0 (Remark 1.2), while optimal brain surgeon yields a discontinuous perturbation like a hard thresholding (see p.165 of [29]). The main advantage of directional pruning is that it can be computed with the g RDA algorithm presented in Section 2, which does not require to estimate the Hessian or its inverse. 1.2 Contributions Our major contribution is to propose the novel directional pruning method (Definition 1.1), and further prove that the algorithm (g RDA) [8] achieves the effect of the directional pruning asymptotically. The (g RDA) has been applied for sparse statistical inference problems with a convex loss and principal component analysis [8]. The connection between the directional pruning and (g RDA) is theoretically proved by leveraging the continuous time approximation developed in [8] under proper assumptions on the gradient flow and the Hessian matrix. It is worth noting that this algorithm does not require to explicitly estimate P0, and it can be implemented like an optimizer in a typical deep learning framework, e.g. Tensorflow or Py Torch. Empirically, we demonstrate that (g RDA) successfully prunes Res Net50 on Image Net, and achieves 73% testing accuracy with only 8% active parameters. Upon benchmarking with other popular algorithms, (g RDA) yields a high accuracy and sparsity tradeoff among many contemporary methods. We also successfully prune deep networks on CIFAR-10/100, and the results are in the appendix. Using VGG16 on CIFAR-10 and WRN28x10 on CIFAR-100, we show that (g RDA) reaches the same valley of minima as the SGD, empirically verifying the directional pruning. Using VGG16 and WRN28x10 on CIFAR-10, we show the proportion of the difference between (g RDA) and the SGD in the leading eigenspace of the Hessian is low, as another evidence for (g RDA) performing the directional pruning. 2 The g RDA algorithm Consider training data Zi = {(Xi, Yi)}N i=1, where Xi is the input variable, e.g. images, and Yi is the response variable, e.g. a vector of real numbers, or labels Yn 2 {0, 1}nl, where nl 2 N. Suppose h(x; w) 2 Rnl is the output of an L-layer feedforward overparameterized DNN, with parameters w 2 Rd. Let L(h; y) : Rnl nl ! R+ be a loss function, e.g. the 2 loss L(h; y) = kh yk2 2 or the cross-entropy loss. Let f(w; Z) := L(h(X; w), Y ), and rf(w; Z) be the gradient of f(w; Z), the loss function (w) and its gradient are defined by (w) := EZ[f(w; Z)], G(w) = r (w) = EZ[rf(w; Z)], (4) where EZ[f(w; Z)] = N 1 PN i=1 f(w; Zi). We adopt the generalized regularized dual averaging (g RDA) algorithms originally proposed in [8]. This algorithm has been successfully applied to the ad click-through rate prediction [39]. Specifically, let {ˆik}1 k=1 be i.i.d. uniform random variables on {1, . . . , N} independent from the training data, wn+1,j = Sg(n,γ) rfj(wk; Zˆik+1) , for j = 1, . . . , d, (g RDA) where Sg : v 7! sign(v)(|v| g)+ is the soft-thresholding operator, w0 is an initializer chosen at random from a distribution; γ is the learning rate; g(n, γ) > 0 is the tuning function, detailed in (5). We can extend (g RDA) to minibatch gradients, by replacing rfj(wk; Zˆik+1) with an average |Sk+1| 1 P i2Sk+1 rf(wk; Zi), where Sk+1 {1, . . . , N} is sampled uniformly. We will focus on (g RDA), i.e. |Sk| = 1 for all k, but our theory can be generalized to any fixed minibatch size. The tuning function g(n, γ) controls the growth rate of penalization. Motivated by [8], g(n, γ) = cγ1/2(nγ)µ, (5) where c, µ > 0 are the two hyperparameters positively related to the strength of penalization. The (nγ)µ is used to match the growing magnitude of SGD. The γ1/2 is an important scaling factor; without it, (g RDA) with µ = 1 reduces to the regularized dual averaging (RDA) algorithm [59] that minimizes (w) + λkwk1 rather than the directional pruning problem in (3). Note that if c = 0, then (g RDA) recovers the stochastic gradient descent: n+1 = w SGD n γrf(w SGD n ; Zˆin+1). (SGD) In this paper, we only consider the constant learning rate. In practice, a constant-and-drop learning rate is often adopted. See Section C.1 and C.2 in the appendix for the algorithms in pseudocode. Remark 2.1 (Selection of µ and c in practice). Our empirical results and theory in later sections suggest µ 2 {0.501, 0.51, 0.55} generally performs well regardless of the task and network used. For a given µ, we recommend to search for the greatest c (starting with e.g. 10 4) such that g RDA yields a comparable test acc. as SGD using 1 5 epochs. 3 Theoretical analysis To show (g RDA) asymptotically achieves the directional pruning in Definition 1.1, we leverage some tools from the continuous time analysis. Define the gradient flow w(t) to be the solution of the ordinary differential equation w = G(w), w(0) = w0, (GF) where w0 is a random initializer, and G is defined in (4). The w(t) can provably find a good global minimizer under various conditions [3, 2, 13, 38, 46, 12]. Throughout this paper, we assume the solution of (GF) is unique. Let H( ) := EZ[r2f( ; Z)] be the Hessian matrix. Let Φ(t, s) 2 Rd d be the solution (termed the principal matrix solution, see Chapter 3.4 of [57]) of the matrix ODE system (s is the initial time): dt = H(w(t))Φ(t, s), Φ(s, s) = Id. (6) Let wγ(t) := wbt/γc and w SGD(t) be the piecewise constant interpolated process of (g RDA) and (SGD), respectively, with the same learning rate, where bac takes the greatest integer that is less than or equal to a. We will make the following assumptions: (A1) G(w) : Rd ! Rd is continuous on Rd. rf(w; Z) G(w) rf(w; Z) G(w) (A2) : Rd ! Rd d is continuous. EZ < 1 for any K > 0 a.s. (A3) H : Rd ! Rd d is continuous, and there exists a non-negative definite matrix H such that R 1 0 k H(w(s)) Hkds < 1 where k k is the spectral norm, and the eigenspace of H associated with the zero eigenvalues matches P0. (A4) 0 sµ 1Φ(t, s) sign(w(s))ds = o(tµ). (A5) There exists T > 0 such that for all t > T: (i) sign{w(t)} = sign{w( T)}; (ii) sign{wj(t)} = sign{w SGD j (t)} for all j. The key theoretical result of this paper shows that (g RDA) performs the directional pruning (Definition 1.1) for a sufficiently large t. Theorem 3.1. Under assumptions (A1)-(A5), and assume µ 2 (0.5, 1) and c > 0 in (5). Then, as γ ! 0, (g RDA) asymptotically performs directional pruning based on w SGD(t); particularly, 2kw SGD(t) wk2 , for any t > T, (8) d means asymptotic in distribution under the empirical probability measure of the gradients, λγ,t = cpγtµ and the sj satisfies limt!1 | sj sj| = 0 for all j. This theorem holds in the asymptotic regime (γ ! 0) with a finite time horizon, i.e. any fixed t T. It is important that λ grows with t, because the magnitude of SGD asymptotically grows like a Gaussian process, i.e., in t0.5. Hence, µ should be slightly greater than 0.5. The proof of Theorem 3.1 is in Section B.2 of the appendix. Remark 3.2 (Condition (A3)). The eigenspace of H associated with the zero eigenvalues and P0 matches when w(t) and SGD converge to the same flat valley of minima. For the 2 loss and in the teacher-student framework, [12, 61, 7] showed w(t) ! w exponentially fast for one hidden layer networks, so the limit H = H(w ) and the condition holds. For the cross-entropy loss, we suspect that H satisfying (A3) is not a zero matrix, but its exact form needs further investigation. Remark 3.3 (Condition (A4)). This condition can be verified (by Problem 3.31 of [57]) if sign(w(t)) is mainly restricted in the eigenspace of H(w(t)) associated with positive eigenvalues as t ! 1. Empirically, this appears to hold as [25, 21] show that w(t) lies mainly in the subspace of H(w(t)) associated with the positive eigenvalues, and Figure 2 suggests the angle between w(t) and sign(w(t)) is very small. Remark 3.4 (Condition (A5)). For (i), under the crossentropy loss, several papers [56, 24, 34, 41] show that w(t)/kw(t)k2 converges to a unique direction while kw(t)k2 ! 1. This implies that sign(w(t)) stabilizes after a finite time. For the 2 loss, [12, 61] show w(t) ! w for one hidden layer networks under regularity conditions, and the condition follows. The (ii) holds if the learning rate is sufficiently small, so that the deviation between the gradient flow and the SGD is small. Figure 2: kwk2/kwk1 is close to its lower bound d 1/2 when the coefficients in w are of similar magnitude, i.e. the direction of w is the same as sign(w). 4 Empirical experiments This section presents the empirical performance of (g RDA), and the evidence that (g RDA) performs the directional pruning (Definition 1.1). Section 4.1 considers Res Net50 with Image Net, and compares with several existing pruning algorithms. To check if (g RDA) performs the directional pruning, Section 4.2 presents the local geometry of the loss around the minimal loss curve that connects the minima found by (SGD) and (g RDA), and Section 4.3 investigates the direction of deviation between the minima found by (SGD) and (g RDA). 4.1 Res Net50 on the Image Net We use (g RDA) to simultaneously prune and train the Res Net50 [31] on the Image Net dataset without any post-processing like retraining. The learning rate schedule usually applied jointly with the SGD with momentum does not work well for (g RDA), so we use either a constant learning rate or dropping the learning rate only once in the later training stage. Please find more implementation details in Section C.1 in the appendix. The results are shown in Figure 3, where µ is the increasing rate of the soft thresholding in the tuning function (5) of (g RDA). Figure 3: Learning trajectories of (SGD) and (g RDA) for Res Net50 [31] on Image Net image recognition task. Left: top 1 training accuracy. Center: top 1 testing accuracy. Right: the ratio between the number of nonzero parameters and the total number of parameters. The number of nonzero weights slightly increases, contradicting with Theorem 3.1. This could be because that Assumption (A5) fails due to the large learning rate. γ = 0.1 for both SGD and g RDA. Minibatch size is 256. Accuracy: g RDAs can perform as accurate as (SGD) after sufficient training. Larger µ (in the tuning function (5)) can perform worse than (SGD) in the early stage of training, but eventually beat (SGD) in the late stage of training. The training accuracy of (SGD) is higher than that of the g RDAs. This may result from a too large learning rate, so the coefficients wj s with sj < 0 (in (3)) overshoot and their magnitudes become too large. Sparsity: Sparsity increases rapidly at the early stage of training. With µ = 0.55 in Figure 3, (g RDA) reaches 92% sparsity, while the testing accuracy is higher than (SGD). Wall time and memory footprint: (g RDA) has a slightly higher wall time than (SGD), but the memory footprint is similar. See Section C.5 for a detailed comparison. The left panel of Figure 4 compares (g RDA) with the magnitude pruning [62] and the variational dropout [42], and (g RDA) is particularly competitive in the high sparsity (90-92%) regime. The right panel of Figure 4 compares different pruning algorithms that do not require expert knowledge for selecting the layerwise pruning level with (g RDA) in terms of the layerwise sparsity. We compare (g RDA) with the Erd os-Rényi-Kernel of [15], variational dropout [42] and a reinforcement-learning based Auto ML method [32]. Our (g RDA) achieves a high sparsity 92% with a competitive testing accuracy. In addition, the layerwise sparsity pattern generated by g RDA is similar to the variational dropout and the Auto ML, as these methods generate higher sparsity in the 3 3 convolutional layers, and lower sparsity in the 1 1 layers and the initial layers, which are less wide than the latter layers. Among these methods, (g RDA) is unique in that its spirit is interweaving with the local loss landscape. Figure 4: Left: A comparison of g RDA with the magnitude pruning [62] and variational dropout [42] with Res Net50 on Image Net, done by [19] with around 100 epochs using SGD with momentum. Our solution is among the high performers in the very sparse regime (90-92%). The numbers next to the red crosses are the epochs. Right: Layerwise sparsity produced by different automatic pruning algorithms. All methods show the pattern that the 3x3 conv layers (on dashed lines) are greatly pruned (valleys), and the 1x1 conv layers are less pruned (peaks). 4.2 Connectivity between the minimizers of g RDA and SGD In this section, we check whether (SGD) and (g RDA) reach the same valley, which implies (g RDA) is performing the directional pruning. Similar analysis has been done for the minima found by (SGD) with different initializers [58, 20, 44, 11, 30, 16, 23]. We train VGG16 [55] on CIFAR-10 and WRN28x10 on CIFAR-100 until nearly zero training loss using both (SGD) and (g RDA). The minima here found by (g RDA) generally have sparsity around 90% or higher for larger µ. We use the method of [20] to search for a quadratic Bézier curve of minimal training loss connecting the minima found by the g RDA and SGD, and then visualize the contour of the training losses and testing errors on the hyperplane containing the minimal loss curve. See Sections C.2 and C.3 for details on implementation. The results are shown for different choices of µ, which is the increasing rate of the soft thresholding in the tuning function (5) of (g RDA). As observed from the contours in Figure 5, the learned parameters of both SGD and g RDA lie in the same valley on the training loss landscape if µ is properly tuned, namely, 0.6 for VGG16 and 0.501 for WRN28x10. This verifies that (g RDA) performs the directional pruning. For large µ, a hill exists on the minimal loss/error path, which may be due to the too large learning rate that leads to large magnitude for the coefficients j with sj < 0. The details (training accuracy, testing accuracy, sparsity) of the endpoints trained on VGG16 and WRN28x10 are shown in Tables 4 and 6 of the Appendix. For the testing error in Figure 5, the g RDA somewhat outperforms SGD when µ is slightly greater than 0.5. Interestingly, the neighborhood of the midpoint on the Bézier curve often has a higher testing accuracy than the both endpoints, except for WRN28x10 on CIFAR-100 with µ = 0.501 and 0.55. This finding resonates with the results of [33]. 4.3 Direction of wg RDA w SGD The directional pruning (Definition 1.1) implies that the vector n := wg RDA n should lie in P0 as n ! 1 if tuned appropriately. Unfortunately, checking this empirically requires estimating P0 which is computationally formidable. Nonetheless, there exists a dominating low dimensional subspace in P? 0 (the subspace orthogonal to P0); particularly, a few studies [53, 54, 21, 47] have empirically shown that for various networks on the CIFAR-10, the magnitude of the ten leading eigenvalues of H(w SGD) are dominating the others. (a) VGG16/CIFAR-10/Train loss (b) VGG16/CIFAR-10/Test error (c) WRN28x10/CIFAR-100/Train loss (d) WRN28x10/CIFAR-100/Test error Figure 5: The upper figure in each panel shows the contour of training loss and testing error on the hyperplane containing the minimal loss Bézier curve (white) interpolating the minimizers found by the SGD and the g RDA. The lower plot of each panel shows the training loss/testing error on the minimal loss Bézier curve interpolating minimizers of SGD and g RDA under different µ. n := span{u1,n, u2,n, . . . , u10,n} be the top subspace spanned by the eigenvectors uj,n associated with the top 10 eigenvalues of H(w SGD u1,n ! u2,n ! ... u10,n ! We train the VGG16 and WRN28x10 on the CIFAR-10, until the training data are nearly interpolated and the training loss is almost zero. During the training process, we fix the initializer and minibatches when we use different optimizers to ensure the comparability. We compute Pn on the training trajectory of VGG16 and WRN28x10. See Section C.4 for details on the computation of these eigenvectors. We test the hypothesis that the proportion of n in Ptop n is low, i.e. k Pn nk/k nk is low. The results from the VGG16 and WRN28x10 in Figure 6 basically confirm this hypothesis, as the magnitude of the proportion of n in Ptop n is very small under the two networks. Particularly, the proportion is always very small for WRN28x10. The results for different µ are similar, showing that n is pointing to the same direction regardless of µ. Figure 6: The fraction of the different between SGD and g RDA on the eigenspace associated with the leading 10 eigenvalues. Left: VGG16. Right: WRN28x10. The k k is the 2 norm. 5 Discussion and future work We propose the novel directional pruning for deep neural networks, that aims to prune DNNs while preserving the training accuracy. For implementation, we show that (g RDA) asymptotically achieves the directional pruning after sufficient epochs of training. Empirical evidence shows that our solution yields a accuracy and sparsity tradeoff within the range of many contemporary pruning techniques. The testing accuracy of (g RDA) is almost always higher than (SGD) if µ is slightly greater than 0.5 when using the Res Nets, and some interpolation between the minima found by (g RDA) and (SGD) often has a better testing accuracy than the two minima; see Figure 5. As suggested by Figure 6, (g RDA) appears to deviate from (SGD) in the flatter directions. These evidences support [30], who argue that the valley of minima is actually asymmetric, and points on the flatter side tend to generalize better. We think a further study of the testing accuracy of (g RDA) along the lines initiated in this work may be an interesting future research topic, as this would shed some light on the mystery of generalization. Broader Impact Our paper belongs to the cluster of works focusing on efficient and resource-aware deep learning. There are numerous positive impacts of these works, including the reduction of memory footprint and computational time, so that deep neural networks can be deployed on devices equipped with less capable computing units, e.g. the microcontroller units. In addition, we help facilitate on-device deep learning, which could replace traditional cloud computation and foster the protection of privacy. Popularization of deep learning, which our research helps facilitate, may result in some negative societal consequences. For example, the unemployment may increase due to the increased automation enabled by the deep learning. Acknowledgments We thank the anonymous reviewers for the helpful comments. Shih-Kang Chao would like to acknowledge the financial support from the Research Council of the University of Missouri. This work was completed while Guang Cheng was a member of the Institute for Advanced Study, Princeton in the fall of 2019. Guang Cheng would like to acknowledge the hospitality of the IAS and the computational resource it has provided. [1] 10 breakthrough technologies, February 2020. [2] Sanjeev Arora, Nadav Cohen, Noah Golowich, and Wei Hu. A convergence analysis of gradient descent for deep linear neural networks. In International Conference on Learning Representations, 2019. [3] Sanjeev Arora, Nadav Cohen, and Elad Hazan. On the optimization of deep networks: Implicit acceleration by overparameterization. In Jennifer Dy and Andreas Krause, editors, Proceedings of the 35th International Conference on Machine Learning, volume 80 of Proceedings of Machine Learning Research, pages 244 253, Stockholmsmässan, Stockholm Sweden, 10 15 Jul 2018. PMLR. [4] Mikhail Belkin, Daniel Hsu, Siyuan Ma, and Soumik Mandal. Reconciling modern machine- learning practice and the classical bias variance trade-off. Proceedings of the National Academy of Sciences, 116(32):15849 15854, 2019. [5] Albert Benveniste, Michel Métivier, and Pierre Priouret. Adaptive Algorithms and Stochastic Approximations. Springer, 1990. [6] J. A. Bucklew, T. G. Kurtz, and W. A. Sethares. Weak convergence and local stability properties of fixed step size recursive algorithms. IEEE Transactions on Information Theory, 39(3):966 978, May 1993. [7] Yuan Cao and Quanquan Gu. Tight sample complexity of learning one-hidden-layer con- volutional neural networks. In H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett, editors, Advances in Neural Information Processing Systems 32, pages 10612 10622. Curran Associates, Inc., 2019. [8] Shih-Kang Chao and Guang Cheng. A generalization of regularized dual averaging and its dynamics. Ar Xiv Preprint ar Xiv:1909.10072, 2019. [9] Y. Cheng, D. Wang, P. Zhou, and T. Zhang. Model compression and acceleration for deep neural networks: The principles, progress, and challenges. IEEE Signal Processing Magazine, 35(1):126 136, Jan 2018. [10] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and F.-F. Li. Image Net: A Large-Scale Hierarchical Image Database. In CVPR09, 2009. [11] Felix Draxler, Kambis Veschgini, Manfred Salmhofer, and Fred Hamprecht. Essentially no barriers in neural network energy landscape. In Jennifer Dy and Andreas Krause, editors, Proceedings of the 35th International Conference on Machine Learning, volume 80 of Proceedings of Machine Learning Research, pages 1309 1318, Stockholmsmässan, Stockholm Sweden, 10 15 Jul 2018. PMLR. [12] Simon S. Du, Jason D. Lee, and Yuandong Tian. When is a convolutional filter easy to learn? In International Conference on Learning Representations, 2018. [13] Simon S. Du, Xiyu Zhai, Barnabas Poczos, and Aarti Singh. Gradient descent provably optimizes over-parameterized neural networks. In International Conference on Learning Representations, 2019. [14] M. S. P. Eastham. The asymptotic solution of linear differential systems, applications of the Levinson theorem. Clarendon Press, Oxford, 1989. [15] Utku Evci, Trevor Gale, Jacob Menick, Pablo Samuel Castro, and Erich Elsen. Rigging the lottery: Making all tickets winners. To appear in ICML 2020, 2020. [16] Utku Evci, Fabian Pedregosa, Aidan Gomez, and Erich Elsen. The difficulty of training sparse neural networks. Ar Xiv Preprint Ar Xiv:1906.10732, 2019. [17] Jonathan Frankle and Michael Carbin. The lottery ticket hypothesis: Finding sparse, trainable neural networks. In International Conference on Learning Representations, 2019. [18] Jonathan Frankle, Gintare Karolina Dziugaite, Daniel M. Roy, and Michael Carbin. Stabilizing the lottery ticket hypothesis. Ar Xiv Preprint ar Xiv:1903.01611, 2019. [19] Trevor Gale, Erich Elsen, and Sara Hooker. The state of sparsity in deep neural networks. Ar Xiv Preprint Arxiv 1902.09574, 2019. [20] Timur Garipov, Pavel Izmailov, Dmitrii Podoprikhin, Dmitry P Vetrov, and Andrew G Wilson. Loss surfaces, mode connectivity, and fast ensembling of dnns. In S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett, editors, Advances in Neural Information Processing Systems 31, pages 8789 8798. Curran Associates, Inc., 2018. [21] Behrooz Ghorbani, Shankar Krishnan, and Ying Xiao. An investigation into neural net opti- mization via hessian eigenvalue density. In Kamalika Chaudhuri and Ruslan Salakhutdinov, editors, Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research, pages 2232 2241, Long Beach, California, USA, 09 15 Jun 2019. PMLR. [22] Noah Golmant, Zhewei Yao, Amir Gholami, Michael Mahoney, and Joseph Gonzalez. pytorch- hessian-eigentings: efficient pytorch hessian eigendecomposition, October 2018. [23] Akhilesh Gotmare, Nitish Shirish Keskar, Caiming Xiong, and Richard Socher. Using mode connectivity for loss landscape analysis. Ar Xiv, abs/1806.06977, 2018. [24] Suriya Gunasekar, Jason Lee, Daniel Soudry, and Nathan Srebro. Characterizing implicit bias in terms of optimization geometry. In Jennifer Dy and Andreas Krause, editors, Proceedings of the 35th International Conference on Machine Learning, volume 80 of Proceedings of Machine Learning Research, pages 1832 1841, Stockholmsmässan, Stockholm Sweden, 10 15 Jul 2018. PMLR. [25] Guy Gur-Ari, Daniel A. Roberts, and Ethan Dyer. Gradient descent happens in a tiny subspace. Ar Xiv Preprint Arxiv 1812.04754, 2018. [26] Song Han, Huizi Mao, and William J. Dally. Deep compression: Compressing deep neural net- works with pruning, trained quantization and huffman coding. Ar Xiv Preprint Arxiv 1510.00149, 2015. cite arxiv:1510.00149Comment: Published as a conference paper at ICLR 2016 (oral). [27] Song Han, Jeff Pool, John Tran, and William J. Dally. Learning both weights and connections for efficient neural networks. In Proceedings of the 28th International Conference on Neural Information Processing Systems - Volume 1, NIPS 15, pages 1135 1143, Cambridge, MA, USA, 2015. MIT Press. [28] B. Hassibi, D. G. Stork, and G. J. Wolff. Optimal brain surgeon and general network pruning. In IEEE International Conference on Neural Networks, pages 293 299 vol.1, 1993. [29] Babak Hassibi and David G. Stork. Second order derivatives for network pruning: Optimal brain surgeon. In S. J. Hanson, J. D. Cowan, and C. L. Giles, editors, Advances in Neural Information Processing Systems 5, pages 164 171. Morgan-Kaufmann, 1993. [30] Haowei He, Gao Huang, and Yang Yuan. Asymmetric valleys: Beyond sharp and flat local minima. In H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett, editors, Advances in Neural Information Processing Systems 32, pages 2549 2560. Curran Associates, Inc., 2019. [31] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 770 778, 2015. [32] Yihui He, Ji Lin, Zhijian Liu, Hanrui Wang, Li-Jia Li, and Song Han. AMC: automl for model compression and acceleration on mobile devices. In Vittorio Ferrari, Martial Hebert, Cristian Sminchisescu, and Yair Weiss, editors, Computer Vision - ECCV 2018 - 15th European Conference, Munich, Germany, September 8-14, 2018, Proceedings, Part VII, volume 11211 of Lecture Notes in Computer Science, pages 815 832. Springer, 2018. [33] Pavel Izmailov, Dmitrii Podoprikhin, Timur Garipov, Dmitry P. Vetrov, and Andrew Gordon Wilson. Averaging weights leads to wider optima and better generalization. In UAI, 2018. [34] Ziwei Ji and Matus Telgarsky. The implicit bias of gradient descent on nonseparable data. In Alina Beygelzimer and Daniel Hsu, editors, Proceedings of the Thirty-Second Conference on Learning Theory, volume 99 of Proceedings of Machine Learning Research, pages 1772 1798, Phoenix, USA, 25 28 Jun 2019. PMLR. [35] Ioannis Karatzas and Steven Shreve. Brownian Motion and Stochastic Calculus, volume 113 of Graduate Texts in Mathematics. Springer, New York, 1998. [36] Alex Krizhevsky. Learning multiple layers of features from tiny images. Technical Report TR-2009, University of Toronto, 2009. [37] Yann Le Cun, John S. Denker, and Sara A. Solla. Optimal brain damage. In D. S. Touretzky, edi- tor, Advances in Neural Information Processing Systems 2, pages 598 605. Morgan-Kaufmann, 1990. [38] Jaehoon Lee, Lechao Xiao, Samuel S. Schoenholz, Yasaman Bahri Roman Novak, Jascha Sohl-Dickstein, and Jeffrey Pennington. Wide neural networks of any depth evolve as linear models under gradient descent. Ar Xiv Preprint Arxiv 1902.06720, 2019. [39] Bin Liu, Chenxu Zhu, Guilin Li, Weinan Zhang, Jincai Lai, Ruiming Tang, Xiuqiang He, Zhengguo Li, and Yong Yu. Auto FIS: Automatic feature interaction selection in factorization models for click-through rate prediction. KDD 2020, 2020. [40] Zhuang Liu, Mingjie Sun, Tinghui Zhou, Gao Huang, and Trevor Darrell. Rethinking the value of network pruning. In International Conference on Learning Representations, 2019. [41] Kaifeng Lyu and Jian Li. Gradient descent maximizes the margin of homogeneous neural networks. In International Conference on Learning Representations, 2020. [42] Dmitry Molchanov, Arsenii Ashukha, and Dmitry Vetrov. Variational dropout sparsifies deep neural networks. In Doina Precup and Yee Whye Teh, editors, Proceedings of the 34th International Conference on Machine Learning, volume 70 of Proceedings of Machine Learning Research, pages 2498 2507, International Convention Centre, Sydney, Australia, 06 11 Aug 2017. PMLR. [43] Pavlo Molchanov, Stephen Tyree, Tero Karras, Timo Aila, and Jan Kautz. Pruning convolutional neural networks for resource efficient inference. In ICLR, 2017. [44] Quynh Nguyen. On connected sublevel sets in deep learning. In Kamalika Chaudhuri and Ruslan Salakhutdinov, editors, Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research, pages 4790 4799, Long Beach, California, USA, 09 15 Jun 2019. PMLR. [45] Francesco Orabona, Koby Crammer, and Nicolò Cesa-Bianchi. A generalized online mirror descent with applications to classification and regression. Machine Learning, 99(3):411 435, Jun 2015. [46] Samet Oymak and Mahdi Soltanolkotabi. Overparameterized nonlinear learning: Gradient descent takes the shortest path? In Kamalika Chaudhuri and Ruslan Salakhutdinov, editors, Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research, pages 4951 4960, Long Beach, California, USA, 09 15 Jun 2019. PMLR. [47] Vardan Papyan. The full spectrum of deepnet hessians at scale: Dynamics with sgd training and sample size. Ar Xiv Preprint Arxiv 1811.07062, 2018. [48] Vardan Papyan. Measurements of three-level hierarchical structure in the outliers in the spectrum of deepnet hessians. In Kamalika Chaudhuri and Ruslan Salakhutdinov, editors, Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research, pages 5012 5021, Long Beach, California, USA, 09 15 Jun 2019. PMLR. [49] Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas Kopf, Edward Yang, Zachary De Vito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. Pytorch: An imperative style, highperformance deep learning library. In H. Wallach, H. Larochelle, A. Beygelzimer, F. d'AlchéBuc, E. Fox, and R. Garnett, editors, Advances in Neural Information Processing Systems 32, pages 8024 8035. Curran Associates, Inc., 2019. [50] A. Pazy. Semigroups of Linear Operators and Applications to Partial Differential Equations. Springer-Verlag New York, 1983. [51] V. I. Piterbarg. Asymptotic methods in the theory of Gaussian processes and fields, volume 148 of Translations of Mathematical Monographs. American Mathematical Society, Providence, RI., 1996. [52] Olga Russakovsky, Jia Deng, Hao Su, Jonathan Krause, Sanjeev Satheesh, Sean Ma, Zhiheng Huang, Andrej Karpathy, Aditya Khosla, Michael Bernstein, Alexander C. Berg, and Li Fei-Fei. Imagenet large scale visual recognition challenge. International Journal of Computer Vision, 115(3):211 252, Dec 2015. [53] Levent Sagun, Leon Bottou, and Yann Le Cun. Eigenvalues of the hessian in deep learning: Singularity and beyond. Ar Xiv Preprint Arxiv 1611.07476, 2016. [54] Levent Sagun, Utku Evci, V. Ugur Guney, Yann Dauphin, and Leon Bottou. Empirical analysis of the hessian of over-parametrized neural networks. ICLR 2018 Workshop, 2018. [55] Karen Simonyan and Andrew Zisserman. Very deep convolutional networks for large-scale image recognition. In International Conference on Learning Representations, 2015. [56] Daniel Soudry, Elad Hoffer, Mor Shpigel Nacson, and Nathan Srebro. The implicit bias of gradient descent on separable data. In International Conference on Learning Representations, 2018. [57] Gerald Teschl. Ordinary Differential Equations and Dynamical Systems, volume 140 of Graduate Studies in Mathematics. American Mathematical Society, Providence, Rhode Island, 2012. [58] Mingwei Wei and David J Schwab. How noise affects the hessian spectrum in overparameterized neural networks. Ar Xiv Preprint Arxiv 1910.00195, 2019. [59] Lin Xiao. Dual averaging methods for regularized stochastic learning and online optimization. Journal of Machine Learning Research, 11:2543 2596, 2010. [60] Sergey Zagoruyko and Nikos Komodakis. Wide residual networks. In Edwin R. Hancock Richard C. Wilson and William A. P. Smith, editors, Proceedings of the British Machine Vision Conference (BMVC), pages 87.1 87.12. BMVA Press, September 2016. [61] Xiao Zhang, Yaodong Yu, Lingxiao Wang, and Quanquan Gu. Learning one-hidden-layer relu networks via gradient descent. In Kamalika Chaudhuri and Masashi Sugiyama, editors, Proceedings of Machine Learning Research, volume 89 of Proceedings of Machine Learning Research, pages 1524 1534. PMLR, 16 18 Apr 2019. [62] Michael H. Zhu and Suyog Gupta. To prune, or not to prune: Exploring the efficacy of pruning for model compression. Co RR, 2017.