# finegrained_optimization_of_deep_neural_networks__d6207198.pdf Fine-grained Optimization of Deep Neural Networks In recent studies, several asymptotic upper bounds on generalization errors on deep neural networks (DNNs) are theoretically derived. These bounds are functions of several norms of weights of the DNNs, such as the Frobenius and spectral norms, and they are computed for weights grouped according to either input and output channels of the DNNs. In this work, we conjecture that if we can impose multiple constraints on weights of DNNs to upper bound the norms of the weights, and train the DNNs with these weights, then we can attain empirical generalization errors closer to the derived theoretical bounds, and improve accuracy of the DNNs. To this end, we pose two problems. First, we aim to obtain weights whose different norms are all upper bounded by a constant number. To achieve these bounds, we propose a two-stage renormalization procedure; (i) normalization of weights according to different norms used in the bounds, and (ii) reparameterization of the normalized weights to set a constant and finite upper bound of their norms. In the second problem, we consider training DNNs with these renormalized weights. To this end, we first propose a strategy to construct joint spaces (manifolds) of weights according to different constraints in DNNs. Next, we propose a finegrained SGD algorithm (FG-SGD) for optimization on the weight manifolds to train DNNs with assurance of convergence to minima. Experimental analyses show that image classification accuracy of baseline DNNs can be boosted using FG-SGD on collections of manifolds identified by multiple constraints. 1 Introduction Understanding generalization behavior of DNNs is an open problem [1]. Recent works [2 8] addressed this problem by extending the early results proposed for shallow linear neural networks (NNs) [9] for a more general class of DNNs (e.g. neural networks with Re LU), recurrent neural networks [10, 11], and convolutional neural networks (CNNs) (see Table 1 for a comparison). The proposed asymptotic bounds were obtained by defining weight matrices of DNNs using random matrices, and applying concentration inequalities on them. Thereby, the bounds were computed by functions of several ℓp norms of these matrices, for 1 p . In this work, we conjecture that if we can impose multiple constraints on weights of DNNs to set upper bounds of the norms of the weight matrices, and train the DNNs with these weights, then the DNNs can achieve empirical generalization errors closer to the proposed theoretical bounds, and we can improve their accuracy in various tasks. We pose two problems in order to achieve this goal; 1. Renormalization of weights to upper bound norms of their matrices. 2. Training DNNs with renormalized weights with assurance to convergence to minima. 1.1 Background Spaces of normalized weights can be identified by different Riemann manifolds [12]; (i) unit norm weights reside on the sphere Sp(Al Bl 1), (ii) orthonormal weights belong to the Stiefel manifold meteozay@gmail.com 33rd Conference on Neural Information Processing Systems (Neur IPS 2019), Vancouver, Canada. St(Al,Bl), and (iii) weights with orthogonal columns reside on the oblique manifold Ob(Al Bl), at each lth layer of a DNN (formal mathematical definitions are given in the supplemental material). Challenge of training DNNs with multiple constraints: DNNs can be trained with multiple constraints using optimization methods proposed for training shallow algorithms [13, 14], and individual manifolds [12, 15]. If we employ these methods on products of weight manifolds (POMs) to train DNNs, then we observe early divergence, vanishing and exploding gradients due to nonlinear geometry of product of different manifolds. More precisely, the assumption of a bound on the operator norm of Hessian of geodesics in POMs, which is required for assurance of convergence, fails, while performing Stochastic Gradient Descent (SGD) with backpropagation on product of different weight manifolds. Therefore, a non-increasing bound on the probability of failure of the optimization algorithm cannot be computed, and a convergence bound cannot be obtained. We consider training DNNs using a more general setting employing groups of weights which can be normalized according to different normalization constraints. Group wise operations are implemented by concatenating weight matrices ωi g,l belonging to each gth group by ωg,l = (ω1 g,l,ω2 g,l,...,ωg g,l), g = 1,2,...,Gl. For the corresponding group, a space of concatenated weights is identified by Cartesian product of manifolds of weights ωi g,l,i = 1,2,...,g. In addition, if we renormalize weights using standard deviation of features obtained at each epoch, then geometry of the manifolds of weights also changes. Therefore, we address the second subproblem which is optimization on dynamically changing product manifolds of renormalized weights. In order to solve these problems, we first propose a mathematical framework to make use of the geometric relationship between weight manifolds determined by different constraints (Section 5). Then, we suggest an approach for training DNNs using multiple constraints on weights to improve their performance under the proposed framework. To this end, we propose a new algorithm that we call fine-grained stochastic gradient descent (FG-SGD) to train DNNs using POMs. We elucidate geometric properties of POMs to assure convergence of FG-SGD to global minima while training nonlinear DNNs with particular assumptions on their architectures, and to local minima while training a more generic class of nonlinear DNNs. 2 A Conceptual Overview of the Contributions (1) Bounding norms of weights: We propose a two-stage renormalization procedure. First, we normalize weights according to the Euclidean, Frobenius and spectral norm, since they are used in the bounds of generalization errors [2 8]. Second, we aim to reparameterize the normalized weights to set a finite and constant upper bound on the weight matrices. For this purpose, we can use a parameter learning approach as utilized in batch normalization (BN) [16]. However, such an approach substantially increases running time of DNNs during training. In addition, it is not efficient to estimate the parameters using small number of samples in batch training. Therefore, we reparameterize weights according to (a) geometric properties of weight spaces, and (b) statistical properties of features (standard deviation) on which the weights are applied. Using (b), we can also decrease computation time of statistical properties especially when we apply separable operations such as channel/depth wise separable convolutions [17 19]. The proposed reparameterization method enables to set upper bound of each different norm of weight matrices to 1.0. In addition, the proposed renormalization procedure enables to control variance of weights during training of DNNs, thereby assures that DNNs do not have spurious local minima [20]. Employment of standard deviation in reparameterization also makes optimization landscapes significantly smoother by bounding amount of change of norms of gradients during training. This property has been recently studied to analyze effect of BN on optimization landscape in [21]. We use this property to develop a new optimization method for weight renormalization in this paper, as explained in the next problem. (2) Training DNNs with renormalized weights: Note that, there is not a single procedure used to normalize weights jointly according to all different norms. Thereby, we normalize weights in groups such that similar or different norms can be used to normalize matrices of weights belonging to each different group. We can mathematically prove that the procedure proposed to solve the previous problem (1) can set an upper bound for all of the aforementioned norms. However, we do not have a mathematical proof to explain whether weights normalized according a single norm can provide the best generalization bound, and to determine its type. We examine this question in various experiments in detail in the supp. mat. Since we cannot mathematically verify this observation, we conjecture that using a diverse set of weights normalized with different constraints improves the generalization error compared to using weights normalized according to single constraint. We consider mathematical characterization of this property as an open problem. Our contributions are summarized as follows: 1. DNNs trained using weights renormalized by the proposed method can achieve tighter bounds for theoretical generalization errors compared to using unnormalized weights (see Proposition 1 in the supp. mat. for derivation). These DNNs do not have spurious local minima [20] (see the next section for a detailed discussion). The proposed scaling method generalizes the scaling method proposed in [22] for weight normalization by incorporating geometric properties of weight manifolds. 2. We explicate the geometry of weight manifolds defined by multiple constraints in DNNs. For this purpose, we explore the relationship between geometric properties of POMs (i.e. sectional curvature), gradients computed at POMs (Theorem 1), and those of component manifolds of weights in DNNs in Section 5 (please see Lemma 1 in the supp. mat. for more precise results). 3. We propose an algorithm (FG-SGD) for optimization on different collections of POMs (Section 5) by generalizing SGD methods employed on weight manifolds [12, 23]. Next, we explore the effect of geometric properties of the POMs on the convergence of the FG-SGD using our theoretical results. In the proof of convergence theorems, we observe that gradients of weights should satisfy a particular normalization requirement and we employ this requirement for adaptive computation of step size of the FG-SGD (see (5) in Section 5.2.2). We also provide an example for computation of a step size function for optimization on POMs identified by the sphere (Corollary 2 in the supp. mat.). 4. We prove that loss functions of DNNs trained using the proposed FG-SGD converges to minima almost surely (see Theorem 2 and Corollary 1 in the supplemental material). 3 Construction of Sets of POMs in DNNs Let S = {si = (Ii,yi)}N i=1 be a set of training samples, where yi is a class label of the ith image Ii. We consider an L-layer DNN consisting of a set of tensors W = {Wl}L l=1, where Wl = {Wd,l RAl Bl Cl}Dl d=1, and Wd,l = [Wc,d,l RAl Bl]Cl c=1 is a tensor2 of weight matrices Wc,d,l, l = 1,2,...,L, for each cth channel c = 1,2,...,Cl and each dth weight d = 1,2,...,Dl. In popular DNNs, weights with Al = 1 and Bl = 1 are used at fully connected layers, and those with Al > 1 or Bl > 1 are used at convolutional layers. At each lth layer, a feature representation fl(Xl;Wl) is computed by compositionally employing non-linear functions by fl(Xl;Wl) = fl( ;Wl) fl 1( ;Wl 1) f1(X1;W1), (1) where Xl = [Xc,l]Cl c=1, and X1 = I is an image at the first layer (l = 1). The cth channel of the data matrix Xc,l is convolved with the kernel Wc,d,l to obtain the dth feature map Xc,l+1 = q( ˆXd,l) by ˆXd,l = Wc,d,l Xc,l, c,d,l, where q( ) is a non-linear function, such as Re LU. Previous works [12, 23] employ SGD using weights each of which reside on a single manifold3 at each layer of a DNN. We extend this approach considering that each weight can reside on an individual manifold or on collections of products of manifolds, which are defined next. Definition 1 (Products of weight manifolds and their collections). Suppose that Gl = {Mι,l ι IGl} is a set of weight manifolds3 Mι,l of dimension nι,l, which is identified by a set of indices IGl, l = 1,2,...,L. More concretely, IGl contains indices each of which represents an identity number (ι) of a weight that resides on a manifold Mι,l at the lth layer. In addition, a subset Ig l IGl,g = 1,2,...,Gl, is used to determine a subset Gg l Gl of weight manifolds which will be 2We use shorthand notation for matrix concatenation such that [Wc,d,l]Cl c=1 [W1,d,l, W2,d,l, , WCl,d,l]. 3In this work, we consider Riemannian manifolds of normalized weights defined in the previous section. Formal definitions are given in the supp. mat. Table 1: Comparison of generalization bounds. O denotes big-O and O is soft-O. δl,F , δl,2, and δl,2 1 denotes upper bounds of the Frobenius norm ωl F δl,F , spectral norm ωl 2 δl,2 and the sum of the Euclidean norms for all rows ωl 2 1 δl,2 1 (ℓ2 1) of weights ωl at the lth layer of an L layer DNN using N samples. DNNs (dynamic group scaling) Neyshabur et al. [24] O( Gl g=1 δg,l,F Bartlett et al. [3] O Gl g=1 δg,l,2 Gl g=1( δg,l,2 1 δg,l,2 ) 2 3 ) Neyshabur et al. [8] O Gl g=1 δg,l,2 Á Á ÀL2ϖ L l=1 δ2 g,l,F δ2 g,l,2 Table 2: Comparison of norms of weights belonging to different weight manifolds. Suppose that weights ωi g,l RAl Bl belonging to the gth group of size g ,g = 1,2,...,Gl, l have the same size Al Bl for simplicity, and σ(ωi g,l) denotes the top singular value of ωi g,l. ωi g,l F , ωi g,l 2, and ωi g,l 2 1, denotes respectively the Frobenius, spectral and ℓ2 1 norms of the weight ωi g,l. Norms (i) Sphere (ii) Stiefel (iii) Oblique ωi g,l 2 σ(ωi g,l) 1.0 σ(ωi g,l) ωi g,l F 1.0 (Bl)1/2 (Bl)1/2 ωi g,l 2 1 1.0 (Bl)1/4 (Bl)1/4 aggregated to construct a product of weight manifolds (POM). Each Mι,l Gg l is called a component manifold of a product of weight manifolds which is denoted by Mg,l. A weight ωg,l Mg,l is obtained by concatenating weights belonging to Mι,l, ι Ig l , using ωg,l = (ω1,ω2, ,ω Ig l ), where Ig l is the cardinality of Ig l . A Gl is called a collection of POMs. We propose three schemes called POMs for input channels (PI), for output channels (PO) and input/output channels (PIO) to construct index sets. Indices of the sets are selected randomly using a hypergeometric distribution without replacement at the initialization of a training step, and fixed in the rest of the training. Implementation details and experimental analyses are given in the supp. mat. 4 Bounding Generalization Errors using Fine-grained Weights Mathematically, norms of concatenated weights ωg,l, g, are lower bounded by products of norms of component weights ωi g,l, i. Weights are rescaled dynamically at each tth epoch of an optimization method proposed to train DNNs using Rt i,l = γi,l λt i,l , where γi,l > 0 is a geometric scaling parameter and λt i,l is the standard deviation of features input to the ith weight in the gth group ωi g,l, i,g. The scaling parameter Rt i,l enables us to upper bound the norms of weights by 1 (see Table 1). Suppose that all layers have the same width ϖ, weights have the same length K and the same stride s. Then, generalization bounds are obtained for DNNs using these fixed parameters by ωl 2 = K s , ωl F = ϖ and ωl 2 1 = ϖ. We compute a concatenated weight matrix ωg,l = (ω1 g,l,ω2 g,l,...,ω g g,l) for the gth weight group of size g ,g = 1,2,...,Gl, l using a weight grouping strategy. Then, we have upper bounds of norms by ωg,l F δg,l,F 1, ωg,l 2 δg,l,2 1 and ωg,l 2 1 δg,l,2 1 1,g = 1,2,...,Gl, which are defined in Table 2. 4.1 The Proof strategy for Computation of the Upper Bounds The upper bounds are computed in Proposition 1 in the supplemental material. The proof strategy is summarized as follows: Let bi,l be multiplication of the number of input channels and the size of the receptive field of the unit that employs ωi g,l, and ˆbi,l be multiplication of the dimension of output feature maps and the number of output channels used at the lth layer, respectively. Then, geometric scaling γi,l of the weight space of ωi g,l is computed by bi,l + ˆbi,l . (2) We can consider that standard deviation of features satisfy λt i,l 1 using two approaches. First, by employing the central limit theory for weighted summation of random variables of features, we can prove that λt i,l converges to 1 asymptotically, as popularly employed in the previous works. Second, we can assume that we apply batch normalization (BN) by setting the re-scaling parameter of the BN to 1. Thereby, we can obtain 1 λt i,l 1. By definition, γ2 i,l < Bl, i,l. In order to show that σ(ωi g,l) (γi,l) 1, i,l, we apply the Bai-Yin law [25, 26]. Thereby, we conclude that norms of concatenated weights belonging to groups given in Table 1 are upper bounded by 1, if the corresponding component weights given in Table 2 are rescaled by Rt i,l, i,l,t during training. Remark 1. We compute norms of weights belonging to each different manifold in Table 2. Fol- lowing Proposition 1 given in the supplemental material, we have ωg,l F ( g i=1 ωi g,l F )1/ g , ωg,l 2 ( g i=1 ωi g,l 2)1/ g and ωg,l 2 1 ( g i=1 ωi g,l 2 1)1/ g . 4.2 Generalization of Scaled Weight Initialization Methods Note that scaling by Rt i,l computed using (2) is different from the scaling method suggested in [12] such that our proposed method assures tighter upper bound for norms of weights. Our method also generalizes the scaling method given in [27] (a.k.a. Xavier initialization)in two ways. First, we use size of input receptive fields and output feature spaces which determine dimension of weight manifolds, as well as number of input and output dimensions which determine number of manifolds used in groups. Second, we perform scaling not just at initialization but also at each tth epoch of the optimization method. Therefore, diversity of weights is controlled and we can obtain weights uniformly distributed on the corresponding manifolds whose geometric properties change dynamically at each epoch. Applying this property with the results given in [20], we can prove that NNs applying the proposed scaling have no spurious local minima4. In addition, our method generalizes the scaling method proposed in [22] for weight normalization by incorporating geometric properties of weight manifolds. 5 Optimization using Fine-Grained Stochastic Gradient Descent in DNNs In this section, we address the problem of optimization of DNNs considering constraints on weight matrices according the upper bounds on their norms given in Section 4. 5.1 Optimization on POMs in DNNs: Challenges Employment of a vanilla SGD on POMs with assurance to convergence to local or global minima for training DNNs using back-propagation (BP) with collections of POMs is challenging. More precisely, we observe early divergence of SGD, and exploding and vanishing gradients in practice, due to the following theoretical properties of collections of POMs: 4We omit the formal theorem and the proof on this result in this work to focus on our main goal and novelty for optimization with multiple weight manifolds. Algorithm 1 Optimization using FG-SGD on products manifolds of fine-grained weights. 1: Input: T (number of iterations), S (training set), Θ (set of hyperparameters), L (a loss function), Il g IGl, g,l. 2: Initialization: Construct a collection of products of weight manifolds Gl, initialize re-scaling parameters Rt l and initialize weights ωt g,l Mg,l with Il g IGl, m,l. 3: for each iteration t = 1,2,...,T do 4: for each layer l = 1,2,...,L do 5: grad L(ωt g,l) = Πωt g,l(grad E L(ωt g,l),Θ,Rt l), Gl. 6: vt = h(grad L(ωt g,l),r(t,Θ)), Gl. 7: ωt+1 g,l = φωt g,l(vt,Rt l), ωt g,l, Gl. 8: end for 9: end for 10: Output: A set of estimated weights {ωT g,l}L l=1, g. Geometric properties of a POM Mg,l can be different from those of its component manifolds Mι, even if the component manifolds are identical. For example, we observe locally varying curvatures when we construct POMs of unit spheres. Weight manifolds with more complicated geometric properties can be obtained using the proposed PIO strategy, especially by constructing collections of POMs of non-identical manifolds. Therefore, assumption on existence of compact weight subsets in POMs may fail due to locally varying metrics within a nonlinear component manifold and among different component manifolds. When we optimize weights using SGD in DNNs, we first obtain gradients computed for each weight ωg,l Mg,l at the lth layer from the (l + 1)st layer using backpropagation (BP). Then, each weight ωg,l moves on Mg,l according to the gradient. However, curvatures and metrics of Mg,l can locally vary, and they may be different from those of component manifolds of Mg,l as explained above. This geometric drawback causes two critical problems. First, weights can be moved incorrectly if we move them using only gradients computed for each individual component of the weights, as popularly employed for the Euclidean linear weight spaces. Second, due to incorrect employment of gradients and movement of weights, probability of failure of the SGD cannot be bounded, and convergence cannot be achieved (see proofs of Theorem 2, Corollary 1 and Corollary 2 for details). In practice, this causes unbounded increase or decrease of values of gradients and weights. 5.2 A Geometric Approach to Optimization on POMs in DNNs In order to address these problems for training DNNs, we first analyze the relationship between geometric properties of POMs and those of their component manifolds in the next theorem. Remark 2. (See Lemma 1 given in the supp. mat. for the complete proof of the following propositions) Our main theoretical results regarding geometric properties of POMs are summarized as follows: 1. Computation of metrics: A metric defined on a product weight manifold Mg,l can be computed by superposition (i.e. linear combination) of Riemannian metrics of its component manifolds. 2. Lower bounds of sectional curvatures: Sectional curvature of a product weight manifold Mg,l is lower bounded by 0. 5.2.1 Development of FG-SGD employing Geometry of POMs We use the first result (1) for projection of Euclidean gradients obtained using BP onto product weight manifolds. More precisely, we can compute norms of gradients at weights on a product weight manifold by linear superposition of those computed on its component manifolds in FG-SGD. Thereby, we can move a weight on a product weight manifold by (i) retraction of components of the weight on component manifolds of the product weight manifold, and (ii) concatenation of projected weight components in FG-SGD. The second result (2) show that some sectional curvatures vanish on a product weight manifold Mg,l. For instance, suppose that each component weight manifold Mι,l of Mg,l is a unit two-sphere S2, ι IGl. Then, Mg,l has unit curvature along two-dimensional subspaces of its tangent spaces, called two-planes. However, Mg,l has zero curvature along all two-planes spanning exactly two distinct spheres. In addition, weights can always move according to a non-negative bound on sectional curvature of compact product weight manifolds on its tangent spaces. Therefore, we do not need to worry about varying positive and negative curvatures observed at its different component manifolds. The second result also suggests that learning rates need to be computed adaptively by a function of norms of gradients and bounds on sectional curvatures at each layer of the DNN and at each epoch of FG-SGD for each weight ω on each product weight manifold Mg,l. We employ these results to analyze convergence of FG-SGD and compute its adaptive step size in the following sections. 5.2.2 Optimization on POMs using FG-SGD in DNNs An algorithmic description of our proposed fine-grained SGD (FG-SGD) is given in Algorithm 1. At the initialization of the FG-SGD, we identify the component weight manifolds Mι,l of each product weight manifold Mg,l according to the constraints that will be applied on the weights ωι Mι,l for each gth group at each lth layer. For t = 1, each manifold Mι,l is scaled by Rt=1 ι,l using λt=1 ι,l = 1, ι,l. For t > 1, each Mι,l is re-scaled by Rt ι,l Rt l computing empirical standard deviation λt ι of features input to each weight of Mι,l, and Rt l is the set of all re-scaling parameters computed at the tth epoch at each lth layer. When we employ a FG-SGD on a product weight manifold Mg,l each weight ωt g,l Mg,l is moved on Mg,l in the descent direction of gradient of loss at each tth step of the FG-SGD by the following steps: Line 5 (Projection of gradients on tangent spaces): The gradient grad E L(ωt g,l), obtained using back-propagation from the upper layer, is projected onto the tangent space Tωt g,l Mg,l = ι Ilg Tωt ι,l Mι,l to compute grad L(ωt g,l) at the weight ωt g,l using the results given in Remark 2, where Tωt ι,l Mι,l is the tangent space at ωt ι,l on the component manifold Mι,l of Mg,l. Line 6 (Movement of weights on tangent spaces): The weight ωt g,l is moved on Tωt g,l Mg,l using h(grad L(ωt g,l),r(t,Θ)) = r(t,Θ) r(ωt g,l)grad L(ωt g,l), (3) where r(t,Θ) is the learning rate that satisfies t=0 r(t,Θ) = + and t=0 r(t,Θ)2 < , (4) r(ωt Gm l ) = max{1,Γt 1} 1 2 (5) Γt 1 = (Rt g,l)2Γt 2, Rt g,l grad L(ωt g,l) 2 is computed using (6), Γt 2 = max{(2ρt g,l + Rt g,l)2,(1 + cg,l(ρt g,l + Rt g,l))}, cg,l is the sectional curvature of Mg,l, ρt g,l ρ(ωt g,l, ˆωg,l) is the geodesic distance between ωt g,l and a local minima ˆωg,l on Mg,l. The following result is used for computation of the ℓ2 norm of gradients. Theorem 1 (Computation of gradients on tangent spaces). The ℓ2 norm grad L(ωt g,l) 2 of the gradient grad L(ωt g,l) residing on Tωt g,l Mg,l at the tth epoch and the lth layer can be computed by grad L(ωt g,l) 2 = ( ι Ilg grad L(ωt ι,l)2) where grad L(ωt ι,l) is the gradient computed for ωt ι,l on the tangent space Tωt ι,l Mι, ι Il g. We compute norms of gradients on tangent spaces of product manifolds by just superposing gradients computed on tangent spaces of component manifolds using (6). Norms of gradients, section curvatures of the product manifolds and their metric properties affect convergence of weights on the product manifolds to local or global minima. More precisely, they define an upper bound on the divergence of the weights from local or global minima (please see proof of Theorem 2 in the supplemental material for details). This upper bound is further bounded by normalizing gradients in (3). Line 7 (Projection of moved weights onto product of manifolds): The moved weight located at vt is projected onto Mg,l re-scaled by Rt l using φωt g,l(vt,Rt l) to compute ωt+1 g,l , where φωt g,l(vt,Rt l) is an exponential map, or a retraction, i.e. an approximation of the exponential map [28]. The function r(ωt g,l) used for computing step size in (3) is employed as a regularizer to control the change of gradient grad L(ωt g,l) at each step of FG-SGD. This property is examined in the experimental analyses in the supp. mat. For computation of r(ωt g,l), we use (6) with Theorem 1. In FG-SGD, weights residing on each POM are moved and projected jointly on the POMs, by which we can employ their interaction using the corresponding gradients considering nonlinear geometry of manifolds unlike SGD methods studied in the literature. G-SGD can consider interactions between component manifolds as well as those between POMs in groups of weights. Employment of (3) and (4) at line 7, and retractions at line 8 are essential for assurance of convergence as explained next. 5.3 Convergence Properties of FG-SGD Convergence properties of the proposed FG-SGD used to train DNNs are summarized as follows: Convergenge to local minima: The loss function of a non-linear DNN, which employs the proposed FG-SGD, converges to a local minimum, and the corresponding gradient converges to zero almost surely (a.s.). The formal theorem and proof are given in Theorem 2 in the supplemental material. Convergenge to global minima: Loss functions of particular DNNs such as linear DNNs, onehidden-layer CNNs, one-hidden-layer Leaky Relu networks, nonlinear DNNs with specific network structures (e.g. pyramidal networks), trained using FG-SGD, converge to a global minimum a.s. under mild assumptions on data (e.g. being distributed from Gaussian distribution, normalized, and realized by DNNs). The formal theorem and proof of this result are given in Corollary 1 in the supp. mat. The proof idea is to use the property that local minima of loss functions of these networks are global minima under these assumptions, by employing the results given in the recent works [29 38]. An example for adaptive computation of step size: Suppose that Mι are identified by nι 2 dimensional unit sphere, or the sphere scaled by the proposed scaling method. If step size is computed using (3) with r(ωt Gm l ) = (max{1,(Rt Gm l )2(2 + Rt Gm l )2}) 1 2 , (7) then the loss function converges to local minima for a generic class of nonlinear DNNs, and to global minima for DNNs characterized in Corollary 1. The formal theorem and proof of this result are given in Corollary 2 in the supp. mat. The proof idea follows the property that the unit sphere has positive sectional curvature, and a product of unit spheres has non-negative curvature. Using these results for the sphere, the normalizing function (5) is computed by the maximum of 1 and a polynomial function of the norm of gradients (6). Note that, variations of this function (7) have been used in practice to train large DNNs successfully, i.e. avoiding exploding and vanishing gradients. Our mathematical framework elucidates the underlying theory behind the success of gradient normalization methods for convergence of DNNs to local and global minima. 6 Conclusion and Discussion We introduced and elucidated a problem of training CNNs using multiple constraints employed on convolution weights with convergence properties. Following our theoretical results, we proposed the FG-SGD algorithm and adaptive step size estimation methods for optimization on collections of POMs that are identified by the constraints. Due to page limit, experimental analyses are given in the supplemental material. In these analyses, we observe that our proposed methods can improve convergence properties and classification performance of CNNs. Overall, the results show that employment of collections of POMs using FG-SGD can boost the performance of various different CNNs on various datasets. We consider a research direction for investigating how far local minima are from global minima in search spaces of FG-SGD using products of weight manifolds with nonlinear DNNs and their convergence rates. We believe that our proposed mathematical framework and results will be useful and inspiring for researchers to study geometric properties of parameter spaces of deep networks, and to improve our understanding of deep feature representations. [1] Chiyuan Zhang, Samy Bengio, Moritz Hardt, Benjamin Recht, and Oriol Vinyals. Understanding deep learning requires rethinking generalization. In Proc. of the 5th Int. Conf. on Learn. Rep. (ICLR), 2017. [2] Behnam Neyshabur, Srinadh Bhojanapalli, David Mcallester, and Nati Srebro. Exploring generalization in deep learning. In Advances in Neural Information Processing Systems (NIPS), pages 5947 5956. 2017. [3] Peter L Bartlett, Dylan J Foster, and Matus J Telgarsky. Spectrally-normalized margin bounds for neural networks. In Advances in Neural Information Processing Systems (NIPS), pages 6240 6249. 2017. [4] Taiji Suzuki. Fast generalization error bound of deep learning from a kernel perspective. In Proc. of the 21st Int. Conf. on Artificial Intelligence and Statistics (AISTATS), volume 84, pages 1397 1406, 2018. [5] Pan Zhou and Jiashi Feng. Understanding generalization and optimization performance of deep CNNs. In Proc. of the 35th Int. Conf. on Mach. Learn. (ICML), volume 80, pages 5960 5969, 2018. [6] Sanjeev Arora, Rong Ge, Behnam Neyshabur, and Yi Zhang. Stronger generalization bounds for deep nets via a compression approach. In Proc. of the 35th Int. Conf. on Mach. Learn. (ICML), volume 80, pages 254 263, 2018. [7] Noah Golowich, Alexander Rakhlin, and Ohad Shamir. Size-independent sample complexity of neural networks. In roc. of the 31st Conf. on Learning Theory (COLT), pages 297 299. 2018. [8] Behnam Neyshabur, Srinadh Bhojanapalli, and Nathan Srebro. A PAC-bayesian approach to spectrally-normalized margin bounds for neural networks. In Proc. of the 6th Int. Conf. on Learn. Rep. (ICLR), 2018. [9] Martin Anthony and Peter L. Bartlett. Neural Network Learning: Theoretical Foundations. Cambridge University Press, New York, NY, USA, 1st edition, 2009. [10] Mario Lezcano-Casado and David Martinez-Rubio. Cheap orthogonal constraints in neural networks: A simple parametrization of the orthogonal and unitary group. In Proc. of the 39th Int. Conf. on Mach. Learn. (ICML), 2019. [11] Mario Lezcano-Casado. Trivializations for gradient-based optimization on manifolds. In Advances in Neural Information Processing Systems (NIPS). 2019. [12] Mete Ozay and Takayuki Okatani. Training cnns with normalized kernels. In AAAI Conference on Artificial Intelligence, 2018. [13] Haoran Chen, Yanfeng Sun, Junbin Gao, Yongli Hu, and Baocai Yin. Partial least squares regression on riemannian manifolds and its application in classifications. Co RR, abs/1609.06434, 2016. [14] Yui Man Lui. Human gesture recognition on product manifolds. J. Mach. Learn. Res., 13(1):3297 3321, Nov 2012. [15] Minhyung Cho and Jaehyung Lee. Riemannian approach to batch normalization. In Advances in Neural Information Processing Systems (NIPS), 2017. [16] Sergey Ioffe and Christian Szegedy. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proc. of the 35th Int. Conf. on Mach. Learn. (ICML), pages 448 456, 2015. [17] Alex Krizhevsky, Ilya Sutskever, and Geoffrey E. Hinton. Imagenet classification with deep convolutional neural networks. In Advances in Neural Information Processing Systems (NIPS), pages 1097 1105, 2012. [18] J. Luo, H. Zhang, H. Zhou, C. Xie, J. Wu, and W. Lin. Thinet: Pruning cnn filters for a thinner net. IEEE Transactions on Pattern Analysis and Machine Intelligence, pages 1 1, 2018. [19] Guotian Xie, Ting Zhang, Kuiyuan Yang, Jianhuang Lai, and Jingdong Wang. Decoupled convolutions for cnns, 2018. [20] Bo Xie, Yingyu Liang, and Le Song. Diverse Neural Network Learns True Target Functions. In Proc. of the 20th Int. Conf. on Artificial Intelligence and Statistics (AISTATS), volume 54, pages 1216 1224, 2017. [21] Shibani Santurkar, Dimitris Tsipras, Andrew Ilyas, and Aleksander Madry. How does batch normalization help optimization? (no, it is not about internal covariate shift). In Advances in Neural Information Processing Systems (NIPS). 2018. [22] Tim Salimans and Diederik P. Kingma. Weight normalization: A simple reparameterization to accelerate training of deep neural networks. In Advances in Neural Information Processing Systems (NIPS), 2016. [23] Minhyung Cho and Jaehyung Lee. Riemannian approach to batch normalization. In Advances in Neural Information Processing Systems (NIPS), pages 5231 5241. 2017. [24] Behnam Neyshabur, Ryota Tomioka, and Nathan Srebro. Norm-based capacity control in neural networks. In Proc. of the 28th Conf. on Learning Theory (COLT), volume 40, pages 1376 1401, 2015. [25] Z. D. Bai and Y. Q. Yin. Limit of the smallest eigenvalue of a large dimensional sample covariance matrix. The Annals of Probability, 21(3):1275 1294, 07 1993. [26] Z. D Bai, Jack W Silverstein, and Y.Q Yin. A note on the largest eigenvalue of a large dimensional sample covariance matrix. Journal of Multivariate Analysis, 26(2):166 168, 1988. [27] Xavier Glorot and Yoshua Bengio. Understanding the difficulty of training deep feedforward neural networks. In Proc. of the 13th Int. Conf. on Artificial Intelligence and Statistics (AISTATS), volume 9, pages 249 256, 2010. [28] P. A. Absil and Jerome Malick. Projection-like retractions on matrix manifolds. SIAM Journal on Optimization, 22(1):135 158, 2012. [29] Kenji Kawaguchi. Deep learning without poor local minima. In D. D. Lee, M. Sugiyama, U. V. Luxburg, I. Guyon, and R. Garnett, editors, Advances in Neural Information Processing Systems (NIPS), pages 586 594. 2016. [30] Alon Brutzkus and Amir Globerson. Globally optimal gradient descent for a convnet with gaussian inputs. In Proc. of the 34th Int. Conf. on Mach. Learn. (ICML), pages 605 614, 2017. [31] Simon S. Du, Jason D. Lee, Yuandong Tian, Barnabás Póczos, and Aarti Singh. Gradient descent learns one-hidden-layer cnn: Don t be afraid of spurious local minima. In Proc. of the 35th Int. Conf. on Mach. Learn. (ICML), 2018. [32] Simon S. Du and Jason D. Lee. On the power of over-parametrization in neural networks with quadratic activation. In Proc. of the 35th Int. Conf. on Mach. Learn. (ICML), 2018. [33] Chulhee Yun, Suvrit Sra, and Ali Jadbabaie. Global optimality conditions for deep neural networks. In Proc. of the 6th Int. Conf. on Learn. Rep. (ICLR), 2018. [34] Moritz Hardt and Tengyu Ma. Identity matters in deep learning. In Proc. of the 5th Int. Conf. on Learn. Rep. (ICLR), 2017. [35] Rong Ge, Jason D. Lee, and Tengyu Ma. Learning one-hidden-layer neural networks with landscape design. In Proc. of the 5th Int. Conf. on Learn. Rep. (ICLR), 2017. [36] Chulhee Yun, Suvrit Sra, and Ali Jadbabaie. A critical view of global optimality in deep learning. Co RR, abs/1802.03487, 2018. [37] Maithra Raghu, Ben Poole, Jon Kleinberg, Surya Ganguli, and Jascha Sohl-Dickstein. On the expressive power of deep neural networks. In Proc. of the 34th Int. Conf. on Mach. Learn. (ICML), volume 70, pages 2847 2854, 2017. [38] Quynh Nguyen and Matthias Hein. The loss surface of deep and wide neural networks. In Proc. of the 34th Int. Conf. on Mach. Learn. (ICML), volume 70, pages 2603 2612, 2017. [39] Jie Hu, Li Shen, and Gang Sun. Squeeze-and-excitation networks. In Proc. IEEE Conf. Comp. Vis. Patt. Recog. (CVPR), 2018. [40] Sergey Ioffe and Christian Szegedy. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proc. of the 32nd Int. Conf. Mach. Learn., volume 37, pages 448 456, 2015.