# principled_architectureaware_scaling_of_hyperparameters__fe3c1c5f.pdf Published as a conference paper at ICLR 2024 PRINCIPLED ARCHITECTURE-AWARE SCALING OF HYPERPARAMETERS Wuyang Chen1 , Junru Wu2 , Zhangyang Wang1, Boris Hanin3 1University of Texas, Austin 2Google Research 3Princeton University 1{wuyang.chen,atlaswang}@utexas.edu 2junru@google.com 3bhanin@princeton.edu Training a high-quality deep neural network requires choosing suitable hyperparameters, which is a non-trivial and expensive process. Current works try to automatically optimize or design principles of hyperparameters, such that they can generalize to diverse unseen scenarios. However, most designs or optimization methods are agnostic to the choice of network structures, and thus largely ignore the impact of neural architectures on hyperparameters. In this work, we precisely characterize the dependence of initializations and maximal learning rates on the network architecture, which includes the network depth, width, convolutional kernel size, and connectivity patterns. By pursuing every parameter to be maximally updated with the same mean squared change in preactivations, we can generalize our initialization and learning rates across MLPs (multi-layer perception) and CNNs (convolutional neural network) with sophisticated graph topologies. We verify our principles with comprehensive experiments. More importantly, our strategy further sheds light on advancing current benchmarks for architecture design. A fair comparison of Auto ML algorithms requires accurate network rankings. However, we demonstrate that network rankings can be easily changed by better training networks in benchmarks with our architecture-aware learning rates and initialization. Our code is available at https: //github.com/VITA-Group/principled_scaling_lr_init. 1 INTRODUCTION An important theme in the success of modern deep learning is the development of novel neural network architectures: Conv Nets (Simonyan & Zisserman, 2014), Res Nets (He et al., 2016), Transformers (Dosovitskiy et al., 2020) to name a few. Whether a given architecture works well in practice, however, depends crucially on having reliable principles for selecting training hyperparameters. Indeed, even using more or less standard architectures at scale often requires considerable hyperparameter tuning. Key examples of such hyperparameters are the learning rate and the network initialization, which play a crucial role in determining the quality of trained models. How to choose them in practice depends non-trivially on the interplay between data, architecture, and optimizer. Principles for matching deep networks to initialization and optimization schemes have been developed in several vibrant lines of work. This includes approaches typically based on analysis in the infinite width limit to developing initialization schemes suited to specific architectures, including fully connected architectures (Jacot et al., 2018; Perrone et al., 2018), Conv Nets (Ilievski & Feng, 2016; Stoll et al., 2020), Res Nets (Horváth et al., 2021; Li et al., 2022b;a), and Transformers (Zhu et al., 2021; Dinan et al., 2023). It also includes several approaches to zero-shot hyperparameter transfer, which seek to establish the functional relationships between a hyperparameter (such as a good learning rate) and the architecture characteristics, such as depth and width. This relation then allows for hyperparameter tuning on small, inexpensive models, to be transferred reliably to larger architectures (Yang et al., 2022; Iyer et al., 2022; Yaida, 2022). Architectures of deep networks, with complicated connectivity patterns and heterogeneous operations, are common in practice and especially important in neural architectures search (Liu et al., 2018; Work done while at Texas A&M University Published as a conference paper at ICLR 2024 Dong & Yang, 2020). This is important because long-range connections and heterogeneous layer types are common in advanced networks (He et al., 2016; Huang et al., 2017; Xie et al., 2019). However, very little prior work develops initialization and learning rate schemes adapted to complicated network architectures. Further, while automated methods for hyperparameter optimization (Bergstra et al., 2013) are largely model agnostic, they cannot discover generalizable principles. As a partial remedy, recent works (Zela et al., 2018; Klein & Hutter, 2019; Dong et al., 2020b) jointly design networks and hyperparameters, but still ignore the dependence of learning rates on the initialization scheme. Moreover, ad-hoc hyperparameters that are not tuned in an intelligent way will further jeopardize the design of novel architectures. Without a fair comparison and evaluation of different networks, benchmarks of network architectures may be misleading as they could be biased toward architectures on which standard initializations and learning rates we use happen to work out well. This article seeks to develop principles for both initialization schemes and zero-shot learning rate selection for general neural network architectures, specified by a directed acyclic graph (DAG). These architectures can be highly irregular (Figure 1), significantly complicating the development of simple intuitions for training hyperparameters based on experience with regular models (e.g. feedforward networks). At the beginning of gradient descent training, we first derive a simple initialization scheme that preserves the variance of pre-activations during the forward propagation of irregular architectures, and then derive architecture-aware learning rates by following the maximal update (µP) prescription from (Yang et al., 2022). We generalize our initialization and learning rate principles across MLPs (multi-layer perception) and CNNs (convolutional neural network) with arbitrary graph-based connectivity patterns. We experimentally verify our principles on a wide range of architecture configurations. More importantly, our strategy further sheds light on advancing current benchmarks for architecture design. We test our principles on benchmarks for neural architecture search, and we can immediately see the difference that our architecture-aware initializations and architecture-aware learning rates could make. Accurate network rankings are required such that different Auto ML algorithms can be fairly compared by ordering their searched architectures. However, we demonstrate that network rankings can be easily changed by better training networks in benchmarks with our architecture-aware learning rates and initialization. Our main contributions are: We derive a simple modified fan-in initialization scheme that is architecture-aware and can provably preserve the flow of information through any architecture s graph topology ( 3.2). Using this modified fan-in initialization scheme, we analytically compute the dependence on the graph topology for how to scale learning rates in order to achieve the maximal update (µP) heuristic (Yang et al., 2022), which asks that in the first step of optimization neuron pre-activations change as much as possible without diverging at large width. For example, we find that ( 3.3) for a Re LU network of a graph topology trained with MSE loss under gradient descent, the µP learning rate scales as PP p=1 L3 p 1/2 , where P is the total number of end-to-end paths from the input to the output, Lp is the depth of each path, p = 1, , P. In experiments, we not only verify the superior performance of our prescriptions, but also further re-evaluate the quality of standard architecture benchmarks. By unleashing the potential of architectures (higher accuracy than the benchmark), we achieve largely different rankings of networks, which may lead to different evaluations of Auto ML algorithms for neural architecture search (NAS). 2 RELATED WORKS 2.1 HYPERPARAMETER OPTIMIZATION The problem of hyperparameter optimization (HPO) has become increasingly relevant as machine learning models have become more ubiquitous (Li et al., 2020; Chen et al., 2022; Yang et al., 2022). Poorly chosen hyperparameters can result in suboptimal performance and training instability. Beyond the basic grid or random search, many works tried to speed up HPO. (Snoek et al., 2012) optimize HPO via Bayesian optimization by jointly formulating the performance of each hyperparameter configuration as a Gaussian process (GP), and further improved the time cost of HPO by replacing the GP with a neural network. Meanwhile, other efforts tried to leverage large-scale parallel infrastructure to speed up hyperparameter tuning (Jamieson & Talwalkar, 2016; Li et al., 2020). Published as a conference paper at ICLR 2024 Our approach is different from the above since we do not directly work on hyperparameter optimization. Instead, we analyze the influence of network depth and architecture on the choice of hyperparameters. Meanwhile, our method is complementary to any HPO method. Given multiple models, typical methods optimize hyperparameters for each model separately. In contrast, with our method, one only needs to apply HPO once (to find the optimal learning rate for a small network), and then directly scale up to larger models with our principle ( 4). 2.2 HYPERPARAMETER TRANSFER Recent works also tried to propose principles of hyperparameter scaling and targeted on better stability for training deep neural networks (Glorot & Bengio, 2010; Schoenholz et al., 2016; Yang & Schoenholz, 2017; Zhang et al., 2019; Bachlechner et al., 2021; Huang et al., 2020; Liu et al., 2020). Some even explored transfer learning of hyperparameters (Yogatama & Mann, 2014; Perrone et al., 2018; Stoll et al., 2020; Horváth et al., 2021). (Smith et al., 2017; Hoffer et al., 2017) proposed to scale the learning rate with batch size while fixing the total epochs of training. (Shallue et al., 2018) demonstrated the failure of finding a universal scaling law of learning rate and batch size across a range of datasets and models. Assuming that the optimal learning rate should scale with batch size, (Park et al., 2019) empirically studied how the optimal ratio of learning rate over batch size scales for MLP and CNNs trained with SGD. More recently, Yang & Hu (2020) found that standard parameterization (SP) and NTK parameterization (NTP) lead to bad infinite-width limits and hence are suboptimal for wide neural networks. Yang et al. (2022) proposed µP initialization, which enabled the tuning of hyperparameters indirectly on a smaller model, and zero-shot transfer them to the full-sized model without any further tuning. In (Iyer et al., 2022), an empirical observation was found that, for fully-connected deep Re LU networks, the maximal initial learning rate follows a power law of the product of width and depth. Jelassi et al. (2023) recently studied the depth dependence of µP learning rates vanilla Re LU-based MLP networks. Two core differences between µP and our method: 1) µP mainly gives width-dependent learning rates in the first and last layer, and is width-independent in other layers (see Table 3 in (Yang et al., 2022)). However, there is a non-trivial dependence on both network depth and topologies; 2) Architectures studied in µP mainly employ sequential structures without complicated graph topologies. 2.3 BECHMARKING NEURAL ARCHITECTURE SEARCH Neural architecture search (NAS), one research area under the broad topic of automated machine learning (Auto ML), targets the automated design of network architecture without incurring too much human inductive bias. It is proven to be principled in optimizing architectures with superior accuracy and balanced computation budgets (Zoph & Le, 2016; Tan et al., 2019; Tan & Le, 2019). To fairly and efficiently compare different NAS methods, many benchmarks have been developed (Ying et al., 2019; Dong & Yang, 2020; Dong et al., 2020a). Meanwhile, many works explained the difficulty of evaluating NAS. For example, (Yang et al., 2019) emphasized the importance of tricks in the evaluation protocol and the inherently narrow accuracy range from the search space. However, rare works point out one obvious but largely ignored defect in each of these benchmarks and evaluations: diverse networks are blindly and equally trained under the same training protocol. It is questionable if the diverse architectures in these benchmarks would share the same set of hyperparameters. As a result, to create a credible benchmark, it s necessary to find optimal hyperparameters for each network accurately and efficiently, which our method aims to solve. This also implied, NAS benchmarks that were previously regarded as gold standard , could be no longer reliable and pose questions on existing NAS algorithms that evaluated upon it. For general network architectures of directed acyclic computational graphs (DAGs), we propose topology-aware initialization and learning rate scaling. Our core advantage lies in that we precisely characterize the non-trivial dependence of learning rate scales on both networks depth, DAG topologies and CNN kernel sizes, beyond previous heuristics, with practical implications ( 3.5). 1. In a DAG architecture ( 3.1), an initialization scheme that depends on a layer s in-degree is required to normalize inputs that flow into this layer ( 3.2). Published as a conference paper at ICLR 2024 2. Analyzing the initial step of SGD allows us to detail how changes in pre-activations depend on the network s depth and learning rate ( B.1. 3. With our in-degree initialization, by ensuring the magnitude of pre-activation changes remains Θ(1), we deduce how the learning rate depends on the network topology ( 3.3) and the CNN kernel size ( 3.4). 3.1 DEFINITION OF DAG NETWORKS a DAG space = {MLP, Res Net, } 𝒛(2) 𝒛(3) 𝒛(4) # end-to-end paths Linear + Re LU Skip-connection (identity) Zero (disabled edge, no forward/backward) 𝑾: operator (layer) Depth (𝑑C): # Linear+Re LU on path 𝑝 Figure 1: A neural network s architecture can be represented by a direct acyclic graph (DAG). x is the input, z(1), z(2), , z(L) are pre-activations (vertices), and z(L+1) is the output. W is the layer operation (edge). Our DAG space includes architectures of different connections and layer types ( Linear + Re LU , Skip-connect , and Zero ). For example, by disabling some edges (gray) or replacing with identity I (skip connection, dashed arrow), a DAG can represent practical networks such as MLP and Res Net (He et al., 2016). We first define the graph topology of complicated neural architectures. This formulation is inspired by network structures designed for practical applications (Xie et al., 2019; You et al., 2020; Dong & Yang, 2020). By definition, a directed acyclic graph is a directed graph G = (V, E) in which the edge set has no directed cycles (a sequence of edges that starts and ends in the same vertex). We will denote by L + 2 := |V | the number of vertices (the input x and pre-activations z) in V and write: V = [0, L + 1] := {0, 1, . . . , L + 1} . We will adopt the convention that edges terminating in vertex ℓoriginate only in vertices ℓ < ℓ. We define a DAG neural network (with computational graph G = (V, E), the number of vertices |V | = L + 2, hidden layer width n, output dimension 1, and Re LU activations) to be a function x Rn 7 z(L+1)(x) R given by ( x, ℓ= 0 P (ℓ ,ℓ) E W (ℓ ,ℓ)σ(z(ℓ )(x)), ℓ= 1, . . . , L + 1 , (1) σ(t) = Re LU(t) = max{0, t}. W (ℓ ,ℓ) are n n matrices of weights initialized as follows: W (ℓ ,ℓ) ij ( N(0, C(ℓ ,ℓ)/n), ℓ= 1, . . . , L N(0, C(ℓ ,ℓ)/n2), ℓ= L + 1 , (2) where C(ℓ ,ℓ) is a scaling variable that we will determine based on the network topology (equation 4). We choose 1/n2 as the variance for the output layer mainly for aligning with µP s desiderata. 3.2 TOPOLOGY-AWARE INITIALIZATION SCHEME A common way to combine features from multiple layers is to directly sum them up, often without normalization (He et al., 2016; Dong & Yang, 2020) as we do in equation 1. This summation implies that the weight magnitude of each pre-activation layer should scale proportionally regarding the number of inputs (in-degree). However, most commonly used weight initialization schemes, such as He initialization (He et al., 2015) and Xavier initialization (Glorot & Bengio, 2010), only consider forward or backward propagation between consecutive layers, ignoring any high-level structures, i.e. how layers are connected with each other (see also (Defazio & Bottou, 2022)). This strategy can be problematic in the presence of a bottleneck layer that has a large number of inputs. Motivated by this issue, we propose an architecture-aware initialization strategy. We target deriving the normalizing factor C(ℓ ,ℓ) in the initialization (equation 2) adapted to potentially very different in-degrees of the target layer. Like many theoretical approaches to neural network initialization, we do this by seeking to preserve the flow of information through the network at initialization. While this dictum can take various guises, we interpret it by asking that the expected norm of changes in pre-activations between consecutive layer have the same typical scale: E z(ℓ) 2 = E z(ℓ ) 2 for all ℓ, ℓ , (3) Published as a conference paper at ICLR 2024 where E [ ] denotes the average over initialization. This condition essentially corresponds to asking for equal scales in any hidden layer representations. Our first result, which is closely related to the ideas developed for MLPs (Theorem 2 in (Defazio & Bottou, 2022)), gives a simple criterion for ensuring that equation 3 holds for a DAG-structured neural network with Re LU layers: Initialization scaling for DAG. For a neural network with directed acyclic computational graph G = (V, E), V = [0, L+1], width n, and Re LU activations, the information flow condition equation 3 is satisfied on average over a centred random input x if C(ℓ ,ℓ) = 2 d(ℓ ) in , (4) where for each vertex ℓ [0, L + 1], we ve written d(ℓ) in for its in-degree in G. This implies that, when we initialize a layer, we should consider how many features flow into it. We include our derivations in Appendix A. 3.3 TOPOLOGY-AWARE LEARNING RATES The purpose of this section is to describe a specific method for choosing learning rates for DAG networks. Our approach follows the maximal update (µP) heuristic proposed in (Yang et al., 2022). Specifically, given a DAG network recall from equation 1 that z(ℓ)(x) = z(ℓ) i (x), i = 1, . . . , n denotes the pre-activations at layer ℓcorresponding to a network input x. The key idea in (Yang et al., 2022) is to consider the change z(ℓ) i of z(ℓ) i from one gradient descent step on training loss L: µ ℓ µz(ℓ) i µ = η X µ ℓ µz(ℓ) i µL, (5) where µ denotes a network weight, µ ℓmeans that µ is a weight in W (ℓ ,ℓ) for some ℓ ℓ 1, µ for the change in µ after one step of GD, η is the learning rate. By symmetry, all weights in layer ℓshould have the same learning rate at initialization and we will take as an objective function the MSE loss: z(L+1)(x; θ) y 2 (6) over a batch B of training datapoints. The µP heuristic posited in (Yang et al., 2022) states that the best learning rate is the largest one for which z (ℓ) i remains O(1) for all ℓ= 1, . . . , L + 1 independent of the network width n. To make this precise, we will mainly focus on analyzing the average squared change (second moment) of z (ℓ) i and ask to find learning rates such that E z(ℓ) i 2 = Θ(1) for all ℓ [1, L + 1] (7) with mean-field initialization, which coincides with equation 2 except that the final layer weights are initialized with variance 1/n2 instead of 1/n. Our next result provides a way to estimate how the µP learning rate (i.e. the one determined by equation 7) depends on the graph topology. To state it, we need two definitions: DAG Width. Given a DAG G = (V, E) with V = {0, . . . , L + 1} we define its width P to be the number of unique directed paths from the input vertex 0 to output vertex L + 1. Path Depth. Given a DAG G = (V, E), the number of Re LU activations in a directed path p from the input vertex 0 to the output vertex L + 1 is called its depth, denoted as Lp (p = 1, , P). We are now ready to state the main result of this section: Published as a conference paper at ICLR 2024 Learning rate scaling in DAG. Consider a Re LU network with associated DAG G = (V, E), hidden layer width n, hidden layer weights initialized as in 3.2 and the output weights initialized with mean-field scaling. Consider a batch with one training datapoint (x, y) R2n sampled independently of network weights and satisfying n x 2 = 1, E [y] = 0, Var[y] = 1. The µP learning rate (η ) for hidden layer weights is independent of n but has a non-trivial dependence of depth, scaling as where the sum is over P unique directed paths through G, and the implicit constant c is independent of the graph topology and of the hidden layer width n. This result indicates that, we should use smaller learning rates for networks with large depth and width (of its graph topology); and vice versa. We include our derivations in Appendix B. 3.4 LEARNING RATES IN DAG NETWORK WITH CONVOLUTIONAL LAYERS In this section, we further expand our scaling principles to convolutional neural networks, again, with arbitrary graph topologies as we discussed in 3.1. Setting. Let x Rn m be the input, where n is the number of input channels and m is the number of pixels. Given z(ℓ) Rn m for ℓ [1, L + 1], we first use an operator ϕ( ) to divide z(ℓ) into m patches. Each patch has size qn and this implies a mapping of ϕ(z(ℓ)) Rqn m. For example, when the stride is 1 and q = 3, we have: z(ℓ) 1,0:2 , . . . , z(ℓ) 1,m 1:m+1 . . . , . . . , . . . z(ℓ) n,0:2 , . . . , z(ℓ) n,m 1:m+1 where we let z(ℓ) :,0 = z(ℓ) :,m+1 = 0, i.e., zero-padding. Let W (ℓ) Rn qn. we have ( x, ℓ= 0 P (ℓ ,ℓ) E W (ℓ ,ℓ)σ(ϕ(z(ℓ )(x))), ℓ= 1, . . . , L + 1 . Learning rate scaling in CNNs. Consider a Re LU-CNN network with associated DAG G = (V, E), kernel size q, hidden layer width n, hidden layer weights initialized as in 3.2, and the output weights initialized with mean-field scaling. The µP learning rate η for hidden layer weights is independent of n but has a non-trivial dependence of depth, scaling as This result implies that in addition to network depth and width, CNNs with a larger kernel size should use a smaller learning rate. We include our derivations in Appendix C. 3.5 IMPLICATIONS ON ARCHITECTURE DESIGN We would like to emphasize that, beyond training deep networks to better performance, our work will imply a much broader impact in questioning the credibility of benchmarks and algorithms for automated machine learning (Auto ML). Designing novel networks is vital to the development of deep learning and artificial intelligence. Specifically, Neural Architecture Search (NAS) dramatically speeds up the discovery of novel architectures in a principled way (Zoph & Le, 2016; Pham et al., 2018; Liu et al., 2018), and Published as a conference paper at ICLR 2024 meanwhile many of them archive state-of-the-art performance. The validation of NAS algorithms largely hinges on the quality of architecture benchmarks. However, in most literature, each NAS benchmark adapts the same set of training protocols when comparing diverse neural architectures, largely due to the prohibitive cost of searching the optimal hyperparameter for each architecture. To our best knowledge, there exist very few in-depth studies of how to train architectures sampled from the NAS search space, thus we are motivated to examine the unexplored hidden gem question: how big a difference can be made if we specifically focus on improving the training hyperparameters? We will answer this question in 4.3. 4 EXPERIMENTS In our experiments, we will apply our learning rate scaling to weights and bias. We study our principles on MLPs, CNNs, and networks with advanced architectures from NAS-Bench-201 (Dong & Yang, 2020). All our experiments are repeated for three random runs. We also include more results in the Appendix ( E.1 for Image Net, E.2 for the Ge LU activation). We adapt our principles as follows: 1. Find the base maximal learning rate of the base architecture with the smallest depth (L = 1 in our case: input hidden output ): conduct a grid search over a range of learning rates, and find the maximal learning rate which achieves the smallest training loss at the end of one epoch1 2. Initialize the target network (of deeper layers or different topologies) according to equation 4. 3. Train the target network by scaling the learning rate based on equation 8 for MLPs or equation 9 for CNNs. The above steps are much cheaper than directly tuning hyperparameters for heavy networks, since the base network is much smaller. We verify this principle for MLPs with different depths in Appendix D. 4.1 MLPS WITH TOPOLOGY SCALING 0.01 0.02 0.03 0.04 0.05 Maximal LR (Experiments) Maximal LR (Our Estimation) MLPs (Topology Scaling) Figure 2: MLPs with different topologies (graph structures). X-axis: ground truth maximal learning rates found by grid search. Y-axis: estimated learning rates by our principle (equation 8). The red line indicates the identity. Based on the ground truth maximal learning rate of the basic MLP with L = 1, we scale up both learning rates and initialization to diverse architectures. The radius of a dot indicates the variance over three random runs. Data: CIFAR-10. We study MLP networks by scaling the learning rate to different graph structures based on equation 8. In addition to the learning rate, variances of initializations are also adjusted based on the in-degree of each layer in a network. To verify our scaling rule of both learning rates and initializations, we first find the ground truth maximal learning rates via grid search on MLPs with different graph structures, and then compare them with our scaled learning rates. As shown in Figure 2, our estimation strongly correlates with the ground truth maximal learning rates (r = 0.838). This result demonstrates that our learning rate scaling principle can be generalized to complicated network architectures. 4.2 CNNS WITH TOPOLOGY SCALING We further extend our study to CNNs. Our base CNN is of L = 1 and kernel size q = 3, and we scale the learning rate to CNNs of both different graph structures and kernel sizes, based on equation 9. Both the learning rate and the variances of initializations are adjusted for each network. 1Rationales for finding maximal learning rates for the 1st epoch: 1) Finding hyperparameters in early training is both practically effective and widely adopted (Tab. 1 and 4 of (Bohdal et al., 2022), Fig. 2 and 3a of (Egele et al., 2023). Analyzing subsequent training epochs will require more computations and careful design of HPO algorithms. 2) µP Yang et al. (2022) finds optimal learning rates in early training (Fig. 4 and 19). Exploiting more training epochs (for HPO purposes) is orthogonal to our method and is beyond the scope of our work. Published as a conference paper at ICLR 2024 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 Maximal LR (Experiments) Maximal LR (Our Estimation) CNNs (Topology & Kernel Scaling) r = 0.856 kernel = 3 kernel = 5 kernel = 7 Figure 3: CNNs with different graph structures and kernel sizes. The x-axis shows the ground truth maximal learning rates found by grid search. The y-axis shows the estimated learning rates by our principle (equation 9). The red line indicates the identity. Based on the ground truth maximal learning rate of the CNN with L = 1 and kernel size as 3, we scale up both learning rates and initialization to diverse architectures. The radius of a dot indicates the variance over three random runs. Data: CIFAR-10. Again, we first collect the ground truth maximal learning rates by grid search on CNNs with different graph structures and kernel sizes, and compare them with our estimated ones. As shown in Figure 3, our estimation strongly correlates with the ground truth maximal learning rates of heterogeneous CNNs (r = 0.856). This result demonstrates that our initialization and learning rate scaling principle can be further generalized to CNNs with sophisticated network architectures and different kernel sizes. 4.3 BREAKING NETWORK RANKINGS IN NAS Networks performance ranking is vital to the comparison of NAS algorithms (Guo et al., 2020; Mellor et al., 2021; Chu et al., 2021; Chen et al., 2021). If the ranking cannot faithfully reflect the true performance of networks, we cannot trust the quality of architecture benchmarks. Here we will show that: by simply better train networks, we can easily improve accuracies and break rankings of architectures (those in benchmarks or searched by Auto ML algorithms). We will use NAS-Bench-201 (Dong & Yang, 2020) as the test bed. NAS-Bench-201 provides a cell-based search space NAS benchmark, the network s accuracy is directly available by querying the database, benefiting the study of NAS methods without network evaluation. It contains five heterogeneous operator types: none (zero), skip connection, conv1 1, conv3 3 convolution, and average pooling3 3. Accuracies are provided for three datasets: CIFAR-10, CIFAR-100, Image Net16-120. However, due to the prohibitive cost of finding the optimal hyperparameter set for each architecture, the performance of all 15,625 architectures in NAS-Bench-201 is obtained with the same protocol (learning rate, initialization, etc.). Sub-optimal hyperparameters may setback the maximally achievable performance of each architecture, and may lead to unreliable comparison between NAS methods. We randomly sample architectures from NAS-Bench-201, and adopt the same training protocol as Dong & Yang (2020) (batch size, warm-up, learning rate decay, etc.). The only difference is that we use our principle ( 3.4) to re-scale learning rates and initializations for each architecture, and train networks till converge. Results are shown in Figure 4. We derive three interesting findings from our plots to contribute to the NAS community from novel perspectives: 1. Compared with the default training protocol on NAS-Bench-201, neural architectures can benefit from our scaled learning rates and initialization with much better performance improvements (left column in Figure 4). This is also a step further towards an improved gold standard NAS benchmark. Specifically, on average we outperform the test accuracy in NAS-Bench-201 by 0.44 on CIFAR-10, 1.24 on CIFAR-100, and 2.09 on Image Net16-120. We also show our superior performance over µP in Appendix E.3. 2. Most importantly, compared with the default training protocol on NAS-Bench-201, architectures exhibit largely different performance rankings, especially for top-performing ones (middle column in Figure 4). This implies: with every architecture better-trained, state-of-the-art NAS algorithms evaluated on benchmarks may also be re-ranked. This is because: many NAS methods (e.g. works summarized in (Dong et al., 2021)) are agnostic to hyperparameters (i.e. search methods do not condition on training hyperparameters). Changes in the benchmark s training settings will not affect architectures they searched, but instead will lead to different training performances and rankings of searched architectures. 3. By adopting our scaling method, the performance gap between architectures is mitigated (right column in Figure 4). In other words, our method enables all networks to be converged similarly well, thus less distinguishable in performance. Published as a conference paper at ICLR 2024 20 40 60 80 Test Accuracy (NAS-Bench-201) Test Accuracy (Ours) 90.0 92.5 88 100 80 60 40 20 10 5 1 Top Architectures (%) Correlation: Ours vs. NAS-Bench-201 (Rankings of Architectures) 100 80 60 40 20 10 5 1 Top Architectures (%) Pairwise Distance of Architectures (Test Accuracy) NAS-Bench-201 Ours 20 40 60 Test Accuracy (NAS-Bench-201) Test Accuracy (Ours) 100 80 60 40 20 10 5 1 Top Architectures (%) Correlation: Ours vs. NAS-Bench-201 (Rankings of Architectures) 100 80 60 40 20 10 5 1 Top Architectures (%) Pairwise Distance of Architectures (Test Accuracy) NAS-Bench-201 Ours 20 30 40 Test Accuracy (NAS-Bench-201) Test Accuracy (Ours) Image Net16-120 100 80 60 40 20 10 5 1 Top Architectures (%) Correlation: Ours vs. NAS-Bench-201 (Rankings of Architectures) Image Net16-120 100 80 60 40 20 10 5 1 Top Architectures (%) Pairwise Distance of Architectures (Test Accuracy) Image Net16-120 NAS-Bench-201 Ours Figure 4: We adopt our scaling principles to existing architecture benchmarks. Left column: better accuracy. For different network architectures, we plot the accuracy trained with different scaling principles (y-axis Re-scaled ) against the accuracy trained with a fixed training recipe (x-axis NAS-Bench-201 ). Each dot represents a unique architecture with different layer types and topologies (Figure 1 in (Dong & Yang, 2020)). Our principle (blue dots) achieves better accuracy compared to the benchmark (red line y = x). Middle column: network rankings in benchmarks are fragile. We compare networks performance rankings at different top K% percentiles (K = 100, 90, , 10, 5, 1; bottom right dots represent networks on the top-right in the left column), trained by our method vs. benchmarks, and find better networks are ranked more differently from the benchmark. This indicates current network rankings in benchmarks (widely used to compare NAS algorithms) can be easily broken by simply better train networks. Right column: less distinguishable architectures. We plot the pairwise performance gaps between different architectures, and find our principle makes networks similar and less distinguishable in terms of their accuracies. 5 CONCLUSION Training a high-quality deep neural network requires choosing suitable hyperparameters, which are non-trivial and expensive. However, most scaling or optimization methods are agnostic to the choice of networks, and thus largely ignore the impact of neural architectures on hyperparameters. In this work, by analyzing the dependence of pre-activations on network architectures, we propose a scaling principle for both the learning rate and initialization that can generalize across MLPs and CNNs, with different depths, connectivity patterns, and kernel sizes. Comprehensive experiments verify our principles. More importantly, our strategy further sheds light on potential improvements in current benchmarks for architecture design. We point out that by appropriately adjusting the learning rate and initializations per network, current rankings in these benchmarks can be easily broken.Our work contributes to both network training and model design for the Machine Learning community. Limitations of our work: 1) Principle of width-dependent scaling of the learning rate. 2) Characterization of depth-dependence of learning rates beyond the first step / early training of networks. 3) Depth-dependence of learning rates with the existence of normalization layers. Published as a conference paper at ICLR 2024 ACKNOWLEDGMENTS B. Hanin and Z. Wang are supported by NSF Scale-Mo DL (award numbers: 2133806, 2133861). Thomas Bachlechner, Bodhisattwa Prasad Majumder, Henry Mao, Gary Cottrell, and Julian Mc Auley. Rezero is all you need: Fast convergence at large depth. In Uncertainty in Artificial Intelligence, pp. 1352 1361. PMLR, 2021. James Bergstra, Daniel Yamins, and David Cox. Making a science of model search: Hyperparameter optimization in hundreds of dimensions for vision architectures. In International conference on machine learning, pp. 115 123. PMLR, 2013. Ondrej Bohdal, Lukas Balles, Beyza Ermis, Cédric Archambeau, and Giovanni Zappella. Pasha: Efficient hpo with progressive resource allocation. ar Xiv preprint ar Xiv:2207.06940, 2022. Wuyang Chen, Xinyu Gong, and Zhangyang Wang. Neural architecture search on imagenet in four gpu hours: A theoretically inspired perspective. ar Xiv preprint ar Xiv:2102.11535, 2021. Yutian Chen, Xingyou Song, Chansoo Lee, Zi Wang, Qiuyi Zhang, David Dohan, Kazuya Kawakami, Greg Kochanski, Arnaud Doucet, Marc aurelio Ranzato, et al. Towards learning universal hyperparameter optimizers with transformers. ar Xiv preprint ar Xiv:2205.13320, 2022. Xiangxiang Chu, Bo Zhang, and Ruijun Xu. Fairnas: Rethinking evaluation fairness of weight sharing neural architecture search. In Proceedings of the IEEE/CVF International Conference on computer vision, pp. 12239 12248, 2021. Aaron Defazio and Léon Bottou. A scaling calculus for the design and initialization of relu networks. Neural Computing and Applications, 34(17):14807 14821, 2022. Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pp. 248 255. Ieee, 2009. Emily Dinan, Sho Yaida, and Susan Zhang. Effective theory of transformers at initialization. ar Xiv preprint ar Xiv:2304.02034, 2023. Xuanyi Dong and Yi Yang. Nas-bench-102: Extending the scope of reproducible neural architecture search. ar Xiv preprint ar Xiv:2001.00326, 2020. Xuanyi Dong, Lu Liu, Katarzyna Musial, and Bogdan Gabrys. Nats-bench: Benchmarking nas algorithms for architecture topology and size. ar Xiv preprint ar Xiv:2009.00437, 2020a. Xuanyi Dong, Mingxing Tan, Adams Wei Yu, Daiyi Peng, Bogdan Gabrys, and Quoc V Le. Autohas: Efficient hyperparameter and architecture search. ar Xiv preprint ar Xiv:2006.03656, 2020b. Xuanyi Dong, David Jacob Kedziora, Katarzyna Musial, and Bogdan Gabrys. Automated deep learning: Neural architecture search is not the end. ar Xiv preprint ar Xiv:2112.09245, 2021. Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. ar Xiv preprint ar Xiv:2010.11929, 2020. Romain Egele, Isabelle Guyon, Yixuan Sun, and Prasanna Balaprakash. Is one epoch all you need for multi-fidelity hyperparameter optimization? ar Xiv preprint ar Xiv:2307.15422, 2023. Xavier Glorot and Yoshua Bengio. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the thirteenth international conference on artificial intelligence and statistics, pp. 249 256. JMLR Workshop and Conference Proceedings, 2010. Published as a conference paper at ICLR 2024 Zichao Guo, Xiangyu Zhang, Haoyuan Mu, Wen Heng, Zechun Liu, Yichen Wei, and Jian Sun. Single path one-shot neural architecture search with uniform sampling. In Computer Vision ECCV 2020: 16th European Conference, Glasgow, UK, August 23 28, 2020, Proceedings, Part XVI 16, pp. 544 560. Springer, 2020. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE international conference on computer vision, pp. 1026 1034, 2015. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770 778, 2016. Elad Hoffer, Itay Hubara, and Daniel Soudry. Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems, 30, 2017. Samuel Horváth, Aaron Klein, Peter Richtárik, and Cédric Archambeau. Hyperparameter transfer learning with adaptive complexity. In International Conference on Artificial Intelligence and Statistics, pp. 1378 1386. PMLR, 2021. Gao Huang, Zhuang Liu, Laurens Van Der Maaten, and Kilian Q Weinberger. Densely connected convolutional networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 4700 4708, 2017. Xiao Shi Huang, Felipe Perez, Jimmy Ba, and Maksims Volkovs. Improving transformer optimization through better initialization. In International Conference on Machine Learning, pp. 4475 4483. PMLR, 2020. Ilija Ilievski and Jiashi Feng. Hyperparameter transfer learning through surrogate alignment for efficient deep neural network training. ar Xiv preprint ar Xiv:1608.00218, 2016. Gaurav Iyer, Boris Hanin, and David Rolnick. Maximal initial learning rates in deep relu networks. ar Xiv preprint ar Xiv:2212.07295, 2022. Arthur Jacot, Franck Gabriel, and Clément Hongler. Neural tangent kernel: Convergence and generalization in neural networks. In Advances in neural information processing systems, pp. 8571 8580, 2018. Kevin Jamieson and Ameet Talwalkar. Non-stochastic best arm identification and hyperparameter optimization. In Artificial intelligence and statistics, pp. 240 248. PMLR, 2016. Samy Jelassi, Boris Hanin, Ziwei Ji, Sashank J Reddi, Srinadh Bhojanapalli, and Sanjiv Kumar. Depth dependence of µp learning rates in relu mlps. ar Xiv preprint ar Xiv:2305.07810, 2023. Aaron Klein and Frank Hutter. Tabular benchmarks for joint architecture and hyperparameter optimization. ar Xiv preprint ar Xiv:1905.04970, 2019. Liam Li, Kevin Jamieson, Afshin Rostamizadeh, Ekaterina Gonina, Jonathan Ben-Tzur, Moritz Hardt, Benjamin Recht, and Ameet Talwalkar. A system for massively parallel hyperparameter tuning. Proceedings of Machine Learning and Systems, 2:230 246, 2020. Yang Li, Yu Shen, Huaijun Jiang, Tianyi Bai, Wentao Zhang, Ce Zhang, and Bin Cui. Transfer learning based search space design for hyperparameter tuning. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pp. 967 977, 2022a. Yang Li, Yu Shen, Huaijun Jiang, Wentao Zhang, Zhi Yang, Ce Zhang, and Bin Cui. Transbo: Hyperparameter optimization via two-phase transfer learning. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pp. 956 966, 2022b. Hanxiao Liu, Karen Simonyan, and Yiming Yang. Darts: Differentiable architecture search. ar Xiv preprint ar Xiv:1806.09055, 2018. Published as a conference paper at ICLR 2024 Liyuan Liu, Xiaodong Liu, Jianfeng Gao, Weizhu Chen, and Jiawei Han. Understanding the difficulty of training transformers. ar Xiv preprint ar Xiv:2004.08249, 2020. Joe Mellor, Jack Turner, Amos Storkey, and Elliot J Crowley. Neural architecture search without training. In International Conference on Machine Learning, pp. 7588 7598. PMLR, 2021. Daniel Park, Jascha Sohl-Dickstein, Quoc Le, and Samuel Smith. The effect of network width on stochastic gradient descent and generalization: an empirical study. In International Conference on Machine Learning, pp. 5042 5051. PMLR, 2019. Valerio Perrone, Rodolphe Jenatton, Matthias W Seeger, and Cédric Archambeau. Scalable hyperparameter transfer learning. Advances in neural information processing systems, 31, 2018. Hieu Pham, Melody Y Guan, Barret Zoph, Quoc V Le, and Jeff Dean. Efficient neural architecture search via parameter sharing. ar Xiv preprint ar Xiv:1802.03268, 2018. Samuel S Schoenholz, Justin Gilmer, Surya Ganguli, and Jascha Sohl-Dickstein. Deep information propagation. ar Xiv preprint ar Xiv:1611.01232, 2016. Christopher J Shallue, Jaehoon Lee, Joseph Antognini, Jascha Sohl-Dickstein, Roy Frostig, and George E Dahl. Measuring the effects of data parallelism on neural network training. ar Xiv preprint ar Xiv:1811.03600, 2018. Karen Simonyan and Andrew Zisserman. Very deep convolutional networks for large-scale image recognition. ar Xiv preprint ar Xiv:1409.1556, 2014. Samuel L Smith, Pieter-Jan Kindermans, Chris Ying, and Quoc V Le. Don t decay the learning rate, increase the batch size. ar Xiv preprint ar Xiv:1711.00489, 2017. Jasper Snoek, Hugo Larochelle, and Ryan P Adams. Practical bayesian optimization of machine learning algorithms. Advances in neural information processing systems, 25, 2012. Danny Stoll, Jörg KH Franke, Diane Wagner, Simon Selg, and Frank Hutter. Hyperparameter transfer across developer adjustments. ar Xiv preprint ar Xiv:2010.13117, 2020. Mingxing Tan and Quoc V Le. Efficientnet: Rethinking model scaling for convolutional neural networks. ar Xiv preprint ar Xiv:1905.11946, 2019. Mingxing Tan, Bo Chen, Ruoming Pang, Vijay Vasudevan, Mark Sandler, Andrew Howard, and Quoc V Le. Mnasnet: Platform-aware neural architecture search for mobile. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2820 2828, 2019. Saining Xie, Alexander Kirillov, Ross Girshick, and Kaiming He. Exploring randomly wired neural networks for image recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 1284 1293, 2019. Sho Yaida. Meta-principled family of hyperparameter scaling strategies. ar Xiv preprint ar Xiv:2210.04909, 2022. Antoine Yang, Pedro M Esperança, and Fabio M Carlucci. Nas evaluation is frustratingly hard. ar Xiv preprint ar Xiv:1912.12522, 2019. Ge Yang and Samuel Schoenholz. Mean field residual networks: On the edge of chaos. In Advances in neural information processing systems, pp. 7103 7114, 2017. Greg Yang and Edward J Hu. Feature learning in infinite-width neural networks. ar Xiv preprint ar Xiv:2011.14522, 2020. Greg Yang, Edward J Hu, Igor Babuschkin, Szymon Sidor, Xiaodong Liu, David Farhi, Nick Ryder, Jakub Pachocki, Weizhu Chen, and Jianfeng Gao. Tensor programs v: Tuning large neural networks via zero-shot hyperparameter transfer. ar Xiv preprint ar Xiv:2203.03466, 2022. Chris Ying, Aaron Klein, Eric Christiansen, Esteban Real, Kevin Murphy, and Frank Hutter. Nasbench-101: Towards reproducible neural architecture search. In International Conference on Machine Learning, pp. 7105 7114. PMLR, 2019. Published as a conference paper at ICLR 2024 Dani Yogatama and Gideon Mann. Efficient transfer learning method for automatic hyperparameter tuning. In Artificial intelligence and statistics, pp. 1077 1085. PMLR, 2014. Jiaxuan You, Jure Leskovec, Kaiming He, and Saining Xie. Graph structure of neural networks. ar Xiv preprint ar Xiv:2007.06559, 2020. Arber Zela, Aaron Klein, Stefan Falkner, and Frank Hutter. Towards automated deep learning: Efficient joint neural architecture and hyperparameter search. ar Xiv preprint ar Xiv:1807.06906, 2018. Hongyi Zhang, Yann N Dauphin, and Tengyu Ma. Residual learning without normalization via better initialization. In International Conference on Learning Representations, volume 3, pp. 2, 2019. Chen Zhu, Renkun Ni, Zheng Xu, Kezhi Kong, W Ronny Huang, and Tom Goldstein. Gradinit: Learning to initialize neural networks for stable and efficient training. Advances in Neural Information Processing Systems, 34:16410 16422, 2021. Barret Zoph and Quoc V Le. Neural architecture search with reinforcement learning. ar Xiv preprint ar Xiv:1611.01578, 2016. A ARCHITECTURE-AWARE INITIALIZATION Derivations for 3.2. Consider a random network input x sampled from a distribution for which x and x have the same distribution. Let us first check that the joint distribution of the pre-activation vectors z(ℓ )(x) with ℓ = 0, . . . , L + 1 is symmetric around 0. We proceed by induction on ℓ. When ℓ= 0 this statement is true by assumption. Next, suppose we have proved the statement for 0, . . . , ℓ. Then, since z(ℓ+1)(x) is a linear function of z(0)(x), . . . , z(ℓ)(x), we conclude that it is also symmetric around 0. A direct consequence of this statement is that, by symmetry, E h σ2 z(ℓ) j (x) i = 1 2E z(ℓ) j (x) 2 , j = 1, . . . , n, ℓ= 0, . . . , L + 1 (10) Let us fix k = 1, . . . , n and a vertex ℓwe calculate CW in equation 2: E z(ℓ) 2 = E j =1 W (ℓ ,ℓ) jj σ z(ℓ ) j n 1 2E z(ℓ ) j 2 1 2C(ℓ,ℓ ) W E z(ℓ ) 2 . (11) We ask E h z(ℓ) 2i = E z(ℓ ) 2 , which yields C(ℓ ,ℓ) W = 2 d(ℓ) in . (12) We now seek to show E z(ℓ) 2 = E z(ℓ ) 2 ℓ, ℓ V. (13) To do so, let us define for each ℓ= 0, . . . , L + 1 d(ℓ, L + 1) := {length of longest directed path in G from ℓto L + 1} . Published as a conference paper at ICLR 2024 The relation equation 13 now follows from a simple argument by induction show that by induction on d(ℓ, L + 1), starting with d(ℓ, L + 1) = 0. Indeed, the case d(ℓ, L + 1) = 0 simply corresponds to ℓ= L + 1. Next, note that if d(ℓ, L + 1) > 0 and if (ℓ , ℓ) E, then d(ℓ, L + 1) < d(ℓ , L + 1) by the maximality of the path length in the definition of d(ℓ, L + 1). Hence, substituting equation 12 into equation 11 completes the proof of the inductive step. B ARCHITECTURE-DEPENDENT LEARNING RATES (FOR 3.3) We start by setting some notation. Specifically, we fix a network input x Rn0 at which we study both the forward and backward pass. We denote by z(ℓ) i := z(ℓ) i (x), z(ℓ) := z(ℓ)(x) the corresponding pre-activations at vertex ℓ. We assume our network has a uniform width over layers (nℓ n), and parameter-dependently learning rates: ηµ = learning rate of µ. We will restrict to the case ηµ = η at the end. Further, recall from equation 6 that we consider the empirical MSE over a batch B with a single element (x, y) and hence our loss is z(L+1) y 2 . The starting point for derivations in 3.3 is the following Lemma Lemma B.1 (Adapted from Lemma 2.1 in Jelassi et al. (2023)). For ℓ= 1, . . . , L, we have E h ( z(ℓ) i )2] i = A(ℓ) + B(ℓ), µ1,µ2 ℓ ηµ1ηµ2 µ1z(ℓ) 1 µ2z(ℓ) 1 (14) 1 d(L+1) in 2 X (ℓ 1,L+1),(ℓ 2,L+1) E µ1z(ℓ 1) j1 µ2z(ℓ 1) j1 z(ℓ 2) j2 2 + 2z(ℓ 1) j1 µ1z(ℓ 2) j1 z(ℓ 1) j2 µ2z(ℓ 2) j2 µ1,µ2 ℓ ηµ1ηµ2 µ1z(ℓ) 1 µ2z(ℓ) 1 j=1 µ1z(ℓ ) j µ2z(ℓ ) j Proof. The proof of this result follows very closely the derivation of Lemma 2.1 in Jelassi et al. (2023) which considered only the case of MLPs. We first expand z(ℓ) i by applying the chain rule: µ ℓ µz(ℓ) i µ, (17) where the sum is over weights µ that belong to some weight matrix W (ℓ ,ℓ ) for which there is a directed path from ℓ to ℓin G and we ve denoted by µ the change in µ after one step of GD. The SGD update satisfies: 2 µ z(L+1) y 2 = ηµ µz(L+1) z(L+1) y , (18) Published as a conference paper at ICLR 2024 where we ve denoted by (x, y) the training datapoint in the first batch. We now combine equation 17 and equation 18 to obtain: µ ℓ ηµ µz(ℓ) i µz(L+1) y z(L+1) . (19) Using equation 19, we obtain E z(ℓ) i 2 =E µ ℓ ηµ µz(ℓ) 1 µz(L+1) 1 z(L+1) y µ1,µ2 ℓ ηµ1ηµ2 µ1z(ℓ) 1 µ2z(ℓ) 1 µ1z(L+1) µ2z(L+1)Ey Here, we ve used that by symmetry the answer is independent of i. Taking into account the distribution of z(L+1) and y, we have z(L+1) y 2 = (z(L+1))2 + 1 (21) We plug equation 21 in equation 20 and obtain E[( z(ℓ) i )2] = A(ℓ) + B(ℓ), (22) µ1,µ2 ℓ ηµ1ηµ2 µ1z(ℓ) µ2z(ℓ) µ1z(L+1) µ2z(L+1) z(L+1) 2 µ1,µ2 ℓ ηµ1ηµ2 µ1z(ℓ) 1 µ2z(ℓ) µ1z(L+1) µ2z(L+1) By definition, we have j=1 W (ℓ ,L+1) j σ z(ℓ ) j , W (ℓ ,L+1) j N 0, 1 Therefore, integrating out weights of the form W (ℓ ,L+1) in equation 24 yields µ1,µ2 ℓ ηµ1ηµ2 µ1z(ℓ) 1 µ2z(ℓ) 1 j=1 µ1σ z(ℓ ) j µ2σ z(ℓ ) j µ1,µ2 ℓ ηµ1ηµ2 µ1z(ℓ) 1 µ2z(ℓ) 1 j=1 µ1z(ℓ ) j µ2z(ℓ ) j where in the last equality we used equation 10. This yields the desired formula for B(ℓ). A similar computation yields the expression for A(ℓ). As in the proof of Theorem 1.1 in Jelassi et al. (2023), we have A(ℓ) = O(n 1) due to the presence of an extra pre-factor of 1/n. Our derivation in 3.3 therefore comes down to finding how B(ℓ) is influenced by the topology of the graph G. Published as a conference paper at ICLR 2024 We then re-write the high-level change of pre-activations (equation 5) in an arbitrary graph topology. We need to accumulate all changes of pre-activations that flow into a layer (summation over in-degrees ℓ L + 1): z(L+1) i = X µ ℓ µz(ℓ ) i µ (25) Therefore, for the change of pre-activation of layer L + 1, we can simplify the analysis to each individual layer ℓ that connects L + 1 (i.e. edge (ℓ L + 1) E), and then sum up all layers that flow into layer ℓ. Next, we focus on deriving B(ℓ) of each individual path in the case of the DAG structure of network architectures. It is important that here we adopt our architecture-aware initialization ( 3.2). µ1,µ2 ℓ ηµ1ηµ2 µ1z(ℓ) i µ2z(ℓ) i X j=1 µ1z(ℓ ) j µ2z(ℓ ) j µ1,µ2 ℓ ηµ1ηµ2 µ1z(ℓ) i µ2z(ℓ) i µ1z(ℓ ) j µ2z(ℓ ) j µ1,µ2 ℓ ηµ1ηµ2 µ1z(ℓ) i µ2z(ℓ) i µ1z(ℓ) j µ2z(ℓ) j Thus, we can see that with our architecture-aware initialization, B(ℓ) reduce back to the basic sequential MLP case in Theorem 1.1 in Jelassi et al. (2023). Therefore, we have: p=1 Θ(η2L3 p). where P is the total number of end-to-end paths that flow from the input to the final output z L+1, and Lp is the number of Re LU layers on each end-to-end path. C LEARNING RATE SCALING FOR CNNS Derivations for 3.4. In addition to B(ℓ) in equation 24, we further refer to the contributing term of B(ℓ) in Jelassi et al. (2023) when µ1 ℓ 1 and µ2 ℓ(or vice versa), denoted as C(ℓ): µ ℓ ηµ 1 n2 z(ℓ) j1 µz(ℓ) j2 2 We start from deriving the recursion of B(ℓ) of each individual end-to-end path for convolutional layers of a kernel size as q. Again, we assume our network has a uniform width over layers (nℓ n). If µ1, µ2 ℓthen the contribution to equation 24 is σ(ℓ 1) j1 σ(ℓ 1) j2 2 = η(ℓ) 2 q2 4 n2 x 2 e5 Pℓ 2 ℓ =1 1 n When µ1 ℓ 1 and µ2 ℓ(or vice versa) the contribution to equation 24 is µ1 ℓ 1 ηµ1 1 n σ(ℓ 1) k 2 1 = η(ℓ)q C(ℓ 1). Published as a conference paper at ICLR 2024 Finally, when µ1, µ2 ℓ 1 we find the contribution to equation 24 becomes µ1,µ2 ℓ 1 ηµ1ηµ2 µ1z(ℓ) 1 µ2z(ℓ) 1 2 + 1 1 µ1z(ℓ) 1 µ2z(ℓ) 1 µ1z(ℓ) 2 µ2z(ℓ) 2 n e B(ℓ 1). Therefore, we have B(ℓ) = η(ℓ) 2 q2 4 n2 x 2 e5 Pℓ 2 ℓ =1 1 n + η(ℓ)q C(ℓ 1) + 1 + 1 1 n e B(ℓ) = η(ℓ) 2 q2 4 x 4 n2 e5 Pℓ 2 ℓ =1 1 n + η(ℓ)q C(ℓ 1) + 1 n e B(ℓ 1) + 2 We further derive C(ℓ). When µ ℓthe contribution to equation 27 is z(ℓ 1) j1 z(ℓ 1) j2 2 = η(ℓ)q x 4 n2 e5 Pℓ 1 ℓ =1 1 n . When µ ℓ 1 the contribution to equation 27 is µz(ℓ) 1 z(ℓ) 1 2 + 1 1 µz(ℓ) 1 2 z(ℓ) 2 2 = C(ℓ 1) + 1 n e C(ℓ 1). n2 e5 Pℓ 1 ℓ =1 1 n + 1 n C(ℓ 1) + 1 + 1 Finally, for each end-to-end path, we have B(ℓ) Θ(η2ℓ3q2). Therefore, together with 3.3, we want D MLPS WITH DEPTH SCALING We verify the depth-wise scaling rule in Jelassi et al. (2023). We scale a vanilla feedforward network by increasing its depth (adding more Linear-Re LU layers). Starting from the most basic feedforward network with L = 3 (an input layer, a hidden layer, plus an output layer), we first scan a wide range of learning rates and find the maximal learning rate. We then scale the learning rate to feedforward networks of different depths according to equation 8. Different feedforward networks share the same initialization since both the out-degree and in-degree of all layers are 1. To verify the scaling results, we also conduct the grid search of learning rates for feedforward networks of different depths, and compare them with the scaled learning rates. As shown in Figure 5, on CIFAR-10 the estimation strongly correlates with the ground truth maximal learning rates (r = 0.962), and the plotted dots are very close to the identity line. This result demonstrates that the depth-wise learning rate scaling principle in Jelassi et al. (2023) is highly accurate across feedforward neural networks of different depths and across different datasets. Published as a conference paper at ICLR 2024 0.005 0.010 0.015 0.020 0.025 Maximal LR (Experiments) Maximal LR (Our Estimation) MLPs (Depth L Scaling) Figure 5: MLP networks of different depths on CIFAR-10. The x-axis shows the ground truth maximal learning rates found by our grid search experiments. The y-axis shows the estimated learning rates by our principle in equation 8. The red line indicates the identity. Based on the true maximal learning rate of the feedforward networks of L = 3, we scale up to L = 10. The radius of a dot indicates the variance over three random runs. E MORE EXPERIMENTS E.1 IMAGENET 0.010 0.015 0.020 0.025 0.030 0.035 Maximal LR (Experiments) Maximal LR (Our Estimation) MLPs (Depth L Scaling) (Image Net) 0.015 0.020 0.025 0.030 0.035 Maximal LR (Experiments) Maximal LR (Our Estimation) MLPs (Topology Scaling) (Image Net) Figure 6: MLP networks (Re LU activation) of different depths (left) and graph topologies (right) on Image Net Deng et al. (2009). The x-axis shows the ground truth maximal learning rates found by our grid search experiments. The y-axis shows the estimated learning rates by our principle in equation 8. The red line indicates the identity. The radius of a dot indicates the variance over three random runs. Similar to Figure 5 and Figure 2, we further verify the learning rate scaling rule on Image Net Deng et al. (2009). We scale up a vanilla feedforward network by increasing its depth (adding more Linear-Re LU layers) or changing its graph topology. As shown in Figure 7, our estimations achieve strong correlations with the ground truth maximal learning rates for both depth-wise scaling (r = 0.856) and topology-wise scaling (r = 0.729). This result demonstrates that our learning rate scaling principle is highly accurate across feedforward neural networks of different depths and across different datasets. E.2 THE GELU ACTIVATION To demonstrate that our learning rate scaling principle can generalize to different activation functions, we further empirically verify MLP networks with GELU layers (on CIFAR-10). We scale up a vanilla feedforward network by increasing its depth (adding more Linear-Re LU layers) or changing its graph topology. Again, our estimations achieve strong correlations with the ground truth maximal learning rates for both depth-wise scaling (r = 0.920) and topology-wise scaling (r = 0.680). Moreover, for CNNs, our estimation can also achieve r = 0.949. E.3 THE µP HEURISTICS We further compare the µP heuristics (Yang et al., 2022) with our architecture-aware initialization and learning rate scaling by training architectures defined in NAS-Bench-201 (Dong & Published as a conference paper at ICLR 2024 0.0025 0.0050 0.0075 0.0100 0.0125 0.0150 0.0175 0.0200 0.0225 Maximal LR (Experiments) Maximal LR (Our Estimation) MLPs (GELU) (Depth L Scaling) 0.0050 0.0075 0.0100 0.0125 0.0150 0.0175 0.0200 0.0225 0.0250 Maximal LR (Experiments) Maximal LR (Our Estimation) MLPs (GELU) (Topology Scaling) 0.01 0.02 0.03 0.04 0.05 0.06 0.07 Maximal LR (Experiments) Maximal LR (Our Estimation) CNNs (GELU) (Topology & Kernel Scaling) r = 0.949 kernel = 3 kernel = 5 kernel = 7 Figure 7: MLP networks with GELU activations of different depths (left) and graph topologies (middle), and CNNs (right), on CIFAR-10. The x-axis shows the ground truth maximal learning rates found by our grid search experiments. The y-axis shows the estimated learning rates by our principle in equation 8. The red line indicates the identity. The radius of a dot indicates the variance over three random runs. Yang, 2020). We follow the usage from https://github.com/microsoft/mup?tab= readme-ov-file#basic-usage to set up the base model, and train with Mu SGD. The learning rate is set as 0.1 following the original setting in NAS-Bench-201. From Figure 8, we can see that µP initialization and scaling strategy yields inferior results. 20 40 60 80 Test Accuracy (µP) Test Accuracy (Ours) 10 20 30 40 50 60 70 Test Accuracy (µP) Test Accuracy (Ours) 5 10 15 20 25 30 35 40 45 Test Accuracy (µP) Test Accuracy (Ours) Image Net16-120 Figure 8: Comparison between the µP heuristics Yang et al. (2022) (x-axis) and our architecture-aware initialization and learning rate scaling (y-axis) on NAS-Bench-201 Dong & Yang (2020). Left: CIFAR-10. Middle: CIFAR-100. Right: Image Net-16-120. E.4 NETWORK RANKINGS INFLUENCED BY RANDOM SEEDS To delve deeper into Figure 4, we analyze how the random seed affects network rankings. Specifically, NAS-Bench-201 reports network performance trained with three different random seeds (seeds = [777, 888, 999]). For CIFAR-100, we created a plot similar to Figure 4 middle column, but it shows pairwise ranking correlations among those three random seeds. As shown in Figure 9, although we observe different ranking correlations between seeds, they are consistently higher (i.e., their rankings are more consistent) than those produced by our method. This confirms that changes in network rankings by our architecture-aware hyperparameters are meaningful, which can train networks to better performance. Published as a conference paper at ICLR 2024 100 80 60 40 20 105 1 Top Architectures (%) Correlation: Ours vs. NAS-Bench-201 (Rankings of Architectures) our vs. bench-201 (all seeds) bench-201 (seed 777 vs. 888) bench-201 (seed 777 vs. 999) bench-201 (seed 888 vs. 999) Figure 9: Comparison of network rankings influenced by our method versus random seeds. NAS-Bench201 Dong & Yang (2020) provides training results with three different random seeds (777, 888, 999). Blue dots represent the Kendall-Tau correlations between networks trained by our method and the accuracy from NASBench-201 (averaged over three seeds). We also plot correlations between random seeds in a pairwise manner. We compare networks performance rankings at different top K% percentiles (K = 100, 90, , 10, 5, 1; bottom right dots represent networks on the top-right in the left column). This indicates that, although network rankings can be influenced by randomness during training, our method leads to significant changes in their rankings while still enabling these networks to achieve better accuracy.