# revisiting_resnets_improved_training_and_scaling_strategies__8b8d8b54.pdf Revisiting Res Nets: Improved Training and Scaling Strategies Irwan Bello Google Brain William Fedus Google Brain Xianzhi Du Google Brain Ekin D. Cubuk Google Brain Aravind Srinivas UC Berkeley Tsung-Yi Lin Google Brain Jonathon Shlens Google Brain Barret Zoph Google Brain Novel computer vision architectures monopolize the spotlight, but the impact of the model architecture is often conflated with simultaneous changes to training methodology and scaling strategies. Our work revisits the canonical Res Net [13] and studies these three aspects in an effort to disentangle them. Perhaps surprisingly, we find that training and scaling strategies may matter more than architectural changes, and further, that the resulting Res Nets match recent state-of-the-art models. We show that the best performing scaling strategy depends on the training regime and offer two new scaling strategies: (1) scale model depth in regimes where overfitting can occur (width scaling is preferable otherwise); (2) increase image resolution more slowly than previously recommended [55]. Using improved training and scaling strategies, we design a family of Res Net architectures, Res Net RS, which are 1.7x - 2.7x faster than Efficient Nets on TPUs, while achieving similar accuracies on Image Net. In a large-scale semi-supervised learning setup, Res Net-RS achieves 86.2% top-1 Image Net accuracy, while being 4.7x faster than Efficient Net-Noisy Student. The training techniques improve transfer performance on a suite of downstream tasks (rivaling state-of-the-art self-supervised algorithms) and extend to video classification on Kinetics-400. We recommend practitioners use these simple revised Res Nets as baselines for future research. 1 Introduction The performance of a vision model is a product of the architecture, training methods and scaling strategy. Novel architectures underlie many advances, but are often simultaneously introduced with other critical and less publicized changes in the details of the training methodology and hyperparameters. Additionally, new architectures enhanced by modern training methods are sometimes compared to older architectures with dated training methods (e.g. Res Net-50 with Image Net Top-1 accuracy of 76.5% [13]). Our work addresses these issues and empirically studies the impact of training methods and scaling strategies on the popular Res Net architecture [13]. We survey the modern training and regularization techniques widely in use today and apply them to Res Nets (Figure 1). In the process, we encounter interactions between training methods and show a benefit of reducing weight decay values when used in tandem with other regularization techniques. An additive study of training methods in Table 1 reveals the significant impact of these decisions: a Correspondence to Irwan Bello and Barret Zoph {ibello,barretzoph}@google.com. Code and checkpoints available in Tensor Flow: https://github.com/tensorflow/models/tree/master/ official/vision/beta and https://github.com/tensorflow/tpu/tree/master/ models/official/resnet/resnet_rs 35th Conference on Neural Information Processing Systems (Neur IPS 2021). 0.0 0.5 1.0 1.5 2.0 2.5 3.0 Time Per Training Step (Sec) Top-1 Image Net Accuracy + Minor Architectural Changes + Improved Training Strategies 2.7x Speedup Speed-Accuracy Pareto Curve Res Net-RS Efficient Net Res Net Figure 1: Improving Res Nets to state-of-the-art performance. We improve on the canonical Res Net [13] with modern training methods (as also used in Efficient Nets [55]), minor architectural changes and improved scaling strategies. The resulting models, Res Net-RS, outperform Efficient Nets on the speed-accuracy Pareto curve with speed-ups ranging from 1.7x - 2.7x on TPUs and 2.1x - 3.3x on GPUs. Res Net ( ) is a Res Net-200 trained at 256 256 resolution. Training times reported on TPUs. canonical Res Net-200 with 79.0% top-1 Image Net accuracy is improved to 82.2% (+3.2%) through improved training methods alone. This is increased further to 83.4% by two small and commonly used architectural improvements: Res Net-D [15] and Squeeze-and-Excitation [21]. Figure 1 traces this refinement over the starting Res Net in a speed-accuracy Pareto curve. We offer new perspectives and practical advice on scaling vision architectures. While prior works extrapolate scaling rules from small models [55] or from short training duration [39], we design scaling strategies by exhaustively training models across a variety of scales for the full training duration (e.g. 350 epochs instead of 10 epochs). In doing so, we uncover strong dependencies between the best performing scaling strategy and the training regime (e.g. number of epochs, model size, dataset size). These dependencies are missed in any of these smaller regimes, leading to suboptimal scaling decisions. Our analysis leads to new scaling strategies summarized as (1) scale the model depth when overfitting can occur (scaling the width is preferable otherwise) and (2) scale the image resolution more slowly than prior works [55]. Using the improved training and scaling strategies, we design a family of re-scaled Res Nets, Res Net RS, across model various scales (Figure 1). Res Net-RS models use less memory during training and are 1.7x - 2.7x faster on TPUs (2.1x - 3.3x faster on GPUs) than the popular Efficient Nets on the speed-accuracy Pareto curve. In a large-scale semi-supervised learning setup, Res Net-RS obtains a 4.7x training speed-up on TPUs (5.5x on GPUs) over Efficient Net-B5 when co-trained on Image Net [30] and an additional 130M pseudo-labeled images. Finally, we conclude with a suite of experiments testing the generality of the improved training and scaling strategies. We first demonstrate that our scaling strategy improves the speed-accuracy Pareto curve of Efficient Net. Next, we show that the improved training strategies yield representations that rival or outperform those from self-supervised algorithms (Sim CLR and Sim CLRv2 [4, 5]) on a suite of downstream tasks. The improved training strategies also extend to video classification, yielding an improvement from 73.4% to 77.4% (+4.0%) on the Kinetics-400 dataset. Through combining lightweight architectural changes (used since 2018) and improved training and scaling strategies, we discover the Res Net architecture sets a state-of-the-art baseline for vision research. This finding highlights the importance of teasing apart each of these factors in order to understand what architectures perform better than others. We summarize our contributions: An empirical study of regularization techniques and their interplay, which leads to a training strategy that achieves strong performance (e.g. +3.2% top-1 Image Net accuracy, +4.0% top-1 Kinetics-400 accuracy) without having to change the model architecture. An empirical study of scaling which uncovers strong dependencies between training and the best performing scaling strategy. We propose a simple scaling strategy: (1) scale depth when overfitting can occur (scaling width can be preferable otherwise) and (2) scale the image resolution more slowly than prior works [55]. This scaling strategy improves the speed-accuracy Pareto curve of both Res Nets and Efficient Nets. Res Net-RS: a Pareto curve of Res Net architectures that are 1.7x - 2.7x faster than Efficient Nets on TPUs (2.1x - 3.3x on GPUs) by applying the training and scaling strategies. Semi-supervised training of Res Net-RS with an additional 130M pseudo-labeled images achieves 86.2% top-1 Image Net accuracy, while being 4.7x faster on TPUs (5.5x on GPUs) than the corresponding Efficient Net-Noisy Student [57]. Empirically show that representations obtained from supervised learning using modern training techniques rival or outperform state-of-the-art self-supervised representations (Sim CLR [4], Sim CLRv2 [5]) on suite of downstream computer vision tasks. 2 Characterizing Improvements on Image Net Since the breakthrough of Alex Net [30] on Image Net [45], a wide variety of improvements have been proposed to further advance image recognition performance. These improvements broadly arise along four orthogonal axes: (a) architecture, (b) training/regularization methodology, (c) scaling strategy and (d) using additional training data. (a) Architecture. The works that perhaps receive the most attention are novel architectures. Notable proposals since Alex Net include VGG [49], Res Net [13], Inception [52, 53], and Res Ne Xt [58]. Automated search strategies for designing architectures have further pushed the state-of-the-art [67, 41, 55]. There have also been efforts in going beyond standard Conv Nets for image classification, by adapting self-attention [56] to the visual domain [2, 40, 20, 47, 8, 1]. (b) Training and Regularization Methods. Image Net progress has simultaneously been boosted by innovations in training (e.g. improved learning rate schedules [34, 12]) and regularization methods, such as dropout [50], label smoothing [53], stochastic depth [22], dropblock [11] and data augmentation [61, 59, 6, 7]. Regularization methods have become especially useful to prevent overfitting when training ever-increasingly larger models [23] on limited data (e.g. 1.2M Image Net images). (c) Scaling Strategies. Increasing the model dimensions (width, depth and resolution) has been another successful axis to improve quality [44, 17]. Res Net architectures are typically scaled up by adding layers (depth): Res Nets-18 to Res Net-200 and beyond [14, 62]. Wide Res Nets [60] and Mobile Nets [19] instead scale the width. Increasing image resolutions consistently improves performance: Efficient Net uses 600 image resolutions [55] while both Res Ne St [62] and TRes Net [43] use 400+ image resolutions for their largest model. In an attempt to systematize these heuristics, Efficient Net proposed the compound scaling rule, which jointly scales network depth, width and image resolution using a constant scaling factor. However, Section 7.1 shows this scaling strategy is sub-optimal for not only Res Nets, but Efficient Nets as well. (d) Additional Training Data. Finally, Image Net accuracy is commonly improved by training on additional sources of data (either labeled, weakly labeled, or unlabeled). Pre-training on large-scale datasets [51, 35, 27] has significantly pushed the state-of-the-art, with Vi T [8] and NFNets [3] recently achieving 88.6% and 89.2% Image Net accuracy respectively. Using pseudo-labels on additional unlabeled images [57, 37] in a semi-supervised learning fashion has also been a fruitful avenue for improving accuracy. We present semi-supervised learning results in Section 7.2. 3 Related Work Improved training methods combined with architectural changes to Res Nets have routinely yielded competitive Image Net performance [15, 31, 43, 62, 1, 3]. [15] achieved 79.2% top-1 Image Net accuracy (a +3% improvement over their Res Net-50 baseline) by modifying the stem and downsampling block and using label smoothing and mixup. [31] further improved the Res Net-50 model with additional architectural modifications such as Squeeze-and-Excitation [21], selective kernel [32], and anti-alias downsampling [63], while also using label smoothing, mixup, and dropblock to achieve 81.4% accuracy. [43, 62] incorporate several architectural modifications to the Res Net architectures along with improved training methodologies to outperform Efficient Net models on the speed-accuracy Pareto curve on GPUs. Many prior works do remark the importance of improved training and regularization methods. However experiments are still largely concerned with architectural changes and the simultaneous introduction of improved training techniques can make it hard to identify where the gains come from1. Additionally, due to the ever-increasing performance of machine learning accelerators, newer architectures are routinely pushed to much larger scales than the original Res Nets [13]. As a result, works that propose novel architectures do not (cannot) compare against properly trained and scaled Res Nets (since such a baseline did not exist), making it challenging to evaluate the significance of the proposed architectural changes compared to simple Res Nets. Lastly, prior work often puts little emphasis on studying scaling strategies or advocates for scaling strategies which we find to be sub-optimal. For example, the largest models in Efficient Net [55], TRes Net [43] and Res Ne St [62] use 600, 448 and 416 image sizes respectively, which our scaling analysis reveals to be excessively large. Reg Net [39] advocates for width scaling, which we find only works well when overfitting does not occur (e.g. 10 epochs). In contrast to other works, we only consider lightweight architectural changes (that are widely used since 2018) and keep the architecture fixed. Instead, we focus exclusively on training and scaling strategies to build a Pareto curve of models. Perhaps surprisingly, we find that doing so suffices to outperform models that were introduced after Res Nets: our improved training and scaling methods lead to Res Nets that are significantly faster than Efficient Nets on TPUs on GPUs (see Section 7.1). We note that our scaling improvements are sometimes orthogonal to the architectural innovations introduced in prior works in which case we expect them to be additive. 4 Methodology Architecture. Our work studies the Res Net architecture, with two widely used architecture changes, the Res Net-D [15] modification and Squeeze-and-Excitation (SE) in all bottleneck blocks [21]. These architectural changes are used in used many architectures, including TRes Net, Res Ne St and Efficient Nets. The exact details of our architecture can be found in Appendix E. In our experiments, we sometimes use the original Res Net implementation without SE (referred to as Res Net) to compare different training methods. Clear denotations are made in table captions when this is the case. Regularization and Data Augmentation. We apply weight decay, label smoothing, dropout and stochastic depth for regularization. Dropout [50] is a common technique used in computer vision and we apply it to the output after the global average pooling occurs in the final layer. Stochastic depth [22] drops out each layer in the network (that has residual connections around it) with a specified probability that is a function of the layer depth. We use Rand Augment [7] data augmentation as an additional regularizer. Rand Augment applies a sequence of random image transformations (e.g. translate, shear, color distortions) to each image independently during training. Our training method closely matches that of Efficient Net, where we train for 350 epochs, but with a few small differences (e.g. we use Momentum with cosine learning rate schedule as opposed to RMSProp with exponential decay). See Appendix D for details. Hyperparameter Tuning. To select the hyperparameters for the various regularization and training methods, we use a held-out validation set comprising 2% of the Image Net training set (20 shards out of 1024). This is referred to as the minival-set and the original Image Net validation set (the one reported in most prior works) is referred to as validation-set. Unless specified otherwise, results are reported on the validation-set. The hyperparameters of all Res Net-RS models are in Table 8 in the Appendix C. 1A notable exception is Reg Net [39] which purposely makes no use of improved training techniques and shows improvements over worsened Efficient Net baselines, but does not demonstrate Image Net accuracies above (a rather low) 81%. While this approach facilitates fair comparisons with prior work, it is unclear whether improvements are sustained at larger scales with improved training setups. For example, our scaling analysis shows that the scaling strategy advocated by Reg Net does not generalize to training regimes where overfitting can occur. 5 Improved Training Methods 5.1 Additive Study of Improvements We present an additive study of training, regularization methods and architectural changes in Table 1 (left). The baseline Res Net-200 gets 79.0% top-1 accuracy. We improve its performance to 82.2% (+3.2%) through improved training methods alone without any architectural changes. Adding two common and simple architectural changes (Squeeze-and-Excitation and Res Net-D) further boosts the performance to 83.4%. Training methods alone cause 3/4 of the total improvement, which demonstrates their critical impact on Image Net performance. Improvements Top-1 Res Net-200 - 256x256 79.0 + Cosine LR Decay 79.3 +0.3 + Increase training epochs 78.8 -0.5 + EMA of weights 79.1 +0.3 + Label Smoothing 80.4 +1.3 + Stochastic Depth 80.6 +0.2 + Rand Augment 81.0 +0.4 + Dropout on FC 80.7 -0.3 + Decrease weight decay 82.2 +1.5 + Squeeze-and-Excitation 82.9 +0.7 + Res Net-D 83.4 +0.5 Model Regularization Weight Decay 1e-4 4e-5 Res Net-50 None 79.7 78.7 (-1.0) Res Net-50 RA-LS 82.4 82.3 (-0.1) Res Net-50 RA-LS-DO 82.2 82.7 (+0.5) Res Net-200 None 82.5 81.7 (-0.8) Res Net-200 RA-LS 85.2 84.9 (-0.3) Res Net-200 RA-LS-SD-DO 85.3 85.5 (+0.2) Table 1: (Left) Additive study of training , regularization and architecture improvements. The baseline Res Net-200 is trained at resolution 256 256 for the standard 90 epochs using a stepwise learning rate decay schedule. All numbers are reported on the Image Net validation-set and averaged over 2 runs. Increasing training duration to 350 epochs only becomes useful once the regularization methods are used, otherwise the accuracy drops due to over-fitting. dropout hurts as we have not yet decreased the weight decay. (Right) Decreasing weight decay improves performance when combining regularization methods such as dropout (DO), stochastic depth (SD), label smoothing (LS) and Rand Augment (RA). Image resolution is 224 224 for Res Net-50 and 256 256 for Res Net-200. All numbers are reported on the Image Net minival-set from an average of two runs. 5.2 Importance of decreasing weight decay when combining regularization methods Table 1 (right) highlights the importance of changing weight decay when combining regularization methods together. When applying Rand Augment and label smoothing, there is no need to change the default weight decay of 1e-4. But when we further add dropout and/or stochastic depth, the performance can decrease unless we further decrease the weight decay. The intuition is that since weight decay acts as a regularizer, its value must be decreased in order to not overly regularize the model when combining many techniques. Furthermore, [65] presents evidence that the addition of data augmentation shrinks the L2 norm of the weights, which renders some of the effects of weight decay redundant. Other works use smaller weight decay values, but do not point out the significance of the effect when using more regularization [54, 55]. 6 Improved Scaling Strategies The prior section demonstrates the significant impact of training methodology and we now show the scaling strategy is similarly important. In order to establish scaling trends, we perform an extensive search on Image Net over width multipliers in [0.25,0.5,1.0,1.5,2.0], depths of [26,50,101,200,300,350,400] and resolutions of [128,160,224,320,448]. We train these architectures for 350 epochs, mimicking the training setup of state-of-the-art Image Net models, and increase regularization with model size in an effort to limit overfitting. See Appendix F for regularization and model hyperparameters. 108 109 1010 1011 FLOPs Image Net Error Scaling Properties of Res Nets 128 160 224 320 448 Figure 2: Scaling properties of Res Nets across varying model scales. Error approximately scales as a power law with FLOPs (linear fit on the log-log curve) in the lower FLOPs regime but the trend breaks for larger FLOPs. We observe diminishing returns of scaling the image resolutions beyond 320 320, which motivates the slow image resolution scaling (Strategy #2). All results are on the Image Net minival-set. FLOPs do not accurately predict performance in the bounded data regime. Prior works on scaling laws observe a power law between error and FLOPs in unbounded data regimes [25, 16]. In order to test whether this also holds in our scenario, we plot Image Net error against FLOPs for all scaling configurations in Figure 2. For the smaller models, we observe an overall power law trend between error and FLOPs, with minor dependency on the scaling configuration (i.e. depth, width and image resolution). However, the trend breaks for larger model sizes and we observe a large variation in Image Net performance for a fixed amount of FLOPs, especially in the higher FLOP regime. Therefore the exact scaling configuration (i.e. depth, width and image resolution) can have a big impact on performance even when controlling for the same amount of FLOPs. The best performing scaling strategy depends on the training regime. We next look directly at latencies2 on the hardware of interest to identify scaling strategies that improve the speed-accuracy Pareto curve. Figure 3 presents accuracies and latencies of models scaled with either width or depth across four image resolutions and three different training regimes (10, 100 and 350 epochs). We observe that the best performing scaling strategy, especially whether to scale depth and/or width, highly depends on the training regime. 6.1 Strategy #1 - Depth Scaling in Regimes Where Overfitting Can Occur Depth scaling outperforms width scaling for longer epoch regimes. In the 350 epochs setup (Figure 3 - right), we observe depth scaling to significantly outperform width scaling across all image resolutions. Scaling the width is subject to overfitting and sometimes hurts performance even with increased regularization. We hypothesize that this is due to the larger increase in parameters when scaling the width. The Res Net architecture maintains constant FLOPs across all block groups and multiplies the number of parameters by 4 every block group. Scaling the depth, especially in the earlier layers, therefore introduces fewer parameters compared to scaling the width. Width scaling outperforms depth scaling for shorter epoch regimes. In contrast, width scaling is better when only training for 10 epochs (Figure 3 - left). For 100 epochs (Figure 3 - middle), the best performing scaling strategy varies between depth scaling and width scaling, depending on the image resolution. The dependency of the scaling strategy on the training regime reveals a pitfall of extrapolating scaling rules. We point out that prior works also choose to scale the width when training for a small number of epochs on large-scale datasets (e.g. 40 epochs on 300M images), consistent with our experimental findings that scaling the width is preferable in shorter epoch regimes. In particular, [27] train a Res Net-152 with 4x filter multiplier while [3] scales the width with 1.5x filter multiplier. 6.2 Strategy #2 - Slow Image Resolution Scaling In Figure 2, we also observe that larger image resolutions yield diminishing returns. We therefore propose to increase the image resolution more gradually than previous works. This contrasts with the compound scaling rule proposed by Efficient Net which leads to very large images (e.g. 600 for Efficient Net-B7, 800 for Efficient Net-L2 [57]). Other works such as Res Ne St [62] and TRes Net [43]) scale the image resolution up to 400+. Our experiments indicate that slower image scaling improves not only Res Net architectures, but also Efficient Nets on a speed-accuracy basis (Section 7.1). 2FLOPs is not a good indicator of latency on modern hardware. See Section 7.1 for a more detailed discussion. 0.05 0.10 0.15 0.20 0.25 0.30 0.35 0.40 Time Per Training Step (Sec) Top-1 Image Net Accuracy Res: 224 Res: 320 Depth Scaling Width Scaling 0.05 0.10 0.15 0.20 0.25 0.30 0.35 0.40 Time Per Training Step (Sec) Top-1 Image Net Accuracy Epochs: 100 Depth Scaling Width Scaling 0.05 0.10 0.15 0.20 0.25 0.30 0.35 0.40 Time Per Training Step (Sec) Top-1 Image Net Accuracy Epochs: 350 Depth Scaling Width Scaling Figure 3: Scaling of Res Nets across depth, width, image resolution and training epochs. We compare depth scaling and width scaling across four different image resolutions [128,160,224,320] when training models for 10, 100 or 350 epochs. We find that the best performing scaling strategy depends on the training regime, which reveals the pitfall of extrapolating scaling rules from small scale regimes. (Left) 10 Epoch Regime: width scaling is the best strategy for the speed-accuracy Pareto curve. (Middle) 100 Epoch Regime: depth scaling is sometimes outperformed by width scaling. (Right) 350 Epoch Regime: depth scaling consistently outperforms width scaling by a large margin. Overfitting remains an issue even when using regularization methods. Model Details: All models start from a depth of 101 and are increased through [101,200,300,400]. All model widths start with a multiplier of 1.0x and are increased through [1.0,1.5,2.0]. For all models, we tune regularization in an effort to limit overfitting (see Appendix F). Accuracies are reported on the Image Net minival-set and training times are measured on TPUs. 6.3 Designing Scaling Strategies Our scaling analysis surfaces two common pitfalls in prior research on scaling strategies. Pitfall #1: Extrapolating scaling strategies from small-scale regimes. Scaling strategies found in small scale regimes (e.g. on small models or with few training epochs) can fail to generalize to larger models or longer training iterations. The dependencies between the best performing scaling strategy and the training regime are missed by prior works which extrapolate scaling rules from either small models [55] or shorter training epochs [39]. We therefore do not recommend generating scaling rules exclusively in a small scale regime because these rules can break down. Pitfall #2: Extrapolating scaling strategies from a single and potentially sub-optimal initial architecture. Beginning from a sub-optimal initial architecture can skew the scaling results. For example, the compound scaling rule derived from a small grid search around Efficient Net-B0, which was obtained by architecture search using a fixed FLOPs budget and a specific image resolution. However, since this image resolution can be sub-optimal for that FLOPs budget, the resulting scaling strategy can be sub-optimal. In contrast, our work designs scaling strategies by training models across a variety of widths, depths and image resolutions. Summary of Improved Scaling Strategies. For image classification, the scaling strategies are summarized as (1) scale the depth in regimes where overfitting can occur (scaling the width is preferable otherwise) and (2) slow image resolution scaling. Experiments indicate that applying these scaling strategies to Res Nets (Res Net-RS) and Efficient Nets (Efficient Net-RS) leads to significant speed-ups over Efficient Nets. We note that similar scaling strategies are also employed in recent works that obtain large speed-ups over Efficient Nets such as Lambda Res Nets [1] and NFNets [3]. For a new task, we recommend running a small subset of models across different scales, for the full training epochs, to gain intuition on which dimensions are the most useful across model scales. While this approach may appear more costly, we point out that the cost is offset by not searching for the architecture. 7 Experiments with Improved Training and Scaling Strategies 7.1 Res Net-RS on a Speed-Accuracy Basis Using the improved training and scaling strategies, we design Res Net-RS, a family of re-scaled Res Nets across a wide range of model scales (see Appendix C and E for experimental and architectural details). Figure 4 and Table 2 compare Efficient Nets against Res Net-RS on a speed-accuracy Pareto curve. We find that Res Net-RS match Efficient Nets performance while being 1.7x - 2.7x faster on TPUs (2.1x - 3.3x faster on GPUs). We point that these speed-ups are superior to those obtained by TRes Nest and Res Ne St3, suggesting that Res Net-RS also outperform TRes Net and Res Ne St. 3TRes Net and Res Ne St report 1.3 - 2.0x speed-ups over Efficient Net on a GPU V100. 0.0 0.5 1.0 1.5 2.0 2.5 3.0 Time Per Training Step (Sec) Top-1 Image Net Accuracy 350-320 420-320 B6 2.7x Speedup 1.9x Speedup Speed-Accuracy Pareto Curve Res Net-RS Efficient Net Figure 4: Speed-Accuracy Pareto curve comparing Res Nets-RS to Efficient Net. Res Net-RS (annotated with depth - image resolution) are 1.7x - 2.7x faster than the popular Efficient Nets when closely matching their training setup. Although Res Net-RS has more parameters and FLOPs, the model employs less memory and runs faster on TPUs and GPUs. See Appendix C and I for more results and profiling details. Model Image Resolution Params (M) FLOPs (B) V100 Latency (s) TPUv3 Latency (ms) Top-1 Efficient Net-B0 224 5.3 0.8 0.47 90 77.1 Efficient Net-B1 240 7.8 1.4 0.82 150 79.1 Res Net-RS-50 160 36 4.6 0.31 70 78.8 Efficient Net-B2 260 9.2 2.0 1.03 210 80.1 Res Net-RS-101 160 64 8.4 0.48 (2.1 ) 120 (1.8 ) 80.3 Efficient Net-B3 300 12 3.6 1.76 340 81.6 Res Net-RS-101 192 64 12 0.70 170 81.2 Res Net-RS-152 192 87 18 0.99 240 82.0 Efficient Net-B4 380 19 8.4 4.0 710 82.9 Res Net-RS-152 224 87 24 1.48 (2.7 ) 320 (2.2 ) 82.8 Res Net-RS-152 256 87 31 1.76 (2.3 ) 410 (1.7 ) 83.0 Efficient Net-B5 456 30 20 8.16 1510 83.7 Res Net-RS-200 256 93 40 2.86 570 83.4 Res Net-RS-270 256 130 54 3.76 (2.2 ) 780 (1.9 ) 83.8 Efficient Net-B6 528 43 38 15.7 3010 84.0 Res Net-RS-350 256 164 69 4.72 (3.3 ) 1100 (2.7 ) 84.0 Efficient Net-B7 600 66 74 29.9 6020 84.7 Res Net-RS-350 320 164 107 8.48 1630 84.2 Res Net-RS-420 320 192 128 10.16 2090 84.4 Table 2: Details of Res Net-RS models in Pareto curve. See Table 8 for hyperparameters and Section I for profiling details. This large speed-up over Efficient Net may be non-intuitive since Efficient Nets have significantly reduced parameters and FLOPs compared to Res Nets. We next discuss why a model with fewer parameters and fewer FLOPs (Efficient Net) is slower and more memory-intensive during training. FLOPs vs Latency. While FLOPs provide a hardware-agnostic metric for assessing computational demand, they may not be indicative of actual latency times for training and inference [19, 18, 39]. In custom hardware architectures (e.g. TPUs and GPUs), FLOPs are an especially poor proxy because operations are often bounded by memory access costs and have different levels of optimization on modern matrix multiplication units [24]. The inverted bottlenecks [46] used in Efficient Nets employ depthwise convolutions with large activations and have a small compute to memory ratio (operational intensity) compared to the Res Net s bottleneck blocks which employ dense convolutions on smaller activations. This makes Efficient Nets less efficient on modern accelerators compared to Res Nets. Figure 4 (table on the right) illustrates this point: a Res Net-RS model with 1.8x more FLOPs than Efficient Net-B6 is 2.7x faster on a TPUv3 hardware accelerator. Parameters vs Memory. Parameter count does not necessarily dictate memory consumption during training because memory is often dominated by the size of the activations4. The large activations used in Efficient Nets also cause larger memory consumption, which is exacerbated by the use of large image resolutions, compared to our re-scaled Res Nets. A Res Net-RS model with 3.8x more parameters than Efficient Net-B6 consumes 2.3x less memory for a similar Image Net accuracy (Table in Figure 4). We emphasize that both memory consumption and latency are tightly coupled to the software and hardware stack (Tensor Flow on TPUv3) due to compiler optimizations such as operation layout assignments and memory padding. Improving scaling of Efficient Nets The scaling analysis from Section 6 reveals that scaling the image resolution results in diminishing returns. This suggests that the compound scaling rule advocated in Efficient Net which jointly increases model depth, width and resolution at a constant rate is sub-optimal. To test this hypothesis, we apply the slow image resolution scaling strategy (Strategy #2) to Efficient Nets and train several versions with reduced image resolutions, without changing the width or depth. Figure 5 (Appendix) demonstrates a marked improvement of the re-scaled Efficient Nets (Efficient Net-RS) on the speed-accuracy Pareto curve over the original Efficient Nets. 7.2 Semi-Supervised Learning with Res Net-RS We next measure how Res Net-RS performs as we scale to larger datasets in a large scale semisupervised learning setup. We train Res Nets-RS on the combination of 1.3M labeled Image Net images and 130M pseudo-labeled images, in a similar fashion to Noisy Student [57]. We use the same dataset of 130M images pseudo-labeled as Noisy Student, where the pseudo labels are generated from an Efficient Net-L2 model with 88.4% Image Net accuracy. Model V100 (s) TPUv3 (ms) Top-1 Efficient Net-B5 8.16 1510 86.1 Res Net-RS-152 1.48 (5.5x) 320 (4.7x) 86.2 Table 3: Res Net-RS are efficient semi-supervised learners. Res Net-RS-152 with image resolution 224 is 4.7x faster on TPU (5.5x on GPU) than Efficient Net-B5 Noisy Student for a similar Image Net accuracy. Models are jointly trained on both the labeled and pseudo-labeled data and training hyperparameters are kept the same. Table 3 reveals that Res Net-RS models are very strong in the semi-supervised learning setup as well, achieving a strong 86.2% top-1 Image Net accuracy while being 4.7x faster on TPU (5.5x on GPU) than the corresponding Efficient Net model. 7.3 Transfer Learning to Downstream Tasks with Res Net-RS We now investigate whether the improved supervised training strategies yield better representations for transfer learning and compare them with self-supervised learning algorithms. Recent selfsupervised learning algorithms claim to surpass the transfer learning performance of supervised learning and create more universal representations [4, 5]. Self-supervised algorithms, however, make several changes to the training methods (e.g training for more epochs, data augmentation) making comparisons to supervised learning difficult. Fairly comparing supervised learning and self-supervised learning. In an effort to closely match Sim CLR s training setup and provide fair comparisons, we restrict the RS training strategies to a subset of its original methods. Specifically, we train for for 400 epochs with cosine learning rate decay, data augmentation (Rand Augment), label smoothing, dropout and decreased weight decay but do not use stochastic depth or exponential moving average (EMA) of the weights. We choose this subset to closely match the training setup of Sim CLR: longer training (800 epochs) with cosine learning rate decay, a tailored data augmentation strategy, a tuned temperature parameter in the contrastive loss and a tuned weight decay. Table 4 compares the transfer performance of supervised learning with or without improved training strategies (respectively denoted RS and Supervised) against Sim CLR/Sim CLRv2 [4, 5] on five downstream tasks: CIFAR-100 Classification [29], Pascal Detection & Segmentation [9], ADE Segmentation [64] and NYU Depth [48]. Our experiments demonstrate that the improved training 4Activations are typically stored during training as they are used in backpropagation. At inference, activations can be discarded and parameter count is a better proxy for actual memory consumption. Model Training Epochs CIFAR-100 Pascal Pascal ADE NYU Method Accuracy Detection Segmentation Segmentation Depth Res Net-152 Supervised 90 85.5 80.0 70.0 40.2 81.2 Res Net-152 Sim CLR 800 87.1 83.3 72.2 41.0 83.5 Res Net-152 Sim CLRv2 800 84.7 79.1 73.1 41.1 84.7 Res Net-152 RS 400 88.1 82.2 78.2 42.2 83.4 Table 4: Representations from supervised learning with improved training strategies rival or outperform representations from state-of-the-art self-supervised learning algorithms. Comparison of supervised training methods (supervised, RS) and self-supervised methods (Sim CLR, Sim CLRv2) on a variety of downstream tasks. The improved training strategies (RS) greatly outperforms the baseline supervised training, which highlights the importance of using improved supervised training techniques when comparing to self-supervised learning algorithms. All models employ the vanilla Res Net architecture and are pre-trained on Image Net. strategies significantly improve transfer performance, in line with works that observe that higher Image Net accuracy strongly correlates with improved transfer learning performance [28]. Furthermore, we find that the improved supervised representations (RS) rival or outperform Sim CLR/Sim CLRv2, even when restricted to a smaller subset. These results challenge the notion that self-supervised algorithms lead to more universal representations than supervised learning when labels are available. 7.4 Revised 3D Res Net for Video Classification Improvements Top-1 3D Res Net-50 73.4 + Dropout on FC 74.4 +1.0 + Label smoothing 74.9 +0.5 + Stochastic depth 76.1 +1.2 + EMA of weights 76.1 + Decrease weight decay 76.3 +0.2 + Increase training epochs 76.4 +0.1 + Scale jittering 77.4 +1.0 + Squeeze-and-Excitation 77.9 +0.5 + Res Net-D 78.2 +0.3 Table 5: Additive study of regularization , training and architecture improvements with 3D-Res Net on video classification. We conclude by applying the training strategies to the Kinetics-400 video classification task [26], using a 3D Res Net as the baseline architecture [38]. Table 5 presents an additive study of the RS training recipe and architectural improvements. The training strategies extend to video classification, yielding a combined improvement from 73.4% to 77.4% (+4.0%). The Res Net D and Squeeze-and-Excitation architectural changes further improve the performance to 78.2% (+0.8%). Similarly to our study on image classification (Table 1), we find that most of the improvement can be obtained without architectural changes. 8 Conclusion By updating the de facto vision baseline with modern training methods and an improved scaling strategy, we have revealed the remarkable durability of the Res Net architecture. Simple architectures set strong baselines for state-of-the-art methods: the accuracy gains that motivate complicated architectural changes may be surpassed with thoughtful scaling and training strategies. Our work suggests that the field has myopically overemphasized architectural innovations at the expense of experimental diligence, and we hope it encourages further scrutiny in maintaining consistent methodology for both proposed innovations and baselines alike. We do not foresee any negative societal impact of our work. We include further discussion in the Appendix B. Acknowledgements. We would like to thank Ashish Vaswani, Prajit Ramachandran, Ting Chen, Thang Luong, Hanxiao Liu, Gabriel Bender, Quoc Le, Neil Houlsby, Mingxing Tan, Andrew Howard, Raphael Gontijo Lopes, Andy Brock and David Berthelot for helpful feedback on this work; Jing Li, Pengchong Jin, Yeqing Li and Yin Cui for the support on open-sourcing and infrastructure. [1] I. Bello. Lambdanetworks: Modeling long-range interactions without attention. International Conference in Learning Representations, 2021. [2] I. Bello, B. Zoph, A. Vaswani, J. Shlens, and Q. V. Le. Attention augmented convolutional networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 3286 3295, 2019. [3] A. Brock, S. De, S. L. Smith, and K. Simonyan. High-performance large-scale image recognition without normalization. ar Xiv preprint ar Xiv: 2102.06171, 2021. [4] T. Chen, S. Kornblith, M. Norouzi, and G. Hinton. A simple framework for contrastive learning of visual representations. In International conference on machine learning, pages 1597 1607. PMLR, 2020. [5] T. Chen, S. Kornblith, K. Swersky, M. Norouzi, and G. Hinton. Big self-supervised models are strong semi-supervised learners. ar Xiv preprint ar Xiv:2006.10029, 2020. [6] E. D. Cubuk, B. Zoph, D. Mane, V. Vasudevan, and Q. V. Le. Autoaugment: Learning augmentation policies from data. ar Xiv preprint ar Xiv:1805.09501, 2018. [7] E. D. Cubuk, B. Zoph, J. Shlens, and Q. V. Le. Randaugment: Practical automated data augmentation with a reduced search space. ar Xiv preprint ar Xiv:1909.13719, 2019. [8] A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. ar Xiv preprint ar Xiv:2010.11929, 2020. [9] M. Everingham, L. Van Gool, C. K. Williams, J. Winn, and A. Zisserman. The pascal visual object classes (voc) challenge. International journal of computer vision, 88(2):303 338, 2010. [10] C. Feichtenhofer, H. Fan, J. Malik, and K. He. Slowfast networks for video recognition. 2019 IEEE/CVF International Conference on Computer Vision (ICCV), pages 6201 6210, 2019. [11] G. Ghiasi, T.-Y. Lin, and Q. V. Le. Dropblock: A regularization method for convolutional networks. In Advances in Neural Information Processing Systems, pages 10727 10737, 2018. [12] P. Goyal, P. Dollár, R. Girshick, P. Noordhuis, L. Wesolowski, A. Kyrola, A. Tulloch, Y. Jia, and K. He. Accurate, large minibatch sgd: Training imagenet in 1 hour. ar Xiv preprint ar Xiv:1706.02677, 2017. [13] K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. ar Xiv preprint ar Xiv:1512.03385, 2015. [14] K. He, X. Zhang, S. Ren, and J. Sun. Identity mappings in deep residual networks. ar Xiv preprint ar Xiv:1603.05027, 2016. [15] T. He, Z. Zhang, H. Zhang, Z. Zhang, J. Xie, and M. Li. Bag of tricks for image classification with convolutional neural networks. ar Xiv preprint ar Xiv:1812.01187, 2018. [16] T. Henighan, J. Kaplan, M. Katz, M. Chen, C. Hesse, J. Jackson, H. Jun, T. B. Brown, P. Dhariwal, S. Gray, et al. Scaling laws for autoregressive generative modeling. ar Xiv preprint ar Xiv:2010.14701, 2020. [17] J. Hestness, S. Narang, N. Ardalani, G. Diamos, H. Jun, H. Kianinejad, M. M. A. Patwary, Y. Yang, and Y. Zhou. Deep learning scaling is predictable, empirically. ar Xiv preprint ar Xiv: 1712.00409, 2017. [18] A. Howard, M. Sandler, G. Chu, L.-C. Chen, B. Chen, M. Tan, W. Wang, Y. Zhu, R. Pang, V. Vasudevan, et al. Searching for mobilenetv3. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 1314 1324, 2019. [19] A. G. Howard, M. Zhu, B. Chen, D. Kalenichenko, W. Wang, T. Weyand, M. Andreetto, and H. Adam. Mobilenets: Efficient convolutional neural networks for mobile vision applications. ar Xiv preprint ar Xiv:1704.04861, 2017. [20] H. Hu, Z. Zhang, Z. Xie, and S. Lin. Local relation networks for image recognition. ar Xiv preprint ar Xiv:1904.11491, 2019. [21] J. Hu, L. Shen, and G. Sun. Squeeze-and-excitation networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 7132 7141, 2018. [22] G. Huang, Y. Sun, Z. Liu, D. Sedra, and K. Q. Weinberger. Deep networks with stochastic depth. In European conference on computer vision, pages 646 661. Springer, 2016. [23] Y. Huang, Y. Cheng, A. Bapna, O. Firat, M. X. Chen, D. Chen, H. Lee, J. Ngiam, Q. V. Le, Y. Wu, et al. Gpipe: Efficient training of giant neural networks using pipeline parallelism. ar Xiv preprint ar Xiv:1811.06965, 2018. [24] N. P. Jouppi, C. Young, N. Patil, D. Patterson, G. Agrawal, R. Bajwa, S. Bates, S. Bhatia, N. Boden, A. Borchers, et al. In-datacenter performance analysis of a tensor processing unit. In Proceedings of the 44th annual international symposium on computer architecture, pages 1 12, 2017. [25] J. Kaplan, S. Mc Candlish, T. Henighan, T. B. Brown, B. Chess, R. Child, S. Gray, A. Radford, J. Wu, and D. Amodei. Scaling laws for neural language models. ar Xiv preprint ar Xiv:2001.08361, 2020. [26] W. Kay, J. Carreira, K. Simonyan, B. Zhang, C. Hillier, S. Vijayanarasimhan, F. Viola, T. Green, T. Back, A. Natsev, M. Suleyman, and A. Zisserman. The kinetics human action video dataset. Ar Xiv, abs/1705.06950, 2017. [27] A. Kolesnikov, L. Beyer, X. Zhai, J. Puigcerver, J. Yung, S. Gelly, and N. Houlsby. Big transfer (bit): General visual representation learning. ar Xiv preprint ar Xiv:1912.11370, 6(2):8, 2019. [28] S. Kornblith, J. Shlens, and Q. V. Le. Do better imagenet models transfer better? In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2661 2671, 2019. [29] A. Krizhevsky, G. Hinton, et al. Learning multiple layers of features from tiny images. 2009. [30] A. Krizhevsky, I. Sutskever, and G. E. Hinton. Imagenet classification with deep convolutional neural networks. In Advances in neural information processing systems, pages 1097 1105, 2012. [31] J. Lee, T. Won, and K. Hong. Compounding the performance improvements of assembled techniques in a convolutional neural network. ar Xiv preprint ar Xiv:2001.06268, 2020. [32] X. Li, W. Wang, X. Hu, and J. Yang. Selective kernel networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 510 519, 2019. [33] T.-Y. Lin, P. Dollár, R. Girshick, K. He, B. Hariharan, and S. Belongie. Feature pyramid networks for object detection. In CVPR, 2017. [34] I. Loshchilov and F. Hutter. Sgdr: Stochastic gradient descent with warm restarts. ar Xiv preprint ar Xiv:1608.03983, 2016. [35] D. Mahajan, R. Girshick, V. Ramanathan, K. He, M. Paluri, Y. Li, A. Bharambe, and L. Van Der Maaten. Exploring the limits of weakly supervised pretraining. In Proceedings of the European Conference on Computer Vision (ECCV), pages 181 196, 2018. [36] T. Norrie, N. Patil, D. H. Yoon, G. Kurian, S. Li, J. Laudon, C. Young, N. Jouppi, and D. Patterson. The design process for google s training chips: Tpuv2 and tpuv3. IEEE Micro, 2021. [37] H. Pham, Q. Xie, Z. Dai, and Q. V. Le. Meta pseudo labels. ar Xiv preprint ar Xiv:2003.10580, 2020. [38] R. Qian, T. Meng, B. Gong, M.-H. Yang, H. Wang, S. J. Belongie, and Y. Cui. Spatiotemporal contrastive video representation learning. ar Xiv: Computer Vision and Pattern Recognition, 2020. [39] I. Radosavovic, R. P. Kosaraju, R. Girshick, K. He, and P. Dollár. Designing network design spaces. ar Xiv preprint ar Xiv: 2003.13678, 2020. [40] P. Ramachandran, N. Parmar, A. Vaswani, I. Bello, A. Levskaya, and J. Shlens. Stand-alone self-attention in vision models. ar Xiv preprint ar Xiv:1906.05909, 2019. [41] E. Real, A. Aggarwal, Y. Huang, and Q. V. Le. Regularized evolution for image classifier architecture search. In Proceedings of the aaai conference on artificial intelligence, volume 33, pages 4780 4789, 2019. [42] S. Ren, K. He, R. Girshick, and J. Sun. Faster r-cnn: Towards real-time object detection with region proposal networks. ar Xiv preprint ar Xiv:1506.01497, 2015. [43] T. Ridnik, H. Lawen, A. Noy, and I. Friedman. Tresnet: High performance gpu-dedicated architecture. ar Xiv preprint ar Xiv:2003.13630, 2020. [44] J. S. Rosenfeld, A. Rosenfeld, Y. Belinkov, and N. Shavit. A constructive prediction of the generalization error across scales. ar Xiv preprint ar Xiv: 1909.12673, 2019. [45] O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A. Karpathy, A. Khosla, M. Bernstein, et al. Imagenet large scale visual recognition challenge. IJCV, 2015. [46] M. Sandler, A. Howard, M. Zhu, A. Zhmoginov, and L.-C. Chen. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 4510 4520, 2018. [47] Z. Shen, I. Bello, R. Vemulapalli, X. Jia, and C.-H. Chen. Global self-attention networks for image recognition. ar Xiv preprint ar Xiv: 2010.03019, 2020. [48] N. Silberman, D. Hoiem, P. Kohli, and R. Fergus. Indoor segmentation and support inference from rgbd images. In European conference on computer vision, pages 746 760. Springer, 2012. [49] K. Simonyan and A. Zisserman. Very deep convolutional networks for large-scale image recognition. ar Xiv preprint ar Xiv:1409.1556, 2014. [50] N. Srivastava, G. Hinton, A. Krizhevsky, I. Sutskever, and R. Salakhutdinov. Dropout: a simple way to prevent neural networks from overfitting. The journal of machine learning research, 2014. [51] C. Sun, A. Shrivastava, S. Singh, and A. Gupta. Revisiting unreasonable effectiveness of data in deep learning era. ar Xiv preprint ar Xiv: 1707.02968, 2017. [52] C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke, and A. Rabinovich. Going deeper with convolutions. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 1 9, 2015. [53] C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens, and Z. Wojna. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 2818 2826, 2016. [54] M. Tan, B. Chen, R. Pang, V. Vasudevan, M. Sandler, A. Howard, and Q. V. Le. Mnasnet: Platform-aware neural architecture search for mobile. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2820 2828, 2019. [55] M. Tan and Q. V. Le. Efficientnet: Rethinking model scaling for convolutional neural networks. ar Xiv preprint ar Xiv:1905.11946, 2019. [56] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. Kaiser, and I. Polosukhin. Attention is all you need. ar Xiv preprint ar Xiv: 1706.03762, 2017. [57] Q. Xie, M.-T. Luong, E. Hovy, and Q. V. Le. Self-training with noisy student improves imagenet classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10687 10698, 2020. [58] S. Xie, R. Girshick, P. Dollár, Z. Tu, and K. He. Aggregated residual transformations for deep neural networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 1492 1500, 2017. [59] S. Yun, D. Han, S. J. Oh, S. Chun, J. Choe, and Y. Yoo. Cutmix: Regularization strategy to train strong classifiers with localizable features. In Proceedings of the IEEE International Conference on Computer Vision, pages 6023 6032, 2019. [60] S. Zagoruyko and N. Komodakis. Wide residual networks. ar Xiv preprint ar Xiv:1605.07146, 2016. [61] H. Zhang, M. Cisse, Y. N. Dauphin, and D. Lopez-Paz. mixup: Beyond empirical risk minimization. ar Xiv preprint ar Xiv:1710.09412, 2017. [62] H. Zhang, C. Wu, Z. Zhang, Y. Zhu, Z. Zhang, H. Lin, Y. Sun, T. He, J. Muller, R. Manmatha, M. Li, and A. Smola. Resnest: Split-attention networks. ar Xiv preprint ar Xiv:2004.08955, 2020. [63] R. Zhang. Making convolutional networks shift-invariant again. ar Xiv preprint ar Xiv:1904.11486, 2019. [64] B. Zhou, H. Zhao, X. Puig, S. Fidler, A. Barriuso, and A. Torralba. Scene parsing through ade20k dataset. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 633 641, 2017. [65] B. Zoph, E. D. Cubuk, G. Ghiasi, T.-Y. Lin, J. Shlens, and Q. V. Le. Learning data augmentation strategies for object detection. In European Conference on Computer Vision, pages 566 583. Springer, 2020. [66] B. Zoph, G. Ghiasi, T.-Y. Lin, Y. Cui, H. Liu, E. D. Cubuk, and Q. V. Le. Rethinking pre-training and self-training. ar Xiv preprint ar Xiv:2006.06882, 2020. [67] B. Zoph, V. Vasudevan, J. Shlens, and Q. V. Le. Learning transferable architectures for scalable image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 8697 8710, 2018.