# acgc_lossy_activation_compression_with_guaranteed_convergence__22001d0b.pdf AC-GC: Lossy Activation Compression with Guaranteed Convergence R. David Evans Dept. of Electrical and Computer Engineering University of British Columbia Vancouver, BC V6T 1Z4 rdevans@ece.ubc.ca Tor M. Aamodt Dept. of Electrical and Computer Engineering University of British Columbia Vancouver, BC V6T 1Z4 aamodt@ece.ubc.ca Parallel hardware devices (e.g., graphics processor units) have limited highbandwidth memory capacity. This negatively impacts the training of deep neural networks (DNNs) by increasing runtime and/or decreasing accuracy when reducing model and/or batch size to fit this capacity. Lossy compression is a promising approach to tackling memory capacity constraints, but prior approaches rely on hyperparameter search to achieve a suitable trade-off between convergence and compression, negating runtime benefits. In this paper we build upon recent developments on Stochastic Gradient Descent convergence to prove an upper bound on the expected loss increase when training with compressed activation storage. We then express activation compression error in terms of this bound, allowing the compression rate to adapt to training conditions automatically. The advantage of our approach, called AC-GC, over existing lossy compression frameworks is that, given a preset allowable increase in loss, significant compression without significant increase in error can be achieved with a single training run. When combined with error-bounded methods, AC-GC achieves 15.1 compression with an average accuracy change of 0.1% on text and image datasets. AC-GC functions on any model composed of the layers analyzed and, by avoiding compression rate search, reduces overall training time by 4.6 over Successive Halving. 1 Introduction Stochastic Gradient Descent (SGD) has proven efficient and effective for optimizing Deep and Convolutional Neural Networks (DNNs and CNNs). However, due to deeper and automatically generated networks [20, 22, 44, 52], improvement of accuracy has caused a rapid increase in training memory requirements, which are dominated by the temporary storage of activations between the forward and backward pass of the back-propagation algorithm [45, 48]. Reducing memory consumption leads to faster training and, thus, more effective research of DNN models and applications. However, doing this by decreasing the batch size has many drawbacks. On parallel processors, such as GPUs, a small batch size can lead to poor compute saturation, and reduced training throughput [47]. Smaller batch sizes also introduce errors that impact convergence and accuracy [18]. Over 50GB of memory is required to train some networks, e.g., GPIPE [22]. Many works have examined reducing activation storage overheads. Lossy compression of activations in memory can reduce memory footprint without network modifications [6, 14, 25, 27]. Error bounded lossy compression (EBC) [27] has bounded activation error, however, it uses an empirical study to select an error target. Activations can also be offloaded to an external memory (e.g. CPU DRAM), using either an uncompressed link [31, 45] or compressed link [14, 46]. Activation compression and offloading have performance overheads from 5% to 60% [7, 14, 25, 27, 45]. Reduced precision 35th Conference on Neural Information Processing Systems (Neur IPS 2021). training has the side effect of reducing activation size [9, 51, 57]. Finally, restructuring networks to be reversible [17] or efficient scheduling of network layers [8] can reduce memory use. Prior lossy and reduced precision approaches [6, 9, 14, 25, 51, 57] utilize automated searches or hand-tuning to determine compression rates, which increase training time and have the potential to select poor compression/accuracy trade-offs. Ideally, an activation compression method has 0 5 20 40 60 80 Training Time (days, RTX 2080 Ti) This Work Automatic Fixpoint Successive Halving Fixpoint Act. Compr. Grid Search Fixpoint Act. Compr. Figure 1: Activation compression rate search cost for Image Net/Res Net50 [20]. Each box indicates a different compression from 1to 16-bit fixpoint. high compression and minimal decrease in trained accuracy. Tuning to achieve this is prohibitively expensive. For instance, selecting a fixed-point integer (fixpoint) compression rate for Image Net/Res Net50 using Grid Search uses 16 training runs (Figure 1). Using Successive Halving [26] can decrease training time, however even with aggressive resource allocations (e.g. 24 GPU-days, Figure 1) total training time is still high. With low resource settings, methods such as Successive Halving [26] and Hyperband [36] allocate little time to some configurations, increasing the likelihood that compression artifacts [6, 25] can be missed, resulting in poor accuracy. Additionally, when tuning hyperparameters, lossy compression makes it difficult to determine the cause of degraded accuracy. Finally, prior compression methods have an unknown impact on convergence behavior. In this work, we present a framework for lossy Activation Compression with Guaranteed Convergence (AC-GC). To our knowledge, our work is the first to prove convergence bounds on SGD with activation compression. AC-GC involves allowing an increase in the bound on the expected loss, which we trade-off for increased compression. We formulate this as a constrained optimization problem: maximizing compression subject to a bounded increase in loss. Doing so allows using a single hyperparameter to correlate convergence bounds with the activation error, which creates compression methods that are iteration, network, and dataset agnostic. Having convergence bounds known a priori allows a user to set a tolerable error rate before training, avoiding compression rate search cost entirely. Our contributions: We prove convergence bounds on SGD under error bounded activation compression with weak assumptions on convexity. We express activation compression and convergence as a constrained optimization problem and analyze the activation error tolerance of common DNN layers within this framework. We combine these error bounds with fixpoint, image, and error bounded compression, to create methods with guaranteed convergence and a compression/accuracy trade-off known prior to training. 2 Preliminaries DNNs are commonly used on problems involving a sum, for instance, minimize the total error on a set of training images. The loss L for such a problem takes the form n f(θ, Xn) (1) where f represents the loss of one example input Xn with weights θ. Stochastic Gradient Descent (SGD) is typically used to optimize these finite sums, using the iteration θ(t+1) = θ(t) α θf(θ(t), Xnt) (2) where α is the learning rate, t is the iteration, θf represents the gradient of f with respect to θ, and nt is a randomly chosen training example index from the distribution over n such that E[ θf(θ(t), Xnt)] = θL(θ(t)). Figure 2a shows the computation graph for the back-propagation [48] algorithm for a single DNN layer without compression. Back-propagation is often used as it allows efficient calculation of Activations C D Stored Activations a) Direct activation storage b) With activation compression Figure 2: Computation graph for training of a DNN layer. Activations are a) stored between the forward (left) and backward pass (right) or b) compressed (C) in the forward pass, and decompressed (D) in the backward pass. Red indicates paths potentially affected by compression errors. parameter gradients at the expense of storing activations [48]. A DNN layer is any linear or non-linear function, e.g., a convolution or Re LU activation. The functions fwd, bwd_param, bwd_act are algorithmic implementations of the layer function and the gradients w.r.t. θ and X. In the forward pass, each layer calculates an output activation Y = fwd(X) which is fed to subsequent layers. These activations are temporarily stored after use to avoid a performance penalty from recalculating them in the backward pass. To our knowledge, all frameworks opt to store activations [42, 53]. In the backward pass, parameter gradients and activation gradients are calculated using bwd_param and bwd_act. Parameter gradients are used to update the parameters (Eqn. (2)), and activation gradients are sent downward to the next layer. Depending on the layer type and its derivatives, the bwd_param and bwd_act functions may require activations to be stored. For example, it is computationally efficient to store the input X for convolution layers [42, 53]. The many layer types place a diverse set of constraints on the activation storage. Activation compression addresses one of the most significant contributors to memory consumption in DNNs. In the forward pass, an activation can be lazily compressed after its last usage (C, Figure 2b). Eagerly compressing activations would require storing both a compressed and uncompressed copy until its last use. In the backward pass, activations are decompressed before their first use (D, Figure 2b). The backward pass begins only after the forward pass is completed for all layers, resulting in a large reuse distance for stored activations. Compression can thus be performed off the critical path in parallel with compute, with low performance overheads from 4%-30% [7, 25], provided that sufficient resources are available. We denote the uncompressed activation as X = (xnchw) RN C H W , where N, C, H, and W represent the batch size, channel, height and width, respectively. In the uncompressed backward pass, gradients are calculated from the saved activations, parameters, and gradients from the upward layer (Figure 2 with a); we write this as θf(θ, X) = bwd_param(X, θ, Y f(θ, X)). Lossy compression involves discarding some information of the activation to increase the compression rate. In information theory, this is referred to as the rate-distortion trade-off. In our model, the rate refers to the activation error, and the distortion refers to resulting impacts on gradient error and thus accuracy after training. We model lossy compression between the forward and backward pass as an independent perturbation on each value in the activation, X RN C H W . Thus, in the compressed backward pass, the perturbed activation is X + X. We denote the approximate gradient resulting from lossy error as ˆ θf(θ, X) := bwd_param(X + X, θ, Y f(θ, X)), and the corresponding gradient error as θf(θ, X) := ˆ θf(θ, X) θf(θ, X). 3 Guaranteed Convergence This section details how gradient error θf(θ, X) impacts convergence of SGD. Following this, the lossy compression error X can be expressed in terms of the gradient error, and the compression rate for many methods can be determined. For example, the bitwidth b of fixpoint compression of an activation with range ( 1, 1) is b log2 | xnchw| + ... (3) Fast Convergence a) Early training Convergence b) Late training c) Late training Lossy Compression Epochs d) CIFAR10/Res Net50 Training 1-bit Fixpoint 2-bit Fixpoint 3-bit Fixpoint Uncompressed Figure 3: SGD convergence behavior. a) and b) without compression, c) this work, and d) Res Net50 training with α = 0.25 and fixpoint activation compression (average over five runs, shaded indicates the minimum and maximum training loss). where the remaining terms are constants determined by the rounding mode, sign, etc. Any compression method with bounded activation error for a given rate can be combined with the error bounds from this work to guarantee convergence (Sections 3.2 and 4). 3.1 SGD Convergence We will briefly summarize uncompressed convergence of SGD from Karimi et al. [28]. Consider training using t iterations of SGD with loss L(θ(t)), with a constant learning rate α and initial point θ(0). We assume that L has an optimal point θ( ) and satisfies E[ θf(θ, Xnt) 2] V 2 for all θ and some V 2. We refer to V 2 as the variance. With some assumptions and problem-defined constants C1 and C2 (Appendix A), Karimi et al. [28] demonstrate that the expected error at iteration t is E[L(θ(t)) L(θ( ))] (1 C1α)t(L(θ(0)) L(θ( ))) + C2αV 2 (4) Initially when training, fast convergence occurs as (1 C1α)t approaches zero (Figure 3a). Later in training, C2αV 2 dominates, resulting in an approximately constant expected error (Figure 3b). The key observation of this result is that the final loss scales with gradient variance, E[L(θ( ))] V 2. Many DNN classes fall under this progress bound as it uses relatively weak assumptions and does not require a convex f. DNNs using Re LU activations are Lipschitz continuous [55]. Furthermore, those with an L2 loss are piecewise strongly convex, which implies that the required Polyak-Łojasiewic condition is satisfied locally [40]. Networks this does not apply to could use another progress bound [3, 28, 41]. 3.2 Compressed Convergence Lossy compression trades accuracy for compression. This trade-off can be empirically observed with fixpoint compression on CIFAR10/Res Net50 (Figure 3d and Section 6). Our method functions by allowing the loss to increase by some multiplicative error (1 + e2), where e2 0, and determining how much compression can be extracted from the change (Figure 3c). The error 1 + e2 is chosen such that the compressed loss converges to the uncompressed loss as e2 0. Our process for translating the loss bound into maximum activation errors is illustrated in Figure 4a. This section outlines how the loss can be bounded by bounding the gradient error or using an intermediate bounding function D( X) to simplify the problem. We follow this in Section 4 by deriving an activation error X( ) for each network layer that satisfies this bound. e2 becomes the sole hyperparameter in our method, and selection determines the maximum increase in loss and the compression rate. Compressed convergence with increased loss can be viewed as an uncompressed problem with an increased gradient variance bound, (1 + e2)V 2. To determine the activation error which satisfies this gradient variance, we must express the variance in terms of a bound on the gradient error θf(θ, X) 2. Theorem 1 demonstrates that a maximum gradient error of e2V 2 satisfies the (1+e2) loss bound. There may be multiple regions where the gradient error is below e2V 2 (Figure 4b), which would require iterative solvers to determine suitable activation errors. To apply our technique in practice we introduce a convex function D( X) θf(θ, X) 2, which provides a flexible Bounding Function, Grad. Error a) Flowchart of Derivations b) 1D Bounding Function Example Gradient Variance Bound Per-Layer Equations Maximum Activation Error & Compression Rate Constrained Optimization Gradient Error Bound Loss Bound, Used to choose Section 3.2 Appendix B Section 4 Figure 4: a) Flowchart of derivations for obtaining the maximum activation error X( ) as a function of the loss bound 1 + e2. b) 1D Example for the relationship between the error bound, bounding function, and gradient error for an activation error x. Shaded area satisfies the loss bound, and the hatched area satisfies D( x) e2V 2. proxy for the gradient error (Figure 4b). Using D( X) allows defining the problem so that it has a unique solution. The gradient variance bound, bounding function, and gradient error are related in Theorem 1. Theorem 1. Given f which obeys (4), and a convex function D( X) which bounds the gradient error from above for all X, θ, and X: θf(θ, X) 2 D( X) (5) then any activation error X( ) where D( X( )) e2V 2 satisfies E[ ˆ θf(θ, Xnt) 2] (1 + e2)V 2 (6) assuming that E[ θf(θ, Xnt) ] = 0 for all θ, with positive value e2, and variance V 2 satisfying E[ θf(θ, Xnt) 2] V 2. All expectations are taken over the training examples nt. Proof: See Appendix A. Figure 4b demonstrates the relationship between the various bounds, as well as the motivation for the bounding function. The gradient error is derived from per-layer equations, and in many cases is highly non-convex. Due to this, local minima can cause difficulties when optimizing with the constraint θf(θ, X) 2 e2V 2 (Figure 4b). D( X) is defined to be convex, making the region defined by the constraint a closed region (hatched, Figure 4b). Although the original constraint could tolerate a higher activation error (and thus compression), using a bounding function provides favorable conditions for obtaining a closed-form solution for the activation error. 4 Framework for Activation Compression For our evaluation of AC-GC, we develop activation error bounds for various commonly used DNN layers and apply them to several recent networks. This section summarizes bounds for common layers, and derivations and additional layer bounds are provided in Appendix B. Calculating compression error from the gradient error bound e2V 2 requires expressing and solving for the compression/convergence trade-off. We tackle this by formulating the trade-off as a constrained optimization problem: maximizing the compression subject to bounded gradient error. The problem must be solved once per layer type and can be formally defined as X = argmax X B( X) s.t. D( X) = e2V 2 (7) where D( X) θf(θ, X) 2 (8) where B( X) is a continuous convex function that measures compression rate as a function of the activation error X RN C H W . A closed-form solution can be found for many systems of this type using the method of Lagrange multipliers. The constraint D( X) e2V 2 defines a convex region of potential activation errors where the convergence constraint is satisfied (hatched, Figure 4b). However, as both the B( X) and D( X) functions are convex, the maximum value must occur along the boundary, hence the equality constraint D( X) = e2V 2 in (7) ( x( ), Figure 4b). The convexity of D( X) also implies that X( ) is the maximum activation error, i.e., any error ( X)2 ( X( ))2 also satisfies the variance bound (6). Many compression methods use a variation of fixpoint. Hence, we select B to measure the number of bits removed from the activation when compressed with reduced precision fixpoint (9). Rounding mode and sign are constant factors that do not affect the result, however, we ignore clipping. The target compression method loosely influences the objective, hence, non-fixpoint methods may fare better with another error objective. n,c,h,w log | xnchw| (9) To derive AC-GC error bounds for DNN layers not presented in this work, one can: 1. Derive θf(θ, X) 2 for the layer type 2. Choose a suitable convex bounding function, D( X) θf(θ, X) 2 3. Obtain the maximum error X( ) by solving (7) using the method of Lagrange multipliers For the sake of brevity, we will summarize the notation, assumptions, and AC-GC error bounds for fully connected, convolution, and batch normalization layers (Table 1). We aim to locate closedform solutions with low computation overheads, although tighter bounding functions likely exist. Henceforth we omit arguments of f and use the following definitions: Batch size N, input channels C, output channels K, input and output activations X and Y , and compression error X. A) Fully Connected: Table 1A relates the error for guaranteed convergence with compression error for a fully connected layer, with weights θ = (θkc) RK C, input activation X = (xnc) RN C, and output activation gradient Y f = ( f/ ynk) RN K. As the error bound (e2V 2/2) decreases, the compression must decrease to compensate. All activations for a fully connected layer have the same error tolerance. B) Convolution: Convolution with no padding follows a similar trend to linear layers, with the addition of stride T, a filter size of R S, and increased dimensions X = (xnchw) RN C H W and Y f = ( f/ ynkhw) RN K H W . We assume an average usage of activations due to stride, as uneven usage leads to uneven compression, which would require tracking per-element compression rates. Comparing linear and convolution reveals that convolutions have a lower error tolerance due to the increased number of activations (HW) and weights (RS). To fully cover cases encountered in CNNs, we also derive error bounds for cases where activation errors affect multiple convolution layers (e.g., in Res Nets [20]). C) Batch Normalization: Batch normalization [24] re-normalizes the activation from per-channel standard deviation σ RC to a learned γ RC. There is a different dependence on activation error from convolution and linear layers. Instead of causing errors in the parameter gradient γf exclusively, activation error propagates to the activation gradient Xf and then to all subsequent layers in the network. To avoid bounding all weights in the network, we isolate the layer and bound the convergence of the batch normalization activations using Xf 2 e2V 2 (10) Arriving at a closed-form solution further requires a bound on the parameter gradient error within the layer using positive values (g2 c) ( ˆ γf)2 (Appendix B). Although not observed for the networks in this work, as network parameter gradients are not directly bounded, the convergence bounds on networks with batch normalization may be violated. Layer Normalization: This layer type is similar to batch normalization and requires a similar set of assumptions and derivations (Appendix B). Re LU, Dropout, Max Pooling, and Summation: Summation does not require storage of any activation, and the remaining layers (Dropout, Max Pooling, and Re LU) only require a bitmask to calculate their respective gradients. For instance, Re LU requires the storage of the bitmask X 0 [14, 25], and Max Pooling requires a bitmask of the locations of maximal values. As these layers have an efficient lossless high compression method available, we do not analyze them. Table 1: Guaranteed convergence equations for common network layers. See Appendices for full derivations and assumptions. Empty sums are over all indicies, i.e. P := PN,C,H,W n,c,h,w . M := NHW QUANTITY A) FULLY CONNECTED B) CONVOLUTION C) BATCH NORMALIZATION D( X) := Y f 2 X 2 RS T 2 Y f 2 X 2 X 2γ2 cg2 c M 2σ4c x2 nchw x( )2 nchw = e2V 2 2NC Y f 2 e2V 2T 2 2RSMC Y f 2 e2V 2Mσ4 c Cγ2cg2c 5 Practical Automatic Lossy Compression Using the AC-GC error bounds, we create convergence bounded compression methods with automatic compression rates, collectively referred to as Auto X. The first two methods (Auto Quant and Auto Cu SZ) adapt scaled fixpoint [14] and error-bounded compression [27], which have bounded errors for a given compression rate. The errors for these methods are unbiased provided that unbiased rounding to fixpoint is used [5, 27]. The third method (Auto JPEG) uses lossy JPEG compression [14]. We bound JPEG error using an empirical error-compression relationship using activations sampled from uncompressed training of CIFAR10/Res Net50 [20]. Samples are used offline with JPEG compression to establish the compression-error relationship. This empirical compression-error relationship is used to calculate the JPEG compression levels, which approximately satisfy the error bounds in Table 1. We chain quantization and JPEG with lossless Zero Value Compression [46] to compress sparse activations better, creating Auto Quant Z and Auto JPEGZ. As all Auto X used some form of fixpoint, we can express activation error in terms of bits. For any layer type, the relationship between bitwidth b and the convergence bound can be described as b log2 | xnchw| + ... = log2 |e| log2 θf + ... (11) Using the results from Table 1, it can be seen that the bitwidth scales additively with the batch size as + log2(N) for convolution layers, and log2(N) for normalization layers. Two issues with using AC-GC in a compression method are: 1) the various norms required are computationally expensive, and 2) the formulation assumes exact gradient information is available during the forward pass. We address both issues by statistically estimating activation error bounds. Instead of evaluating at every iteration, statistics are calculated at the end of every recalculation interval, specified in iterations. In the forward pass, a summary (mean or maximum) of the last ten recalculations is used when calculating the errors. This also allows approximating V 2 θf 2. Although some quantities are available in the forward pass, we estimate all of the activations, parameters, and gradients to avoid performing norm calculations at every iteration. Despite norms being estimated, we do not observe that the convergence bound (6) is violated for any network examined. A few training iterations can be used to verify correctness of the norm estimates (Figure 5c). 6 Evaluation We examine activation compression by modifying the Chainer framework [53] to compress and decompress activations during training. We measure compression rates every 100 iterations, and otherwise perform paired compression/decompression to maintain the highest performance for our experiments. We focus our analysis on CNNs with image and text datasets, as they have large activation memory requirements, but avoid the largest networks [22, 52] due to limited resources. We create a performance implementation based off Chen et al. [7] to measure throughput. For Image Net [11], CIFAR10 [2] and Div2K [1], we use SGD and 0.9 momentum for VGG16 [50], Res Nets (RN18 and RN50) [20], Wide Res Net (WRN) [59], and VDSR [29]. IMDB [39] and Text Copy [4] are trained using ADAM with CNN [53], RNN [53], and transformer heads [54]. All image datasets are augmented with random sizing, flip, and crop, as well as whitening and PCA for Image Net [30], and 8 8 cutout for CIFAR10 [12]. Learning rates, batch sizes, and epochs are 0.05, 128, 300 (CIFAR10, [49]), 0.1, 64, 105 (Image Net, [58]), 0.1, 32, 110 (Div2K, grid search), 2.0, 64, 100 (Text Copy, [4]), and 0.001, 64, 20 (IMDB, [53]). 4 2 0 2 Change in AQ bitwidth ( log|e|) Loss @ Epoch 10 Recalculation Interval (iterations) Loss @ Epoch 10 0 10 20 30 40 50 60 70 80 90100 Variance Ratio ˆ θf 2 / θf 2 Baseline Loss (L) (1 + e2)L AQ max AQ mean Figure 5: a-b) Average loss over the 10th epoch on MNIST/Le Net, where shaded regions indicate the min/max loss over five runs. a) Loss as function of e. Under compression, the empirical loss (AQ max: Auto Quant with a maximum summary) falls below the theoretical loss bound (1 + e2)L. b) Loss where statistics are calculated every recalculation interval iterations. The max or mean of the last ten intervals is used to calculate the bitwidth. c) Ratio of compressed to uncompressed gradients for the first convolution for Image Net/Res Net18, where the compressed gradient is used to update the parameters. Clusters indicate the first 50 iterations of every ten epochs. All forward and backward pass calculations use floating-point, using activations that have been compressed and decompressed between the two passes. Baseline refers to uncompressed training, i.e., 32-bit floating-point activations. Grid Quant Z uses the same implementation of Auto Quant Z, however, it uses grid search over eight bit-widths of 2, 3, 4, 6, 8, 10, 12 and 16 bits, and then chooses the lowest with accuracy within 0.1% of the baseline. These grid points were selected to give good coverage of low and medium bit-widths, and are approximately logarithmic-spaced. Successive Halving [26] and Hyperband [36] produce similar accuracy/compression to grid search but take less time. Unless otherwise stated, all experiments use e2 = 0.5, parameter estimates from the mean of a ten entry window, and a recalculation interval of 100 iterations. A value of e2 = 0.5 allows a small increase in loss over the baseline (+50%), but other values could be chosen depending on the desired error tolerance. Appendices C and D contain additional detail on hyperparameters and implementations. 6.1 Parameter Sensitivity We isolate the impacts of parameter estimation and e selection by training Le Net [33] on MNIST [32], and RN18 [20] on Image Net [11]. We train MNIST to convergence by using SGD with no momentum, a learning rate of 0.001, a batch size of 64, and 10 epochs. The effect of e, recalculation interval, and summary method are evaluated by training MNIST/Lenet with different configurations of Auto Quant (AQ) (Figures 5a and 5b). Decreasing e (which increases bitwidth) causes the loss to increase, however, the average loss does not violate the bound in Theorem 1 (Figure 5a). In general, all networks examined with e2 = 0.5 have loss changes below 2% and validation score changes below 0.5% (Sections 6.2 and 6.3), which demonstrates that AC-GC error bounds are not violated. Both interval and summary method have a minimal impact on the training loss for MNIST (Figure 5b). As there is an insignificant change in loss, we use the mean for AC-GC as it has higher compression. We evaluate Image Net/Res Net18 using a dual training approach to ensure the correctness of parameter estimation (Figure 5c). This involves training using Auto Quant, while evaluating the true weight gradients θf offline and comparing against their compressed counterpart ˆ θf. The ratio is 1 for the duration of training, and does not violate the bound for e2 = 0.5 (i.e., 1.5). The mean ratio is less than one, likely due to decreased activation variance from compression to a discrete set of values. We observe similar behavior for the other layers in the network (not shown). 6.2 CIFAR10, Div2K and IMDB We compare the Auto X methods with fixpoint grid search (Grid Quant Z), and with prior works on lossy JPEG compression [14] (Table 2). Grid search requires oracle knowledge of the baseline, and 8 the training iterations of any other method in Table 2. Compared to Grid Quant Z, Auto Quant Z uses a single run of training, and provides a similar compression rate of 7.5 . Auto Cu SZ has a Table 2: Test/validation score and compression rate (bracketed) for fixpoint with grid search (Grid Quant Z), Auto X methods, and JPEG-ACT (opt L5H from [14]), averaged over 3 runs (Image Net) or 5 runs (remainder). The highest accuracy and compression are bolded. Trained using 900 GPU-days (RTX 2080 Ti). N/A indicates either not run in [14], or lack of spatial activations. MODEL BASE AUTO QUANTZ AUTO CUSZ AUTO JPEGZ JPEGACT[14] GRID QUANTZ CIFAR10 % TOP-1 TEST ACCURACY VGG 93.6 93.5 ( 7.4 ) 93.5 ( 9.4 ) 92.9 (12.5 ) 92.4 (11.9 ) 93.5 ( 6.3 ) RN50 94.9 95.0 ( 4.2 ) 94.7 (15.5 ) 94.3 ( 9.2 ) 94.4 ( 7.5 ) 95.0 ( 5.7 ) WRN 95.8 95.9 ( 6.5 ) 95.8 (14.6 ) 95.3 (11.7 ) 94.2 (10.9 ) 96.0 ( 7.6 ) DIV2K BEST VAL. PSNR VDSR 36.1 36.1 ( 5.1 ) 35.8 (25.2 ) 36.1 ( 7.9 ) 35.4 ( 9.1 ) 36.0 ( 6.7 ) IMDB % BEST VAL. ACCURACY CNN 61.4 61.6 (12.2 ) 61.8 (19.3 ) 61.4 (11.2 ) N/A 61.7 (16.5 ) LSTM 60.3 60.1 (10.0 ) 60.9 ( 8.8 ) N/A N/A 60.4 (14.7 ) TEXT COPY % BEST TEST ACCURACY TRANS 98.8 98.6 ( 7.1 ) 98.3 (12.7 ) N/A N/A 98.9 ( 5.1 ) IMAGENET % TOP-1 CENTER CROP VAL. ACCURACY RN18 68.6 68.5 ( 4.2 ) 68.1 ( 6.8 ) 68.1 ( 8.1 ) 67.3 ( 7.2 ) 68.5 ( 2.9 ) RN50 72.3 72.7 ( 4.8 ) 72.5 (10.1 ) 71.5 ( 8.5 ) 71.6 ( 5.9 ) 72.5 ( 4.9 ) AVERAGE %-POINT CHANGE AND COMPRESSION RATIO ALL 0 +0.0 ( 7.5 ) 0.1 (15.1 ) 0.6 (10.5 ) 1.0 ( 7.4 ) +0.1 ( 7.8 ) compression 2.0 higher than JPEG-ACT, while maintaining accuracy to within 0.1 of the baseline on average. With a suitable error bound, Cu SZ can extract significant compression from zeros and spatial information. On non-spatial data and non-image datasets (Text Copy and IMDB, Table 2) we observe that Auto Cu SZ extracts similarly high compression with little accuracy change. Finally, Auto JPEGZ vs. JPEG-ACT demonstrates that using AC-GC error bounds gives higher accuracy and compression than using heuristics to select JPEG hyperparameters. In general, we find that using e2 > 0.5 decreases accuracy, leaving little reason to modify it. 6.3 Image Net On Image Net training with Res Nets (Table 2), the Auto X methods obtain a high accuracy. Our Image Net accuracies are lower than other works as we do not use random scaling (which improves performance), and we report 1-crop accuracy. Our 10-crop accuracy for the Res Net50 baseline is 75.2%. Reduced compression rates on Res Net18 vs. Res Net50 are due to a lower sparsity in Re LU activations, which we hypothesize is caused by the different bottleneck structures of the two networks [20]. The higher-than-baseline accuracy of Auto Quant Z (Table 2) is caused by a large standard deviation for the Image Net baseline ( 0.21). On Image Net, Auto Cu SZ gives a high compression in exchange for a small decrease in accuracy, 0.15%-points. Auto Quant Z provides high accuracy, at a moderate compression rate of 0.7 , with a 1.2%-point better accuracy than JPEG-ACT. However, the primary advantage of Auto X methods is that the pre-training bound on loss increases. 6.4 Overheads The Auto X methods require error bound calculation (common to all methods) and compression. Our unoptimized Auto Quant Z implementation achieves throughput of 1.66 (N = 128) vs. naive swapping to the CPU (N = 128), and 0.64 vs. uncompressed training (N = 32) (Image Net/ Res Net50). Our primary contribution, AC-GC error bound calculation, uses 0.4% of total training time. This is negligible when compared to compression overhead, e.g., 4% for fixpoint [14, 25], 17% for Cu SZ [27], 33% for Act NN-L3 [7], or 13% for hardware accelerated JPEG [14]. Unoptimized Auto Quant Z is 23% slower compared to Act NN-L3 [7]. Compression rate search time is decreased by 4.6 when compared to Successive Halving (Figure 1). Table 3: Comparison with prior works on CIFAR10 (C10) and Image Net (IN). indicates standard deviation, if available. Accuracy is presented relative to the baseline accuracy of each work. Does not include 2 memory reduction from recalculating activations, which is orthogonal to this work. AUTO CUSZ WAGE [57] BAA [6] JPEGACT[14] ULP [51] EBC [27] ACTNN -L3[7] DATASET IN C10 IN IN IN IN IN MODEL RN50 VGG16 RN152 RN50 RN50 RN50 RN152 METHOD AUTO 8-BIT 4-BIT JPEG 4-BIT CUSZ 2-BIT -CUSZ +RECALC. +ZVC MIX. PREC. ACC. (%) 0.2 0.2 0.3 0.5 0.1 0.3 0.9 0.2 COMPR. 10.1 4 8 5.9 8 11.0 12 7 Related Works We compare Auto Cu SZ with the most recent works in activation compression [6, 14, 25, 27] and reduced precision training [51, 57] (Table 3). Auto Cu SZ obtains better accuracy and compression than most works, with the added benefit of not requiring a search over compression rates. BAA [6] presents a technique of recalculating activations similar to Chen et al. [8], which could be combined with any method in Table 3, including the Auto X methods. Auto Cu SZ obtains higher compression and accuracy than EBC [27], demonstrating that AC-GC error bounds are better than hand-tuning in this case. WAGE [57] and ULP [51] and reduced precision training with theoretical convergence bounds [35, 37] reduce training compute through reduced precision, however, have lower accuracy due to accumulating errors in both the forward and backward pass. Auto X methods can potentially be combined with low precision training and other techniques to reduce memory requirements further. Act NN is most similar to this work in that it invokes gradient variance to partition bits among compressed activations [7]. However, Act NN approaches the problem in an opposite manner to this work: it finds the highest accuracy for a given compression rate. The Act NN approach works well when the goal is to fit within a memory budget. Our approach allows for selecting the target loss increase a priori, and gaining information on the uncompressed accuracy from a single training run. AC-GC will also converge similarly to the uncompressed case on an unknown model, providing assurances that loss changes are due to model/hyperparameter configuration, not compression. Additionally, Act NN is specific to group-wise fixpoint compression, whereas AC-GC is quick to adapt to any lossy compression method (including group-wise fixpoint). On Image Net/Res Nets, Auto Cu SZ and Auto Quant Z obtain half the accuracy change of Act NN-L3, albeit at 0.9 and 0.4 the compression, respectively. Works targeting inference, such as precision reduction [10, 15, 23, 34], compression [13, 19, 38, 56], and sparsification [16, 43] increase memory requirements due to tracking additional state. Other works that directly address memory can be grouped into scheduling [8, 25], offloading [31, 45], and restructuring [17, 21]. Generally, these come with a performance or model-flexibility penalty. Any non-lossy method is partially orthogonal to compression and can be potentially combined with AC-GC for increased memory reduction. 8 Conclusions The AC-GC automatic compression methods described in this work provide high compression rates with error trade-offs known before training. Avoiding compression rate search comes at a computational overhead of only 0.4%, many times less than tuning techniques. Provided that the assumptions of AC-GC hold, it can be further combined with any lossy compression method, layer, or network to guarantee convergence with high activation compression. Although convergence is theoretically guaranteed, some factors could impact convergence, such as floating-point errors. By lowering training costs, this framework allows for training larger models and faster exploration of the machine learning field s landscape. Detailed derivations and implementations can be found in the supplemental material. Code is available https://github.com/rdevans0/acgc. Funding Transparency Statement This research was funded in part by the Computing Hardware for Emerging Intelligent Sensory Applications (COHESA) project financed under the National Sciences and Engineering Research Council of Canada (NSERC); grant number NETGP485577-15. [1] Agustsson, E. and Timofte, R. NTIRE 2017 Challenge on Single Image Super-Resolution: Dataset and Study. In 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 1122 1131, July 2017. doi: 10.1109/CVPRW.2017.150. [2] Alex Krizhevsky and Geoffrey Hinton. Learning Multiple Layers of Features from Tiny Images. Ph D thesis, University of Toronto, Toronto, Canada, 2009. [3] Bach, F. and Moulines, E. Non-strongly-convex smooth stochastic approximation with convergence rate O(1/n). Advances in Neural Information Processing Systems, 26:773 781, 2013. [4] Bartz, C. chainer-transformer. https://github.com/Bartzi/chainer-transformer/blob/master/LIC ENSE, January 2021. [5] Burks, A. W., Goldstine, H. H., and von Neumann, J. Preliminary Discussion of the Logical Design of an Electronic Computing Instrument. In Randell, B. (ed.), The Origins of Digital Computers: Selected Papers, Texts and Monographs in Computer Science, pp. 399 413. Springer, Berlin, Heidelberg, 1982. ISBN 978-3-642-61812-3. doi: 10.1007/978-3-642-61812-3_32. [6] Chakrabarti, A. and Moseley, B. Backprop with Approximate Activations for Memory-efficient Network Training. Advances in Neural Information Processing Systems, 32, 2019. [7] Chen, J., Zheng, L., Yao, Z., Wang, D., Stoica, I., Mahoney, M. W., and Gonzalez, J. E. Actnn: Reducing training memory footprint via 2-bit activation compressed training. ar Xiv:2104.14129 [cs.LG], April 2021. [8] Chen, T., Xu, B., Zhang, C., and Guestrin, C. Training Deep Nets with Sublinear Memory Cost. ar Xiv:1604.06174 [cs], April 2016. [9] Courbariaux, M., Bengio, Y., and David, J.-P. Training deep neural networks with low precision multiplications. ar Xiv:1412.7024 [cs], December 2014. [10] Courbariaux, M., Bengio, Y., and David, J.-P. Binary Connect: Training Deep Neural Networks with binary weights during propagations. In Cortes, C., Lawrence, N. D., Lee, D. D., Sugiyama, M., and Garnett, R. (eds.), Advances in Neural Information Processing Systems 28, pp. 3123 3131. Curran Associates, Inc., 2015. [11] Deng, J., Dong, W., Socher, R., Li, L., Kai Li, and Li Fei-Fei. Image Net: A large-scale hierarchical image database. In 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248 255, June 2009. doi: 10.1109/CVPR.2009.5206848. [12] De Vries, T. and Taylor, G. W. Improved Regularization of Convolutional Neural Networks with Cutout. ar Xiv:1708.04552 [cs], November 2017. [13] Ding, C., Liao, S., Wang, Y., Li, Z., Liu, N., Zhuo, Y., Wang, C., Qian, X., Bai, Y., Yuan, G., Ma, X., Zhang, Y., Tang, J., Qiu, Q., Lin, X., and Yuan, B. Cir CNN: Accelerating and Compressing Deep Neural Networks Using Block-circulant Weight Matrices. In Proceedings of the 50th Annual IEEE/ACM International Symposium on Microarchitecture, MICRO-50 17, pp. 395 408, New York, NY, USA, 2017. ACM. ISBN 978-1-4503-4952-9. doi: 10.1145/3123939. 3124552. [14] Evans, R. D., Liu, L., and Aamodt, T. M. JPEG-ACT: Accelerating Deep Learning via Transform-based Lossy Compression. In 2020 ACM/IEEE 47th Annual International Symposium on Computer Architecture (ISCA), pp. 860 873, May 2020. doi: 10.1109/ISCA45697.2020. 00075. [15] Fang, J., Shafiee, A., Abdel-Aziz, H., Thorsley, D., Georgiadis, G., and Hassoun, J. H. Posttraining Piecewise Linear Quantization for Deep Neural Networks. In Vedaldi, A., Bischof, H., Brox, T., and Frahm, J.-M. (eds.), Computer Vision ECCV 2020, Lecture Notes in Computer Science, pp. 69 86, Cham, 2020. Springer International Publishing. ISBN 978-3-030-58536-5. doi: 10.1007/978-3-030-58536-5_5. [16] Frankle, J. and Carbin, M. The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks. ar Xiv:1803.03635 [cs], March 2019. [17] Gomez, A. N., Ren, M., Urtasun, R., and Grosse, R. B. The Reversible Residual Network: Backpropagation Without Storing Activations. In Guyon, I., Luxburg, U. V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., and Garnett, R. (eds.), Advances in Neural Information Processing Systems 30, pp. 2214 2224. Curran Associates, Inc., 2017. [18] Goyal, P., Dollár, P., Girshick, R., Noordhuis, P., Wesolowski, L., Kyrola, A., Tulloch, A., Jia, Y., and He, K. Accurate, Large Minibatch SGD: Training Image Net in 1 Hour. ar Xiv:1706.02677 [cs], June 2017. [19] Han, S., Mao, H., and Dally, W. J. Deep Compression: Compressing Deep Neural Networks with Pruning, Trained Quantization and Huffman Coding. ar Xiv:1510.00149 [cs], October 2015. [20] He, K., Zhang, X., Ren, S., and Sun, J. Deep Residual Learning for Image Recognition. ar Xiv:1512.03385 [cs], December 2015. [21] Howard, A. G., Zhu, M., Chen, B., Kalenichenko, D., Wang, W., Weyand, T., Andreetto, M., and Adam, H. Mobile Nets: Efficient Convolutional Neural Networks for Mobile Vision Applications. ar Xiv:1704.04861 [cs], April 2017. [22] Huang, Y., Cheng, Y., Bapna, A., Firat, O., Chen, D., Chen, M., Lee, H., Ngiam, J., Le, Q. V., Wu, Y., and Chen, z. GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism. In Wallach, H., Larochelle, H., Beygelzimer, A., d\textquotesingle Alché-Buc, F., Fox, E., and Garnett, R. (eds.), Advances in Neural Information Processing Systems 32, pp. 103 112. Curran Associates, Inc., 2019. [23] Hubara, I., Courbariaux, M., Soudry, D., El-Yaniv, R., and Bengio, Y. Binarized Neural Networks. In Lee, D. D., Sugiyama, M., Luxburg, U. V., Guyon, I., and Garnett, R. (eds.), Advances in Neural Information Processing Systems 29, pp. 4107 4115. Curran Associates, Inc., 2016. [24] Ioffe, S. and Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. In International Conference on Machine Learning, pp. 448 456. PMLR, June 2015. [25] Jain, A., Phanishayee, A., Mars, J., Tang, L., and Pekhimenko, G. Gist: Efficient Data Encoding for Deep Neural Network Training. In 2018 ACM/IEEE 45th Annual International Symposium on Computer Architecture (ISCA), pp. 776 789, June 2018. doi: 10.1109/ISCA.2018.00070. [26] Jamieson, K. and Talwalkar, A. Non-stochastic Best Arm Identification and Hyperparameter Optimization. ar Xiv:1502.07943 [cs, stat], February 2015. [27] Jin, S., Li, G., Song, S. L., and Tao, D. A Novel Memory-Efficient Deep Learning Training Framework via Error-Bounded Lossy Compression. ar Xiv:2011.09017 [cs], November 2020. [28] Karimi, H., Nutini, J., and Schmidt, M. Linear Convergence of Gradient and Proximal-Gradient Methods Under the Polyak-Łojasiewicz Condition. In Frasconi, P., Landwehr, N., Manco, G., and Vreeken, J. (eds.), Machine Learning and Knowledge Discovery in Databases, Lecture Notes in Computer Science, pp. 795 811, Cham, 2016. Springer International Publishing. ISBN 978-3-319-46128-1. doi: 10.1007/978-3-319-46128-1_50. [29] Kim, J., Lee, J. K., and Lee, K. M. Accurate Image Super-Resolution Using Very Deep Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1646 1654, 2016. [30] Krizhevsky, A., Sutskever, I., and Hinton, G. E. Image Net classification with deep convolutional neural networks. In Proceedings of the 25th International Conference on Neural Information Processing Systems - Volume 1, NIPS 12, pp. 1097 1105, Red Hook, NY, USA, December 2012. Curran Associates Inc. [31] Kwon, Y. and Rhu, M. Beyond the memory wall: A case for memory-centric HPC system for deep learning. In Proceedings of the 51st Annual IEEE/ACM International Symposium on Microarchitecture, MICRO-51, pp. 148 161, Fukuoka, Japan, October 2018. IEEE Press. ISBN 978-1-5386-6240-3. doi: 10.1109/MICRO.2018.00021. [32] Le Cun, Y. and Cortes, C. MNIST handwritten digit database. http://yann.lecun.com/exdb/mnist/, 2010. [33] Lecun, Y., Bottou, L., Bengio, Y., and Haffner, P. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278 2324, November 1998. ISSN 1558-2256. doi: 10.1109/5.726791. [34] Li, F., Zhang, B., and Liu, B. Ternary Weight Networks. ar Xiv:1605.04711 [cs], May 2016. [35] Li, H., De, S., Xu, Z., Studer, C., Samet, H., and Goldstein, T. Training quantized nets: A deeper understanding. In Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS 17, pp. 5813 5823, Red Hook, NY, USA, December 2017. Curran Associates Inc. ISBN 978-1-5108-6096-4. [36] Li, L., Jamieson, K., De Salvo, G., Rostamizadeh, A., and Talwalkar, A. Hyperband: A novel bandit-based approach to hyperparameter optimization. The Journal of Machine Learning Research, 18(1):6765 6816, January 2017. ISSN 1532-4435. [37] Li, Z. and De Sa, C. M. Dimension-Free Bounds for Low-Precision Training. In Advances in Neural Information Processing Systems, volume 32. Curran Associates, Inc., 2019. [38] Lin, Y., Han, S., Mao, H., Wang, Y., and Dally, W. J. Deep Gradient Compression: Reducing the Communication Bandwidth for Distributed Training. ar Xiv:1712.01887 [cs, stat], December 2017. [39] Maas, A. L., Daly, R. E., Pham, P. T., Huang, D., Ng, A. Y., and Potts, C. Learning word vectors for sentiment analysis. In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies - Volume 1, HLT 11, pp. 142 150, USA, June 2011. Association for Computational Linguistics. ISBN 978-1-932432-87-9. [40] Milne, T. Piecewise Strong Convexity of Neural Networks. Advances in Neural Information Processing Systems, 32:12973 12983, 2019. [41] Nedi c, A. and Bertsekas, D. Convergence Rate of Incremental Subgradient Algorithms. In Uryasev, S. and Pardalos, P. M. (eds.), Stochastic Optimization: Algorithms and Applications, Applied Optimization, pp. 223 264. Springer US, Boston, MA, 2001. ISBN 978-1-4757-6594-6. doi: 10.1007/978-1-4757-6594-6_11. [42] Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z., Gimelshein, N., Antiga, L., Desmaison, A., Kopf, A., Yang, E., De Vito, Z., Raison, M., Tejani, A., Chilamkurthy, S., Steiner, B., Fang, L., Bai, J., and Chintala, S. Py Torch: An Imperative Style, High-Performance Deep Learning Library. Advances in Neural Information Processing Systems, 32:8026 8037, 2019. [43] Raihan, M. A. and Aamodt, T. M. Sparse Weight Activation Training. ar Xiv:2001.01969 [cs, stat], October 2020. [44] Real, E., Aggarwal, A., Huang, Y., and Le, Q. V. Regularized Evolution for Image Classifier Architecture Search. Proceedings of the AAAI Conference on Artificial Intelligence, 33(01): 4780 4789, July 2019. ISSN 2374-3468. doi: 10.1609/aaai.v33i01.33014780. [45] Rhu, M., Gimelshein, N., Clemons, J., Zulfiqar, A., and Keckler, S. W. v DNN: Virtualized Deep Neural Networks for Scalable, Memory-Efficient Neural Network Design. ar Xiv:1602.08124 [cs], July 2016. [46] Rhu, M., O Connor, M., Chatterjee, N., Pool, J., Kwon, Y., and Keckler, S. W. Compressing DMA Engine: Leveraging Activation Sparsity for Training Deep Neural Networks. In 2018 IEEE International Symposium on High Performance Computer Architecture (HPCA), pp. 78 91, February 2018. doi: 10.1109/HPCA.2018.00017. [47] Rogers, T. G., Johnson, D. R., O Connor, M., and Keckler, S. W. A variable warp size architecture. In 2015 ACM/IEEE 42nd Annual International Symposium on Computer Architecture (ISCA), pp. 489 501, June 2015. doi: 10.1145/2749469.2750410. [48] Rumelhart, D. E., Hinton, G. E., and Williams, R. J. Learning representations by backpropagating errors. Nature, 323(6088):533 536, October 1986. ISSN 1476-4687. doi: 10.1038/ 323533a0. [49] Saito, S. Chainer-cifar10: Various CNN models including for CIFAR10 with Chainer. https://github.com/mitmul /chainer-cifar10, May 2018. [50] Simonyan, K. and Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. ar Xiv:1409.1556 [cs], September 2014. [51] Sun, X., Wang, N., Chen, C.-Y., Ni, J., Agrawal, A., Cui, X., Venkataramani, S., El Maghraoui, K., Srinivasan, V. V., and Gopalakrishnan, K. Ultra-low precision 4-bit training of deep neural networks. In Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M. F., and Lin, H. (eds.), Advances in Neural Information Processing Systems, volume 33, pp. 1796 1807. Curran Associates, Inc., 2020. [52] Tan, M. and Le, Q. Efficient Net: Rethinking Model Scaling for Convolutional Neural Networks. In International Conference on Machine Learning, pp. 6105 6114. PMLR, May 2019. [53] Tokui, S., Okuta, R., Akiba, T., Niitani, Y., Ogawa, T., Saito, S., Suzuki, S., Uenishi, K., Vogel, B., and Yamazaki Vincent, H. Chainer: A Deep Learning Framework for Accelerating the Research Cycle. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, KDD 19, pp. 2002 2011, New York, NY, USA, July 2019. Association for Computing Machinery. ISBN 978-1-4503-6201-6. doi: 10.1145/3292500. 3330756. [54] Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, Ł., and Polosukhin, I. Attention is All you Need. In Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc., 2017. [55] Virmaux, A. and Scaman, K. Lipschitz regularity of deep neural networks: Analysis and efficient estimation. Advances in Neural Information Processing Systems, 31:3835 3844, 2018. [56] Wang, Y., Xu, C., You, S., Tao, D., and Xu, C. CNNpack: Packing Convolutional Neural Networks in the Frequency Domain. In Lee, D. D., Sugiyama, M., Luxburg, U. V., Guyon, I., and Garnett, R. (eds.), Advances in Neural Information Processing Systems 29, pp. 253 261. Curran Associates, Inc., 2016. [57] Wu, S., Li, G., Chen, F., and Shi, L. Training and Inference with Integers in Deep Neural Networks. In International Conference on Learning Representations, February 2018. [58] Wu, W. Reproduce Res Net-v2 (Identity Mappings in Deep Residual Networks) with MXNet. https://github.com/ tornadomeet/Res Net, January 2021. [59] Zagoruyko, S. and Komodakis, N. Wide Residual Networks. ar Xiv:1605.07146 [cs], May 2016. 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] See Section 6 (b) Did you describe the limitations of your work? [Yes] See Section 6 (c) Did you discuss any potential negative societal impacts of your work? [No] This work has the positive impact of decreasing training times, which decreases overall carbon footprint and allows for research of models which benefit society. However, depending on how these models are used there is a possibility of negative impacts such as surveillance, privacy concerns, or job loss. (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [Yes] See Section 3, Appendix A, and the assumptions from Karimi et al. [28], Section 3.3 (b) Did you include complete proofs of all theoretical results? [Yes] See Appendix A 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] See Section 6, Appendices C and D, and the code available at https://github.com/rdevans0/ acgc (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] See Section 6 and Appendix D (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [Yes] See Appendix E due to space limitations (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] See Table 2. This work used an internal cluster. 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [Yes] We cite the original works for all datasets, models and hyperparameters in Section 6 (b) Did you mention the license of the assets? [No] All are public use (c) Did you include any new assets either in the supplemental material or as a URL? [No] (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [No] This is discussed in the original publications on these datasets (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [No] This is discussed in the original publications on these datasets 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] This work does not involve human subjects (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]