# on_distributed_adaptive_optimization_with_gradient_compression__7a761987.pdf Published as a conference paper at ICLR 2022 ON DISTRIBUTED ADAPTIVE OPTIMIZATION WITH GRADIENT COMPRESSION Xiaoyun Li, Belhal Karimi, Ping Li Cognitive Computing Lab Baidu Research 10900 NE 8th St. Bellevue, WA 98004, USA {xiaoyunli,belhalkarimi,liping11}@baidu.com We study COMP-AMS, a distributed optimization framework based on gradient averaging and adaptive AMSGrad algorithm. Gradient compression with error feedback is applied to reduce the communication cost in the gradient transmission process. Our convergence analysis of COMP-AMS shows that such compressed gradient averaging strategy yields same convergence rate as standard AMSGrad, and also exhibits the linear speedup effect w.r.t. the number of local workers. Compared with recently proposed protocols on distributed adaptive methods, COMP-AMS is simple and convenient. Numerical experiments are conducted to justify the theoretical findings, and demonstrate that the proposed method can achieve same test accuracy as the full-gradient AMSGrad with substantial communication savings. With its simplicity and efficiency, COMP-AMS can serve as a useful distributed training framework for adaptive gradient methods. 1 INTRODUCTION Deep neural network has achieved the state-of-the-art learning performance on numerous AI applications, e.g., computer vision and natural language processing (Graves et al., 2013; Goodfellow et al., 2014; He et al., 2016; Young et al., 2018; Zhang et al., 2018), reinforcement learning (Mnih et al., 2013; Levine et al., 2016; Silver et al., 2017), recommendation systems (Covington et al., 2016), computational advertising (Zhao et al., 2019; Xu et al., 2021; Zhao et al., 2022), etc. With the increasing size of data and growing complexity of deep neural networks, standard single-machine training procedures encounter at least two major challenges: Due to the limited computing power of a single-machine, processing the massive number of data samples takes a long time training is too slow. Many real-world applications cannot afford spending days or even weeks on training. In many scenarios, data are stored on multiple servers, possibly at different locations, due to the storage constraints (massive user behavior data, Internet images, etc.) or privacy reasons (Chang et al., 2018). Hence, transmitting data among servers might be costly. Distributed learning framework has been commonly used to tackle the above two issues. Consider the distributed optimization task where n workers jointly solve the following optimization problem min θ f(θ) := min θ 1 n i=1 fi(θ) = 1 i=1 Ex Xi[Fi(θ; x)], (1) where the non-convex function fi represents the average loss over the local data samples for worker i [n], and θ Rd the global model parameter. Xi is the data distribution on each local node. In the classical centralized distributed setting, in each iteration the central server uniformly randomly assigns the data to n local workers (Xi s are the same), at which the gradients of the model are computed in parallel. Then the central server aggregates the local gradients, updates the global model (e.g., by stochastic gradient descent (SGD)), and transmits back the updated model to the Published as a conference paper at ICLR 2022 local nodes for subsequent gradient computation. The scenario where Xi s are different gives rise to the recently proposed Federated Learning (FL) (Mc Mahan et al., 2017) framework, which will not be the major focus of this work. As we can see, distributed training naturally solves aforementioned issues: 1) We use n computing nodes to train the model, so the time per training epoch can be largely reduced; 2) There is no need to transmit the local data to central server. Besides, distributed training also provides stronger error tolerance since the training process could continue even one local machine breaks down. As a result of these advantages, there has been a surge of study and applications on distributed systems (Nedic & Ozdaglar, 2009; Boyd et al., 2011; Duchi et al., 2012; Goyal et al., 2017; Hong et al., 2017; Koloskova et al., 2019; Lu et al., 2019). Gradient compression. Among many optimization strategies, SGD is still the most popular prototype in distributed training for its simplicity and effectiveness (Chilimbi et al., 2014; Agarwal et al., 2018; Mikami et al., 2018). Yet, when the deep learning model is very large, the communication between local nodes and central server could be expensive, and the burdensome gradient transmission would slow down the whole training system. Thus, reducing the communication cost in distributed SGD has become an active topic, and an important ingredient of large-scale distributed systems (e.g., Seide et al. (2014)). Solutions based on quantization, sparsification and other compression techniques of the local gradients have been proposed, e.g., Aji & Heafield (2017); Alistarh et al. (2017); Sa et al. (2017); Wen et al. (2017); Bernstein et al. (2018); Stich et al. (2018); Wangni et al. (2018); Ivkin et al. (2019); Yang et al. (2019); Haddadpour et al. (2020). However, it has been observed both theoretically and empirically (Stich et al., 2018; Ajalloeian & Stich, 2020), that directly updating with the compressed gradients usually brings non-negligible performance downgrade in terms of convergence speed and accuracy. To tackle this problem, studies (e.g., Stich et al. (2018); Karimireddy et al. (2019)) show that the technique of error feedback can to a large extent remedy the issue of such gradient compression, achieving the same convergence rate as full-gradient SGD. Adaptive optimization. In recent years, adaptive optimization algorithms (e.g., Ada Grad (Duchi et al., 2010), Adam (Kingma & Ba, 2015) and AMSGrad (Reddi et al., 2018)) have become popular because of their superior empirical performance. These methods use different implicit learning rates for different coordinates that keep changing adaptively throughout the training process, based on the learning trajectory. In many cases, adaptive methods have been shown to converge faster than SGD, sometimes with better generalization as well. Nevertheless, the body of literature that extends adaptive methods to distributed training is still fairly limited. In particular, even the simple gradient averaging approach, though appearing standard, has not been analyzed for adaptive optimization algorithms. Given that distributed SGD with compressed gradient averaging can match the performance of standard SGD, one natural question is: is it also true for adaptive methods? In this work, we fill this gap formally, by analyzing COMP-AMS, a distributed adaptive optimization framework using the gradient averaging protocol, with communication-efficient gradient compression. Our method has been implemented in the Paddle Paddle platform (www.paddlepaddle.org.cn). Our contributions. We study a simple algorithm design leveraging the adaptivity of AMSGrad and the computational virtue of local gradient compression: We propose COMP-AMS, a synchronous distributed adaptive optimization framework based on global averaging with gradient compression, which is efficient in both communication and memory as no local moment estimation is needed. We consider the Block Sign and Top-k compressors, coupled with the error-feedback technique to compensate for the bias implied by the compression step for fast convergence. We provide the convergence analysis of distributed COMP-AMS (with n workers) in smooth non-convex optimization. In the special case of n = 1 (single machine), similar to SGD, gradient compression with error feedback in adaptive method achieves the same convergence rate O( 1 T ) as the standard full-gradient counterpart. Also, we show that with a properly chosen learning rate, COMP-AMS achieves O( 1 n T ) convergence, implying a linear speedup in terms of the number of local workers to attain a stationary point. Experiments are conducted on various training tasks on image classification and sentiment analysis to validate our theoretical findings on the linear speedup effect. Our results show that COMP-AMS has comparable performance with other distributed adaptive methods, and approaches the accuracy of full-precision AMSGrad with a substantially reduced communication cost. Thus, it can serve as a convenient distributed training strategy in practice. Published as a conference paper at ICLR 2022 2 RELATED WORK 2.1 DISTRIBUTED SGD WITH COMPRESSED GRADIENTS Quantization. To reduce the expensive communication in large-scale distributed SGD training systems, extensive works have considered various compression techniques applied to the gradient transaction procedure. The first strategy is quantization. Dettmers (2016) condenses 32-bit floating numbers into 8-bits when representing the gradients. Seide et al. (2014); Bernstein et al. (2018; 2019); Karimireddy et al. (2019) use the extreme 1-bit information (sign) of the gradients, combined with tricks like momentum, majority vote and memory. Other quantization-based methods include QSGD (Alistarh et al., 2017; Zhang et al., 2017; Wu et al., 2018) and LPC-SVRG (Yu et al., 2019b), leveraging unbiased stochastic quantization. Quantization has been successfully applied to industrial-level applications, e.g., Xu et al. (2021). The saving in communication of quantization methods is moderate: for example, 8-bit quantization reduces the cost to 25% (compared with 32-bit full-precision). Even in the extreme 1-bit case, the largest compression ratio is around 1/32 3.1%. Sparsification. Gradient sparsification is another popular solution which may provide higher compression rate. Instead of commuting the full gradient, each local worker only passes a few coordinates to the central server and zeros out the others. Thus, we can more freely choose higher compression ratio (e.g., 1%, 0.1%), still achieving impressive performance in many applications (Lin et al., 2018). Stochastic sparsification methods, including uniform and magnitude based sampling (Wangni et al., 2018), select coordinates based on some sampling probability, yielding unbiased gradient compressors with proper scaling. Deterministic methods are simpler, e.g., Random-k, Top-k (Stich et al., 2018; Shi et al., 2019) (selecting k elements with largest magnitude), Deep Gradient Compression (Lin et al., 2018), but usually lead to biased gradient estimation. More applications and analysis of compressed distributed SGD can be found in Alistarh et al. (2018); Jiang & Agrawal (2018); Jiang et al. (2018); Shen et al. (2018); Basu et al. (2019), among others. Error Feedback (EF). Biased gradient estimation, which is a consequence of many aforementioned methods (e.g., sign SGD, Top-k), undermines the model training, both theoretically and empirically, with slower convergence and worse generalization (Ajalloeian & Stich, 2020; Beznosikov et al., 2020). The technique of error feedback is able to correct for the bias and fix the convergence issues. In this procedure, the difference between the true stochastic gradient and the compressed one is accumulated locally, which is then added back to the local gradients in later iterations. Stich et al. (2018); Karimireddy et al. (2019) prove the O( 1 T ) and O( 1 T ) convergence rate of EF-SGD in strongly convex and non-convex setting respectively, matching the rates of vanilla SGD (Nemirovski et al., 2009; Ghadimi & Lan, 2013). More recent works on the convergence rate of SGD with error feedback include Stich & Karimireddy (2019); Zheng et al. (2019); Richtárik et al. (2021), etc. 2.2 ADAPTIVE OPTIMIZATION Algorithm 1 AMSGRAD (Reddi et al., 2018) 1: Input: parameters β1, β2, ϵ, learning rate ηt 2: Initialize: θ1 Rd, m0 = v0 = 0 Rd 3: for t = 1, . . . , T do 4: Compute stochastic gradient gt at θt 5: mt = β1mt 1 + (1 β1)gt 6: vt = β2vt 1 + (1 β2)g2 t 7: ˆvt = max(ˆvt 1, vt) 8: θt+1 = θt ηt mt ˆvt+ϵ 9: end for In each SGD update, all the coordinates share the same learning rate, which is either constant or decreasing through the iterations. Adaptive optimization methods cast different learning rates on each dimension. For instance, Ada Grad, developed in Duchi et al. (2010), divides the gradient elementwise by q PT t=1 g2 t Rd, where gt Rd is the gradient vector at time t and d is the model dimensionality. Thus, it intrinsically assigns different learning rates to different coordinates throughout the training elements with smaller previous gradient magnitudes tend to move a larger step via larger learning rate. Other adaptive methods include Ada Delta (Zeiler, 2012) and Adam (Kingma & Ba, 2015), which introduce momentum and moving average of second moment estimation into Ada Grad hence leading to better performances. AMSGrad (Reddi et al., 2018) (Algorithm 1, which is the prototype in our paper), fixes the potential convergence issue of Adam. Wang et al. (2021) and Zhou et al. Published as a conference paper at ICLR 2022 (2020) improve the convergence and generalization of AMSGrad through optimistic acceleration and differential privacy. Adaptive optimization methods have been widely used in training deep learning models in language, computer vision and advertising applications, e.g., Choi et al. (2019); You et al. (2020); Zhang et al. (2021); Zhao et al. (2022). In distributed setting, Nazari et al. (2019); Chen et al. (2021b) study decentralized adaptive methods, but communication efficiency was not considered. Mostly relevant to our work, Chen et al. (2021a) proposes a distributed training algorithm based on Adam, which requires every local node to store a local estimation of the moments of the gradient. Thus, one has to keep extra two more tensors of the model size on each local worker, which may be less feasible in terms of memory particularly with large models. More recently, Tang et al. (2021) proposes an Adam pre-conditioned momentum SGD method. Chen et al. (2020); Karimireddy et al. (2020); Reddi et al. (2021) proposed local/global adaptive FL methods, which can be further accelerated via layer-wise adaptivity (Karimi et al., 2021). 3 COMMUNICATION-EFFICIENT ADAPTIVE OPTIMIZATION 3.1 GRADIENT COMPRESSORS In this paper, we mainly consider deterministic q-deviate compressors defined as below. Assumption 1. The gradient compressor C : Rd 7 Rd is q-deviate: for x Rd, 0 q < 1 such that C(x) x q x . Larger q indicates heavier compression, while smaller q implies better approximation of the true gradient. q = 0 implies C(x) = x, i.e., no compression. In the following, we give two popular and efficient q-deviate compressors that will be adopted in this paper. Definition 1 (Top-k). For x Rd, denote S as the size-k set of i [d] with largest k magnitude |xi|. The Top-k compressor is defined as C(x)i = xi, if i S; C(x)i = 0 otherwise. Definition 2 (Block-Sign). For x Rd, define M blocks indexed by Bi, i = 1, ..., M, with di := |Bi|. The Block-Sign compressor is defined as C(x) = [sign(x B1) x B1 1 d1 , ..., sign(x BM ) x BM 1 d M ], where x Bi is the sub-vector of x at indices Bi. Remark 1. It is well-known (Stich et al., 2018) that for Top-k, q2 = 1 k d. For Block-Sign, by Cauchy-Schwartz inequality we have q2 = 1 mini [M] 1 di where M and di are defined in Definition 2 (Zheng et al., 2019). The intuition behind Top-k is that, it has been observed empirically that when training many deep models, most gradients are typically very small, and gradients with large magnitude contain most information. The Block-Sign compressor is a simple extension of the 1-bit SIGN compressor (Seide et al., 2014; Bernstein et al., 2018), adapted to different gradient magnitude in different blocks, which, for neural nets, are usually set as the distinct network layers. The scaling factor in Definition 2 is to preserve the (possibly very different) gradient magnitude in each layer. In principle, Top-k would perform the best when the gradient is effectively sparse, while Block-Sign compressor is favorable by nature when most gradients have similar magnitude within each layer. 3.2 COMP-AMS: DISTRIBUTED ADAPTIVE TRAINING BY GRADIENT AGGREGATION We present in Algorithm 2 the proposed communication-efficient distributed adaptive method in this paper, COMP-AMS. This framework can be regarded as an analogue to the standard synchronous distributed SGD: in each iteration, each local worker transmits to the central server the compressed stochastic gradient computed using local data. Then the central server takes the average of local gradients, and performs an AMSGrad update. In Algorithm 2, lines 7-8 depict the error feedback operation at local nodes. et,i is the accumulated error from gradient compression on the i-th worker up to time t 1. This residual is added back to gt,i to get the corrected gradient. In Section 4 and Section 5, we will show that error feedback, similar to the case of SGD, also brings good convergence behavior under gradient compression in distributed AMSGrad. Published as a conference paper at ICLR 2022 Algorithm 2 Distributed COMP-AMS with error feedback (EF) 1: Input: parameters β1, β2, ϵ, learning rate ηt 2: Initialize: central server parameter θ1 Rd Rd; e1,i = 0 the error accumulator for each worker; m0 = 0, v0 = 0, ˆv0 = 0 3: for t = 1, . . . , T do 4: parallel for worker i [n] do: 5: Receive model parameter θt from central server 6: Compute stochastic gradient gt,i at θt 7: Compute the compressed gradient gt,i = C(gt,i + et,i) 8: Update the error et+1,i = et,i + gt,i gt,i 9: Send gt,i back to central server 10: end parallel 11: Central server do: 12: gt = 1 n Pn i=1 gt,i 13: mt = β1mt 1 + (1 β1) gt 14: vt = β2vt 1 + (1 β2) g2 t 15: ˆvt = max(vt, ˆvt 1) 16: Update the global model θt+1 = θt ηt mt ˆvt+ϵ 17: end for Comparison with related methods. Next, we discuss the differences between COMP-AMS and two recently proposed methods also trying to solve compressed distributed adaptive optimization. Comparison with Chen et al. (2021a). Chen et al. (2021a) develops a quantized variant of Adam (Kingma & Ba, 2015), called QAdam . In this method, each worker keeps a local copy of the moment estimates, commonly noted m and v, and compresses and transmits the ratio m v as a whole to the server. Their method is hence very much like the compressed distributed SGD, with the exception that the ratio m v plays the role of the gradient vector g communication-wise. Thus, two local moment estimators are additionally required, which have same size as the deep learning model. In our COMP-AMS, the moment estimates m and v are kept and updated only at the central server, thus not introducing any extra variable (tensor) on local nodes during training (except for the error accumulator). Hence, COMP-AMS is not only effective in communication reduction, but also efficient in terms of memory (space), which is feasible when training large-scale learners like BERT and CTR prediction models, e.g., Devlin et al. (2019); Xu et al. (2021), to lower the hardware consumption in practice. Additionally, the convergence rate in Chen et al. (2021a) does not improve linearly with n, while we prove the linear speedup effect of COMP-AMS. Comparison with Tang et al. (2021) The recent work (Tang et al., 2021) proposes 1Bit Adam . They first run some warm-up training steps using standard Adam, and then store the second moment moving average v. Then, distributed Adam training starts with v frozen. Thus, 1Bit Adam is actually more like a distributed momentum SGD with some pre-conditioned coordinate-wise learning rates. The number of warm-up steps also needs to be carefully tuned, otherwise bad pre-conditioning may hurt the learning performance. Our COMP-AMS is simpler, as no pre-training is needed. Also, 1Bit Adam requires extra tensors for m locally, while COMP-AMS does not need additional local memory. 4 CONVERGENCE ANALYSIS For the convergence analysis of COMP-AMS we will make following additional assumptions. Assumption 2. (Smoothness) For i [n], fi is L-smooth: fi(θ) fi(ϑ) L θ ϑ . Assumption 3. (Unbiased and bounded stochastic gradient) For t > 0, i [n], the stochastic gradient is unbiased and uniformly bounded: E[gt,i] = fi(θt) and gt,i G. Assumption 4. (Bounded variance) For t > 0, i [n]: (i) the local variance of the stochastic gradient is bounded: E[ gt,i fi(θt) 2] < σ2; (ii) the global variance is bounded by 1 n Pn i=1 fi(θt) f(θt) 2 σ2 g. Published as a conference paper at ICLR 2022 In Assumption 3, the uniform bound on the stochastic gradient is common in the convergence analysis of adaptive methods, e.g., Reddi et al. (2018); Zhou et al. (2018); Chen et al. (2019). The global variance bound σ2 g in Assumption 4 characterizes the difference among local objective functions, which, is mainly caused by different local data distribution Xi in (1). In classical distributed setting where all the workers can access the same dataset and local data are assigned randomly, σ2 g 0. While typical federated learning (FL) setting with σ2 g > 0 is not the focus of this present paper, we consider the global variance in our analysis to shed some light on the impact of non-i.i.d. data distribution in the federated setting for broader interest and future investigation. We derive the following general convergence rate of COMP-AMS in the distributed setting. The proof is deferred to Appendix B. Theorem 1. Denote C0 = q (1 q2)2 G2 + ϵ, C1 = β1 1 β1 + 2q 1 q2 , θ = arg min f(θ) defined as (1). Under Assumptions 1 to 4, with ηt = η ϵ 3C0 2L max{2L,C1}, Algorithm 2 satisfies t=1 E[ f(θt) 2] 2C0 E[f(θ1) f(θ )] nϵ + 3η2LC0C2 1σ2 + 12η2q2LC0σ2 g (1 q2)2ϵ2 + (1 + C1)G2d T ϵ + η(1 + 2C1)C1LG2d The LHS of Theorem 1 is the expected squared norm of the gradient from a uniformly chosen iterate t [T], which is a common convergence measure in non-convex optimization. From Theorem 1, we see that the more compression we apply to the gradient vectors (i.e., larger q), the larger the gradient magnitude is, i.e., the slower the algorithm converges. This is intuitive as heavier compression loses more gradient information which would slower down the learner to find a good solution. Note that, COMP-AMS with n = 1 naturally reduces to the single-machine (sequential) AMSGrad (Algorithm 1) with compressed gradients instead of full-precision ones. Karimireddy et al. (2019) specifically analyzed this case for SGD, showing that compressed single-machine SGD with error feedback has the same convergence rate as vanilla SGD using full gradients. In alignment with the conclusion in Karimireddy et al. (2019), for adaptive AMSGrad, we have a similar result. Corollary 1. When n = 1, under Assumption 1 to Assumption 4, setting the stepsize as η = min{ ϵ 3C0 2L max{2L,C1}, 1 T }, Algorithm 2 satisfies t=1 E[ f(θt) 2] O( 1 Corollary 1 states that with error feedback, single machine AMSGrad with biased compressed gradients can also match the convergence rate O( 1 T ) of standard AMSGrad (Zhou et al., 2018) in non-convex optimization. It also achieves the same rate O( 1 T ) of vanilla SGD (Karimireddy et al., 2019) when T is sufficiently large. In other words, error feedback also fixes the convergence issue of using compressed gradients in AMSGrad. Linear Speedup. In Theorem 1, the convergence rate is derived by assuming a constant learning rate. By carefully choosing a decreasing learning rate dependent on the number of workers, we have the following simplified statement. Corollary 2. Under the same setting as Theorem 1, set η = min{ ϵ 3C0 2L max{2L,C1}, n COMP-AMS iterates admit t=1 E[ f(θt) 2] O( 1 n T + n(σ2 + σ2 g) T ). (2) In Corollary 2, we see that the global variance σ2 g appears in the O( 1 T ) term, which says that it asymptotically has no impact on the convergence. This matches the result of momentum SGD (Yu Published as a conference paper at ICLR 2022 et al., 2019a). When T O(n3) is sufficiently large, the third term in (2) vanishes, and the convergence rate becomes O( 1 n T ). Therefore, to reach an O(δ) stationary point, one worker (n = 1) needs T = O( 1 δ2 ) iterations, while distributed training with n workers requires only T = O( 1 nδ2 ) iterations, which is n times faster than single machine training. That is, COMP-AMS has a linear speedup in terms of the number of the local workers. Such acceleration effect has also been reported for compressed SGD (Jiang & Agrawal, 2018; Zheng et al., 2019) and momentum SGD (Yu et al., 2019a) with error feedback. 5 EXPERIMENTS In this section, we provide numerical results on several common datasets. Our main objective is to validate the theoretical results, and demonstrate that the proposed COMP-AMS can approach the learning performance of full-precision AMSGrad with significantly reduced communication costs. 5.1 DATASETS, MODELS AND METHODS Our experiments are conducted on various image and text datasets. The MNIST (Le Cun et al., 1998) contains 60000 training samples of 28 18 gray-scale hand-written digits from 10 classes, and 10000 test samples. We train MNIST with a Convolutional Neural Network (CNN), which has two convolutional layers followed by two fully connected layers with Re Lu activation. Dropout is applied after the max-pooled convolutional layer with rate 0.5. The CIFAR-10 dataset (Krizhevsky & Hinton, 2009) consists of 50000 32 32 RGB natural images from 10 classes for training and 10000 images for testing, which is trained by Le Net-5 (Le Cun et al., 1998). Moreover, we also implement Res Net-18 (He et al., 2016) on this dataset. The IMDB movie review (Maas et al., 2011) is a popular binary classification dataset for sentiment analysis. Each movie review is tokenized by top-2000 most frequently appeared words and transformed into integer vectors, which is of maximal length 500. We train a Long-Short Term Memory (LSTM) network with a 32-dimensional embedding layer and 64 LSTM cells, followed by two fully connected layers before output. Cross-entropy loss is used for all the tasks. Following the classical distributed training setting, in each training iteration, data samples are uniformly randomly assigned to the workers. We compare COMP-AMS with full-precision distributed AMSGrad, QAdam (Chen et al., 2021a) and 1Bit Adam (Tang et al., 2021). For COMP-AMS, Top-k picks top 1% gradient coordinates (i.e., sparsity 0.01). QAdam and 1Bit Adam both use 1-bit quantization to achieve high compression. For MNIST and CIFAR-10, the local batch size on each worker is set to be 32. For IMDB, the local batch size is 16. The hyper-parameters in COMP-AMS are set as default β1 = 0.9, β2 = 0.999 and ϵ = 10 8, which are also used for QAdam and 1Bit Adam. For 1Bit Adam, the epochs for warm-up training is set to be 1/20 of the total epochs. For all methods, we tune the initial learning rate over a fine grid (see Appendix A) and report the best results averaged over three independent runs. Our experiments are performed on a GPU cluster with NVIDIA Tesla V100 cards. 5.2 GENERAL EVALUATION AND COMMUNICATION EFFICIENCY The training loss and test accuracy on MNIST + CNN, CIFAR-10 + Le Net and IMDB + LSTM are reported in Figure 1. We provide more results on larger Res Net-18 model in Appendix A. On CIFAR-10, we deploy a popular decreasing learning rate schedule, where the step size η is divided by 10 at the 40-th and 80-th epoch, respectively. We observe: On MNIST, all the methods can approach the training loss and test accuracy of fullprecision AMSGrad. The 1Bit Adam seems slightly better, but the gap is very small. On CIFAR-10, COMP-AMS with Block-Sign performs the best and matches AMSGrad in terms of test accuracy. On IMDB, COMP-AMS with Top-k has both the fastest convergence and best generalization compared with other compressed methods. This is because the IMDB text data is more sparse (with many padded zeros), where Top-k is expected to work better than sign. The 1Bit Adam converges slowly. We believe one possible reason is that 1Bit Adam is quite sensitive to the quality of the warm-up training. For sparse text data, the estimation of second moment v is more unstable, making the strategy of freezing v by warm-up less effective. Published as a conference paper at ICLR 2022 0 5 10 15 20 Epochs Dist-AMS COMP-AMS-Top K COMP-AMS-Bk Sign QAdam 1Bit Adam 0 25 50 75 100 Epochs Dist-AMS COMP-AMS-Top K COMP-AMS-Bk Sign QAdam 1Bit Adam 0 5 10 15 20 Epochs Dist-AMS COMP-AMS-Top K COMP-AMS-Bk Sign QAdam 1Bit Adam 0 5 10 15 20 Epochs Test Accuracy Dist-AMS COMP-AMS-Top K COMP-AMS-Bk Sign QAdam 1Bit Adam 0 25 50 75 100 Epochs Test Accuracy Dist-AMS COMP-AMS-Top K COMP-AMS-Bk Sign QAdam 1Bit Adam 0 5 10 15 20 Epochs Test Accuracy Dist-AMS COMP-AMS-Top K COMP-AMS-Bk Sign QAdam 1Bit Adam Figure 1: Training loss and test accuracy vs. epochs, on MNIST + CNN, CIFAR-10 + Le Net and IMDB + LSTM with n = 16 local workers. Communication Efficiency. In Figure 2, we plot the training loss and test accuracy against the number of bits transmitted to the central server during the distributed training process, where we assume that the full-precision gradient is represented using 32 bits per floating number. As we can see, COMP-AMS-Top-0.01 achieves around 100x communication reduction, to attain similar accuracy as the full-precision distributed AMSGrad. The saving of Block-Sign is around 30x, but it gives slightly higher accuracy than Top-0.01 on MNIST and CIFAR-10. In all cases, COMPAMS can substantially reduce the communication cost compared with full-precision distributed AMSGrad, without losing accuracy. 20 25 30 log(No. bits) Dist-AMS COMP-AMS-Top K COMP-AMS-Bk Sign QAdam 1Bit Adam 25 30 35 40 log(No. bits) Dist-AMS COMP-AMS-Top K COMP-AMS-Bk Sign QAdam 1Bit Adam 20 25 30 log(No. bits) Dist-AMS COMP-AMS-Top K COMP-AMS-Bk Sign QAdam 1Bit Adam 20 25 30 log(No. bits) Test Accuracy Dist-AMS COMP-AMS-Top K COMP-AMS-Bk Sign QAdam 1Bit Adam 25 30 35 40 log(No. bits) Test Accuracy Dist-AMS COMP-AMS-Top K COMP-AMS-Bk Sign QAdam 1Bit Adam 20 25 30 log(No. bits) Test Accuracy Dist-AMS COMP-AMS-Top K COMP-AMS-Bk Sign QAdam 1Bit Adam Figure 2: Train loss and Test accuracy vs. No. bits transmitted, on MNIST + CNN, CIFAR-10 + Le Net and IMDB + LSTM with n = 16 local workers. Published as a conference paper at ICLR 2022 5.3 LINEAR SPEEDUP OF COMP-AMS 0 5000 10000 Iterations n = 2 n = 4 n = 8 n = 16 0 5000 10000 15000 Iterations n = 2 n = 4 n = 8 n = 16 Figure 3: The linear speedup of COMP-AMS with varying n. Left: MNIST with Block-Sign compressor on CNN. Right: CIFAR-10 with Top-k-0.01 compression on Le Net. Corollary 2 reveals the linear speedup of COMP-AMS in distributed training. In Figure 3, we present the training loss on MNIST and CIFAR-10 against the number of iterations, with varying number of workers n. We use COMP-AMS with Block-Sign on MNIST, and Top-k-0.01 on CIFAR. As suggested by the theory, we use 5 10 4 n as the learning rate. From Figure 3, we see the number of iterations to achieve a certain loss exhibits a strong linear relationship with n it (approximately) decreases by half whenever we double n, which justifies the linear speedup of COMP-AMS. 5.4 DISCUSSION We provide a brief summary of our empirical observations. The proposed COMP-AMS is able to match the learning performance of full-gradient AMSGrad in all the presented experiments. In particular, for data/model involving some sparsity structure, COMP-AMS with the Top-k compressor could be more effective. Also, our results reveal that 1Bit Adam might be quite sensitive to the pre-conditioning quality, while COMP-AMS can be more easily tuned and implemented in practice. We would like to emphasize that, the primary goal of the experiments is to show that COMP-AMS is able to match the performance of full-precision AMSGrad, but not to argue that COMP-AMS is always better than the other algorithms. Since different methods use different underlying optimization algorithms (e.g., AMSGrad, Adam, momentum SGD), comparing COMP-AMS with other distributed training methods would be largely determined by the comparison among these optimization protocols, which is typically data and task dependent. Our results say that: whenever one wants to use AMSGrad to train a deep neural network, she/he can simply employ the distributed COMPAMS scheme to gain a linear speedup in training time with learning performance as good as the full-precision training, taking little communication cost and memory consumption. 6 CONCLUSION In this paper, we study the simple, convenient, yet unexplored gradient averaging strategy for distributed adaptive optimization called COMP-AMS. Top-k and Block-Sign compressor are incorporated for communication efficiency, whose biases are compensated by the error feedback strategy. We develop the convergence rate of COMP-AMS, and show that same as the case of SGD, for AMSGrad, compressed gradient averaging with error feedback matches the convergence of full-gradient AMSGrad, and linear speedup can be obtained in the distributed training. Numerical experiments are conducted to justify the theoretical findings, and demonstrate that COMP-AMS provides comparable performance with other distributed adaptive methods, and achieves similar accuracy as full-precision AMSGrad with significantly reduced communication overhead. Given the simple architecture and hardware (memory) efficiency, we expect COMP-AMS shall be able to serve as a useful and convenient distributed adaptive optimization framework in practice. Published as a conference paper at ICLR 2022 Naman Agarwal, Ananda Theertha Suresh, Felix X. Yu, Sanjiv Kumar, and Brendan Mc Mahan. cp SGD: Communication-efficient and differentially-private distributed SGD. In Advances in Neural Information Processing Systems (Neur IPS), pp. 7575 7586, Montréal, Canada, 2018. Ahmad Ajalloeian and Sebastian U Stich. Analysis of SGD with biased gradient estimators. ar Xiv preprint ar Xiv:2008.00051, 2020. Alham Fikri Aji and Kenneth Heafield. Sparse communication for distributed gradient descent. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 440 445, Copenhagen, Denmark, 2017. Dan Alistarh, Demjan Grubic, Jerry Li, Ryota Tomioka, and Milan Vojnovic. QSGD: communication-efficient SGD via gradient quantization and encoding. In Advances in Neural Information Processing Systems (NIPS), pp. 1709 1720, Long Beach, CA, 2017. Dan Alistarh, Torsten Hoefler, Mikael Johansson, Nikola Konstantinov, Sarit Khirirat, and Cédric Renggli. The convergence of sparsified gradient methods. In Advances in Neural Information Processing Systems (Neur IPS), pp. 5977 5987, Montréal, Canada, 2018. Debraj Basu, Deepesh Data, Can Karakus, and Suhas N. Diggavi. Qsparse-local-sgd: Distributed SGD with quantization, sparsification and local computations. In Advances in Neural Information Processing Systems (Neur IPS), pp. 14668 14679, Vancouver, Canada, 2019. Jeremy Bernstein, Yu-Xiang Wang, Kamyar Azizzadenesheli, and Animashree Anandkumar. SIGNSGD: compressed optimisation for non-convex problems. In Proceedings of the 35th International Conference on Machine Learning (ICML), pp. 559 568, Stockholmsmässan, Stockholm, Sweden, 2018. Jeremy Bernstein, Jiawei Zhao, Kamyar Azizzadenesheli, and Anima Anandkumar. sign SGD with majority vote is communication efficient and fault tolerant. In Proceedings of the 7th International Conference on Learning Representations (ICLR), New Orleans, LA, 2019. Aleksandr Beznosikov, Samuel Horváth, Peter Richtárik, and Mher Safaryan. On biased compression for distributed learning. ar Xiv preprint ar Xiv:2002.12410, 2020. Stephen P. Boyd, Neal Parikh, Eric Chu, Borja Peleato, and Jonathan Eckstein. Distributed optimization and statistical learning via the alternating direction method of multipliers. Found. Trends Mach. Learn., 3(1):1 122, 2011. Ken Chang, Niranjan Balachandar, Carson K. Lam, Darvin Yi, James M. Brown, Andrew Beers, Bruce R. Rosen, Daniel L. Rubin, and Jayashree Kalpathy-Cramer. Distributed deep learning networks among institutions for medical imaging. J. Am. Medical Informatics Assoc., 25(8): 945 954, 2018. Congliang Chen, Li Shen, Hao-Zhi Huang, and Wei Liu. Quantized adam with error feedback. ACM Trans. Intell. Syst. Technol., 12(5):56:1 56:26, 2021a. Xiangyi Chen, Sijia Liu, Ruoyu Sun, and Mingyi Hong. On the convergence of A class of adam-type algorithms for non-convex optimization. In Proceedings of the 7th International Conference on Learning Representations (ICLR), New Orleans, LA, 2019. Xiangyi Chen, Xiaoyun Li, and Ping Li. Toward communication efficient adaptive gradient method. In Proceedings of the ACM-IMS Foundations of Data Science Conference (FODS), pp. 119 128, Seattle, WA, 2020. Xiangyi Chen, Belhal Karimi, Weijie Zhao, and Ping Li. On the convergence of decentralized adaptive gradient methods. ar Xiv preprint ar Xiv:2109.03194, 2021b. Trishul M. Chilimbi, Yutaka Suzue, Johnson Apacible, and Karthik Kalyanaraman. Project adam: Building an efficient and scalable deep learning training system. In Proceedings of the 11th USENIX Symposium on Operating Systems Design and Implementation (OSDI), pp. 571 582, Broomfield, CO, 2014. Published as a conference paper at ICLR 2022 Dami Choi, Christopher J Shallue, Zachary Nado, Jaehoon Lee, Chris J Maddison, and George E Dahl. On empirical comparisons of optimizers for deep learning. ar Xiv preprint ar Xiv:1910.05446, 2019. Paul Covington, Jay Adams, and Emre Sargin. Deep neural networks for youtube recommendations. In Proceedings of the 10th ACM Conference on Recommender Systems, pp. 191 198, Boston, MA, 2016. Tim Dettmers. 8-bit approximations for parallelism in deep learning. In Proceedings of the 4th International Conference on Learning Representations (ICLR), San Juan, Puerto Rico, 2016. Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. BERT: pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT), pp. 4171 4186, Minneapolis, MN, 2019. John C. Duchi, Elad Hazan, and Yoram Singer. Adaptive subgradient methods for online learning and stochastic optimization. In Proceedings of the 23rd Conference on Learning Theory (COLT), pp. 257 269, Haifa, Israel, 2010. John C. Duchi, Alekh Agarwal, and Martin J. Wainwright. Dual averaging for distributed optimization: Convergence analysis and network scaling. IEEE Trans. Autom. Control., 57(3):592 606, 2012. Saeed Ghadimi and Guanghui Lan. Stochastic firstand zeroth-order methods for nonconvex stochastic programming. SIAM J. Optim., 23(4):2341 2368, 2013. Ian J. Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron C. Courville, and Yoshua Bengio. Generative adversarial nets. In Advances in Neural Information Processing Systems (NIPS), pp. 2672 2680, Montreal, Canada, 2014. Priya Goyal, Piotr Dollár, Ross Girshick, Pieter Noordhuis, Lukasz Wesolowski, Aapo Kyrola, Andrew Tulloch, Yangqing Jia, and Kaiming He. Accurate, large minibatch sgd: Training imagenet in 1 hour. ar Xiv preprint ar Xiv:1706.02677, 2017. Alex Graves, Abdel-rahman Mohamed, and Geoffrey E. Hinton. Speech recognition with deep recurrent neural networks. In Proceedings of IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 6645 6649, Vancouver, Canada, 2013. Farzin Haddadpour, Belhal Karimi, Ping Li, and Xiaoyun Li. Fedsketch: Communication-efficient and private federated learning via sketching. ar Xiv preprint ar Xiv:2008.04975, 2020. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770 778, Las Vegas, NV, 2016. Mingyi Hong, Davood Hajinezhad, and Ming-Min Zhao. Prox-pda: The proximal primal-dual algorithm for fast distributed nonconvex optimization and learning over networks. In Proceedings of the 34th International Conference on Machine Learning (ICML), pp. 1529 1538, Sydney, Australia, 2017. Nikita Ivkin, Daniel Rothchild, Enayat Ullah, Vladimir Braverman, Ion Stoica, and Raman Arora. Communication-efficient distributed SGD with sketching. In Advances in Neural Information Processing Systems (Neur IPS), pp. 13144 13154, Vancouver, Canada, 2019. Jiawei Jiang, Fangcheng Fu, Tong Yang, and Bin Cui. Sketchml: Accelerating distributed machine learning with data sketches. In Proceedings of the 2018 ACM International Conference on Management of Data (SIGMOD), pp. 1269 1284, Houston, TX, 2018. Peng Jiang and Gagan Agrawal. A linear speedup analysis of distributed deep learning with sparse and quantized communication. In Advances in Neural Information Processing Systems (Neur IPS), pp. 2530 2541, Montréal, Canada, 2018. Published as a conference paper at ICLR 2022 Belhal Karimi, Xiaoyun Li, and Ping Li. Fed-LAMB: Layerwise and dimensionwise locally adaptive optimization algorithm. ar Xiv preprint ar Xiv:2110.00532, 2021. Sai Praneeth Karimireddy, Quentin Rebjock, Sebastian U. Stich, and Martin Jaggi. Error feedback fixes signsgd and other gradient compression schemes. In Proceedings of the 36th International Conference on Machine Learning (ICML), pp. 3252 3261, Long Beach, CA, 2019. Sai Praneeth Karimireddy, Martin Jaggi, Satyen Kale, Mehryar Mohri, Sashank J Reddi, Sebastian U Stich, and Ananda Theertha Suresh. Mime: Mimicking centralized stochastic algorithms in federated learning. ar Xiv preprint ar Xiv:2008.03606, 2020. Diederik P. Kingma and Jimmy Ba. Adam: A method for stochastic optimization. In Proceedings of the 3rd International Conference on Learning Representations (ICLR), San Diego, CA, 2015. Anastasia Koloskova, Sebastian U. Stich, and Martin Jaggi. Decentralized stochastic optimization and gossip algorithms with compressed communication. In Proceedings of the 36th International Conference on Machine Learning (ICML), pp. 3478 3487, Long Beach, CA, 2019. A. Krizhevsky and G. Hinton. Learning multiple layers of features from tiny images. Master s thesis, Department of Computer Science, University of Toronto, 2009. Yann Le Cun, Léon Bottou, Yoshua Bengio, and Patrick Haffner. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278 2324, 1998. Sergey Levine, Chelsea Finn, Trevor Darrell, and Pieter Abbeel. End-to-end training of deep visuomotor policies. J. Mach. Learn. Res., 17:39:1 39:40, 2016. Yujun Lin, Song Han, Huizi Mao, Yu Wang, and Bill Dally. Deep gradient compression: Reducing the communication bandwidth for distributed training. In Proceedings of the 6th International Conference on Learning Representations (ICLR), Vancouver, Canada, 2018. Songtao Lu, Xinwei Zhang, Haoran Sun, and Mingyi Hong. GNSD: a gradient-tracking based nonconvex stochastic algorithm for decentralized optimization. In Proceedings of the 2019 IEEE Data Science Workshop (DSW), pp. 315 321, 2019. Andrew L. Maas, Raymond E. Daly, Peter T. Pham, Dan Huang, Andrew Y. Ng, and Christopher Potts. Learning word vectors for sentiment analysis. In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT), pp. 142 150, Portland, OR, 2011. Brendan Mc Mahan, Eider Moore, Daniel Ramage, Seth Hampson, and Blaise Agüera y Arcas. Communication-efficient learning of deep networks from decentralized data. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics (AISTATS), pp. 1273 1282, Fort Lauderdale, FL, 2017. Hiroaki Mikami, Hisahiro Suganuma, Yoshiki Tanaka, and Yuichi Kageyama. Massively distributed SGD: Imagenet/resnet-50 training in a flash. ar Xiv preprint ar Xiv:1811.05233, 2018. Volodymyr Mnih, Koray Kavukcuoglu, David Silver, Alex Graves, Ioannis Antonoglou, Daan Wierstra, and Martin Riedmiller. Playing atari with deep reinforcement learning. ar Xiv preprint ar Xiv:1312.5602, 2013. Parvin Nazari, Davoud Ataee Tarzanagh, and George Michailidis. Dadam: A consensus-based distributed adaptive gradient method for online optimization. ar Xiv preprint ar Xiv:1901.09109, 2019. Angelia Nedic and Asuman E. Ozdaglar. Distributed subgradient methods for multi-agent optimization. IEEE Trans. Autom. Control., 54(1):48 61, 2009. Arkadi Nemirovski, Anatoli B. Juditsky, Guanghui Lan, and Alexander Shapiro. Robust stochastic approximation approach to stochastic programming. SIAM J. Optim., 19(4):1574 1609, 2009. Sashank J. Reddi, Satyen Kale, and Sanjiv Kumar. On the convergence of adam and beyond. In Proceedings of the 6th International Conference on Learning Representations (ICLR), Vancouver, Canada, 2018. Published as a conference paper at ICLR 2022 Sashank J. Reddi, Zachary Charles, Manzil Zaheer, Zachary Garrett, Keith Rush, Jakub Koneˇcný, Sanjiv Kumar, and Hugh Brendan Mc Mahan. Adaptive federated optimization. In Proceedings of the 9th International Conference on Learning Representations (ICLR), Virtual Event, 2021. Peter Richtárik, Igor Sokolov, and Ilyas Fatkhullin. EF21: A new, simpler, theoretically better, and practically faster error feedback. In Advances in Neural Information Processing Systems (Neur IPS), virtual, 2021. Christopher De Sa, Matthew Feldman, Christopher Ré, and Kunle Olukotun. Understanding and optimizing asynchronous low-precision stochastic gradient descent. In Proceedings of the 44th Annual International Symposium on Computer Architecture (ISCA), pp. 561 574, Toronto, Canada, 2017. Frank Seide, Hao Fu, Jasha Droppo, Gang Li, and Dong Yu. 1-bit stochastic gradient descent and its application to data-parallel distributed training of speech dnns. In Proceedings of the 15th Annual Conference of the International Speech Communication Association (ISCA), pp. 1058 1062, Singapore, 2014. Zebang Shen, Aryan Mokhtari, Tengfei Zhou, Peilin Zhao, and Hui Qian. Towards more efficient stochastic decentralized learning: Faster convergence and sparse communication. In Proceedings of the 35th International Conference on Machine Learning (ICML), pp. 4631 4640, Stockholmsmässan, Stockholm, Sweden, 2018. Shaohuai Shi, Kaiyong Zhao, Qiang Wang, Zhenheng Tang, and Xiaowen Chu. A convergence analysis of distributed SGD with communication-efficient gradient sparsification. In Proceedings of the 28th International Joint Conference on Artificial Intelligence (IJCAI), pp. 3411 3417, Macao, China, 2019. David Silver, Julian Schrittwieser, Karen Simonyan, Ioannis Antonoglou, Aja Huang, Arthur Guez, Thomas Hubert, Lucas Baker, Matthew Lai, Adrian Bolton, Yutian Chen, Timothy P. Lillicrap, Fan Hui, Laurent Sifre, George van den Driessche, Thore Graepel, and Demis Hassabis. Mastering the game of go without human knowledge. Nat., 550(7676):354 359, 2017. Sebastian U Stich and Sai Praneeth Karimireddy. The error-feedback framework: Better rates for sgd with delayed gradients and compressed communication. ar Xiv preprint ar Xiv:1909.05350, 2019. Sebastian U. Stich, Jean-Baptiste Cordonnier, and Martin Jaggi. Sparsified SGD with memory. In Advances in Neural Information Processing Systems (Neur IPS), pp. 4447 4458, Montréal, Canada, 2018. Hanlin Tang, Shaoduo Gan, Ammar Ahmad Awan, Samyam Rajbhandari, Conglong Li, Xiangru Lian, Ji Liu, Ce Zhang, and Yuxiong He. 1-bit adam: Communication efficient large-scale training with adam s convergence speed. In Proceedings of the 38th International Conference on Machine Learning (ICML), pp. 10118 10129, Virtual Event, 2021. Jun-Kun Wang, Xiaoyun Li, Belhal Karimi, and Ping Li. An optimistic acceleration of amsgrad for nonconvex optimization. In Proceedings of Asian Conference on Machine Learning (ACML), volume 157, pp. 422 437, Virtual Event, 2021. Jianqiao Wangni, Jialei Wang, Ji Liu, and Tong Zhang. Gradient sparsification for communicationefficient distributed optimization. In Advances in Neural Information Processing Systems (Neur IPS), pp. 1299 1309, Montréal, Canada, 2018. Wei Wen, Cong Xu, Feng Yan, Chunpeng Wu, Yandan Wang, Yiran Chen, and Hai Li. Terngrad: Ternary gradients to reduce communication in distributed deep learning. In Advances in Neural Information Processing Systems (NIPS), pp. 1509 1519, Long Beach, CA, 2017. Jiaxiang Wu, Weidong Huang, Junzhou Huang, and Tong Zhang. Error compensated quantized SGD and its applications to large-scale distributed optimization. In Proceedings of the 35th International Conference on Machine Learning (ICML), pp. 5321 5329, Stockholmsmässan, Stockholm, Sweden, 2018. Published as a conference paper at ICLR 2022 Zhiqiang Xu, Dong Li, Weijie Zhao, Xing Shen, Tianbo Huang, Xiaoyun Li, and Ping Li. Agile and accurate CTR prediction model training for massive-scale online advertising systems. In Proceedings of the International Conference on Management of Data (SIGMOD), pp. 2404 2409, Virtual Event, China, 2021. Guandao Yang, Tianyi Zhang, Polina Kirichenko, Junwen Bai, Andrew Gordon Wilson, and Christopher De Sa. SWALP : Stochastic weight averaging in low precision training. In Proceedings of the 36th International Conference on Machine Learning (ICML), pp. 7015 7024, Long Beach, CA, 2019. Yang You, Jing Li, Sashank J. Reddi, Jonathan Hseu, Sanjiv Kumar, Srinadh Bhojanapalli, Xiaodan Song, James Demmel, Kurt Keutzer, and Cho-Jui Hsieh. Large batch optimization for deep learning: Training BERT in 76 minutes. In Proceedings of the 8th International Conference on Learning Representations (ICLR), Addis Ababa, Ethiopia, 2020. Tom Young, Devamanyu Hazarika, Soujanya Poria, and Erik Cambria. Recent trends in deep learning based natural language processing. IEEE Comput. Intell. Mag., 13(3):55 75, 2018. Hao Yu, Rong Jin, and Sen Yang. On the linear speedup analysis of communication efficient momentum SGD for distributed non-convex optimization. In Proceedings of the 36th International Conference on Machine Learning (ICML), pp. 7184 7193, Long Beach, CA, 2019a. Yue Yu, Jiaxiang Wu, and Junzhou Huang. Exploring fast and communication-efficient algorithms in large-scale distributed networks. In Proceedings of the 22nd International Conference on Artificial Intelligence and Statistics (AISTATS), pp. 674 683, Naha, Okinawa, Japan, 2019b. Matthew D Zeiler. Adadelta: an adaptive learning rate method. ar Xiv preprint ar Xiv:1212.5701, 2012. Hantian Zhang, Jerry Li, Kaan Kara, Dan Alistarh, Ji Liu, and Ce Zhang. Zip ML: Training linear models with end-to-end low precision, and a little bit of deep learning. In Proceedings of the 34th International Conference on Machine Learning (ICML), pp. 4035 4043, Sydney, Australia, 2017. Lei Zhang, Shuai Wang, and Bing Liu. Deep learning for sentiment analysis: A survey. Wiley Interdiscip. Rev. Data Min. Knowl. Discov., 8(4), 2018. Tianyi Zhang, Felix Wu, Arzoo Katiyar, Kilian Q. Weinberger, and Yoav Artzi. Revisiting fewsample BERT fine-tuning. In Proceedings of the 9th International Conference on Learning Representations (ICLR), Virtual Event, 2021. Weijie Zhao, Jingyuan Zhang, Deping Xie, Yulei Qian, Ronglai Jia, and Ping Li. AIBox: CTR prediction model training on a single node. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management (CIKM), pp. 319 328, Beijing, China, 2019. Weijie Zhao, Xuewu Jiao, Mingqing Hu, Xiaoyun Li, Xiangyu Zhang, and Ping Li. Communicationefficient terabyte-scale model training framework for online advertising. ar Xiv preprint ar Xiv:2201.05500, 2022. Shuai Zheng, Ziyue Huang, and James T. Kwok. Communication-efficient distributed blockwise momentum SGD with error-feedback. In Advances in Neural Information Processing Systems (Neur IPS), pp. 11446 11456, Vancouver, Canada, 2019. Dongruo Zhou, Jinghui Chen, Yuan Cao, Yiqi Tang, Ziyan Yang, and Quanquan Gu. On the convergence of adaptive gradient methods for nonconvex optimization. ar Xiv preprint ar Xiv:1808.05671, 2018. Yingxue Zhou, Belhal Karimi, Jinxing Yu, Zhiqiang Xu, and Ping Li. Towards better generalization of adaptive gradient methods. In Advances in Neural Information Processing Systems (Neur IPS), virtual, 2020. Published as a conference paper at ICLR 2022 A TUNING DETAILS AND MORE RESULTS ON RESNET-18 The search grids of the learning rate of each method can be found in Table 1. Empirically, Dist AMS, COMP-AMS and 1Bit Adam has similar optimal learning rate, while QAdam usually needs larger step size to reach its best performance. Table 1: Search grids for learning rate tuning. Learning rate range Dist-AMS [0.00001, 0.00003, 0.00005, 0.0001, 0.0003, 0.0005, 0.001, 0.003, 0.005, 0.01] Comp-AMS [0.00001, 0.00003, 0.00005, 0.0001, 0.0003, 0.0005, 0.001, 0.003, 0.005, 0.01] QAdam [0.0001, 0.0003, 0.0005, 0.001, 0.003, 0.005, 0.01, 0.03, 0.05, 0.1, 0.3, 0.5] 1Bit Adam [0.00001, 0.00003, 0.00005, 0.0001, 0.0003, 0.0005, 0.001, 0.003, 0.005, 0.01] We provide more experimental results on CIFAR-10 dataset, trained with Res Net-18 (He et al., 2016). For reference, we also present the result of distributed SGD. As we can see from Figure 4, again COMP-AMS can achieve similar accuracy as AMSGrad, and the Top-k compressor gives the best accuracy, with substantial communication reduction. Note that distributed SGD converges faster than adaptive methods, but the generalization error is slightly worse. This experiment again confirms that COMP-AMS can serve as a simple and convenient distributed adaptive training framework with fast convergence, reduced communication and little performance drop. 0 25 50 75 100 Epochs Dist-AMS COMP-AMS-Top K COMP-AMS-Bk Sign QAdam Dist-SGD 0 25 50 75 100 Epochs Test Accuracy Dist-AMS COMP-AMS-Top K COMP-AMS-Bk Sign QAdam Dist-SGD 30 35 40 45 log(No. bits) Test Accuracy Dist-AMS COMP-AMS-Top K COMP-AMS-Bk Sign QAdam Dist-SGD Figure 4: Training loss and test accuracy of different distributed training methods on CIFAR-10 with Res Net-18 (He et al., 2016). Published as a conference paper at ICLR 2022 B PROOF OF CONVERGENCE RESULTS In this section, we provide the proof of our main result. B.1 PROOF OF THEOREM 1 Theorem. Denote C0 = q (1 q2)2 G2 + ϵ, C1 = β1 1 β1 + 2q 1 q2 . Under Assumption 1 to Assumption 4, with ηt = η ϵ 3C0 2L max{2L,C2}, for any T > 0, COMP-AMS satisfies t=1 E[ f(θt) 2] 2C0 E[f(θ1) f(θ )] nϵ + 3η2LC0C2 1σ2 + 12η2q2LC0σ2 g (1 q2)2ϵ2 + (1 + C1)G2d T ϵ + η(1 + 2C1)C1LG2d Proof. We first clarify some notations. At time t, let the full-precision gradient of the i-th worker be gt,i, the error accumulator be et,i, and the compressed gradient be gt,i = C(gt,i + et,i). Slightly different from the notations in the algorithm, we denote gt = 1 n Pn i=1 gt,i, gt = 1 n Pn i=1 gt,i and et = 1 n Pn i=1 et,i. The second moment computed by the compressed gradients is denoted as vt = β2vt 1 + (1 β2) g 2 t, and ˆvt = max{ˆvt 1, vt}. Also, the first order moving average sequence mt = β1mt 1 + (1 β1) gt and m t = β1m t 1 + (1 β1) gt, where m t represents the first moment moving average sequence using the uncompressed stochastic gradients. By construction we have m t = (1 β1) Pt τ=1 βt τ 1 gτ. Our proof will use the following auxiliary sequences, Et+1 := (1 β1) τ=1 βt+1 τ 1 eτ, θ t+1 := θt+1 η Et+1 ˆvt + ϵ. Then, we can write the evolution of θ t as θ t+1 = θt+1 η Et+1 ˆvt + ϵ = θt η (1 β1) Pt τ=1 βt τ 1 gτ + (1 β1) Pt+1 τ=1 βt+1 τ 1 eτ ˆvt + ϵ = θt η (1 β1) Pt τ=1 βt τ 1 ( gτ + eτ+1) + (1 β)βt 1 e1 ˆvt + ϵ = θt η (1 β1) Pt τ=1 βt τ 1 eτ ˆvt + ϵ η m t ˆvt + ϵ = θt η Et p ˆvt 1 + ϵ η m t ˆvt + ϵ + η( 1 p ˆvt 1 + ϵ 1 ˆvt + ϵ)Et (a) = θ t η m t ˆvt + ϵ + η( 1 p ˆvt 1 + ϵ 1 ˆvt + ϵ)Et := θ t η m t ˆvt + ϵ + ηDt Et, where (a) uses the fact that for every i [n], gt,i + et+1,i = gt,i + et,i, and et,1 = 0 at initialization. Further define the virtual iterates: xt+1 := θ t+1 η β1 1 β1 m t ˆvt + ϵ, Published as a conference paper at ICLR 2022 which follows the recurrence: xt+1 = θ t+1 η β1 1 β1 m t ˆvt + ϵ = θ t η m t ˆvt + ϵ η β1 1 β1 m t ˆvt + ϵ + ηDt Et = θ t η β1m t 1 + (1 β1) gt + β2 1 1 β1 m t 1 + β1 gt ˆvt + ϵ + ηDt Et = θ t η β1 1 β1 m t 1 ˆvt + ϵ η gt ˆvt + ϵ + ηDt Et = xt η gt ˆvt + ϵ + η β1 1 β1 Dtm t 1 + ηDt Et. When summing over t = 1, ..., T, the difference sequence Dt satisfies the bounds of Lemma 5. By the smoothness Assumption 2, we have f(xt+1) f(xt) + f(xt), xt+1 xt + L 2 xt+1 xt 2. Taking expectation w.r.t. the randomness at time t, we obtain E[f(xt+1)] f(xt) ηE[ f(xt), gt ˆvt + ϵ ] + ηE[ f(xt), β1 1 β1 Dtm t 1 + Dt Et ] 2 E[ gt ˆvt + ϵ β1 1 β1 Dtm t 1 Dt Et 2] = ηE[ f(θt), gt ˆvt + ϵ ] | {z } I + ηE[ f(xt), β1 1 β1 Dtm t 1 + Dt Et ] | {z } II 2 E[ gt ˆvt + ϵ β1 1 β1 Dtm t 1 Dt Et 2] | {z } III + ηE[ f(θt) f(xt), gt ˆvt + ϵ ] | {z } IV In the following, we bound the terms separately. Bounding term I. We have I = ηE[ f(θt), gt p ˆvt 1 + ϵ ] ηE[ f(θt), ( 1 ˆvt + ϵ 1 p ˆvt 1 + ϵ ) gt ] ηE[ f(θt), f(θt) p ˆvt 1 + ϵ ] + ηG2E[ Dt ]. (1 q2)2 G2 + ϵ E[ f(θt) 2] + ηG2E[ Dt 1], (4) where we use Assumption 3, Lemma 4 and the fact that l2 norm is no larger than l1 norm. Bounding term II. By the definition of Et, we know that Et (1 β1) Pt τ=1 βt τ 1 et 2q 1 q2 G. Then we have II η(E[ f(θt), β1 1 β1 Dtm t 1 + Dt Et ] + E[ f(xt) f(θt), β1 1 β1 Dtm t 1 + Dt Et ]) ηE[ f(θt) β1 1 β1 Dtm t 1 + Dt Et ] + η2 LE[ β1 1 β1 m t 1 + Et p ˆvt 1 + ϵ β1 1 β1 Dtm t 1 + Dt Et ] ηC1G2E[ Dt 1] + η2C2 1LG2 ϵ E[ Dt 1], (5) Published as a conference paper at ICLR 2022 where C1 := β1 1 β1 + 2q 1 q2 . The second inequality is because of smoothness of f(θ), and the last inequality is due to Lemma 2, Assumption 3 and the property of norms. Bounding term III. This term can be bounded as follows: III η2LE[ gt ˆvt + ϵ 2] + η2LE[ β1 1 β1 Dtm t 1 Dt Et 2]] i=1 gt,i f(θt) + f(θt) 2] + η2LE[ Dt( β1 1 β1 m t 1 Et) 2] ϵ E[ f(θt) 2] + η2Lσ2 nϵ + η2C2 1LG2E[ Dt 2], (6) where (a) follows from f(θt) = 1 n Pn i=1 fi(θt) and Assumption 4 that gt,i is unbiased of fi(θt) and has bounded variance σ2. Bounding term IV. We have IV = ηE[ f(θt) f(xt), gt p ˆvt 1 + ϵ ] + ηE[ f(θt) f(xt), ( 1 ˆvt + ϵ 1 p ˆvt 1 + ϵ ) gt ] ηE[ f(θt) f(xt), f(θt) p ˆvt 1 + ϵ ] + η2LE[ β1 1 β1 m t 1 + Et p ˆvt 1 + ϵ Dtgt ] 2ϵ E[ f(θt) 2] + η 2ρE[ f(θt) f(xt) 2] + η2C1LG2 2ϵ E[ f(θt) 2] + η3L β1 1 β1 m t 1 + Et p ˆvt 1 + ϵ 2] + η2C1LG2 ϵ E[ Dt 1], (7) where (a) is due to Young s inequality and (b) is based on Assumption 2. Regarding the second term in (7), by Lemma 3 and Lemma 1, summing over t = 1, ..., T we have β1 1 β1 m t 1 + Et p ˆvt 1 + ϵ 2] 2ρϵ E[ β1 1 β1 m t 1 + Et 2] h β2 1 (1 β1)2 E[ m t 2] + E[ Et 2] i nρ(1 β1)2ϵ + η3β2 1L ρ(1 β1)2ϵ t=1 E[ f(θt) 2] + 4Tη3q2L ρ(1 q2)2ϵ(σ2 + σ2 g) + 4η3q2L ρ(1 q2)2ϵ t=1 E[ f(θt) 2] nρϵ + 4Tη3q2Lσ2 g ρ(1 q2)2ϵ + η3LC2 t=1 E[ f(θt) 2], (8) with C2 := β2 1 (1 β1)2 + 4q2 (1 q2)2 . Now integrating (4), (5), (6), (7) and (8) into (3), taking the telescoping summation over t = 1, ..., T, we obtain E[f(x T +1) f(x1)] t=1 E[ f(θt) 2] + Tη2Lσ2 nϵ + Tη3LC2σ2 nρϵ + 4Tη3q2Lσ2 g ρ(1 q2)2ϵ + (η(1 + C1)G2 + η2(1 + C1)C1LG2 t=1 E[ Dt 1] + η2C2 1LG2 T X t=1 E[ Dt 2. Published as a conference paper at ICLR 2022 with C0 := q (1 q2)2 G2 + ϵ. Setting η ϵ 3C0 2L max{2L,C2} and choosing ρ = ϵ 3C0 , we further E[f(x T +1) f(x1)] t=1 E[ f(θt) 2] + Tη2Lσ2 nϵ + 3Tη3LC0C2σ2 nϵ2 + 12Tη3q2LC0σ2 g (1 q2)2ϵ2 + η(1 + C1)G2d ϵ + η2(1 + 2C1)C1LG2d where the inequality follows from Lemma 5. Re-arranging terms, we get that t=1 E[ f(θt) 2] 2C0 E[f(x1) f(x T +1)] nϵ + 3η2LC0C2σ2 + 12η2q2LC0σ2 g (1 q2)2ϵ2 + (1 + C1)G2d T ϵ + η(1 + 2C1)C1LG2d 2C0 E[f(θ1) f(θ )] nϵ + 3η2LC0C2 1σ2 + 12η2q2LC0σ2 g (1 q2)2ϵ2 + (1 + C1)G2d T ϵ + η(1 + 2C1)C1LG2d where C0 = q (1 q2)2 G2 + ϵ, C1 = β1 1 β1 + 2q 1 q2 . The last inequality is because x1 = θ1, θ := arg minθ f(θ) and the fact that C2 C2 1. This completes the proof. B.2 INTERMEDIATE LEMMATA The lemmas used in the proof of Theorem 1 are given as below. Lemma 1. Under Assumption 1 to Assumption 4 we have: t=1 E m t 2 Tσ2 t=1 E[ f(θt) 2]. Proof. For the first part, it is easy to see that by Assumption 3, m t = (1 β1) τ=1 βt τ 1 gt G. For the second claim, the expected squared norm of average stochastic gradient can be bounded by E[ g2 t ] = E[ 1 i=1 gt,i f(θt) + f(θt) 2] i=1 (gt,i fi(θt)) 2] + E[ f(θt) 2] n + E[ f(θt) 2], Published as a conference paper at ICLR 2022 where we use Assumption 4 that gt,i is unbiased with bounded variance. Let gt,j denote the j-th coordinate of gt. By the updating rule of COMP-AMS, we have E[ m t 2] = E[ (1 β1) τ=1 βt τ 1 gτ 2] (1 β1)2 d X τ=1 βt τ 1 gτ,j)2] (a) (1 β1)2 d X τ=1 βt τ 1 )( τ=1 βt τ 1 g2 τ,j)] τ=1 βt τ 1 E[ gτ 2] τ=1 βt τ 1 E[ f(θt) 2], where (a) is due to Cauchy-Schwartz inequality. Summing over t = 1, ..., T, we obtain t=1 E m t 2 Tσ2 t=1 E[ f(θt) 2]. This completes the proof. Lemma 2. Under Assumption 4, we have for t and each local worker i [n], (1 q2)2 G2, E[ et+1,i 2] 4q2 (1 q2)2 σ2 + 2q2 τ=1 (1 + q2 2 )t τE[ fi(θτ) 2]. Proof. We start by using Assumption 1 and Young s inequality to get et+1,i 2 = gt,i + et,i C(gt,i + et,i) 2 q2 gt,i + et,i 2 q2(1 + ρ) et,i 2 + q2(1 + 1 2 et,i 2 + 2q2 1 q2 gt,i 2, (9) where (9) is derived by choosing ρ = 1 q2 2q2 and the fact that q < 1. Now by recursion and the initialization e1,i = 0, we have E[ et+1,i 2] 2q2 τ=1 (1 + q2 2 )t τE[ gτ,i 2] (1 q2)2 σ2 + 2q2 τ=1 (1 + q2 2 )t τE[ fi(θτ) 2], which proves the second argument. Meanwhile, the absolute bound et,i 2 4q2 (1 q2)2 G2 follows directly from (9). Lemma 3. For the moving average error sequence Et, it holds that t=1 E[ Et 2] 4Tq2 (1 q2)2 (σ2 + σ2 g) + 4q2 t=1 E[ f(θt) 2]. Published as a conference paper at ICLR 2022 Proof. Denote Kt,i := Pt τ=1( 1+q2 2 )t τE[ fi(θτ) 2]. Using the same technique as in the proof of Lemma 1, denoting et,j as the j-th coordinate of et, it follows that E[ Et 2] = E[ (1 β1) τ=1 βt τ 1 eτ 2] (1 β1)2 d X τ=1 βt τ 1 eτ,j)2] (a) (1 β1)2 d X τ=1 βt τ 1 )( τ=1 βt τ 1 e2 τ,j)] τ=1 βt τ 1 E[ eτ 2] τ=1 βt τ 1 E[ 1 i=1 eτ,i 2] (1 q2)2 σ2 + 2q2(1 β1) τ=1 βt τ 1 ( 1 where (a) is due to Cauchy-Schwartz and (b) is a result of Lemma 2. Summing over t = 1, ..., T and using the technique of geometric series summation leads to t=1 E[ Et 2] 4Tq2 (1 q2)2 σ2 + 2q2(1 β1) τ=1 βt τ 1 ( 1 (1 q2)2 σ2 + 2q2 τ=1 (1 + q2 i=1 fi(θτ) 2] (1 q2)2 σ2 + 4q2 i=1 fi(θt) 2] (1 q2)2 σ2 + 4q2 i=1 fi(θt) 2 + 1 i=1 fi(θt) f(θt) 2] (1 q2)2 (σ2 + σ2 g) + 4q2 t=1 E[ f(θt) 2], where (a) is derived by the variance decomposition and the last inequality holds due to Assumption 4. The desired result is obtained. Lemma 4. It holds that t [T], i [d], ˆvt,i 4(1+q2)3 (1 q2)2 G2. Proof. For any t, by Lemma 2 and Assumption 3 we have gt 2 = C(gt + et) 2 C(gt + et) (gt + et) + (gt + et) 2 2(q2 + 1) gt + et 2 4(q2 + 1)(G2 + 4q2 (1 q2)2 G2) = 4(1 + q2)3 (1 q2)2 G2. It s then easy to show by the updating rule of ˆvt, there exists a j [t] such that ˆvt = vj. Then ˆvt,i = (1 β2) τ=1 βj τ 2 g2 τ,i 4(1 + q2)3 (1 q2)2 G2, Published as a conference paper at ICLR 2022 which concludes the claim. Lemma 5. Let Dt := 1 ˆvt 1+ϵ 1 ˆvt+ϵ be defined as above. Then, t=1 Dt 1 d ϵ, Proof. By the updating rule of COMP-AMS, ˆvt 1 ˆvt for t. Therefore, by the initialization ˆv0 = 0, we have ˆvt 1,i + ϵ 1 p ˆv0,i + ϵ 1 p ˆv T,i + ϵ) For the sum of squared l2 norm, note the fact that for a b > 0, it holds that (a b)2 (a b)(a + b) = a2 b2. ˆvt 1,i + ϵ 1 p ˆvt,i + ϵ)2 i=1 ( 1 ˆvt 1,i + ϵ 1 ˆvt,i + ϵ) which gives the desired result.