# drive_onebit_distributed_mean_estimation__f508f6ff.pdf DRIVE: One-bit Distributed Mean Estimation Shay Vargaftik VMware Research shayv@vmware.com Ran Ben Basat University College London r.benbasat@cs.ucl.ac.uk Amit Portnoy Ben-Gurion University amitport@post.bgu.ac.il Gal Mendelson Stanford University galmen@stanford.edu Yaniv Ben-Itzhak VMware Research ybenitzhak@vmware.com Michael Mitzenmacher Harvard University michaelm@eecs.harvard.edu We consider the problem where n clients transmit d-dimensional real-valued vectors using dp1 op1qq bits each, in a manner that allows the receiver to approximately reconstruct their mean. Such compression problems naturally arise in distributed and federated learning. We provide novel mathematical results and derive computationally efficient algorithms that are more accurate than previous compression techniques. We evaluate our methods on a collection of distributed and federated learning tasks, using a variety of datasets, and show a consistent improvement over the state of the art. 1 Introduction In many computational settings, one wishes to transmit a d-dimensional real-valued vector. For example, in distributed and federated learning scenarios, multiple participants (a.k.a. clients) in distributed SGD send gradients to a parameter server that averages them and updates the model parameters accordingly [1]. In these applications and others (e.g., traditional machine learning methods such K-Means and power iteration [2] or other methods such as geometric monitoring [3]), sending approximations of vectors may suffice. Moreover, the vectors dimension d is often large (e.g., in neural networks, d can exceed a billion [4, 5, 6]), so sending compressed vectors is appealing. Indeed, recent works have studied how to send vector approximations using representations that use a small number of bits per entry (e.g., [2, 7, 8, 9, 10, 11]). Further, recent work has shown direct training time reduction from compressing the vectors to one bit per coordinate [12]. Most relevant to our work are solutions that address the distributed mean estimation problem. For example, [2] uses the randomized Hadamard transform followed by stochastic quantization (a.k.a. randomized rounding). When each of the n clients transmits Opdq bits, their Normalized Mean Squared Error (NMSE) is bounded by O log d n . They also show a Op 1 nq bound with Opdq bits via variable-length encoding, albeit at a higher computational cost. The sampling method of [10] yields an O r R n NMSE bound using dp1 op1qq bits in expectation, where r is each coordinate s representation length and R is the normalized average variance of the sent vectors. Recently, researchers proposed to use Kashin s representation [11, 13, 14]. Broadly speaking, it allows representing a d-dimensional vector in (a higher) dimension λ d for some λ ą 1 using small coefficients. This results in an O λ2 NMSE bound, where each client transmits λ dp1 op1qq bits [14]. A recent work [15] suggested an algorithm where if all clients vectors have pairwise distances of at most y P R (i.e., for any client pair c1, c2, it holds that xpc1q xpc2q 2 ď y), the resulting MSE is Opy2q (which is tight with respect to Equal Contribution. 35th Conference on Neural Information Processing Systems (Neur IPS 2021). y) using Op1q bits per coordinate on average. This solution provides a stronger MSE bound when vectors are sufficiently close (and thus y is small) but does not improve the worst-case guarantee. We step back and focus on approximating d-dimensional vectors using dp1 op1qq bits (e.g., one bit per dimension and a lower order overhead). We develop novel biased and unbiased compression techniques based on (uniform as well as structured) random rotations in high-dimensional spheres. Intuitively, after a rotation, the coordinates are identically distributed, allowing us to estimate each coordinate with respect to the resulting distribution. Our algorithms do not require expensive operations, such as variable-length encoding or computing the Kashin s representation, and are fast and easy to implement. We obtain an O 1 n NMSE bound using dp1 op1qq bits, regardless of the coordinates representation length, improving over previous works. Evaluation results indicate that this translates to a consistent improvement over the state of the art in different distributed and federated learning tasks. 2 Problem Formulation and Notation 1b - Vector Estimation. We start by formally defining the 1b - vector estimation problem. A sender, called Buffy, gets a real-valued vector x P Rd and sends it using a dp1 op1qq bits message (i.e., asymptotically one bit per coordinate). The receiver, called Angel, uses the message to derive an estimate px of the original vector x. We are interested in the quantity x px 2 2, which is the sum of squared errors (SSE), and its expected value, the Mean Squared Error (MSE). For ease of exposition, we hereafter assume that x 0 as this special case can be handled with one additional bit. Our goal is to minimize the vector-NMSE (denoted v NMSE), defined as the normalized MSE, i.e., Er x px 2 2s x 2 2 . 1b - Distributed Mean Estimation. The above problem naturally generalizes to the 1b - Distributed Mean Estimation problem. Here, we have a set of n P N clients and a server. Each client c P t1, . . . , nu has its own vector xpcq P Rd, which it sends using a dp1 op1qq-bits message to the server. The server then produces an estimate y xavg P Rd of the average xavg 1 n řn c 1 xpcq with the goal of minimizing its NMSE, defined as the average estimate s MSE normalized by the average norm of the clients original vectors, i.e., Er xavg y xavg 2 2s 1 n řn c 1 xpcq 2 Notation. We use the following notation and definitions throughout the paper: Subscripts. xi denotes the i th coordinate of the vector x, to distinguish it from client c s vector xpcq. Binary-sign. For a vector x P Rd, we denote its binary-sign function as signpxq, where signpxqi 1 if xi ě 0 and signpxqi 1 if xi ă 0. Unit vector. For any (non-zero) real-valued vector x P Rd, we denote its normalized vector by x x x 2 . That is, x and x has the same direction and it holds that x 2 1. Rotation Matrix. A matrix R P Rdˆd is a rotation matrix if RT R I. The set of all rotation matrices is denoted as Opdq. It follows that @R P Opdq: detp Rq P t 1, 1u and @x P Rd: x 2 Rx 2. Random Rotation. A random rotation R is a distribution over all random rotations in Opdq. For ease of exposition, we abuse the notation and given x P Rd denote the random rotation of x by Rpxq Rx, where R is drawn from R. Similarly, R 1pxq R 1x RT x is the inverse rotation. Rotation Property. A quantity that determines the guarantees of our algorithms is Ld R,x Rp xq 2 1 d (note the use of the L1 norm). We show that rotations with high Ld R,x values yield better estimates. Shared Randomness. We assume that Buffy and Angel have access to shared randomness, e.g., by agreeing on a common PRNG seed. Shared randomness is studied both in communication complexity (e.g., [16]) and in communication reduction in machine learning systems (e.g., [2, 7]). In our context, it means that Buffy and Angel can generate the same random rotations without communication. 3 The DRIVE Algorithm We start by presenting DRIVE (Deterministically Round Ing randomly rotated VEctors), a novel 1b - Vector Estimation algorithm. Later, we extend DRIVE to the 1b - Distributed Mean Estimation Algorithm 1 DRIVE Buffy: 1: Compute Rpxq, S. 2: Send S, signp Rpxqq to Angel. z Rpxq 1: Compute z Rpxq S sign Rpxq . 2: Estimate px R 1 z Rpxq . problem. In DRIVE, Buffy uses shared randomness to sample a rotation matrix R R and rotates the vector x P Rd by computing Rpxq Rx. Buffy then calculates S, a scalar quantity we explain below. Buffy then sends S, signp Rpxqq to Angel. As we discuss later, sending S, signp Rpxqq requires dp1 op1qq bits. In turn, Angel computes z Rpxq S signp Rpxqq P t S, Sud. It then uses the shared randomness to generate the same rotation matrix and employs the inverse rotation, i.e., estimates px R 1p z Rpxqq. The pseudocode of DRIVE appears in Algorithm 1. The properties of DRIVE depend on the rotation R and the scale parameter S. We consider both uniform rotations, that provide stronger guarantees, and structured rotations that are orders of magnitude faster to compute. As for the scale S Spx, Rq, its exact formula determines the characteristics of DRIVE s estimate, e.g., having minimal v NMSE or being unbiased. The latter allows us to apply DRIVE to the 1b - Distributed Mean Estimation (Section 4.2) and get an NMSE that decreases proportionally to the number of clients. We now prove a general result on the SSE of DRIVE that applies to any random rotation R and any vector x P Rd. In the following sections, we use this result to obtain the v NMSE when considering specific rotations and scaling methods as well as analyzing their guarantees. Theorem 1. The SSE of DRIVE is: x px 2 2 x 2 2 2 S Rpxq 1 d S2 . Proof. The SSE in estimating Rpxq using z Rpxq equals that of estimating x using px. Therefore, x px 2 2 Rpx pxq 2 2 Rpxq Rppxq 2 2 Rpxq z Rpxq 2 2 Rpxq 2 2 2 A Rpxq, z Rpxq E z Rpxq 2 2 x 2 2 2 A Rpxq, z Rpxq E z Rpxq 2 2 . (1) Next, we have that, A Rpxq, z Rpxq E řd i 1 Rpxqi z Rpxqi S řd i 1 Rpxqi sign Rpxqi S Rpxq 1 , (2) z Rpxq 2 2 řd i 1 z Rpxq 2 i d S2 . (3) Substituting Eq. (2) and Eq. (3) in Eq. (1) yields the result. 4 DRIVE With a Uniform Random Rotation We first consider the thoroughly studied uniform random rotation (e.g., [17, 18, 19, 20, 21]), which we denote by RU. The sampled matrix is denoted by RU RU, that is, RUpxq RU x. An appealing property of a uniform random rotation is that, as we show later, it admits a scaling that results in a low constant v NMSE even with unbiased estimates. 4.1 1b - Vector Estimation Using Theorem 1, we obtain the following result. The result holds for any rotation, including RU. Lemma 1. For any x P Rd, DRIVE s SSE is minimized by S Rpxq 1 d (that is, S Rx 1 determined after R R is sampled). This yields a v NMSE of Er x px 2 2s x 2 2 1 E Ld R,x . Proof. By Theorem 1, to minimize the SSE we require B BS x 2 2 2 S Rpxq 1 d S2 2 Rpxq 1 2 d S 0 , leading to S Rpxq 1 d . Then, the SSE of DRIVE becomes: x px 2 2 x 2 2 2 S Rpxq 1 d S2 x 2 2 2 Rpxq 2 1 d d Rpxq 2 1 d2 x 2 2 Rpxq 2 1 d x 2 2 x 2 2 Rp xq 2 1 d x 2 2 1 Rp xq 2 1 d . Thus, the normalized SSE is x px 2 2 x 2 2 1 Ld R,x. Taking expectation yields the result. Interestingly, for the uniform random rotation, Ld RU,x follows the same distribution for all x. This is because, by the definition of RU, it holds that RUp xq is distributed uniformly over the unit sphere for any x. Therefore DRIVE s v NMSE depends only on the dimension d. We next analyze the v NMSE attainable by the best possible S, as given in Lemma 1, when the algorithm uses RU and is not required to be unbiased. In particular, we state the following theorem whose proof appears in Appendix A.1 (all appendices appear in the Supplementary Material and the extended paper version [22]). Theorem 2. For any x P Rd, the v NMSE of DRIVE with S RUpxq 1 4.2 1b - Distributed Mean Estimation An appealing property of DRIVE with a uniform random rotation, established in this section, is that with a proper scaling parameter S, the estimate is unbiased. That is, for any x P Rd, our scale guarantees that E rpxs x. Unbiasedness is useful when generalizing to the Distributed Mean Estimation problem. Intuitively, when n clients send their vectors, any biased algorithm would result in an NMSE that may not decrease with respect to n. For example, if they have the same input vector, the bias would remain after averaging. Instead, an unbiased encoding algorithm has the property that when all clients act (e.g., use different PRNG seeds) independently, the NMSE decreases proportionally to 1 Another useful property of uniform random rotation is that its distribution is unchanged when composed with other rotations. We use it in the following theorem s proof, given in Appendix A.2. Theorem 3. For any x P Rd, set S x 2 2 RUpxq 1 . Then DRIVE satisfies Erpxs x. Now, we proceed to obtain v NMSE guarantees for DRIVE s unbiased estimate. Lemma 2. For any x P Rd, DRIVE with S x 2 2 RUpxq 1 has a v NMSE of E 1 Ld RU ,x Proof. By Theorem 1, the SSE of the algorithm satisfies: x 2 2 2 S RU x 1 d S2 x 2 2 2 x 2 2 RU x 1 RU x 1 d x 2 2 RU x 1 x 2 2 x 2 RU x 1 2 x 2 2 d x 2 2 RU x 2 1 x 2 2 x 2 2 Normalizing by x 2 2 and taking expectation over RU concludes the proof. Our goal is to derive an upper bound on the above expression and thus upper-bound the v NMSE. Most importantly, we show that even though the estimate is unbiased and we use only a single bit per coordinate, the v NMSE does not increase with the dimension and is bounded by a small constant. In particular, in Appendix A.3, we prove the following: Theorem 4. For any x P Rd, the v NMSE of DRIVE with S x 2 2 RUpxq 1 satisfies: piq For all d ě 2, it is at most 2.92. piiq For all d ě 135, it is at most π p6π3 12π2q ln d 1 This theorem yields strong bounds on the v NMSE. For example, the v NMSE is lower than 1 for d ě 4096 and lower than 0.673 for d ě 105. Finally, we obtain the following corollary, Corollary 1. For any x P Rd, the v NMSE tends to π 2 1 0.571 as d Ñ 8 . Recall that DRIVE s above scale S is a function of both x and the sampled RU. An alternative approach is to deterministically set S to x 2 2 Er RUpxq 1s. As we prove in Appendix A.4, the resulting scale is x 2 pd 1q Bp 1 2 q 2d , where B is the Beta function. Interestingly, this scale no longer depends on x but only on its norm. In the appendix, we also prove that the resulting v NMSE is bounded by π 2 1 for any d. In practice, we find that the benefit is marginal. Finally, with a v NMSE guarantee for the unbiased estimate by DRIVE, we obtain the following key result for the 1b - Distributed Mean Estimation problem, whose proof appears in Appendix A.5. We note that this result guarantees (e.g., see [23]) that distributed SGD, where the participants gradients are compressed with DRIVE, converges at the same asymptotic rate as without compression. Theorem 5. Assume n clients, each with its own vector xpcq P Rd. Let each client independently sample RU,c RU and set its scale to xpcq 2 2 RU,c xpcq 1 . Then, the server average estimate s NMSE satisfies: Er xavg y xavg 2 2s 1 n řn c 1 xpcq 2 n , where v NMSE is given by Lemma 2 and is bounded by Theorem 4. To the best of our knowledge, DRIVE is the first algorithm with a provable NMSE of Op 1 nq for the 1b - Distributed Mean Estimation problem (i.e., with dp1 op1qq bits). In practice, we use only d Op1q bits to implement DRIVE. We use the dp1 op1qq notation to ensure compatibility with the theoretical results; see Appendix B for a discussion. 5 Reducing the v NMSE with DRIVE To reduce the v NMSE further, we introduce the DRIVE algorithm. In DRIVE , we also use a scale parameter, denoted S S px, Rq to differentiate it from the scale S of DRIVE. Here, instead of reconstructing the rotated vector in a symmetric manner, i.e., z Rpxq P S t 1, 1ud, we have that z Rpxq P S tc1, c2ud where c1, c2 are computed using K-Means clustering with K 2 over the d entries of the rotated vector Rpxq. That is, c1, c2 are chosen to minimize the SSE over any choice of two values. This does not increase the (asymptotic) time complexity over the random rotations considered in this paper as solving K-Means for the special case of one-dimensional data is deterministically solvable in Opd log dq (e.g., [24]). Notice that DRIVE still requires dp1 op1qq bits as we communicate p S c1, S c2q and a single bit per coordinate, indicating its nearest centroid. We defer the pseudocode and analyses of DRIVE to Appendix C. We show that with proper scaling, for both the 1b - Vector Estimation and 1b - Distributed Mean Estimation problems, DRIVE yields guarantees that are at least as strong as those of DRIVE. 6 DRIVE with a Structured Random Rotation Uniform random rotation generation usually relies on QR factorization (e.g., see [25]), which requires Opd3q time and Opd2q space. Therefore, uniform random rotation can only be used in practice to rotate low-dimensional vectors. This is impractical for neural network architectures with many millions of parameters. To that end, we continue to analyze DRIVE and DRIVE with the (randomized) Hadamard transform, a.k.a. structured random rotation [2, 26], that admits a fast inplace, parallelizable, Opd log dq time implementation [27, 28, 29]. We start with a few definitions. Definition 1. The Walsh-Hadamard matrix ([30]) H2k P t 1, 1u2kˆ2k is recursively defined via: H2k ˆ H2k 1 H2k 1 H2k 1 H2k 1 and H1 p1q. Also, p 1 ? d Hq T I and detp 1 ? d Hq P r 1, 1s. Definition 2. Let RH denote the rotation matrix HD ? d P Rdˆd, where H is a Walsh-Hadamard matrix and D is a diagonal matrix whose diagonal entries are i.i.d. Rademacher random variables (i.e., taking values uniformly in 1). Then RHpxq RH x 1 ? d H px1 D11, . . . , xd Dddq T is the randomized Hadamard transform of x and R 1 H pxq RT H x DH ? d x is the inverse transform. 6.1 1b - Vector Estimation Recall that the v NMSE of DRIVE, when minimized using S Rpxq 1 d , is 1 E Ld R,x (see Lemma 1). We now bound this quantity of DRIVE with a structured random rotation. Lemma 3. For any dimension d ě 2 and vector x P Rd, the v NMSE of DRIVE with a structured random rotation and scale S RHpxq 1 d is: 1 E Ld RH,x ď 1 Proof. Observe that for all i, E r|RHpxqi|s E ˇˇřd j 1 xj ? d Hij Djj ˇˇ . Since t Hij Djj | j P rdsu are i.i.d. Rademacher random variables we can use the Khintchine inequality [31, 32] which implies that 1 ? 2d x 2 ď E r|RHpxqi|s ď 1 ? d x 2 (see [33, 34] for simplified proofs). We conclude that: E Ld RH,x 1 d E RHp xq 2 1 ı ě 1 d E r RHp xq 1s2 ě 1 d přd i 1 1 ? This bound is sharp since for d ě 2 we have that Ld RH,x 1 2 for x p 1 ? 2, 0, . . . , 0q T . Observe that unlike for the uniform random rotation, E Ld RH,x depends on x. We also note that this bound of 1 2 applies to DRIVE (with scale S 1) as we show in Appendix C.2. 6.2 1b - Distributed Mean Estimation For an arbitrary x P Rd and R, and in particular for RH, the estimates of DRIVE cannot be made unbiased. For example, for x p 2 3q T we have that signp RHpxqq p D11, D11q T and thus { RHpxq S p D11, D11q T . This implies that px R 1 H p { RHpxqq 1 ? 2 D H S p D11, D11q T ? 2 S D p D11, 0q T ? 2 S p D2 11, 0q T p ? 2 S, 0q T . Therefore, Erpxs x regardless of the scale. Nevertheless, we next provide evidence for why when the input vector is high dimensional and admits finite moments, a structured random rotation performs similarly to a uniform random rotation, yielding all the appealing aforementioned properties. Indeed, it is a common observation that the distribution of machine learning workloads and, in particular, neural network gradients are governed by such distributions (e.g., lognormal [35] or normal [36, 37]). We seek to show that at high dimensions, the distribution of RHpxq is sufficiently similar to that of RUpxq. By definition, the distribution of RUpxq is that of a uniformly at random distributed point on a sphere. Previous studies of this distribution for high dimensions (e.g., [38, 39, 40, 41]) have shown that individual coordinates of RUpxq converge to the same normal distribution and that these coordinates are weakly dependent in the sense that the joint distribution of every Op1q-sized subset of coordinates is similar to that of independent normal variables for large d. We hereafter assume that x px1, . . . , xdq, where the xis are i.i.d. and that Erx2 js σ2 and Er|xj|3s ρ ă 8 for all j. We show that RHpxqi converges to the same normal distribution for all i. Let Fi,dpxq be the cumulative distribution function (CDF) of 1 σ RHpxqi and Φ be the CDF of the standard normal distribution. The following lemma, proven in Appendix D.1, shows the convergence. Lemma 4. For all i, RHpxqi converges to a normal variable: supx PR |Fi,dpxq Φpxq| ď 0.409 ρ With this result, we continue to lay out evidence for the weak dependency among the coordinates. We do so by calculating the moments of their joint distribution in increasing subset sizes showing that these moments converge to those of independent normal variables. Previous work has shown that a structured random rotation on vectors with specific distributions results in weakly dependent normal variables. This line of research [42, 43, 44] utilized the Hadamard transform for a different purpose. Their goal was to develop a computationally cheap method to generate independent normally distributed variables from simpler (e.g., uniform) distributions. We apply their analysis to our setting. We partially rely on the following observation that the Hadamard matrix satisfies. Observation 1. ([42]) The Hadamard product (coordinate-wise product), Hxiy Hxℓy, of two rows Hxiy, Hxℓy in the Hadamard matrix yields another row at the matrix Hxiy Hxℓy Hx1 pi 1q pℓ 1qy. Here, pi 1q pℓ 1q is the bitwise xor of the plog dq-sized binary representation of pi 1q and pℓ 1q. It follows that řd j 1 Hij Hℓj řd j 1p Hxiy Hxℓyqj řd j 1 H1 pi 1q pℓ 1q,j . 25 28 211 Dimension Empirical NMSE (n=10 clients) Normal(0,1) 25 28 211 Dimension Lognormal(0,1) 25 28 211 Dimension Exponential(1) DRIVE (Hadamard) DRIVE+ (Hadamard) (π/2 1)/10 0.0571 DRIVE (Uniform) DRIVE+ (Uniform) Figure 1: Distributed mean estimation comparison: each data point is averaged over 104 trials. In each trial, the same (randomly sampled) vector is sent by n 10 clients. We now analyze the moments of the rotated variables, starting with the following observation. It follows from the sign-symmetry of D and matches the joint distribution of i.i.d. normal variables. Observation 2. All odd moments containing RHpxq entries are 0. That is, @q P N, @i1, . . . , i2q 1 P t1, . . . , du : E RHpxqi1 . . . RHpxqi2q 1 0. Therefore, we need to examine only even moments. We start with showing that the second moments also match with the distribution of independent normal variables. Lemma 5. For all i ℓit holds that E rp RH xqi p RH xqℓs 0, whereas E p RH xq2 i σ2. Proof. Since t Djj | j P t1, . . . , duu are sign-symmetric and i.i.d., E p RH xqi p RH xqℓ E 1 dpřd j 1 xj Hij Djjq přd j 1 xj Hℓj Djjq E x2 j 1 d řd j 1 Hij Hℓj. Notice that řd j 1 H1j d and řd j 1 Hij 0 for all i ą 1. Thus, by Observation 1 we get 0 if i ℓand σ2 otherwise. We have established that the coordinates are pairwise uncorrelated. Similar but more involved analysis shows that the same trend continues under the assumption of the existence of x s higher moments. In Appendix D.2 we analyze the 4th moments showing that they indeed approach the 4th moments of independent normal variables with a rate of 1 d; the reader is referred to [42] for further intuition and higher moments analysis. We therefore expect that using DRIVE and DRIVE with Hadamard transform will yield similar results to that of a uniform random rotation at high dimensions and when the input vectors respect the finite moments assumption. In addition to the theoretical evidence, in Figure 1, we show experimental results comparing the measured NMSE for the 1b - Distributed Mean Estimation problem with n 10 clients (all given the same vector so biases do not cancel out) for DRIVE and DRIVE using both uniform and structured random rotations over three different distributions. The results indicate that all variants yield similar NMSEs in reasonable dimensions, in line with the theoretical guarantee of Corollary 1 and Theorem 5. 7 Evaluation We evaluate DRIVE and DRIVE , comparing them to standard and recent state-of-the-art techniques. We consider classic distributed learning tasks as well as federated learning tasks (e.g., where the data distribution is not i.i.d. and clients may change over time). All the distributed tasks are implemented over Py Torch [45] and all the federated tasks are implemented over Tensor Flow Federated [46]. We focus our comparison on vector quantization algorithms and recent sketching techniques and exclude sparsification methods (e.g., [47, 48, 49, 50]) and methods that involve client-side memory since these can often work in conjunction with our algorithms. Datasets. We use MNIST [51, 52], EMNIST [53], CIFAR-10 and CIFAR-100 [54] for image classification tasks; a next-character-prediction task using the Shakespeare dataset [55]; and a next-wordprediction task using the Stack Overflow dataset [56]. Additional details appear in Appendix E.1. Algorithms. Since our focus is on the distributed mean estimation problem and its federated and distributed learning applications, we run DRIVE and DRIVE with the unbiased scale quantities.1 We compare against several alternative algorithms: (1) Fed Avg [1] that uses the full vectors (i.e., each coordinate is represented using a 32-bit float); (2) Hadamard transform followed by 1-bit stochastic Dimension (d) Hadamard + 1-bit SQ Kashin + 1-bit SQ Drive (Uniform) Drive (Uniform) Drive (Hadamard) Drive (Hadamard) 128 0.5308, 0.34 0.2550, 2.12 0.0567, 40.4 0.0547, 40.7 0.0591, 0.36 0.0591, 0.72 8,192 1.3338, 0.57 0.3180, 3.42 0.0571, 5088 0.0571, 5101 0.0571, 0.60 0.0571, 1.06 524,288 2.1456, 0.79 0.3178, 4.69 0.0571, 0.82 0.0571, 1.35 33,554,432 2.9332, 27.1 0.3179, 332 0.0571, 27.2 0.0571, 37.8 Table 1: Empirical NMSE and average per-vector encoding time (in milliseconds, on an RTX 3090 GPU) for distributed mean estimation with n 10 clients (same as in Figure 1) and Lognormal(0,1) distribution. Each entry is a (NMSE, time) tuple and the most accurate result is highlighted in bold. quantization (SQ) [2, 10]; (3) Kashin s representation followed by 1-bit stochastic quantization [11]; (4) Tern Grad [8], which clips coordinates larger than 2.5 times the standard deviation, then performs 1-bit stochastic quantization on the absolute values and separately sends their signs and the maximum coordinate for scale (we note that Tern Grad is a low-bit variant of a well-known algorithm called QSGD [9], and we use Tern Grad since we found it to perform better in our experiments);2 and (5-6) Sketched-SGD [57] and Fetch SGD [58], which are both count-sketch [59] based algorithms designed for distributed and federated learning, respectively. We note that Hadamard with 1-bit stochastic quantization is our most fair comparison, as it uses the same number of bits as DRIVE (and slightly more than DRIVE) and has similar computational costs. This contrasts with Kashin s representation, where both the number of bits and the computational costs are higher. For example, a standard Tensor Flow Federated implementation (e.g., see CLASS KASHINHADAMARDENCODINGSTAGE hyperparameters at [60]) uses a minimum of 1.17 bits per coordinate, and three iterations of the algorithm resulting in five Hadamard transforms for each vector. Also, note that Tern Grad uses an extra bit per coordinate for sending the sign. Moreover, the clipping performed by Tern Grad is a heuristic procedure, which is orthogonal to our work. For each task, we use a subset of datasets and the most relevant competition. Detailed configuration information and additional results appear in Appendix E. We first evaluate the v NMSE-Speed tradeoffs and then proceed to federated and distributed learning experiments. v NMSE-Speed Tradeoff. Appearing in Table 1, the results show that our algorithms offer the lowest NMSE and that the gap increases with the dimension. As expected, DRIVE and DRIVE with uniform rotation are more accurate for small dimensions but are significantly slower. Similarly, DRIVE is as accurate as DRIVE , and both are significantly more accurate than Kashin (by a factor of 4.4ˆ-5.5ˆ) and Hadamard (9.3ˆ-51ˆ) with stochastic quantization. Additionally, DRIVE is 5.7ˆ-12ˆ faster than Kashin and about as fast as Hadamard. In Appendix E.3 we discuss the results, give the complete experiment specification, and provide measurements on a commodity machine. We note that the above techniques, including DRIVE, are more computationally expensive than lineartime solutions like Tern Grad. Nevertheless, DRIVE s computational overhead becomes insignificant for modern learning tasks. For example, our measurements suggest that it can take 470 ms for computing the gradient on a Res Net18 architecture (for CIFAR100, batch size = 128, using NVIDIA Ge Force GTX 1060 (6GB) GPU) while the encoding of DRIVE (Hadamard) takes 2.8 ms. That is, the overall computation time is only increased by 0.6% while the error reduces significantly. Taking the transmission and model update times into consideration would reduce the importance of the compression time further. Federated Learning. We evaluate over four tasks: (1) EMNIST over customized CNN architecture with two convolutional layers with 1.2M parameters [61]; (2) CIFAR-100 over Res Net-18 [54] ; (3) a next-character-prediction task using the Shakespeare dataset [1]; (4) a next-word-prediction task using the Stack Overflow dataset [62]. Both (3) and (4) use LSTM recurrent models [63] with 820K and 4M parameters, respectively. We use code, client partitioning, models, hyperparameters, and validation metrics from the federated learning benchmark of [62]. 1For DRIVE the scale is S x 2 2 Rpxq 1 (see Theorem 3). For DRIVE the scale is S x 2 2 c 2 2 , where c P tc1, c2ud is the vector indicating the nearest centroid to each coordinate in Rpxq (see Section 5). 2When restricted to two quantization levels, Tern Grad is identical to QSGD s max normalization variant with clipping (slightly better due to the ability to represent 0). 2000 4000 6000 0 500 1000 1500 400 800 1200 .42 .58 Shakespeare 500 1000 1500 Stack Overflow 5960 5980 6000 .46 1460 1480 1500 .842 1160 1180 1200 .545 1460 1480 1500 .23 Fed Avg (32-bit floats) Kashin + 1-bit SQ Tern Grad Drive (Hadamard) Fetch SGD Drive+ (Hadamard) Hadamard + 1-bit SQ Figure 2: Accuracy per round on various federated learning tasks. Smoothing is done using a rolling mean with a window size of 150. The second row zooms-in on the last 50 rounds. 100 125 150 CIFAR-10 - Zoom In .75 CIFAR-100 200 225 250 .6 CIFAR-100 - Zoom In Mini-batch SGD (32-bit floats) Kashin + 1-bit SQ Tern Grad Drive (Hadamard) Sketched-SGD Drive+ (Hadamard) Hadamard + 1-bit SQ Figure 3: Accuracy per round on distributed learning tasks, with a zoom-in on the last 50 rounds. The results are depicted in Figure 2. We observe that in all tasks, DRIVE and DRIVE have accuracy that is competitive with that of the baseline, Fed Avg. In CIFAR-100, Tern Grad and DRIVE provide the best accuracy. For the other tasks, DRIVE and DRIVE have the best accuracy, while the best alternative is either Kashin + 1-bit SQ or Tern Grad, depending on the task. Hadamard + 1-bit SQ, which is the most similar to our algorithms (in terms of both bandwidth and compute), provides lower accuracy in all tasks. Additional details and hyperparameter configurations are presented in Appendix E.4. Distributed CNN Training. We evaluate distributed CNN training with 10 clients in two configurations: (1) CIFAR-10 dataset with Res Net-9; (2) CIFAR-100 with Res Net-18 [54, 64]. In both tasks, DRIVE and DRIVE have similar accuracy to Fed Avg, closely followed by Kashin + 1-bit SQ. The other algorithms are less accurate, with Hadamard + 1-bit SQ being better than Sketched-SGD and Tern Grad for both tasks. Additional details and hyperparameter configurations are presented in Appendix E.5. Figure 3 depicts the results. Evaluation Summary. Overall, it is evident that DRIVE and DRIVE consistently offer markedly favorable results in comparison to the alternatives in our setting. Kashin s representation appears to offer the best competition, albeit at somewhat higher computational complexity and bandwidth requirements. The lesser performance of the sketch-based techniques is attributed to the high noise of the sketch under such a low (dp1 op1qq bits) communication requirement. This is because the number of counters they can use is too low, making too many coordinates map into each counter. In Appendix E.6, we also compare DRIVE and DRIVE to state of the art techniques over K-Means and Power Iteration tasks for 10, 100, and 1000 clients, yielding similar trends. 8 Discussion Proven Error Bounds. We summarize the proven error bounds in Table 2. Since DRIVE (Hadamard) is generally not unbiased (as discussed in Section 6.2), we cannot establish a formal guarantee for the 1b - DME problem when using Hadamard. It is a challenging research question whether there exists other structured rotations with low computational complexity and stronger guarantees. Scale Rotation Problem S Uniform Hadamard 1b - VE Rpxq 1 d v NMSE 1 2 1b - DME x 2 2 Rpxq 1 NMSE ď 1 n 2.92; d ě 135 ùñ NMSE ď 1 n ˆ π 2 1 b p6π3 12π2q ln d 1 Table 2: Summary of the proven error bounds for DRIVE. Input Distribution Assumption. The distributed mean estimation analysis of our Hadamardbased variants is based on an assumption (Section 6.2) about the vector distributions. While machine learning workloads, and DNN gradients in particular (e.g., [35, 36, 37]), were observed to follow such distributions, this assumption may not hold for other applications. For such cases, we note that DRIVE is compatible with the error feedback (EF) mechanism [65, 66] that ensured convergence and recovery of the convergence rate of non-compressed SGD. Specifically, as evident by Lemma 3, any scale Rpxq 1 d ď S ď 2 Rpxq 1 d is sufficient to respect the compressor (i.e., bounded variance) assumption. For completeness, in Appendix E.2, we perform EF experiments comparing DRIVE and DRIVE to other compression techniques that use EF. Varying Communication Budget. Unlike some previous works, our algorithms guarantees with more than one bit per coordinate are not established. It is thus an interesting future work to extend DRIVE to other communication budgets and understand what are the resulting guarantees. We refer the reader to [67] for initial steps towards that direction. Entropy Encoding. Entropy encoding methods (such as Huffman coding) can further compress vectors of values when the values are not uniformly distributed. We have compared DRIVE against stochastic quantization methods using entropy encoding for the challenging setting for DRIVE where all vectors are the same (see Table 1 for further description). The results appear in Appendix E.7, where DRIVE still outperforms these methods. We also note that, when computation allows and when using DRIVE with multiple bits per entry, DRIVE can also be enhanced by entropy encoding techniques. We describe some initial results for this setting in [67]. Structured Data. When the data is highly sparse, skewed, or otherwise structured, one can leverage that for compression. We note that some techniques that exploit sparsity or structure can be use in conjunction with our techniques. For example, one may transmit only non-zero entries or Top-K entries while compressing these using DRIVE to reduce communication overhead even further. Compatibility With Distributed All-Reduce Techniques. Quantization techniques, including DRIVE, may introduce overheads in the context of All-Reduce (depending on the network architecture and communication patterns). In particular, if every node in a cluster uses a different rotation, DRIVE will not allow for efficient in-path aggregation without decoding the vectors. Further, the computational overhead of the receiver increases by a log d factor as each vector has to be decoded separately before an average can be computed. It is an interesting future direction for DRIVE to understand how to minimize such potential overheads. For example, one can consider bucketizing co-located workers and apply DRIVE s quantization only for cross-rack traffic. 9 Conclusions In this paper, we studied the vector and distributed mean estimation problems. These problems are applicable to distributed and federated learning, where clients communicate real-valued vectors (e.g., gradients) to a server for averaging. To the best of our knowledge, our algorithms are the first with a provable error of Op 1 nq for the 1b - Distributed Mean Estimation problem (i.e., with dp1 op1qq bits). As shown in [14], any algorithm that uses Opdq shared random bits (e.g., our Hadamard-based variant) has a v NMSE of Ωp1q, i.e., DRIVE and DRIVE are asymptotically optimal; additional discussion is given in Appendix F. Our experiments, carried over various tasks and datasets, indicate that our algorithms improve over the state of the art. All the results presented in this paper are fully reproducible by our source code, available at [29]. Acknowledgments and Disclosure of Funding MM was supported in part by NSF grants CCF-2101140, CCF-2107078, CCF-1563710, and DMS2023528. MM and RBB were supported in part by a gift to the Center for Research on Computation and Society at Harvard University. AP was supported in part by the Cyber Security Research Center at Ben-Gurion University of the Negev. We thank Moshe Gabel, Mahmood Sharif, Yuval Filmus and, the anonymous reviewers for helpful comments and suggestions. [1] H. Brendan Mc Mahan, Eider Moore, Daniel Ramage, Seth Hampson, and Blaise Agüera y Arcas. Communication-Efficient Learning of Deep Networks from Decentralized Data. In Artificial Intelligence and Statistics, pages 1273 1282, 2017. [2] Ananda Theertha Suresh, X Yu Felix, Sanjiv Kumar, and H Brendan Mc Mahan. Distributed Mean Estimation With Limited Communication. In International Conference on Machine Learning, pages 3329 3337. PMLR, 2017. [3] Yuval Alfassi, Moshe Gabel, Gal Yehuda, and Daniel Keren. A Distance-Based Scheme for Reducing Bandwidth in Distributed Geometric Monitoring. In 2021 IEEE 37th International Conference on Data Engineering (ICDE), 2021. [4] Jeffrey Dean, Greg Corrado, Rajat Monga, Kai Chen, Matthieu Devin, Mark Mao, Marc' aurelio Ranzato, Andrew Senior, Paul Tucker, Ke Yang, Quoc Le, and Andrew Ng. Large Scale Distributed Deep Networks. In F. Pereira, C. J. C. Burges, L. Bottou, and K. Q. Weinberger, editors, Advances in Neural Information Processing Systems, volume 25. Curran Associates, Inc., 2012. [5] Mohammad Shoeybi, Mostofa Patwary, Raul Puri, Patrick Le Gresley, Jared Casper, and Bryan Catanzaro. Megatron-lm: Training multi-billion parameter language models using model parallelism. ar Xiv preprint ar Xiv:1909.08053, 2019. [6] Yanping Huang, Youlong Cheng, Ankur Bapna, Orhan Firat, Dehao Chen, Mia Chen, Hyouk Joong Lee, Jiquan Ngiam, Quoc V Le, Yonghui Wu, and zhifeng Chen. GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism. In H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 32. Curran Associates, Inc., 2019. [7] Ran Ben-Basat, Michael Mitzenmacher, and Shay Vargaftik. How to Send a Real Number Using a Single Bit (and Some Shared Randomness). ar Xiv preprint ar Xiv:2010.02331, 2020. [8] Wei Wen, Cong Xu, Feng Yan, Chunpeng Wu, Yandan Wang, Yiran Chen, and Hai Li. Tern Grad: Ternary Gradients to Reduce Communication in Distributed Deep Learning. In Advances in neural information processing systems, pages 1509 1519, 2017. [9] Dan Alistarh, Demjan Grubic, Jerry Li, Ryota Tomioka, and Milan Vojnovic. QSGD: Communication-Efficient Sgd via Gradient Quantization and Encoding. Advances in Neural Information Processing Systems, 30:1709 1720, 2017. [10] Jakub Koneˇcn y and Peter Richtárik. Randomized Distributed Mean Estimation: Accuracy vs. Communication. Frontiers in Applied Mathematics and Statistics, 4:62, 2018. [11] Sebastian Caldas, Jakub Koneˇcný, H Brendan Mc Mahan, and Ameet Talwalkar. Expanding the Reach of Federated Learning by Reducing Client Resource Requirements. ar Xiv preprint ar Xiv:1812.07210, 2018. [12] Youhui Bai, Cheng Li, Quan Zhou, Jun Yi, Ping Gong, Feng Yan, Ruichuan Chen, and Yinlong Xu. Gradient compression supercharged high-performance data parallel dnn training. In The 28th ACM Symposium on Operating Systems Principles (SOSP 2021), 2021. [13] Yurii Lyubarskii and Roman Vershynin. Uncertainty Principles and Vector Quantization. IEEE Transactions on Information Theory, 56(7):3491 3501, 2010. [14] Mher Safaryan, Egor Shulgin, and Peter Richtárik. Uncertainty Principle for Communication Compression in Distributed and Federated Learning and the Search for an Optimal Compressor. ar Xiv preprint ar Xiv:2002.08958, 2020. [15] Peter Davies, Vijaykrishna Gurunanthan, Niusha Moshrefi, Saleh Ashkboos, and Dan Alistarh. New Bounds For Distributed Mean Estimation and Variance Reduction. In International Conference on Learning Representations, 2021. [16] Ilan Newman. Private vs. Common Random Bits in Communication Complexity. Information processing letters, 39(2):67 71, 1991. [17] Francesco Mezzadri. How to Generate Random Matrices From the Classical Compact Groups. ar Xiv preprint math-ph/0609050, 2006. [18] RWM Wedderburn. Generating Random Rotations. Research Report, 1975. [19] Richard M Heiberger. Generation of Random Orthogonal Matrices. Journal of the Royal Statistical Society: Series C (Applied Statistics), 27(2):199 206, 1978. [20] Gilbert W Stewart. The Efficient Generation of Random Orthogonal Matrices With an Application to Condition Estimators. SIAM Journal on Numerical Analysis, 17(3):403 409, 1980. [21] Martin A Tanner and Ronald A Thisted. Remark as r42: A Remark on as 127. Generation of Random Orthogonal Matrices. Journal of the Royal Statistical Society. Series C (Applied Statistics), 31(2):190 192, 1982. [22] Shay Vargaftik, Ran Ben Basat, Amit Portnoy, Gal Mendelson, Yaniv Ben-Itzhak, and Michael Mitzenmacher. Drive: One-bit distributed mean estimation. ar Xiv preprint ar Xiv:2105.08339, 2021. [23] Aleksandr Beznosikov, Samuel Horváth, Peter Richtárik, and Mher Safaryan. On biased compression for distributed learning. ar Xiv preprint ar Xiv:2002.12410, 2020. [24] Allan Grønlund, Kasper Green Larsen, Alexander Mathiasen, Jesper Sindahl Nielsen, Stefan Schneider, and Mingzhou Song. Fast Exact K-Means, K-Medians and Bregman Divergence Clustering in 1D. ar Xiv preprint ar Xiv:1701.07204, 2017. [25] Mario Lezcano Casado. Geo Torch s Documentation. A Library for Constrained Optimization and Manifold Optimization for Deep Learning in Pytorch. https://geotorch.readthedocs. io/en/latest/_modules/geotorch/so.html, 2021. accessed 17-May-21. [26] Nir Ailon and Bernard Chazelle. The Fast Johnson Lindenstrauss Transform and Approximate Nearest Neighbors. SIAM Journal on computing, 39(1):302 322, 2009. [27] Bernard J. Fino and V. Ralph Algazi. Unified matrix treatment of the fast walsh-hadamard transform. IEEE Transactions on Computers, 25(11):1142 1146, 1976. [28] Chunyuan Li, Heerad Farkhoor, Rosanne Liu, and Jason Yosinski. Measuring the intrinsic dimension of objective landscapes. In International Conference on Learning Representations, 2018. Code available at: https://github.com/uber-research/intrinsic-dimension. [29] Shay Vargaftik, Ran Ben Basat, Amit Portnoy, Gal Mendelson, Yaniv Ben-Itzhak, and Michael Mitzenmacher. DRIVE open source code. https://github.com/amitport/ DRIVE-One-bit-Distributed-Mean-Estimation, 2021. [30] Kathy J Horadam. Hadamard Matrices and Their Applications. Princeton university press, 2012. [31] Aleksandr Khintchine. Über Dyadische Brüche. Mathematische Zeitschrift, 18(1):109 116, 1923. [32] S Szarek. On the Best Constants in the Khinchin Inequality. Studia Mathematica, 2(58):197 208, 1976. [33] Yuval Filmus. Khintchine-Kahane using Fourier Analysis. Posted at http: // www. cs. toronto. edu/ yuvalf/ KK. pdf , 2012. [34] Rafał Latała and Krzysztof Oleszkiewicz. On the Best Constant In the Khinchin-Kahane Inequality. Studia Mathematica, 109(1):101 104, 1994. [35] Brian Chmiel, Liad Ben-Uri, Moran Shkolnik, Elad Hoffer, Ron Banner, and Daniel Soudry. Neural Gradients Are Near-Lognormal: Improved Quantized and Sparse Training. ar Xiv preprint ar Xiv:2006.08173, 2020. [36] Ron Banner, Yury Nahshan, Elad Hoffer, and Daniel Soudry. Post-Training 4-Bit Quantization of Convolution Networks for Rapid-Deployment. ar Xiv preprint ar Xiv:1810.05723, 2018. [37] Xucheng Ye, Pengcheng Dai, Junyu Luo, Xin Guo, Yingjie Qi, Jianlei Yang, and Yiran Chen. Accelerating CNN Training by Pruning Activation Gradients. In European Conference on Computer Vision, pages 322 338. Springer, 2020. [38] Marcus Spruill et al. Asymptotic Distribution of Coordinates on High Dimensional Spheres. Electronic communications in probability, 12:234 247, 2007. [39] Persi Diaconis and David Freedman. A Dozen de Finetti-Style Results in Search of a Theory. In Annales de l IHP Probabilités et statistiques, volume 23, pages 397 423, 1987. [40] Svetlozar T Rachev, L Ruschendorf, et al. Approximate Independence of Distributions on Spheres and Their Stability Properties. The Annals of Probability, 19(3):1311 1337, 1991. [41] Adriaan J Stam. Limit Theorems for Uniform Distributions on Spheres in High-Dimensional Euclidean Spaces. Journal of Applied probability, 19(1):221 228, 1982. [42] Charles M Rader. A New Method of Generating Gaussian Random Variables by Computer. Technical report, Massachusetts Institute of Technology, Lincoln Lab, 1969. [43] David B Thomas. Parallel Generation of Gaussian Random Numbers Using the Table-Hadamard Transform. In 2013 IEEE 21st Annual International Symposium on Field-Programmable Custom Computing Machines, pages 161 168. IEEE, 2013. [44] Tamás Herendi, Thomas Siegl, and Robert F Tichy. Fast Gaussian Random Number Generation Using Linear Transformations. Computing, 59(2):163 181, 1997. [45] Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas Kopf, Edward Yang, Zachary De Vito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. Py Torch: An Imperative Style, High-Performance Deep Learning Library. In H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett, editors, Advances in Neural Information Processing Systems 32, pages 8026 8037. Curran Associates, Inc., 2019. [46] Google. Tensor Flow Federated: Machine Learning on Decentralized Data. https://www. tensorflow.org/federated, 2020. accessed 25-Mar-20. [47] Jakub Koneˇcný, H. Brendan Mc Mahan, Felix X. Yu, Peter Richtárik, Ananda Theertha Suresh, and Dave Bacon. Federated Learning: Strategies for Improving Communication Efficiency, 2017. [48] Hongyi Wang, Scott Sievert, Shengchao Liu, Zachary B. Charles, Dimitris S. Papailiopoulos, and Stephen Wright. ATOMO: Communication-efficient Learning via Atomic Sparsification. In Neur IPS, pages 9872 9883, 2018. [49] Thijs Vogels, Sai Praneeth Karimireddy, and Martin Jaggi. Power SGD: Practical Low-Rank Gradient Compression for Distributed Optimization. In H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 32, 2019. [50] Sebastian U Stich, Jean-Baptiste Cordonnier, and Martin Jaggi. Sparsified SGD with Memory. In S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 31. Curran Associates, Inc., 2018. [51] Yann Le Cun, Léon Bottou, Yoshua Bengio, and Patrick Haffner. Gradient-Based Learning Applied to Document Recognition. Proceedings of the IEEE, 86(11):2278 2324, 1998. [52] Yann Le Cun, Corinna Cortes, and CJ Burges. Mnist handwritten digit database. ATT Labs [Online]. Available: http://yann.lecun.com/exdb/mnist, 2, 2010. [53] Gregory Cohen, Saeed Afshar, Jonathan Tapson, and Andre Van Schaik. EMNIST: Extending MNIST to Handwritten Letters. In 2017 International Joint Conference on Neural Networks (IJCNN), pages 2921 2926. IEEE, 2017. [54] Alex Krizhevsky, Geoffrey Hinton, et al. Learning Multiple Layers of Features From Tiny Images. 2009. [55] William Shakespeare. The Complete Works of William Shakespeare. https://www. gutenberg.org/ebooks/100. [56] Stack Overflow Data. https://www.kaggle.com/stackoverflow/stackoverflow. accessed 01-Mar-21. [57] Nikita Ivkin, Daniel Rothchild, Enayat Ullah, Vladimir Braverman, Ion Stoica, and Raman Arora. Communication-Efficient Distributed Sgd With Sketching. ar Xiv preprint ar Xiv:1903.04488, 2019. [58] Daniel Rothchild, Ashwinee Panda, Enayat Ullah, Nikita Ivkin, Ion Stoica, Vladimir Braverman, Joseph Gonzalez, and Raman Arora. Fetch SGD: Communication-Efficient Federated Learning With Sketching. In International Conference on Machine Learning, pages 8253 8265. PMLR, 2020. [59] Moses Charikar, Kevin Chen, and Martin Farach-Colton. Finding frequent items in data streams. In International Colloquium on Automata, Languages, and Programming, pages 693 703. Springer, 2002. [60] The Tensor Flow Authors. Tensor Flow Federated: Compression via Kashin s representation from Hadamard transform. https://github.com/tensorflow/model-optimization/blob/ 9193d70f6e7c9f78f7c63336bd68620c4bc6c2ca/tensorflow_model_optimization/ python/core/internal/tensor_encoding/stages/research/kashin.py#L92. accessed 16-May-21. [61] Sebastian Caldas, Sai Meher Karthik Duddu, Peter Wu, Tian Li, Jakub Koneˇcný, H. Brendan Mc Mahan, Virginia Smith, and Ameet Talwalkar. LEAF: A Benchmark for Federated Settings, 2019. [62] Sashank Reddi, Zachary Charles, Manzil Zaheer, Zachary Garrett, Keith Rush, Jakub Koneˇcný, Sanjiv Kumar, and H. Brendan Mc Mahan. Adaptive Federated Optimization, 2020. [63] S. Hochreiter and J. Schmidhuber. Long Short-Term Memory. Neural Computation, 9:1735 1780, 1997. [64] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770 778, 2016. [65] Frank Seide, Hao Fu, Jasha Droppo, Gang Li, and Dong Yu. 1-Bit Stochastic Gradient Descent and Its Application to Data-Parallel Distributed Training of Speech DNNs. In Fifteenth Annual Conference of the International Speech Communication Association, 2014. [66] Sai Praneeth Karimireddy, Quentin Rebjock, Sebastian Stich, and Martin Jaggi. Error Feedback Fixes Sign SGD and other Gradient Compression Schemes. In International Conference on Machine Learning, pages 3252 3261, 2019. [67] Shay Vargaftik, Ran Ben Basat, Amit Portnoy, Gal Mendelson, Yaniv Ben-Itzhak, and Michael Mitzenmacher. Communication-efficient federated learning via robust distributed mean estimation. ar Xiv preprint ar Xiv:2108.08842, 2021. [68] Mervin E Muller. A Note on a Method for Generating Points Uniformly on N-Dimensional Spheres. Communications of the ACM, 2(4):19 20, 1959. [69] George Bennett. Probability Inequalities for the Sum of Independent Random Variables. Journal of the American Statistical Association, 57(297):33 45, 1962. [70] Andreas Maurer et al. A Bound on the Deviation Probability for Sums of Non-Negative Random Variables. J. Inequalities in Pure and Applied Mathematics, 4(1):15, 2003. [71] Paweł J Szabłowski. Uniform Distributions on Spheres in Finite Dimensional Lα and Their Generalizations. Journal of multivariate analysis, 64(2):103 117, 1998. [72] Carl-Gustav Esseen. A Moment Inequality With an Application to the Central Limit Theorem. Scandinavian Actuarial Journal, 1956(2):160 170, 1956. [73] Google. Tensor Flow Model Optimization Toolkit. https://www.tensorflow.org/model_ optimization, 2021. accessed 19-Mar-21. [74] Apache license, version 2.0. https://www.apache.org/licenses/LICENSE-2.0. accessed 19-Mar-21. [75] Pytorch license. https://github.com/pytorch/pytorch/blob/ aaccdc39965ade4b61b7852329739e777e244c25/LICENSE. accessed 19-Mar-21. [76] Creative Commons Attribution-Share Alike 3.0 license. https://creativecommons.org/ licenses/by-sa/3.0/. accessed 01-Mar-21. [77] The Tensor Flow Authors. Tensor Flow Federated: Encoded Sum Factory class. https://github.com/tensorflow/federated/blob/v0.19.0/tensorflow_ federated/python/aggregators/encoded.py#L40-L142. accessed 19-May-21. 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] (b) Did you describe the limitations of your work? [Yes] See Section 9. (c) Did you discuss any potential negative societal impacts of your work? [No] We do not believe there is an inherit negative societal impact in this work. (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [Yes] (b) Did you include complete proofs of all theoretical results? [Yes] 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] In the Supplemental Material. It will be released as open-source after publication. (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] see Section 7 and Appendix E. (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [No] Most experiments can be easily reproduced but require considerable execution time and compute resources. We did include error bars in Figure 1, where it was feasible. (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] See Section 7 and Appendix E. 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [Yes] See Section 7 and Appendix E. (b) Did you mention the license of the assets? [Yes] See Section 7 and Appendix E. (c) Did you include any new assets either in the supplemental material or as a URL? [Yes] We include the source code of our algorithms in the Supplemental Material. It will be released as open-source after publication. (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [N/A] We used well-known datasets whose creators use either publicly available data or received proper consent. (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [N/A] We used well-known datasets, which do not contain personally identifiable information or offensive content. 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] We did not use crowdsourcing or conducted research with human subjects. (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] We did not use crowdsourcing or conducted research with human subjects. (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A] We did not use crowdsourcing or conducted research with human subjects.