# distributed_principal_component_analysis_with_limited_communication__55b8a106.pdf Distributed Principal Component Analysis with Limited Communication Foivos Alimisis Department of Mathematics University of Geneva Peter Davies Department of Computer Science University of Surrey Bart Vandereycken Department of Mathematics University of Geneva Dan Alistarh IST Austria & Neural Magic, Inc. We study efficient distributed algorithms for the fundamental problem of principal component analysis and leading eigenvector computation on the sphere, when the data are randomly distributed among a set of computational nodes. We propose a new quantized variant of Riemannian gradient descent to solve this problem, and prove that the algorithm converges with high probability under a set of necessary spherical-convexity properties. We give bounds on the number of bits transmitted by the algorithm under common initialization schemes, and investigate the dependency on the problem dimension in each case. 1 Introduction The need to process ever increasing quantities of data has motivated significant research interest into efficient distributed algorithms for standard tasks in optimization and data analysis; see, e.g., [16, 27, 11]. In this context, one of the fundamental metrics of interest is communication cost, measured as the number of bits sent and received by the processors throughout computation. Communicationefficient variants are known for several classical tasks and algorithms, from basic ones such as mean estimation [28, 20, 8] to variants of gradient descent [19, 12] or even higher-order optimization [2, 15]. By contrast, much less is known concerning the communication cost for classical problems in data analysis, such as Principal Component Analysis (PCA) [13, 5], or spectral clustering [23]. In this paper, we will focus on the closely-related problem of computing the leading eigenvector of the data covariance matrix, i.e., determining the first principal component, in the setting where d-dimensional data comes from an unknown distribution, and is split among n machines. Given the fundamental nature of PCA, there is a rich body of work on solving sequential variants, both in the Euclidean domain [24, 25, 26] and, more recently, based on Riemannian optimization. This is because maximizing the Rayleigh quotient can be re-interpreted in terms of a hyper-sphere, leading to an unconstrained Riemannian optimization problem [1, 32, 31]. These references have exclusively focused on the sequential setting, where the data can be stored and processed by a single node. However, in many settings the large volume of data requires processing by multiple nodes, which has inspired significant work on reducing the distribution overheads of PCA, e.g. [4, 11], specifically on obtaining efficient algorithms in terms of round complexity. More precisely, an impressive line of work considered the deterministic setting [17, 4, 6, 10], where the input is partitioned arbitrarily between machines, for which both round-efficient and communication-efficient solutions are now known. The general approach is to first perform a careful low-rank approximation of the local data and then communicate iteratively to reconstruct a faithful 35th Conference on Neural Information Processing Systems (Neur IPS 2021). estimate of the covariance matrix. Due to the worst-case setting, the number of communication rounds scales polynomially in the eigengap δ and in the precision ε, making them costly in practice. Thus, on the practical side, it is common to assume a setting with stochastic rather than worst-case data partitioning. Specifically, Garber et al. [11] consider this setting, and note that standard solutions based on distributed Power Method or Lanczos would require a number of communication rounds which increases with the sample size, which is impractical at scale. They circumvent this issue by replacing explicit Power Method iterations with a set of convex optimization problems, which can be solved in a round-efficient manner. Huang and Pan [14] proposed a round-efficient solution, by leveraging the connection to Riemannian optimization. Their proposal relies on a procedure which allows the update of node-local variables via a surrogate gradient, whose deviation from the global gradient should be bounded. Thus, communication could only be used to reach consensus on the local variables among the nodes. Unfortunately, upon close inspection, we have noted significant technical gaps in their analysis, which we discuss in detail in the next section. The line of work in the stochastic setting focused primarily on latency cost, i.e., the number of communication rounds for solving this problem. For highly-dimensional data and/or highly-distributed settings, it is known that bandwidth cost, i.e., the number of bits which need to be transmitted per round, can be a significant bottleneck, and an important amount of work deals with it for other classic problems, e.g. [28, 12, 15]. By contrast, much less is known concerning the communication complexity of distributed PCA. Contribution. In this paper, we provide a new algorithm for distributed leading eigenvector computation, which specifically minimizes the total number of bits sent and received by the nodes. Our approach leverages the connection between PCA and Riemannian optimization; specifically, we propose the first distributed variant of Riemannian gradient descent, which supports quantized communication. We prove convergence of this algorithm utilizing a set of necessary convexity-type properties, and bound its communication complexity under different initializations. At the technical level, our result is based on a new sequential variant of Riemannian gradient descent on the sphere. This may be of general interest, and the algorithm can be generalized to Riemannian manifolds of bounded sectional curvatures using more sophisticated geometric bounds. We strengthen the gradient dominance property proved in [31], by proving that our objective is weakly-quasi-convex in a ball of some minimizer and combining that with a known quadratic-growth condition. We leverage these properties by adapting results of [7] in the Riemannian setting and choosing carefully the learning-rate in order to guarantee convergence for the single-node version. This algorithm is specifically-designed to be amenable to distribution: specifically, we port it to the distributed setting, and add communication-compression in the tangent spaces of the sphere, using a variant of the lattice quantization technique of [8], which we port from the context of standard gradient descent. We provide a non-trivial analysis for the communication complexity of the resulting algorithm under different initializations. We show that, under random initialization, a solution can be reached with total communication that is quadratic in the dimension d, and linear in 1/δ, where δ is the gap between the two biggest eigenvalues; when the initialization is done to the leading eigenvector of one of the local covariance matrices, this dependency can be reduced to linear in d. 2 Setting and Related Work Setting. We assume to be given m total samples coming from some distribution D, organized as a global m d data matrix M, which is partitioned row-wise among n processors, with node i being assigned the matrix Mi, consisting of mi consecutive rows, such that Pn i=1 mi = m. As is common, let A := M T M = Pd i=1 M T i Mi be the global covariance matrix, and Ai := M T i Mi the local covariance matrix owned by node i. We denote by λ1, λ2, ..., λd the eigenvalues of A in descending order and by v1, v2, ..., vd the corresponding eigenvectors. Then, it is well-known that, if the gap δ = λ1 λ2 between the two leading eigenvalues of A is positive, we can approximate the leading eigenvector by solving the following empirical risk minimization problem up to accuracy ε: x = argminx Rd x T Ax = argminx Rd, x 2=1 x T Ax = argminx Sd 1 x T Ax , (1) where Sd 1 is the (d 1)-dimensional sphere. We define f : Sd 1 R, with f(x) = x T Ax and fi : Sd 1 R, with fi(x) = x T Aix. Since the inner product is bilinear, we can write the global cost as the sum of the local costs: f(x) = Pn i=1 fi(x). In our analysis we use notions from the geometry of the sphere, which we recall in Appendix A. Related Work. As discussed, there has been a significant amount of research on efficient variants of PCA and related problems [5, 24, 25, 26, 1, 32, 31]. Due to space constraints, we focus on related work on communication-efficient algorithms. In particular, we discuss the relationship to previous round-efficient algorithms; to our knowledge, this is the first work to specifically focus on the bit complexity of this problem in the setting where data is randomly partitioned. More precisely, previous work on this variant implicitly assumes that algorithms are able to transmit real numbers at unit cost. The straightforward approach to solve the minimization problem (1) in a distributed setting, where the data rows are partitioned, would be to use a distributed version of the Power Method, Riemannian gradient descent (RGD), or the Lanczos algorithm. In order to achieve an ε-approximation of the minimizer x , the latter two algorithms require O(λ1 log(1/ε)/δ) rounds, where the O notation hides poly-logarithmic factors in d. Distributed Lanczos and accelerated RGD would improve this by an O( p λ1/δ) factor. However, Garber et al. [11] point out that, when δ is small, e.g. δ = Θ(1/ m), then unfortunately the number of communication rounds would increase with the sample size, which renders these algorithms non-scalable. Standard distributed convex approaches, e.g. [16, 27], do not directly extend to our setting due to non-convexity and the unit-norm constraint. Garber et al. [11] proposed a variant of the Power Method called Distributed Shift-and-Invert (DSI) which converges in roughly O q b δ m log2(1/ε) log(1/δ) rounds, where b is a bound on the squared ℓ2-norm of the data. Huang and Pan [14] aimed to improve the dependency of the algorithm on ε and δ, and proposed an algorithm called Communication-Efficient Distributed Riemannian Eigensolver (CEDRE). This algorithm is shown to have round complexity O( b δ m log(1/ε)), which does not scale with the sample size for δ = Ω(1/ m), and has only logarithmic dependency on the accuracy ε. Technical issues in [14]. Huang and Pan [14] proposed an interesting approach, which could provide the most round-efficient distributed algorithm to date. Despite the fact that we find the main idea of this paper very creative, we have unfortunately identified a significant gap in their analysis, which we now outline. Specifically, one of their main results, Theorem 3, uses the local gradient dominance shown in [31]; yet, the application of this result is invalid, as it is done without knowing in advance that the iterates of the algorithms continue to remain in the ball of initialization. This is compounded by another error on the constant of gradient dominance (Lemma 2), which we believe is caused by a typo in the last part of the proof of Theorem 4 in [31]. This typo is unfortunately propagated into their proof. More precisely, the objective is indeed gradient dominated, but with a constant which vanishes when we approach the equator, in contrast with the 2/δ which is claimed globally. (We reprove this result independently in Proposition 4). Thus, starting from a point lying in some ball of the minimizer can lead to a new point where the objective is more poorly gradient dominated. This is a non-trivial technical problem which, in the case of gradient descent, can be addressed by a choice of the learning rate depending on the initialization. Given these issues, we perform a new and formally-correct analysis for gradient descent based on new convexity properties which guarantee convergence directly in terms of iterates, and not just function values. We would like however to note that our focus in this paper is on bit and not round complexity. 3 Step 1: Computing the Leading Eigenvector in One Node 3.1 Convexity-like Properties and Smoothness We first analyze the case that all the matrices Ai are simultaneously known to all machines. This practically means that we can run gradient descent in one of the machines (let it be the master), since we have all the data stored there. All the proofs for this section are in Appendix B. Our problem reads min x Sd 1 x T Ax where A = M T M is the global covariance matrix. If δ = λ1 λ2 > 0, this problem has exactly two global minima: v1 and v1. Let x Sd 1 be an arbitrary point. Then x can be written in the form x = Pd i=1 αivi. Fixing the minimizer v1, we have that a ball in Sd 1 around v1 is of the form Ba = {x Sd 1 | α1 a} = {x Sd 1 | x, v1 a} (2) for some a. Without loss of generality, we may assume a > 0 (otherwise, consider the ball around v1 to establish convergence to v1). In this case, the region A as defined in Def. 11 is identical to Ba and x := v1. We begin by investigating the convexity properties of the function x T Ax. In particular, we prove that this function is weakly-quasi-convex (see Def. 13) with constant 2a in the ball Ba. See also Appendix A for a definition of the Riemannian gradient gradf and the logarithm log. Proposition 1. The function f(x) = x T Ax satisfies 2a(f(x) f ) gradf(x), logx(x ) for any x Ba with a > 0. We continue by providing a quadratic growth condition for our cost function, which can also be found in [30, Lemma 2] in a slightly different form. Here dist is the intrinsic distance in the sphere, that we also define in Appendix A. Proposition 2. The function f(x) = x T Ax satisfies 4dist2(x, x ) for any x Ba with a > 0. According to Def. 14, this means that f satisfies quadratic growth with constant µ = δ/2. Next, we prove that quadratic growth and weak-quasi-convexity imply a strong-weak-convexity property. This is similar to [7, Proposition 3.1]. Proposition 3. A function f : Sd 1 R which satisfies quadratic growth with constant µ > 0 and is 2a-weakly-quasi-convex with respect to a minimizer x in a ball of this minimizer, satisfies for any x in the same ball the inequality a gradf(x), logx(x ) µ 2 dist2(x, x ). Interestingly, using Proposition 3, we can recover the gradient dominance property proved in [31, Theorem 4]. Proposition 4. The function f(x) = x T Ax satisfies gradf(x) 2 δa2(f(x) f ) for any x Ba with a > 0. The proof of Proposition 4 can be done independently based only on Proposition 3 and [7, Lemma 3.2]. This verifies that our results until now are optimal regarding the quality of initialization since gradient dominance proved in [31] is optimal in that sense (i.e. the inequalities in the proof are tight). Proposition 5. The function f(x) = x T Ax is geodesically 2λ1-smooth on the sphere. Thus, the smoothness constant of f, as defined in Def. 12, equals γ = 2λ1. Similarly, let γi denote the smoothness constant of fi, which equals twice the largest eigenvalue of Ai. In order to estimate γ in a distributed fashion using only the local data matrices Ai, we shall use the over-approximation γ n maxi=1,...,n γi. 3.2 Convergence We now consider Riemannian gradient descent with learning rate η > 0 starting from a point x(0) Ba: x(t+1) = expx(t)( η gradf(x(t))), where exp is the exponential map of the sphere, defined in Appendix A. Using Proposition 3 and a proper choice of η, we can establish a convergence rate for the instrinsic distance of the iterates to the minimizer. Proposition 6. An iterate of Riemannian gradient descent for f(x) = x T Ax starting from a point x(t) Ba and with step-size η a/γ where γ 2λ1, produces a point x(t+1) that satisfies dist2(x(t+1), x ) (1 aµη) dist2(x(t), x ) for µ = δ/2. Note that this result implies directly that if our initialization x(0) lies in the ball Ba, the distance of x(1) to the center x decreases and thus all subsequent iterates continue being in the initialization ball. This is essential since it guarantees that the convexity-like properties for f continue to hold during the whole optimization process. The proof of Proposition 6 is in Appendix B. This result is also valid in a more general setting when applied to an arbitrary f : Sd 1 R that is 2a-weakly-quasi convex and γ-smooth, and that satisfies quadratic growth with constant µ. The Euclidean analogue of this result can be found in [7, Lemma 4.4]. 4 Step 2: Computing the Leading Eigenvector in Several Nodes We now present our version of distributed gradient descent for leading eigenvector computation and measure its bit complexity until reaching accuracy ϵ in terms of the intrinsic distance of an iterate x(T ) from the minimizer x (x is the leading eigenvector closest to the initialization point). Lattice Quantization. For estimating the Riemannian gradient in a distributed manner with limited communication, we use a quantization procedure developed in [8]. The original quantization scheme involves randomness, but we use a deterministic version of it, by picking up the closest point to the vector that we want to encode. This is similar to the quantization scheme used by [3] and has the following properties. Proposition 7. [8, 3] Denoting by b the number of bits that each machine uses to communicate, there exists a quantization function Q: Rd 1 Rd 1 R+ R+ Rd 1, which, for each w, y > 0, consists of an encoding function encw,y : Rd 1 {0, 1}b and a decoding one decw,y : {0, 1}b Rd 1 Rd 1, such that, for all x, x Rd 1, decw,y(encw,y(x), x ) = Q(x, x , y, w), if x x y. Q(x, x , y, w) x w, if x x y. If y/w > 1, the cost of the quantization procedure in number of bits satisfies b = O((d 1)log2 y w) = O(dlog y In the following, the quantization takes place in the tangent space of each iterate Tx(t)Sd 1, which is linearly isomorphic to Rd 1. We denote by Qx the specification of the function Q at Tx Sd 1. The vector inputs of the function Qx are represented in the local coordinate system of the tangent space that the quantization takes place at each step. For decoding at t > 0, we use information obtained in the previous step, that we need to translate to the same tangent space. We do that using parallel transport Γ, which we define in Appendix A. We present now our main algorithm, which is inspired by quantized gradient descent firstly designed by [22], and its similar version in [3]. 1. Choose an arbitrary machine to be the master node, let it be i0. 2. Choose x(0) Sd 1 (we analyze later specific ways to do that). 3. Consider the following parameters σ := 1 cos(D)µη, K := 2 σ , θ := σ(1 σ) ξ := θK + σ, R(t) = γK( p where D is an over-approximation for dist(x(0), x ) (the way to practically choose D is discussed later). We assume that cos(D)µη 1 2, otherwise we run the algorithm with σ = 1 2. In Tx(0)Sd 1: 4. Compute the local Riemannian gradient gradfi(x(0)) at x(0) in each node. 5. Encode gradfi(x(0)) in each node and decode in the master node using its local information: q(0) i = Qx(0) gradfi(x(0)), gradfi0(x(0)), 2γπ, θR(0) 6. Sum the decoded vectors in the master node: i=1 gradfi(x(0)). 7. Encode the sum in the master and decode in each machine i using its local information: q(0) = Qx(0) r(0), gradfi(x(0)), θR(0) 2 + 2γπ, θR(0) For t 0: 8. Take a gradient step using the exponential map: x(t+1) = expx(t)( ηq(t)) with step-size η (the step-size is discussed later). In Tx(t+1)Sd 1: 9. Compute the local Riemannian gradient gradfi(x(t+1)) at x(t+1) in each node. 10. Encode gradfi(x(t+1)) in each node and decode in the master node using its (parallelly transported) local information from the previous step: q(t+1) i = Qx(t+1) gradfi(x(t+1)), Γx(t+1) x(t) q(t) i , R(t+1) n , θR(t+1) 11. Sum the decoded vectors in the master node: i=1 gradfi(x(t+1)). 12. Encode the sum in the master and decode in each machine using its local information in the previous step after parallel transport: q(t+1) = Qx(t+1) r(t+1), Γx(t+1) x(t) q(t), 1 + θ R(t+1), θR(t+1) Convergence We first control the convergence of iterates simultaneously with the convergence of quantized gradients. Note that p 2 + σ = 1 + σ Lemma 8. If η cos(D) γ , the previous quantized gradient descent algorithm produces iterates x(t) and quantized gradients q(t) that satisfy dist2(x(t), x ) ξt D2, q(t) i gradfi(x(t)) θR(t) 2n , q(t) gradf(x(t)) θR(t). The proof is a Riemannian adaptation of the similar one in [22] and [3] and is presented in Appendix C. We recall that since the sphere is positively curved, it provides a landscape easier for optimization. It is quite direct to derive a general Riemannian method for manifolds of bounded curvature using more advanced geometric bounds, however this exceeds the scope of this work which focuses on leading eigenvector computation. We now move to our main complexity result. Theorem 9. Let η cos(D) γ . Then, the previous quantized gradient descent algorithm needs at most b = O nd 1 cos(D)δη log n cos(D)δη bits in total to estimate the leading eigenvector with an accuracy ϵ measured in intrinsic distance. The proof is based on the previous Lemma 8 in order to count the number of steps that the algorithm needs to estimate the minimizer with accuracy ϵ and Proposition 7 to count the quantization cost in each round. We present the proof again in Appendix C. 5 Step 3: Dependence on Initialization 5.1 Uniformly random initialization The cheapest choice to initialize quantized gradient descent is a point in the sphere chosen uniformly at random. According to Theorem 4 in [31], such a random point x(0) will lie, with probability at least 1 p, in a ball Ba (see Def. 2) where d dist(x(0), x ) arccos c p Here, c is a universal constant (we estimated numerically that c is lower bounded by 1, see Appendix D, thus we can use c = 1). By choosing the step-size η as and the parameter D as D = arccos c p we are guaranteed that η = cos(D) γ , and dist(x(0), x ) D with probability at least 1 p. Our analysis above therefore applies (up to probability 1 p) and the general communication complexity result becomes O nd cos(D)δη log n cos(D)δη log D ηγδη log n ηγδη log D η2γδ log n η2γδ log D Substituting η2 = p2c2 dγ2 , the number of bits satisfies (up to probability 1 p) the upper bound p2 γ δ log ndγ 5.2 Warm Start A reasonable strategy to get a more accurate initialization is to perform an eigenvalue decomposition to one of the local covariance matrices, for instance Ai0 (in the master node i0), and compute its leading eigenvector, let it be vi0. For simplicity we will assume here that each machine hosts the same number of data points mi = m n . Then we communicate vi0 to all machines in order to use the normalized quantized approximation x(0) as initialization. We define: x(0) = Q vi0, vi, vi0 vi , vi0, x x(0) = x(0) where Q is the lattice quantization scheme in Rd (i.e. we quantize the leading eigenvector of the master node as a vector in Rd and then project back to the sphere). The input and output variance in this quantization can be bounded by constants that we can practically estimate (see Appendix D). Proposition 10. Assume that our data are i.i.d. and sampled from a distribution D bounded in ℓ2 norm by a constant h. Given that the eigengap δ, the number of machines n and the total number of data points m satisfy we have that the previous quantization costs O(nd) many bits and x(0), x is lower bounded by a constant with probability at least 1 p. Thus, if bound 4 is satisfied, then the communication complexity becomes many bits in total with probability at least 1 p (notice that this can be further simplified using bound 4). This is because D in Theorem 9 is upper bounded by a constant and the communication cost of quantizing vi0 does not affect the total communication cost. If we can estimate the specific relation in bound 4 (the constant hidden inside Ω), then we can compute estimations of the quantization parameters in the definition of x(0). Condition 4 is quite typical in this literature; see [14] and references therein (beginning of page 2), as we also briefly discussed in the introduction. Notice that m appears in the numerator and not the denominator, only because we deal with the sum of local covariance matrices and not the average, thus our eigengap is m times larger than the eigengap of the normalized covariance matrix. Denoting by δ the eigengap of the normalized covariance matrix, bound 4 can be written equivalently as where mi0 is the number of data points owned by the master node (and any other machine) and Ω hides logarithmic factors from the lower bound. 6 Experiments We evaluate our approach experimentally, comparing the proposed method of Riemannian gradient quantization against three other benchmark methods: Full-precision Riemannian gradient descent: Riemannian gradient descent, as described in Section 3.2, is performed with the vectors communicated at full (64-bit) precision. Euclidean gradient difference quantization: the naïve approach to quantizing Riemannian gradient descent. Euclidean gradients are quantized and averaged before being projected to Riemannian gradients and used to take a step. To improve performance, rather than quantizing Euclidean gradients directly, we quantize the difference between the current local gradient and the previous local gradient, at each node. Since these differences are generally smaller than the gradients themselves, we expect this quantization to introduce lower error. Quantized power iteration: we also use as a benchmark a quantized version of power iteration, a common method for leading-eigenvector computation given by the update rule x(t+1) Ax(t) Ax(t) . Ax(t) can be computed in distributed fashion by communicating and summing the vectors Aix(t), i n. It is these vectors that we quantize. All three of the quantized methods use the same vector quantization routine, for fair comparison. Figure 1 Convergence results on real datasets We show convergence results (Figure 1) for the methods on four real datasets: Human Activity from the MATLAB Statistics and Machine Learning Toolbox, and Mice Protein Expression, Spambase, and Libras Movement from the UCI Machine Learning Repository [9]. All results are averages over 10 runs of the cost function x T Ax (for each method, the iterates x are normalized to lie in the 1-ball, so a lower value of x T Ax corresponds to being closer to the principal eigenvector). All four datasets display similar behavior: our approach of Riemannian gradient quantization outperforms naïve Euclidean gradient quantization, and essentially matches the performance of the full-precision method while communicating only 4 bits per coordinate, for a 16 compression. Power iteration converges slightly faster (as would be expected from the theoretical convergence guarantees), but is much more adversely affected by quantization, and reaches a significantly suboptimal result. We also tested other compression levels, though we do not present figures here due to space constraints: for fewer bits per coordinate (1-3), the performance differences between the four methods are in the same order but more pronounced; however, their convergence curves become more erratic due to increased variance from quantization. For increased number of bits per coordinate, the performance of the quantized methods reaches that of the full-precision methods at between 8-10 bits per coordinate depending on input, and the observed effect of quantized power iteration converging to a suboptimal result gradually diminishes. Our code is publicly available 1. 1https://github.com/IST-DASLab/QRGD 7 Discussion and Future Work We proposed a communication efficient algorithm to compute the leading eigenvector of a covariance data matrix in a distributed fashion using only limited amount of bits. To the best of our knowledge, this is the first algorithm for this problem with bounded bit complexity. We focused on gradient descent since it provides robust convergence in terms of iterates, and can circumvent issues that arise due to a cost function that only has local convexity-like properties; second, it known to be amenable to quantization. We would like briefly to discuss the round complexity of our method, emphasizing inherent round complexity trade-offs due to quantization. Indeed, the round complexity of our quantized RGD algorithm for accuracy ϵ is O( γ a2δ), where ϵ appears inside a logarithm. Specifically, if the initialization x(0) is uniformly random in the sphere (section 5.1), then a = O(1/ d) . If we have warm start (section 5.2), then a = O(1). Both statements hold with high probability. For random initialization, the round complexity of distributed power method is O( λ1 δ ) (see the table at the beginning of page 4 in [11]). This is because the convergence rate of power method is tan(dist(xk, x )) (1 λ1/δ)k tan(dist(x0, x )). The dimension is again present in that rate, because tan(dist(x0, x )) = Θ( d), but it appears inside a logarithm in the final iteration complexity. Since we want to insert quantization, the standard way to do so is by starting from a rate which is contracting (see e.g. [22]). It may in theory be possible to add quantization in a non-contracting algorithm, but we are unaware of such a technique. Thus, we firstly prove that RGD with step-size a γ is contracting (based on weak-convexity and quadratic growth) and then we add quantization, paying the price of having an extra O(1/a2) = O(d) in our round complexity. Thus, our round complexity is slightly worse than distributed power method, but this is in some sense inherent due to the introduction of quantization. Despite this, our algorithm can achieve low bit complexity and it is more robust in perturbations caused by bit limitations, as our experiments suggest. We conclude with some interesting points for future work. First, our choice of step-size is quite conservative since the local convexity properties of the cost function improve when approaching the optimum. Thus, instead of setting the step-size to be η = a/γ, where a is a lower bound for x(0), x , we can re-run the whole analysis with an adaptive step-size η = x(t), x /γ based on an improved value for a from x(t), x . This should improve the dependency on the dimension of our complexity results. Second, the version of Riemannian gradient descent that we have chosen is expensive, since one needs to calculate the exponential map in each step using sin and cos calculations. Practitioners usually use cheaper retraction-based versions, and it would be of interest to study how our analysis can be modified if the exponential map is substituted by a retraction (i.e., any first-order approximation). Third, it would be interesting to see whether the ideas and techniques presented in this work can also be used for other data science problems in a distributed fashion through communication-efficient Riemannian methods. This could be, for example, PCA for the top k eigenvectors using optimization on the Stiefel or Grassmann manifold [1] and low-rank matrix completion on the fixed-rank manifold [29]. Acknowledgments and Disclosure of Funding We would like to thank the anonymous reviewers for helpful comments and suggestions. We also thank Aurelien Lucchi and Antonio Orvieto for fruitful discussions at an early stage of this work. FA is partially supported by the SNSF under research project No. 192363 and conducted part of this work while at IST Austria under the European Union s Horizon 2020 research and innovation programme (grant agreement No. 805223 Scale ML). PD partly conducted this work while at IST Austria and was supported by the European Union s Horizon 2020 programme under the Marie Skłodowska-Curie grant agreement No. 754411. Bibliography [1] P-A Absil, Robert Mahony, and Rodolphe Sepulchre. Optimization algorithms on matrix manifolds. Princeton University Press, 2009. [2] Foivos Alimisis, Peter Davies, and Dan Alistarh. Communication-efficient distributed optimization with quantized preconditioners. In International Conference on Machine Learning. PMLR, 2021. [3] Dan Alistarh and Janne H. Korhonen. Improved communication lower bounds for distributed optimisation. ar Xiv preprint ar Xiv:2010.08222, 2020. [4] Maria-Florina Balcan, Vandana Kanchanapally, Yingyu Liang, and David Woodruff. Improved distributed principal component analysis. ar Xiv preprint ar Xiv:1408.5823, 2014. [5] Christopher M Bishop. Pattern recognition and machine learning. Springer, 2006. [6] Christos Boutsidis, David P Woodruff, and Peilin Zhong. Optimal principal component analysis in distributed and streaming models. In Proceedings of the forty-eighth annual ACM symposium on Theory of Computing, pages 236 249, 2016. [7] Jingjing Bu and Mehran Mesbahi. A note on nesterov s accelerated method in nonconvex optimization: a weak estimate sequence approach, 2020. [8] Peter Davies, Vijaykrishna Gurunathan, Niusha Moshrefi, Saleh Ashkboos, and Dan Alistarh. New bounds for distributed mean estimation and variance reduction. In International Conference on Learning Representations, 2021. [9] Dheeru Dua and Casey Graff. UCI machine learning repository, 2017. [10] Jianqing Fan, Dong Wang, Kaizheng Wang, and Ziwei Zhu. Distributed estimation of principal eigenspaces. Annals of statistics, 47(6):3009, 2019. [11] Dan Garber, Ohad Shamir, and Nathan Srebro. Communication-efficient algorithms for distributed stochastic principal component analysis. In International Conference on Machine Learning, pages 1203 1212. PMLR, 2017. [12] Samuel Horváth, Dmitry Kovalev, Konstantin Mishchenko, Sebastian Stich, and Peter Richtárik. Stochastic distributed learning with gradient quantization and variance reduction. ar Xiv preprint ar Xiv:1904.05115, 2019. [13] Harold Hotelling. Analysis of a complex of statistical variables into principal components. Journal of educational psychology, 24(6):417, 1933. [14] Long-Kai Huang and Sinno Pan. Communication-efficient distributed PCA by Riemannian optimization. In International Conference on Machine Learning, pages 4465 4474. PMLR, 2020. [15] Rustem Islamov, Xun Qian, and Peter Richtárik. Distributed second order methods with fast rates and compressed communication. ar Xiv preprint ar Xiv:2102.07158, 2021. [16] Martin Jaggi, Virginia Smith, Martin Takáˇc, Jonathan Terhorst, Sanjay Krishnan, Thomas Hofmann, and Michael I Jordan. Communication-efficient distributed dual coordinate ascent. ar Xiv preprint ar Xiv:1409.1458, 2014. [17] Ravi Kannan, Santosh Vempala, and David Woodruff. Principal component analysis and higher correlations for distributed data. In Conference on Learning Theory, pages 1040 1057. PMLR, 2014. [18] Hamed Karimi, Julie Nutini, and Mark Schmidt. Linear convergence of gradient and proximalgradient methods under the polyak-łojasiewicz condition, 2020. [19] Jakub Koneˇcn y, H Brendan Mc Mahan, Felix X Yu, Peter Richtárik, Ananda Theertha Suresh, and Dave Bacon. Federated learning: Strategies for improving communication efficiency. ar Xiv preprint ar Xiv:1610.05492, 2016. [20] Jakub Koneˇcn y and Peter Richtárik. Randomized distributed mean estimation: Accuracy vs. communication. Frontiers in Applied Mathematics and Statistics, 4:62, 2018. [21] S. Li. Concise Formulas for the Area and Volume of a Hyperspherical Cap. Asian Journal of Mathematics & Statistics, 4(1):66 70, 2011. [22] Sindri Magnússon, Hossein Shokri-Ghadikolaei, and Na Li. On maintaining linear convergence of distributed learning and optimization under limited communication. IEEE Transactions on Signal Processing, 68:6101 6116, 2020. [23] Andrew Ng, Michael Jordan, and Yair Weiss. On spectral clustering: Analysis and an algorithm. Advances in neural information processing systems, 14:849 856, 2001. [24] Erkki Oja and Juha Karhunen. On stochastic approximation of the eigenvectors and eigenvalues of the expectation of a random matrix. Journal of mathematical analysis and applications, 106(1):69 84, 1985. [25] Ohad Shamir. A stochastic pca and svd algorithm with an exponential convergence rate. In International Conference on Machine Learning, pages 144 152. PMLR, 2015. [26] Ohad Shamir. Fast stochastic algorithms for svd and pca: Convergence properties and convexity. In International Conference on Machine Learning, pages 248 256. PMLR, 2016. [27] Ohad Shamir, Nati Srebro, and Tong Zhang. Communication-efficient distributed optimization using an approximate newton-type method. In International conference on machine learning, pages 1000 1008. PMLR, 2014. [28] Ananda Theertha Suresh, X Yu Felix, Sanjiv Kumar, and H Brendan Mc Mahan. Distributed mean estimation with limited communication. In International Conference on Machine Learning, pages 3329 3337. PMLR, 2017. [29] Bart Vandereycken. Low-rank matrix completion by riemannian optimization. SIAM Journal on Optimization, 23(2):1214 1236, 2013. [30] Zhiqiang Xu, Xin Cao, and Xin Gao. Convergence analysis of gradient descent for eigenvector computation. International Joint Conferences on Artificial Intelligence, 2018. [31] Hongyi Zhang, Sashank J Reddi, and Suvrit Sra. Riemannian svrg: Fast stochastic optimization on riemannian manifolds. ar Xiv preprint ar Xiv:1605.07147, 2016. [32] Hongyi Zhang and Suvrit Sra. First-order methods for geodesically convex optimization. In Conference on Learning Theory, pages 1617 1638. PMLR, 2016.