# federated_multiobjective_learning__7caf1bd3.pdf Federated Multi-Objective Learning Haibo Yang Dept. of Comput. & Info. Sci. Rochester Institute of Technology Rochester, NY 14623 hbycis@rit.edu Zhuqing Liu Dept. of ECE The Ohio State University Columbus,OH 43210 liu.9384@osu.edu Jia Liu Dept. of ECE The Ohio State University Columbus,OH 43210 liu@ece.osu.edu Chaosheng Dong Amazon.com Inc. Seattle, WA 98109 chaosd@amazon.com Michinari Momma Amazon.com Inc. Seattle, WA 98109 michi@amazon.com In recent years, multi-objective optimization (MOO) emerges as a foundational problem underpinning many multi-agent multi-task learning applications. However, existing algorithms in MOO literature remain limited to centralized learning settings, which do not satisfy the distributed nature and data privacy needs of such multi-agent multi-task learning applications. This motivates us to propose a new federated multi-objective learning (FMOL) framework with multiple clients distributively and collaboratively solving an MOO problem while keeping their training data private. Notably, our FMOL framework allows a different set of objective functions across different clients to support a wide range of applications, which advances and generalizes the MOO formulation to the federated learning paradigm for the first time. For this FMOL framework, we propose two new federated multi-objective optimization (FMOO) algorithms called federated multi-gradient descent averaging (FMGDA) and federated stochastic multi-gradient descent averaging (FSMGDA). Both algorithms allow local updates to significantly reduce communication costs, while achieving the same convergence rates as those of their algorithmic counterparts in the single-objective federated learning. Our extensive experiments also corroborate the efficacy of our proposed FMOO algorithms. 1 Introduction In recent years, multi-objective optimization (MOO) has emerged as a foundational problem underpinning many multi-agent multi-task learning applications, such as training neural networks for multiple tasks [1], hydrocarbon production optimization [2], recommendation system [3], tissue engineering [4], and learning-to-rank [5, 6, 7]. MOO aims at optimizing multiple objectives simultaneously, which can be mathematically cast as: min x D F(x) := [f1(x), , f S(x)], (1) where x D Rd is the model parameter, and fs : Rd R, s [S] is one of the objective functions. Compared to conventional single-objective optimization, one key difference in MOO is the coupling and potential conflicts between different objective functions. As a result, there may not exist a common x-solution that minimizes all objective functions. Rather, the goal in MOO is to find a Pareto stationary solution that is not improvable for all objectives without sacrificing some objectives. For example, in recommender system designs for e-commerce, the platform needs to consider different 37th Conference on Neural Information Processing Systems (Neur IPS 2023). customers with substantially conflicting shopping objectives (price, brand preferences, delivery speed, etc.). Therefore, the platform s best interest is often to find a Pareto-stationary solution, where one cannot deviate to favor one consumer group further without hurting any other group. MOO with conflicting objectives also has natural incarnations in many competitive game-theoretic problems, where the goal is to determine an equilibrium among the conflicting agents in the Pareto sense. Since its inception dating back to the 1950s, MOO algorithm design has evolved into two major categories: gradient-free and gradient-based methods, with the latter garnering increasing attention in the learning community in recent years due to their better performances (see Section 2 for more detailed discussions). However, despite these advances, all existing algorithms in the current MOO literature remain limited to centralized settings (i.e., training data are aggregated and accessible to a centralized learning algorithm). Somewhat ironically, such centralized settings do not satisfy the distributed nature and data privacy needs of many multi-agent multi-task learning applications, which motivates application of MOO in the first place. This gap between the existing MOO approaches and the rapidly growing importance of distributed MOO motivates us to make the first attempt to pursue a new federated multi-objective learning (FMOL) framework, with the aim to enable multiple clients to distributively solve MOO problems while keeping their computation and training data private. So far, however, developing distributed optimization algorithms for FMOL with provable Paretostationary convergence remains uncharted territory. There are several key technical challenges that render FMOL far from being a straightforward extension of centralized MOO problems. First of all, due to the distributed nature of FMOL problems, one has to consider and model the objective heterogeneity (i.e., different clients could have different sets of objective functions) that is unseen in centralized MOO. Moreover, with local and private datasets being a defining feature in FMOL, the impacts of data heterogeneity (i.e., datasets are non-i.i.d. distributed across clients) also need to be mitigated in FMOL algorithm design. Last but not least, under the combined influence of objective and data heterogeneity, FMOL algorithms could be extremely sensitive to small perturbations in the determination of common descent direction among all objectives. This makes the FMOL algorithm design and the associated convergence analysis far more complicated than those of the centralized MOO. Toward this end, a fundamental question naturally arises: Under both objective and data heterogeneity in FMOL, is it possible to design effective and efficient algorithms with Pareto-stationary convergence guarantees? In this paper, we give an affirmative answer to the above question. Our key contribution is that we propose a new FMOL framework that captures both objective and data heterogeneity, based on which we develop two gradient-based algorithms with provable Pareto-stationary convergence rate guarantees. To our knowledge, our work is the first systematic attempt to bridge the gap between federated learning and MOO. Our main results and contributions are summarized as follows: We formalize the first federated multi-objective learning (FMOL) framework that supports both objective and data heterogeneity across clients, which significantly advances and generalizes the MOO formulation to the federated learning paradigm. As a result, our FMOL framework becomes a generic model that covers existing MOO models and various applications as special cases (see Section 3.2 for further details). This new FMOL framework lays the foundation to enable us to systematically develop FMOO algorithms with provable Pareto-stationary convergence guarantees. For the proposed FMOL framework, we first propose a federated multi-gradient descent averaging (FMGDA) algorithm based on the use of local full gradient evaluation at each client. Our analysis reveals that FMGDA achieves a linear O(exp( µT)) and a sublinear O(1/T) Pareto-stationary convergence rates for µ-strongly convex and non-convex settings, respectively. Also, FMGDA employs a two-sided learning rates strategy to significantly lower communication costs (a key concern in the federated learning paradigm). It is worth pointing out that, in the single-machine special case where FMOL degenerates to a centralized MOO problem and FMGDA reduces to the traditional MGD method [8], our results improve the state-of-the-art analysis of MGD by eliminating the restrictive assumptions on the linear search of learning rate and extra sequence convergence. Thus, our results also advance the state of the art in general MOO theory. To alleviate the cost of full gradient evaluation in the large dataset regime, we further propose a federated stochastic multi-gradient descent averaging (FSMGDA) algorithm based on the use of stochastic gradient evaluations at each client. We show that FSMGDA achieves O(1/T) and O(1/ T) Pareto-stationary convergence rate for µ-strongly convex and non-convex settings, respectively. We establish our convergence proof by proposing a new (α, β)-Lipschitz continuous Table 1: Convergence rate results (shaded parts are our results) comparisons. Methods Strongly Convex Non-convex Rate Assumption Rate Assumption MGD [8] O(r T ) # Linear search & sequence convergence O(1/T ) Linear search & sequence convergence SMGD [9] O(1/T ) First moment bound & Lipschitz continuity of λ Not provided Not provided FMGDA O(exp( µT )) # Not needed O(1/T ) Not needed FSMGDA O(1/T ) (α, β)-Lipschitz continuous stochastic gradient O(1/ T ) (α, β)-Lipschitz continuous stochastic gradient #Notes on constants: µ is the strong convexity modulus; r is a constant depends on µ, s.t., r (0, 1). Assumption short-hands: Linear search : learning rate linear search [8]; Sequence convergence : {xt} converges to x [8]; First moment bound (Asm. 5.2(b) [9]): E[ f(x, ξ) f(x) ] η(a + b f(x) ); Lipschitz continuity of λ (Asm. 5.4 [9]): λk λt β ( f1(xk) f1(xt))T , . . . , ( f S(xk) f S(xt))T ; (α, β)-Lipschitz continuous stochastic gradient : see Asm. 4. stochastic gradient assumption (cf. Assumption 4), which relaxes the strong assumptions on first moment bound and Lipschitz continuity on common descent directions in [9]. We note that this new (α, β)-Lipschitz continuous stochastic gradient assumption can be viewed as a natural extension of the classical Lipschitz-continuous gradient assumption and could be of independent interest. The rest of the paper is organized as follows. In Section 2, we review related works. In Section 3, we introduce our FMOL framework and two gradient-based algorithms (FMGDA and FSMGDA), which are followed by their convergence analyses in Section 4. We present the numerical results in Section 5 and conclude the work in Section 6. Due to space limitations, we relegate all proofs and some experiments to supplementary material. 2 Related work In this section, we will provide an overview on algorithm designs for MOO and federated learning (FL), thereby placing our work in a comparative perspective to highlight our contributions and novelty. 1) Multi-objective Optimization (MOO): As mentioned in Section 1, since federated/distributed MOO has not been studied in the literature, all existing works we review below are centralized MOO algorithms. Roughly speaking, MOO algorithms can be grouped into two main categories. The first line of works are gradient-free methods (e.g., evolutionary MOO algorithms and Bayesian MOO algorithms [10, 11, 12, 13]). These methods are more suitable for small-scale problems but less practical for high-dimensional MOO models (e.g., deep neural networks). The second line of works focus on gradient-based approaches [14, 15, 8, 16, 9, 17], which are more practical for high-dimensional MOO problems. However, while having received increasing attention from the community in recent years, Pareto-stationary convergence analysis of these gradient-based MOO methods remains in its infancy. Existing gradient-based MOO methods can be further categorized as i) multi-gradient descent (MGD) algorithms with full gradients and ii) stochastic multi-gradient descent (SMGD) algorithms. It has been shown in [8] that MGD methods achieve O(r T ) for some r (0, 1) and O(1/T) Paretostationary convergence rates for µ-strongly convex and non-convex functions, respectively. However, these results are established under the unconventional linear search of learning rate and sequence convergence assumptions, which are difficult to verify in practice. In comparison, FMGDA achieves a linear rate without needing such assumptions. For SMGD methods, the Pareto-stationary convergence analysis is further complicated by the stochastic gradient noise. Toward this end, an O(1/T) rate analysis for SMGD was provided in [9] based on rather strong assumptions on a first-moment bound and Lipschtiz continuity of common descent direction. As a negative result, it was shown in [9] and [18] that the common descent direction needed in the SMGD method is likely to be a biased estimation, which may cause divergence issues. In contrast, our FSMGDA achieves state-of-the-art O(1/T) and O(1/ T) convergence rates for strongly-convex and non-convex settings, respectively, under a much milder assumption on Lipschtiz continuous stochastic gradients. For easy comparisons, we summarize our results and the existing works in Table 1. It is worth noting recent works [18, 19, 20] established faster convergence rates in the centralized MOO setting by using acceleration techniques, such as momentum, regularization and bi-level formulation. However, due to different settings and focuses, these results are orthogonal to ours and thus not directly comparable. Also, since acceleration itself is a non-trivial topic and could be quite brittle if not done right, in this paper, we focus on the basic and more robust stochastic gradient approach in FMOL. But for a comprehensive comparison on assumptions and main results of accelerated centralized MOO, we refer readers to Appendix A for further details. Federated Learning (FL) : Since the seminal work by [21], FL has emerged as a popular distributed learning paradigm. Traditional FL aims at solving single-objective minimization problems with a large number of clients with decentralized data. Recent FL algorithms enjoy both high communication efficiency and good generalization performance [21, 22, 23, 24, 25, 26]. Theoretically, many FL methods have the same convergence rates as their centralized counterparts under different FL settings [27, 28, 29, 30]. Recent works have also considered FL problems with more sophisticated problem structures, such as min-max learning [31, 32], reinforcement learning [33], multi-armed bandits [34], and bilevel and compositional optimization [35]. Although not directly related, classic FL has been reformulated in the form of MOO[36], which allows the use of a MGD-type algorithm instead of vanilla local SGD to solve the standard FL problem. We will show later that this MOO reformulation is a special case of our FMOL framework. So far, despite a wide range of applications (see Section 3.2 for examples), there remains a lack of a general FL framework for MOO. This motivates us to bridge the gap by proposing a general FMOL framework and designing gradient-based methods with provable Pareto-stationary convergence rates. 3 Federated multi-objective learning 3.1 Multi-objective optimization: A primer As mentioned in Section 1, due to potential conflicts among the objective functions in MOO problem in (1), MOO problems adopt the the notion of Pareto optimality: Definition 1 ((Weak) Pareto Optimality). For any two solutions x and y, we say x dominates y if and only if fs(x) fs(y), s [S] and fs(x) < fs(y), s [S]. A solution x is Pareto optimal if it is not dominated by any other solution. One solution x is weakly Pareto optimal if there does not exist a solution y such that fs(x) > fs(y), s [S]. Similar to solving single-objective non-convex optimization problems, finding a Pareto-optimal solution in MOO is NP-Hard in general. As a result, it is often of practical interest to find a solution satisfying Pareto-stationarity (a necessary condition for Pareto optimality) stated as follows [14, 37]: Definition 2 (Pareto Stationarity). A solution x is said to be Pareto stationary if there is no common descent direction d Rd such that fs(x) d < 0, s [S]. Note that for strongly convex functions, Pareto stationary solutions are also Pareto optimal. Following Definition 2, gradient-based MOO algorithms typically search for a common descent direction d Rd such that fs(x) d 0, s [S]. If no such a common descent direction exists at x, then x is a Pareto stationary solution. For example, MGD [15] searches for an optimal weight λ of gradients F(x) { fs(x), s [S]} by solving λ (x) = argminλ C λ F(x) 2. Then, a common descent direction can be chosen as: d = λ F(x). MGD performs the iterative update rule: x x ηd until a Pareto stationary point is reached, where η is a learning rate. SMGD [9] also follows the same process except for replacing full gradients by stochastic gradients. For MGD and SMGD methods, it is shown in [8] and [18] show that if λ F(x) = 0 for some λ C, where C {y [0, 1]S, P s [S] ys = 1}, then x is a Pareto stationary solution. Thus, d 2 = λ F(x) 2 can be used as a metric to measure the convergence of non-convex MOO algorithms [8, 18, 19]. On the other hand, for more tractable strongly convex MOO problems, the optimality gap P s [S] λs [fs(x) fs(x )] is typically used as the metric to measure the convergence of an algorithm [9], where x denotes the Pareto optimal point. We summarize and compare different convergence metrics as well as assumptions in MOO, detailed in Appendix A. 3.2 A general federated multi-objective learning framework With the MOO preliminaries in Section 3.1, we now formalize our general federated multi-objective learning (FMOL) framework. For a system with M clients and S tasks (objectives), our FMOL framework can be written as: min x Diag(FA ), (2) f1,1 f1,M ... ... ... f S,1 f S,M a1,1 a1,M ... ... ... a S,1 a S,M where matrix F groups all potential objectives fs,i(x) for each task s at each client i, and A {0, 1}S M is a binary objective indicator matrix, with each element as,i = 1 if task s is of client i s interest and as,i = 0 otherwise. For each task s [S], the global objective function fs(x) is the average of local objectives over all related clients, i.e., fs(x) 1 |Rs| P i Rs fs,i(x), where Rs = {i : as,i = 1, i [M]}. Note that, for notation simplicity, here we use simple average in fs(x), which corresponds to the balanced dataset setting. Our FMLO framework can be directly extended to imbalanced dataset settings by using weighted average proportional to dataset sizes of related clients. For a client i [M], its objectives of interest are {fs,i(x):as,i =1, s [S]}, which is a subset of [S]. We note that FMOL generalizes MOO to the FL paradigm, which includes many existing MOO problems as special cases and corresponds to a wide range of applications. If each client has only one distinct objective, i.e., A = IM, S = M, then Diag(FA ) = [f1(x), . . . , f S(x)] , where each objective fs(x), s [S] is optimized only by client s. This special FMOL setting corresponds to the conventional multi-task learning and federated learning. Indeed, [1] and [38] formulated a multi-task learning problem as MOO and considered Pareto optimal solutions with various trade-offs. [36] also formulated FL as as distributed MOO problems. Other examples of this setting include bi-objective formulation of offline reinforcement learning [39] and decentralized MOO [40]. If all clients share the same S objectives, i.e., A is an all-one matrix, then Diag(FA ) = 1 M P i [M] f1,i(x), . . . , 1 M P i [M] f S,i(x) . In this case, FMOL reduces to federated MOO problems with decentralized data that jointly optimizing fairness, privacy, and accuracy [41, 42, 43], as well as MOO with decentralized data under privacy constraints (e.g., machine reassignment among data centres [44] and engineering problems [45, 46, 47, 48]). If each client has a different subset of objectives (i.e., objective heterogeneity), FMLO allows distinct preferences at each client. For example, each customer group on a recommender system in e-commerce platforms might have different combinations of shopping preferences, such as product price, brand, delivery speed, etc. 3.3 Federated Multi-Objective Learning Algorithms Upon formalizing our FMOL framework, our next goal is to develop gradient-based algorithms for solving large-scale high-dimensional FMOL problems with provable Pareto stationary convergence guarantees and low communication costs. To this end, we propose two FMOL algorithms, namely federated multiple gradient descent averaging (FMGDA) and federated stochastic multiple gradient descent averaging (FSMGDA) as shown in Algorithm 1. We summarize our key notation in Table 3 in Appendix to allow easy references for readers. As shown in Algorithm 1, in each communication round t [T], each client synchronizes its local model with the current global model xt from the server (cf. Step 1). Then each client runs K local steps based on local data for all effective objectives (cf. Step 2) with two options: i) for FMGDA, each local step performs local full gradient descent, i.e., xt,k+1 s,i = xt,k s,i ηL fs,i(xt,k s,i), s Si; ii) For FSMGDA, the local step performs stochastic gradient descent, i.e., xt,k+1 s,i = xt,k s,i ηL fs,i(xt,k s,i, ξt,k i ), s Si, where ξt,k i denotes a random sample in local step k and round t at client i. Upon finishing K local updates, each client returns the accumulated update t s,i for each effective objective to the server (cf. Step 3). Then, the server aggregates all returned -updates from Algorithm 1 Federated (Stochastic) Multiple Gradient Descent Averaging (FMGDA/FSMGDA). At Each Client i: 1. Synchronize local models xt,0 s,i = xt, s Si. 2. Local updates: for all s Si, for k = 1, . . . , K, (FMGDA): xt,k s,i = xt,k 1 s,i ηL fs,i(xt,k 1 s,i ). (FSMGDA): xt,k s,i = xt,k 1 s,i ηL fs,i(xt,k 1 s,i , ξt,k i ). 3. Return accumulated updates to server { t s,i, s Si}: (FMGDA): t s,i = P k [K] fs,i(xt,k s,i). (FSMGDA): t s,i = P k [K] fs,i(xt,k s,i, ξt,k i ). At the Server: 4. Receive accumulated updates { t s,i, s Si, i [M]}. 5. Compute t s = 1 |Rs| P i Rs t s,i, s [S], where Rs = {i : as,i = 1, i [M]}. 6. Compute λ t [0, 1]S by solving s [S] λt s t s 2 , s.t. X s [S] λt s = 1. (3) 7. Let dt = P s [S] λt, s t s and update the global model as: xt+1 = xt ηtdt, with a global learning rate ηt. the clients to obtain the overall updates t s for each objective s [S] (cf. Steps 4 and 5), which will be used in solving a convex quadratic optimization problem with linear constraints to obtain an approximate common descent direction dt (cf. Step 6). Lastly, the global model is updated following the direction dt with global learning rate ηt (cf. Step 7). Two remarks on Algorithm 1 are in order. First, we note that a two-sided learning rates strategy is used in Algorithm 1, which decouples the update schedules of local and global model parameters at clients and server, respectively. As shown in Section 4 later, this two-sided learning rates strategy enables better convergence rates by choosing appropriate learning rates. Second, to achieve low communication costs, Algorithm 1 leverages K local updates at each client and infrequent periodic communications between each client and the server. By adjusting the two-sided learning rates appropriately, the K-value can be made large to further reduce communication costs. 4 Pareto stationary convergence analysis In this section, we analyze the Pareto stationary convergence performance for our FMGDA and FSMGDA algorithms in Sections 4.1 and 4.2, respectively, each of which include non-convex and strongly convex settings. 4.1 Pareto stationary convergence of FMGDA In what follows, we show FMGDA enjoys linear rate O(exp( µT)) for µ-strongly convex functions and sub-linear rate O( 1 T ) for non-convex functions. 1) FMGDA: The Non-convex Setting. Before presenting our Pareto stationary convergence results for FMGDA, we first state serveral assumptions as follows: Assumption 1. (L-Lipschitz continuous) There exists a constant L > 0 such that fs(x) fs(y) L x y , x, y Rd, s [S]. Assumption 2. (Bounded Gradient) The gradient of each objective at any client is bounded, i.e., there exists a constant G > 0 such that fs,i(x) 2 G2, s [S], i [M]. With the assumptions above, we state the Pareto stationary convergence of FMGDA for non-convex FMOL as follows: Theorem 1 (FMGDA for Non-convex FMOL). Let ηt = η 3 2(1+L). Under Assumptions 1 and 2, if at least one function fs, s [S] is bounded from below by f min s , then the sequence {xt} output by FMGDA satisfies: mint [T ] dt 2 16(f 0 s f min s ) T η + δ, where δ 16η2 LK2L2G2(1+S2) In non-convex functions, we use dt 2 as the metrics for FMOO, where dt = λT t (Diag(FA )) and λt is calculated by the quadratic programming problem 3 based on accumulated (stochastic) gradients t. We compare different metrics for MOO in Appendix A. The convergence bound in Theorem 1 contains two parts. The first part is an optimization error, which depends on the initial point and vanishes as T increases. The second part is due to local update steps K and data heterogeneity G, which can be mitigated by carefully choosing the local learning rate ηL. Specifically, the following Pareto stationary convergence rate of FMGDA follows immediately from Theorem 1 with an appropriate choice of local learning rate ηL: Corollary 2. With a constant global learning rate ηt = η, t, and a local learning rate ηL = O(1/ T), the Pareto stationary convergence rate of FMGDA is (1/T) P t [T ] dt 2 = O(1/T). Several interesting insights of Theorem 1 and Corollary 2 are worth pointing out: 1) We note that FMGDA achieves a Pareto stationary convergence rate O(1/T) for non-convex FMOL, which is the same as the Pareto stationary rate of MGD for centralized MOO and the same convergence rate of gradient descent (GD) for single objective problems. This is somewhat surprising because FMGDA needs to handle more complex objective and data heterogeneity under FMOL; 2) The two-sided learning rates strategy decouples the operation of clients and server by utilizing different learning rate schedules, thus better controlling the errors from local updates due to data heterogeneity; 3) Note that in the single-client special case, FMGDA degenerates to the basic MGD algorithm. Hence, Theorem 1 directly implies a Pareto stationary convergence bound for MGD by setting δ = 0 due to no local updates in centralized MOO. This convergence rate bound is consistent with that in [8]. However, we note that our result is achieved without using the linear search step for learning rate [8], which is much easier to implement in practice (especially for deep learning models); 4) Our proof is based on standard assumptions in first-order optimization, while previous works require strong and unconventional assumptions. For example, a convergence of x-sequence is assumed in [8]. 2) FMGDA: The Strongly Convex Setting. Now, we consider the strongly convex setting for FMOL, which is more tractable but still of interest in many learning problems in practice. In the strongly convex setting, we have the following additional assumption: Assumption 3. (µ-Strongly Convex Function) Each objective fs(x), s [S] is a µ-strongly convex function, i.e., fs(y) fs(x) + fs(x)(y x) + µ 2 y x 2 for some µ > 0. For more tractable strongly-convex FMOL problems, we show that FMGDA achieves a stronger Pareto stationary convergence performance as follows: Theorem 3 (FMGDA for µ-Strongly Convex FMOL). Let ηt = η such that η 3 2(1+L), η 1 2L+µ and η 1 µT . Under Assumptions 13, pick xt as the final output of the FMGDA algorithm with weights wt = (1 µη 2 )1 t. Then, it holds that E[ t Q] x0 x 2µ exp( ηµT 2 ) + δ, where s [S] λt, s [fs(xt) fs(x )] and δ = 8η2 LK2L2G2S2 µ + 2η2 LK2L2G2. Theorem 3 immediately implies following Pareto stationary convergence rate for FMGDA with a proper choice of local learning rate: Corollary 4. If ηL is chosen sufficiently small such that δ = O(µ exp( µT)), then the Pareto stationary convergence rate of FMGDA is E[ t Q] = O(µ exp( µT)). Again, several interesting insights can be drawn from Theorem 3 and Corollary 4. First, for strongly convex FMOL, FMGDA achieves a linear convergence rate O(µ exp( µT)), which again matches those of MGD for centralized MOO and GD for single-objective problems. Second, compared with the non-convex case, the convergence bounds suggest FMGDA could use a larger local learning rate for non-convex functions thanks to our two-sided learning rates design. A novel feature of FMGDA for strongly convex FMOL is the randomly chosen output xt with weight wt from the xt-trajectory, which is inspired by the classical work in stochastic gradient descent (SGD) [49]. Note that, for implementation in practice, one does not need to store all xt-values. Instead, the algorithm can be implemented by using a random clock for stopping [49]. 4.2 Pareto stationary convergence of FSMGDA While enjoying strong performances, FMGDA uses local full gradients at each client, which could be costly in the large dataset regime. Thus, it is of theoretical and practical importance to consider the stochastic version of FMGDA, i.e., federated stochastic multi-gradient descent averaging (FSMGDA). 1) FSMGDA: The Non-convex Setting. A fundamental challenge in analyzing the Pareto stationarity convergence of FSMGDA and other stochastic multi-gradient descent (SMGD) methods stems from bounding the error of the common descent direction estimation, which is affected by both λ t (obtained by solving a quadratic programming problem) and the stochastic gradient variance. In fact, it is shown in [9] and [18] that the stochastic common descent direction in SMGD-type methods could be biased, leading to divergence issues. To address these challenges, in this paper, we propose to use a new assumption on the stochastic gradients, which is stated as follows: Assumption 4 ((α, β)-Lipschitz Continuous Stochastic Gradient). A function f has (α, β)-Lipschitz continuous stochastic gradients if there exist two constants α, β > 0 such that, for any two independent training samples ξ and ξ , E[ f(x, ξ) f(y, ξ ) 2] α x y 2 + βσ2. In plain language, Assumption 4 says that the stochastic gradient estimation of an objective does not change too rapidly. We note that the (α, β)-Lipschitz continuous stochastic gradient assumption is a natural extension of the classic L-Lipschitz continuous gradient assumption (cf. Assumption 1) and generalizes several assumptions of SMGD convergence analysis in previous works. We note that Assumption 1 is not necessarily too hard to satisfy in practice. For example, when the underlying distribution of training samples ξ has a bounded support (typically a safe assumption for most applications in practice due to the finite representation limit of computing systems), suppose that Assumption 1 holds (also a common assumption in the optimization literature), then for any given x and y, the left-hand-side of the inequality in Assumption 4 is bounded due to the L-smoothness in Assumption 1. In this case, there always exist a sufficiently large α and a β such that the righthand-side of the inequality in Assumption 1 holds. Please see Appendix A for further details. In addition, we need the following assumptions for the stochastic gradients, which are commonly used in standard SGD-based analyses [49, 50, 51, 52]. Assumption 5. (Unbiased Stochastic Estimation) The stochastic gradient estimation is unbiased for each objective among clients, i.e., E[ fs,i(x, ξ)] = fs,i(x), s [S], i [M]. Assumption 6. (Bounded Stochastic Gradient) The stochastic gradients satisfiy E[ fs,i(x, ξ) 2] D2, s [S], i [M] for some constant D > 0. With the assumptions above, we now state the Pareto stationarity convergence of FSMGDA as follows: Theorem 5 (FSMGDA for Non-convex FMOL). Let ηt = η 3 2(1+L). Under Assumptions 4 6, if an objective fs is bounded from below by f min s , then the sequence {xt} output by FSMGDA satisfies: mint [T ] E dt 2 8(f 0 s f min s ) ηT + δ, where δ = (2S2 + 4)(αη2 LK2D2 + βσ2). Theorem 5 immediately implies an O(1/ T) convergence rate of FSMGDA for non-convex FMOL: Corollary 6. With a constant global learning rate ηt = η = O(1/ T), t and a local learning rate ηL = O 1/T 1/4 , and if β = O(η), the Pareto stationarity convergence rate of FSMGDA is mint [T ] E dt 2 = O(1/ 2) The Strongly Convex Setting: For more tractable strongly convex FMOL problems, we can show that FSMGDA achieve stronger convergence results as follows: Theorem 7 (FSMGDA for µ-Strongly Convex FMOL). Let ηt = η = Ω( 1 µT ). Under Assumptions 3, 5 and 6, pick xt as the final output of the FSMGDA algorithm with weight wt = (1 µη 2 )1 t. Then, it holds that: E[ t Q] x0 x 2µ exp( η 2µT) + δ, where t Q = P s [S] λt, s [fs(xt) fs(x )] µS2(αη2 LK2D2 + βσ2) + ηS2D2 The following Pareto station convergence rate of FSMGDA follows immediately from Theorem 7: Corollary 8. Choose ηL = O( 1 T ) and η = Θ( log(max(1,µ2T )) µT ). If β = O(η), then the Pareto stationary convergence rate of FSMGDA is E[ t Q] O(1/T). Corollary 8 says that, With proper learning rates, FSMGDA achieves O(1/T) Pareto stationary convergence rate (i.e., ignoring logarithmic factors) for strongly convex FMOL. Also, in the singleclient special case with no local updates, FSMGDA reduces to the SMGD algorithm and δ = 4 µβS2σ2 + ηS2D2 2 in this case. Then, Theorem 7 implies an O( 1 T ) Pateto stationarity convergence rate for SMGD for strongly convex MOO problems, which is consistent with previous works [9]. However, our convergence rate proof uses a more conventional (α, β)-Lipschitz stochastic gradient assumption, rather than the unconventional assumptions on the first moment bound and Lipschitz continuity of common descent directions in [9]. 5 Numerical results In this section, we show the main numerical experiments of our FMGDA and FSMGDA algorithms in different datasets, while relegating the experimental settings and details to the appendix. 0 20 40 60 80 100 Communication rounds Batch size = 16 Batch size = 64 Batch size = 128 Batch size = 256 0 20 40 60 80 100 Communication rounds Batch size = 16 Batch size = 32 Batch size = 128 Batch size = 256 (a) Training loss convergence in terms of communication rounds with different batch-sizes under non-i.i.d. data partition in Multi MNIST. 0 20 40 60 80 100 Communication rounds K = 1 K = 5 K = 10 K = 20 K = 50 0 20 40 60 80 100 Communication rounds K = 1 K = 5 K = 10 K = 20 K = 50 (b) The impacts of local update number K on training loss convergence in terms of communication rounds. Figure 1: Training loss convergence comparison. 1) Ablation Experiments on Two-Tasks FMOL: 1-a) Impacts of Batch Size on Convergence: First, we compare the convergence results in terms of the number of communication rounds using the Multi MNIST dataset [53] with two tasks (L and R) as objectives. We test our algorithms with four different cases with batch sizes being [16, 64, 128, 256]. To reduce computational costs in this experiment, the dataset size for each client is limited to 256. Hence, the batch size 256 corresponds to FMGDA and all other batch sizes correspond to FSMGDA. As shown in Fig. 1(a), under non-i.i.d. data partition, both FMGDA and FSMGDA algorithms converge. Also, the convergence speed of the FSMGDA algorithm increases as the batch size gets larger. These results are consistent with our theoretical analyses as outlined in Theorems 1 and 5. 1-b) Impacts of Local Update Steps on Convergence: Next, we evaluate our algorithms with different numbers of local update steps K. As shown in Fig. 1(b) and Table 2, both algorithms converge faster as the number of the local steps K increases. This is because both algorithms effectively run more iterative updates as K gets large. 1-c) Comparisons between FMOL and Centralized MOO: Since this work is the first that investigates FMOL, it is also interesting to empirically compare the differences between FMOL and centralized MOO methods. In Fig. 2(a), we compare the training loss of FMGDA and FSMGDA with those of the centralized MGD and SMGD methods after 100 communication rounds. For fair comparisons, the centralized MGD and SMGD methods use PM i |Si| batch-sizes and run K T iterations. Our results indicate that FMGDA and MGD produce similar results, while the performance of FSMGDA is slightly worse than that of SMGD due to FSMGDA s sensitivity to objective and data heterogeneity in stochastic settings. These numerical results confirm our theoretical convergence analysis. Table 2: Communication rounds needed for 10 2 loss. i.i.d. non-i.i.d. Task L Task R Task L Task R K = 1 82 84 96 82 K = 5 18(4.6 ) 20(4.2 ) 24(4.0 ) 20(4.1 ) K = 10 10(8.2 ) 9(9.3 ) 13(7.4 ) 10(8.2 ) K = 20 5(16.4 ) 5(16.8 ) 6(16.0 ) 5(16.4 ) MGD K=1 K=5 K=10 K=20 K=50 SMGD K=1 K=5 K=10 K=20 K=50 (a) 100 communication rounds with various local steps K, corresponding federated and centralized settings share the same marker shape. Normalized Loss K=1 K=5 K=10 K=20 Normalized Loss Batch size=16 Batch size=64 Batch size=128 Batch size=256 (b) Normalized loss with the River Flow datasets. Figure 2: Training losses comparison 2) Experiments on Larger FMOL: We further test our algorithms on FMOL problems of larger sizes. In this experiment, we use the River Flow dataset[54], which contains eight tasks in this problem. To better visualize 8 different tasks, we illustrate the normalized loss in radar charts in Fig. 2(b). In this 8-task setting, we can again verify that more local steps K and a larger training batch size lead to faster convergence. In the appendix, we also verify the effectiveness of our FMGDA and FSMGDA algorithms in Celeb A [55] (40 tasks), alongside with other hyperparmeter tuning results. 6 Conclusion and discussions In this paper, we proposed the first general framework to extend multi-objective optimization to the federated learning paradigm, which considers both objective and data heterogeneity. We showed that, even under objective and data heterogeneity, both of our proposed algorithms enjoy the same Pareto stationary convergence rate as their centralized counterparts. In our future work, we will go beyond the limitation in the analysis of MOO that an extra assumption on the stochastic gradients (and λ). In this paper, we have proposed a weaker assumption (Assumption 4). We conjecture that using acceleration techniques, e.g., momentum, variance reduction, and regularization, could relax such assumption and achieve better convergence rate, which is a promising direction for future works. In addition, MOO in distributed learning gives rise to substantially expensive communication costs, which scales linearly with the number of clients and the number of objectives in each client. Developing communication-efficient MOO beyond typical gradient compression methods for distributed learning is also a promising direction for future works. Acknowledgments and Disclosure of Funding This work has been supported in part by NSF grants CAREER CNS-2110259 and CNS-2112471. [1] O. Sener and V. Koltun, Multi-task learning as multi-objective optimization, Advances in neural information processing systems, vol. 31, 2018. [2] J. You, W. Ampomah, and Q. Sun, Development and application of a machine learning based multi-objective optimization workflow for co2-eor projects, Fuel, vol. 264, p. 116758, 2020. [3] T. Zhou, M. Momma, C. Dong, F. Yang, C. Guo, J. Shang, and J. K. Liu, Multi-task learning on heterogeneous graph neural network for substitute recommendation, in 19th International Workshop on Mining and Learning with Graphs, 2023. [4] J. Shi, J. Song, B. Song, and W. F. Lu, Multi-objective optimization design through machine learning for drop-on-demand bioprinting, Engineering, vol. 5, no. 3, pp. 586 593, 2019. [5] D. Mahapatra, C. Dong, Y. Chen, and M. Momma, Multi-label learning to rank through multiobjective optimization, in Proceedings of the 29th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2023. [6] D. Mahapatra, C. Dong, and M. Momma, Querywise fair learning to rank through multiobjective optimization, in Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, 2023. [7] M. Momma, A. Bagheri Garakani, N. Ma, and Y. Sun, Multi-objective ranking via constrained optimization, in Companion Proceedings of the Web Conference 2020, 2020, pp. 111 112. [8] J. Fliege, A. I. F. Vaz, and L. N. Vicente, Complexity of gradient descent for multiobjective optimization, Optimization Methods and Software, vol. 34, no. 5, pp. 949 959, 2019. [9] S. Liu and L. N. Vicente, The stochastic multi-gradient algorithm for multi-objective optimization and its application to supervised machine learning, Annals of Operations Research, pp. 1 30, 2021. [10] Q. Zhang and H. Li, Moea/d: A multiobjective evolutionary algorithm based on decomposition, IEEE Transactions on evolutionary computation, vol. 11, no. 6, pp. 712 731, 2007. [11] K. Deb, A. Pratap, S. Agarwal, and T. Meyarivan, A fast and elitist multiobjective genetic algorithm: Nsga-ii, IEEE transactions on evolutionary computation, vol. 6, no. 2, pp. 182 197, 2002. [12] S. Belakaria, A. Deshwal, N. K. Jayakodi, and J. R. Doppa, Uncertainty-aware search framework for multi-objective bayesian optimization, in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, 2020, pp. 10 044 10 052. [13] M. Laumanns and J. Ocenasek, Bayesian optimization algorithms for multi-objective optimization, in International Conference on Parallel Problem Solving from Nature. Springer, 2002, pp. 298 307. [14] J. Fliege and B. F. Svaiter, Steepest descent methods for multicriteria optimization, Mathematical methods of operations research, vol. 51, no. 3, pp. 479 494, 2000. [15] J.-A. D esid eri, Multiple-gradient descent algorithm (mgda) for multiobjective optimization, Comptes Rendus Mathematique, vol. 350, no. 5-6, pp. 313 318, 2012. [16] S. Peitz and M. Dellnitz, Gradient-based multiobjective optimization with uncertainties, in NEO 2016. Springer, 2018, pp. 159 182. [17] M. Momma, C. Dong, and J. Liu, A multi-objective / multi-task learning framework induced by pareto stationarity, in Proceedings of the 39th International Conference on Machine Learning, 2022. [18] S. Zhou, W. Zhang, J. Jiang, W. Zhong, J. GU, and W. Zhu, On the convergence of stochastic multi-objective gradient manipulation and beyond, in Advances in Neural Information Processing Systems, A. H. Oh, A. Agarwal, D. Belgrave, and K. Cho, Eds., 2022. [Online]. Available: https://openreview.net/forum?id=Scwf Q7hdwy P [19] H. Fernando, H. Shen, M. Liu, S. Chaudhury, K. Murugesan, and T. Chen, Mitigating gradient bias in multi-objective learning: A provably convergent stochastic approach, ar Xiv preprint ar Xiv:2210.12624, 2022. [20] P. Xiao, H. Ban, and K. Ji, Direction-oriented multi-objective learning: Simple and provable stochastic algorithms, ar Xiv preprint ar Xiv:2305.18409, 2023. [21] B. Mc Mahan, E. Moore, D. Ramage, S. Hampson, and B. A. y Arcas, Communication-efficient learning of deep networks from decentralized data, in Artificial intelligence and statistics. PMLR, 2017, pp. 1273 1282. [22] T. Li, A. K. Sahu, M. Zaheer, M. Sanjabi, A. Talwalkar, and V. Smith, Federated optimization in heterogeneous networks, in Proceedings of Machine Learning and Systems, I. Dhillon, D. Papailiopoulos, and V. Sze, Eds., vol. 2, 2020, pp. 429 450. [23] D. A. E. Acar, Y. Zhao, R. M. Navarro, M. Mattina, P. N. Whatmough, and V. Saligrama, Federated learning based on dynamic regularization, in International Conference on Learning Representations, 2021. [24] J. Wang, Q. Liu, H. Liang, G. Joshi, and H. V. Poor, Tackling the objective inconsistency problem in heterogeneous federated optimization, Advances in Neural Information Processing Systems, vol. 33, 2020. [25] T. Lin, S. U. Stich, K. K. Patel, and M. Jaggi, Don t use large mini-batches, use local sgd, in International Conference on Learning Representations, 2020. [Online]. Available: https://openreview.net/forum?id=B1ey O1BFPr [26] H. Yang, P. Qiu, and J. Liu, Taming fat-tailed ( heavier-tailed with potentially infinite variance) noise in federated learning, in Advances in Neural Information Processing Systems, A. H. Oh, A. Agarwal, D. Belgrave, and K. Cho, Eds., 2022. [Online]. Available: https://openreview.net/forum?id=8Sil FGu Xgmk [27] S. P. Karimireddy, S. Kale, M. Mohri, S. Reddi, S. Stich, and A. T. Suresh, SCAFFOLD: Stochastic controlled averaging for federated learning, in Proceedings of the 37th International Conference on Machine Learning, ser. Proceedings of Machine Learning Research, H. D. III and A. Singh, Eds., vol. 119. PMLR, 13 18 Jul 2020, pp. 5132 5143. [28] H. Yang, M. Fang, and J. Liu, Achieving linear speedup with partial worker participation in non-IID federated learning, in International Conference on Learning Representations, 2021. [29] H. Yang, X. Zhang, P. Khanduri, and J. Liu, Anarchic federated learning, in International Conference on Machine Learning. PMLR, 2022, pp. 25 331 25 363. [30] X. Zhang, M. Fang, Z. Liu, H. Yang, J. Liu, and Z. Zhu, Net-fleet: achieving linear convergence speedup for fully decentralized federated learning with heterogeneous data, Proceedings of the Twenty-Third International Symposium on Theory, Algorithmic Foundations, and Protocol Design for Mobile Networks and Mobile Computing, 2022. [31] H. Yang, Z. Liu, X. Zhang, and J. Liu, SAGDA: Achieving O(ϵ 2) communication complexity in federated min-max learning, in Advances in Neural Information Processing Systems, A. H. Oh, A. Agarwal, D. Belgrave, and K. Cho, Eds., 2022. [Online]. Available: https://openreview.net/forum?id=w Tp4Kg VIJ5 [32] P. Sharma, R. Panda, G. Joshi, and P. Varshney, Federated minimax optimization: Improved convergence analyses and algorithms, in International Conference on Machine Learning. PMLR, 2022, pp. 19 683 19 730. [33] S. Khodadadian, P. Sharma, G. Joshi, and S. T. Maguluri, Federated reinforcement learning: Linear speedup under markovian sampling, in International Conference on Machine Learning. PMLR, 2022, pp. 10 997 11 057. [34] C. Shi, C. Shen, and J. Yang, Federated multi-armed bandits with personalization, in International Conference on Artificial Intelligence and Statistics. PMLR, 2021, pp. 2917 2925. [35] D. A. Tarzanagh, M. Li, C. Thrampoulidis, and S. Oymak, Fed Nest: Federated bilevel, minimax, and compositional optimization, in Proceedings of the 39th International Conference on Machine Learning, ser. Proceedings of Machine Learning Research, K. Chaudhuri, S. Jegelka, L. Song, C. Szepesvari, G. Niu, and S. Sabato, Eds., vol. 162. PMLR, 17 23 Jul 2022, pp. 21 146 21 179. [36] Z. Hu, K. Shaloudegi, G. Zhang, and Y. Yu, Federated learning meets multi-objective optimization, IEEE Transactions on Network Science and Engineering, 2022. [37] K. Miettinen, Nonlinear multiobjective optimization. Springer Science & Business Media, 2012, vol. 12. [38] X. Lin, H.-L. Zhen, Z. Li, Q.-F. Zhang, and S. Kwong, Pareto multi-task learning, Advances in neural information processing systems, vol. 32, 2019. [39] Y. Yang, J. Jiang, T. Zhou, J. Ma, and Y. Shi, Pareto policy pool for model-based offline reinforcement learning, in International Conference on Learning Representations, 2022. [Online]. Available: https://openreview.net/forum?id=Oqc Zu8JIIz S [40] M. J. Blondin and M. Hale, A decentralized multi-objective optimization algorithm, Journal of Optimization Theory and Applications, vol. 189, no. 2, pp. 458 485, 2021. [41] L. T. Bui, H. A. Abbass, and D. Essam, Local models an approach to distributed multiobjective optimization, Computational Optimization and Applications, vol. 42, no. 1, pp. 105 139, 2009. [42] S. Cui, W. Pan, J. Liang, C. Zhang, and F. Wang, Addressing algorithmic disparity and performance inconsistency in federated learning, Advances in Neural Information Processing Systems, vol. 34, pp. 26 091 26 102, 2021. [43] N. Mehrabi, C. de Lichy, J. Mc Kay, C. He, and W. Campbell, Towards multi-objective statistically fair federated learning, ar Xiv preprint ar Xiv:2201.09917, 2022. [44] T. Saber, X. Gandibleux, M. O Neill, L. Murphy, and A. Ventresque, A comparative study of multi-objective machine reassignment algorithms for data centres, Journal of Heuristics, vol. 26, no. 1, pp. 119 150, 2020. [45] L. Yin, T. Wang, and B. Zheng, Analytical adaptive distributed multi-objective optimization algorithm for optimal power flow problems, Energy, vol. 216, p. 119245, 2021. [46] Y. Jin, Multi-objective machine learning. Springer Science & Business Media, 2006, vol. 16. [47] A. Mansoor, X. Diao, and C. Smidts, A method for backward failure propagation in conceptual system design, Nuclear Science and Engineering, 2023. [48] A. Mansoor, X. Diao, and Smidts, Backward failure propagation for conceptual system design using isfa, 11 2021. [49] S. Ghadimi and G. Lan, Stochastic first-and zeroth-order methods for nonconvex stochastic programming, SIAM Journal on Optimization, vol. 23, no. 4, pp. 2341 2368, 2013. [50] L. Bottou, F. E. Curtis, and J. Nocedal, Optimization methods for large-scale machine learning, Siam Review, vol. 60, no. 2, pp. 223 311, 2018. [51] H. B. Mc Mahan et al., Advances and open problems in federated learning, Foundations and Trends in Machine Learning, vol. 14, no. 1, 2021. [52] J. Wang, Z. Charles, Z. Xu, G. Joshi, H. B. Mc Mahan, M. Al-Shedivat, G. Andrew, S. Avestimehr, K. Daly, D. Data et al., A field guide to federated optimization, ar Xiv preprint ar Xiv:2107.06917, 2021. [53] S. Sabour, N. Frosst, and G. E. Hinton, Dynamic routing between capsules, Advances in neural information processing systems, vol. 30, 2017. [54] L. Nie, K. Wang, W. Kang, and Y. Gao, Image retrieval with attribute-associated auxiliary references, in 2017 International Conference on Digital Image Computing: Techniques and Applications (DICTA). IEEE, 2017, pp. 1 6. [55] Z. Liu, P. Luo, X. Wang, and X. Tang, Deep learning face attributes in the wild, in Proceedings of the IEEE international conference on computer vision, 2015, pp. 3730 3738. [56] Q. Mercier, F. Poirion, and J.-A. D esid eri, A stochastic multiple gradient descent algorithm, European Journal of Operational Research, vol. 271, no. 3, pp. 808 817, 2018. [57] Y. Le Cun, C. Cortes, and C. Burges, Mnist handwritten digit database, Available: http://yann. lecun. com/exdb/mnist, 1998. [58] K. He, X. Zhang, S. Ren, and J. Sun, Deep residual learning for image recognition. cvpr. 2016, ar Xiv preprint ar Xiv:1512.03385, 2016. A Gradient-based methods in MOO (Stochastic) Gradient-based methods in MOO have attracted much attention owing to simple update rules and less intensive computation recently, thus rendering them perfect candidates to underpin MOO applications in deep learning under first-oracle. However, their theoretical understandings remain less explored relative to their counterparts of single objective optimization. Hence, we highlight the existing works and corresponding assumptions alongside with convergence metrics. Existing Works. Various works managed to explore the convergence rates under different assumptions in strongly-convex, convex, and non-convex functions as listed in Table 4. Using full gradient, MGD [8] could achieve tight convergence rates in strongly-convex and non-convex cases, i.e., linear rate O(r T ), r (0, 1) and sub-linear rate O( 1 T ). However, it requires linear search of learning rate in the algorithm and sequence convergence ({xt} converges to x ). The linear search of learning rate is a classic technique, but does not fits in gradient-based algorithms in deep learning. Moreover, sequence convergence assumption is a too strong assumption. With no local step, our FMGDA degenerates to MGD. As a result, our analysis also provide the same order convergence rates in both strongly-convex and non-convex functions while avoiding strong and unpractical assumptions. If using stochastic gradient, SMGD methods makes a further complicated case. The stochastic gradient noise would complicate the analysis and thus it is still unclear whether SMGD is guaranteed to converge. [9] provided convergence rate for SMGD but extra assumptions and/or unreasonably large batch requirements were needed. On the other hand, [9] and [18] showed that the common descent direction provided by SMGD method is likely to be a biased estimation, rendering non-convergence issues. Recently, by utilizing momentum, Mo Co [19] and CR-MOGM [18] were proposed with corresponding convergence guarantees. [20] utilized direction-oriented approach by a preference direction. However, these analyses do not shed light on pure SMGD despite its widespread application. Assumptions. When applying stochastic gradient to MOO, common descent direction estimation λT F(x, ξ) (F(x) = [f1(x), , f S(x)]) is a biased estimation and thus rendering potential nonconvergence issues [9, 18]. This is a limitation for SMGD. However, SMGD does work well with a wide range of applications in practice. Understanding under what conditions can SMGD have convergence guarantee is thus an important problem. [56] assumes convexity property(H5): f(x, ξ) f(x , ξ) c 2 x x 2 almost sure. [9] utilizes weaker assumptions but still needs first moment bound (Assumption 5.2(b)): E[ f(x, ξ) f(x) ] η(a + b f(x) ) and Lipschitz continuity of λ (Assumption 5.4): λk λt β ( f1(xk) f1(xt))T , . . . , ( f S(xk) f S(xt))T . In this paper, we use (α, β)-Lipschitz continuous stochastic gradient (Assumption 4). In essence, we need the stochastic gradient estimation satisfying E[ f(x, ξ) f(y, ξ ) 2] α x y 2 + βσ2 for any two independent samples ξ and ξ . For the inequality E[ f(x, ξ) f(y, ξ ) 2] α x y 2 + βσ2 in Assumption 4, the notation σ2 just represents a general positive constant. This σ2 does not denote the variance of the stochastic gradient variance. Thus, this inequality does not Table 3: List of key notation. Notation Definition i Client index M Total number of clients s Objective/task index S Total number of Objectives/tasks Si Number of objectives/tasks of client i s interest k Local step index K Total number of local steps t Communication round index T Total number of communication rounds x Rd Global model parameters of FMOL in Problem (2) x0 Rd Initial solution of FMOL in Problem (2) x Rd A Pareto optimal solution of FMOL in Problem (2) ηL The learning rate on the client side ηt The learning rate on the server side in round t depend on the batch size of the stochastic gradient. More specifically, unlike the assumption in [9] that characterizes the difference between a stochastic gradient and its full gradient (hence depending on the batch size), our Assumption 4 only measures the average norm square of two stochastic gradient difference f(x, ξ) f(y, ξ ) given any two points x and y and any two samples ξ and ξ . In other words, Assumption 4 does not involve any full gradient, and hence no dependence on batch size. It is a natural extension of the classic Lipschitz continuous gradient assumption and could generalize existing assumptions. 1. If ξ and ξ are the whole dataset, by setting α = L2 and β = 0, (α, β)-Lipschitz continuous stochastic gradient condition generalizes the traditional Lipschitz continuous gradient assumption f(x) f(y) L x y . 2. If ξ is one data sample, ξ are the whole dataset and x = y, by setting α = 0 and β = 1, (α, β)-Lipschitz continuous stochastic gradient condition generalizes the traditional bounded variance assumption f(x, ξ) f(x) 2 σ2. 3. If ξ is one data sample, ξ are the whole dataset and x = y, by setting β = αk, (α, β)-Lipschitz continuous stochastic gradient condition generalizes the bound on the first moment assumption (assumption 5.2(b)) and bounded sets assumption (assumption 5.3) [9] (E[ f(x, ξ) f(x) ] αk(Ci + ˆCi fi(xk) ) and fi(x) M + LΘ). Metrics. For strongly-convex functions, we use t Q = P s [S] λt, s [fs(xt) fs(x )] as the metrics. We note similar metrics are used in other works. For example, [9] uses mint=1,...,T P s [S] λt, s fs(xt) λT fs(x ) where λT = PT t=1 t PT t=1 tλt. Here λt, s is calculated by the quadratic programming problem 3 with stochastic gradients. Rigorously speaking, the left-hand side is not guaranteed to be positive. But if we impose stronger assumptions as shown in [9, 56], we can have the same convergence metric as that in single objective optimization as an direct extension. In non-convex functions, dt 2 are used as the metrics for FMOO, where dt = λT t F(xt) and λt is calculated based on accumulated (stochastic) gradients t. We note, directly extended from MOO [18, 19] , d t = ˆλ T t F(xt) could also be used as the metrics in FMOO, where ˆλ t is calculated based on full gradients F(xt). However, we prefer dt for the following reasons: i). For applying gradient descent with no local steps, dt degenerates to d t . ii). Clearly, dt 2 d t 2 as ˆλ t is calculated based on gradients F(xt). Hence, dt 2 is stronger convergence measure for FMOO. iii). λt is calculated in the algorithm and thus being more practical to use in practice, while ˆλ t is unknown. Also, the convergence of dt implicitly indicates λt converges to ˆλ t . B Proof of gradient descent type methods For gradient descent type methods, each step utilizes a full gradient to update and the corresponding parameter λ is deterministic. For clarity of notation, we drop for λ, that is, we use λs t to represent the solution of quadratic problem (Step 6 in the algorithm) for task s in the t-th round. Lemma 1. Under bounded gradient assumption, the local model updates for any client s could be bounded Gt,k s,i = xt,k s,i xt 2 4η2 LK2G2, (4) Ht,s = fs(xt) t s 2 4η2 LK2L2G2. (5) Proof. For one task s [S] and one client i Rs, the local update xt xt,k s,i 2 could be further bounded. xt xt,k s,i 2 = xt xt,k 1 s,i + ηL fs,i(xt,k 1 s,i ) 2 (6) (1 + 1 K 1) xt xt,k 1 s,i 2 + η2 LK fs,i(xt,k 1 s,i ) 2 (7) Table 4: Convergence rate (shaded parts are our results) for strongly-convex and non-convex functions, respectively: Methods Rate Assumption Setting Algorithm SC NC Vanilla Gradient MGD [8] O(r T ), r (0, 1) O( 1 T ) Sequence convergence MGD O(exp( µT)) O( 1 T ) - SMGD [9] O( 1 T ) - Lipschitz continuity of λ SMGD [56] O( 1 T ) - Convexity property SMGD [39] - O( 1 T ) Given exact solution λ Momentum Mo Co [19] - O( 1 T ) - CR-MOGM [18] - O( 1 Federated Settings FMGDA O(exp( µT)) O( 1 T ) - FSMGDA O( 1 Assumptions. Linear search [8]: stepsize linear search; sequence convergence [8]: {xt} converges to x ; first moment bound (Asm. 5.2(b) [9]): E[ f(x, ξ) f(x) ] η(a + b f(x) ); Lipschitz continuity of λ (Asm. 5.4 [9]): λk λs β ( f1(xk) f1(xt))T , . . . , ( fm(xk) fm(xt))T ; convexity property(H5) [56]: f(x, ξ) f(x , ξ) c 2 x x 2 almost sure; (α, β)-Lipschitz continuous stochastic gradient (Asm. 4). (1 + 1 K 1) xt xt,k 1 s,i 2 + η2 LKG2 (8) 2η2 LKG2 1 + 1 K 1 " 1 + 1 K 1 (η2 LKG2) (10) 4η2 LK2G2, (11) where the first inequality comes from Young s inequality, the second inequality follows from bounded gradient assumption, and the last inequality follows if 1 + 1 K 1 K 1 4 for K > 1. We have the bound for local update for each task s, Ht,s, as follows: Ht,s = fs(xt) t s 2 (12) h fs,i(xt) fs,i(xt,k s,i) i fs,i(xt) fs,i(xt,k s,i) 2 (14) xt xt,k s,i 2 (15) 4η2 LK2L2G2. (16) Lemma 2. For general L-smooth functions {fs, s [S]}, choose the learning rate ηt s.t. ηt 3 2(1+L), the update dt of the algorithm satisfies: ηt 4 dt 2 fs(xt+1) + fs(xt) + 6η2 LK2L2G2 (17) fs(xt+1) fs(xt) + fs(xt), ηtdt + 1 2L ηtdt 2 (18) = fs(xt) + fs(xt) t s, ηtdt ηt t s, dt + 1 2L ηtdt 2 (19) fs(xt) + fs(xt) t s, ηtdt ηt dt 2 + 1 2L ηtdt 2 (20) 2 fs(xt) t s 2 + 1 2η2 t dt 2 ηt dt 2 + 1 2Lη2 t dt 2 (21) = fs(xt) + 1 2 fs(xt) t s 2 ηt fs(xt) + 2η2 LK2L2G2 ηt The third inequality follows from t s, dt dt 2 since dt is a general solution in the convex hull of the family of vectors { t s, s [S]} (see Lemma 2.1 [15]). Here dt = P s [S] λt, s t s and λt, s is calculated by t s, but we drop the of λ for simplicity. By setting 1 1 4, that is, ηt 3 2(1+L), we have ηt 4 dt 2 fs(xt+1) + fs(xt) + 2η2 LK2L2G2. (24) B.1 Strongly Convex Functions Theorem 3 (FMGDA for µ-Strongly Convex FMOL). Let ηt = η such that η 3 2(1+L), η 1 2L+µ and η 1 µT . Under Assumptions 13, pick xt as the final output of the FMGDA algorithm with weights wt = (1 µη 2 )1 t. Then, it holds that E[ t Q] x0 x 2µ exp( ηµT 2 ) + δ, where s [S] λt, s [fs(xt) fs(x )] and δ = 8η2 LK2L2G2S2 µ + 2η2 LK2L2G2. fs(xt+1) fs(xt) + fs(xt), ηtdt + 1 2L ηtdt 2 (25) fs(x ) + fs(xt), xt x µ 2 xt x 2 (26) + fs(xt), ηtdt + 1 2L ηtdt 2, (27) where the first inequality is due to L-smoothness, the second inequality follows from µ-strongly convex. s [S] λs t [fs(xt+1) fs(x )] (28) s [S] λs t fs(xt), xt x s [S] λs t fs(xt), ηtdt s [S] λs t fs(xt), xt x ηtdt 2 xt x 2 + 1 2L ηtdt 2 (30) = dt, xt x ηtdt µ 2 xt x 2 + 1 2L ηtdt 2 + s [S] λs t fs(xt) dt, xt x ηtdt = dt, xt x ηt dt 2 µ 2 xt x 2 + 1 2Lη2 t dt 2 + s [S] λs t fs(xt) dt, xt+1 x xt x 2 xt+1 x 2 1 2 xt x 2 + 1 2Lη2 t dt 2 (33) s [S] λs t fs(xt) dt +ϵ xt+1 x 2 (34) xt x 2 xt+1 x 2 1 2 xt x 2 + 1 2Lη2 t dt 2 (35) 4ϵHt + ϵ 2 xt x 2 + 2η2 t dt 2 (36) 2 ηt) xt x 2 xt+1 x 2 1 where xt x 2 xt+1 x 2 = η2 t dt 2 + 2 ηtdt, xt x , and we choose ϵ = µ 8 in the last inequality. From Lemma 2, it is clear that |fs(xt+1) fs(xt)| |2η2 LK2L2G2 ηt 4 dt 2| (38) 2η2 LK2L2G2 + ηt 4 dt 2. (39) s [S] λs t [fs(xt) fs(x )] X s [S] λs t [fs(xt+1) fs(x )] + |fs(xt+1) fs(xt)| (40) 2 ηt) xt x 2 xt+1 x 2 1 µHt + 2η2 LK2L2G2. s [S] λs t fs(xt) dt s [S] (λs t)2Ht,s (43) 4η2 LK2L2G2S2. (44) By setting ηt 1 2L+µ, we have s [S] λs t [fs(xt+1) fs(x )] (45) 2 ηt) xt x 2 xt+1 x 2 + 8η2 LK2L2G2S2 µ + 2η2 LK2L2G2 Averaging using weight wt = (1 µη 2 )1 t and using such weight to pick output x. By using Lemma 1 in [27] with η 1 u R, we ahve E[ Q] x0 x 2µ exp( ηµT 2 ) + δ (47) = O(µ exp( µT)) + O(δ). (48) If we set ηL sufficiently small such that δ = O(µ exp( µT)), then we have the convergence rate E[ Q] = O(µ exp( µT)). B.2 Non-Convex Functions Theorem 1 (FMGDA for Non-convex FMOL). Let ηt = η 3 2(1+L). Under Assumptions 1 and 2, if at least one function fs, s [S] is bounded from below by f min s , then the sequence {xt} output by FMGDA satisfies: mint [T ] dt 2 16(f 0 s f min s ) T η + δ, where δ 16η2 LK2L2G2(1+S2) Proof. From Lemma 2, we have ηt 4 dt 2 fs(xt+1) + fs(xt) + 2η2 LK2L2G2. (49) With constant learning rate ηt = η, t [T ] dt 2 4(f 0 s f min s ) Tη + 8η2 LK2L2G2 Note that dt 2 are used as the metrics for FMOO, where dt = λT t (Diag(FA )) and λt is calculated based on accumulated (stochastic) gradients t. Then we have s [S] λs t fs(xt) dt + 2 dt 2 . (51) t [T ] dt 2 16(f 0 s f min s ) Tη + 16η2 LK2L2G2(1 + S2) With constant learning rate η and local learning rate ηL = O( 1 T KLGS ), we have t [T ] dt 2 O( 1 C Proof of stochastic gradient descent type methods For stochastic gradient descent type methods, each step utilizes a stochastic gradient to update and the corresponding parameter λ is stochastic, depending on the random samples in each client. For clarity of notation, we drop for λ, that is, we use λs t to represent the solution of quadratic problem (Step 6 in the algorithm) for task s in the t-th round. Lemma 3. Under bounded stochastic gradient assumption, the local model updates could be bounded Gt,k s,i = E xt,k s,i xt 2 6η2 Lk2 fs,i(xt) 2 , (54) s [S] λt s t s Further with assumption 4, we have Ht,s = E fs(xt, ξt) t s 2 αη2 LK2D2 + βσ2. (56) Proof. For one task s [S] and one client i Rs, the local update xt xt,k s,i 2 could be further bounded. Gt,k s,i = E xt xt,k s,i 2 (57) τ [k] ηL fs,i(xt,τ s,i, ξt,τ s,i ) η2 Lk2D2. (59) s [S] λt s t s s [S] E h (λt s)2 t s 2i (60) s [S] E h t s 2i (61) τ [K] fs,i(xt,τ s,i, ξt,τ s,i ) τ [K] E fs,i(xt,τ s,i, ξt,τ s,i ) 2 (63) Ht,s = E fs(xt, ξt) t s 2 (65) fs,i(xt, ξt) fs,i(xt,k s,i, ξt,k s,i ) i Rs E fs,i(xt, ξt) fs,i(xt,k s,i, ξt,k s,i ) 2 (67) αE xt xt,k s,i 2 + βσ2 (68) αη2 LK2D2 + βσ2. (69) C.1 Strongly Convex Functions Theorem 7 (FSMGDA for µ-Strongly Convex FMOL). Let ηt = η = Ω( 1 µT ). Under Assumptions 3, 5 and 6, pick xt as the final output of the FSMGDA algorithm with weight wt = (1 µη 2 )1 t. Then, it holds that: E[ t Q] x0 x 2µ exp( η 2µT) + δ, where t Q = P s [S] λt, s [fs(xt) fs(x )] µS2(αη2 LK2D2 + βσ2) + ηS2D2 Proof. Taking expectation over random samples conditioning on xt, we have E xt+1 x 2 = E s [S] λt s t s x xt x , 2ηt X s [S] λt s t s s [S] λt s t s xt x , 2ηt X s [S] λt s fs(xt, ξt) xt x , 2ηt X s [S] λt s( fs(xt, ξt) t s) s [S] λt s t s xt x , 2ηt X s [S] E[λt s] fs(xt) xt x , 2ηt X s [S] λt s( fs(xt, ξt) t s) s [S] λt s t s 2 xt x 2 + X s [S] E[λt s](fs(xt) fs(x )) + ϵ xt x 2 (76) s [S] λt s( fs(xt, ξt) t s) s [S] λt s t s 2 xt x 2 + X s [S] E[λt s](fs(xt) fs(x )) + ϵ xt x 2 (78) 4ϵ4η2 t S X s [S] E h (λt s)2 ( fs(xt, ξt) t s) 2i + η2 t E s [S] λt s t s 2 xt x 2 + X s [S] E[λt s](fs(xt) fs(x )) + ϵ xt x 2 (80) 4ϵ4η2 t S X s [S] E fs(xt, ξt) t s 2 + η2 t E s [S] λt s t s 2 xt x 2 + X s [S] E[λt s](fs(xt) fs(x )) + ϵ xt x 2 (82) 4ϵ4η2 t S2(αη2 LK2D2 + βσ2) + η2 t S2D2 (83) 2 ) xt x 2 2ηt s [S] E[λt s](fs(xt) fs(x )) µηt S2(αη2 LK2D2 + βσ2) + η2 t S2D2, (85) where the first equality is due to strongly-convex objective functions, and we set ϵ = ηtµ s [S] E[λt s](fs(x) fs(x )) 1 2ηt (1 ηtµ 2 ) xt x 2 1 2ηt xt+1 x 2 (86) µS2(αη2 LK2D2 + βσ2) + ηt S2D2 Averaging using weight wt = (1 µηt 2 )1 t and using such weight to pick output x. By using Lemma 1 in [27] with constant learning rate ηt = η = Ω( 1 µT ), we have E[ Q] x0 x 2µ exp( η 2µT) + O(δ) (88) where δ = 1 µS2(αη2 LK2D2 + βσ2) + ηS2D2 By letting β = η, ηL = O( 1 T ) and η = Θ( log(max(1,µ2T )) C.2 Non-convex Functions Theorem 5 (FSMGDA for Non-convex FMOL). Let ηt = η 3 2(1+L). Under Assumptions 4 6, if an objective fs is bounded from below by f min s , then the sequence {xt} output by FSMGDA satisfies: mint [T ] E dt 2 8(f 0 s f min s ) ηT + δ, where δ = (2S2 + 4)(αη2 LK2D2 + βσ2). Proof. Similar to Lemma 2 and taking expectation on the random data samples conditioning on xt, we have Efs(xt+1) fs(xt) + E fs(xt), ηtdt + 1 2LE ηtdt 2 (90) = fs(xt) + E fs(xt) t s, ηtdt ηt E t s, dt + 1 2LE ηtdt 2 (91) fs(xt) + E fs(xt) t s, ηtdt ηt E dt 2 + 1 2LE ηtdt 2 (92) 2E fs(xt) t s 2 + 1 2η2 t E dt 2 ηt E dt 2 + 1 2LEη2 t dt 2 (93) = fs(xt) + 1 2E fs(xt) t s 2 ηt E dt 2, (94) where dt = P s [S] λt, s t s and λt, s is calculated by the accumulated stochastic gradients t s, s [S], but we drop the of λ for simplicity. With ηt 3 2(1+L), we have 4 E dt 2 fs(xt+1) + fs(xt) + 1 2E fs(xt) t s 2 (95) fs(xt+1) + fs(xt) + 1 2(αη2 LK2D2 + βσ2) (96) With constant learning rate ηt = η, 1 T t [T ] E dt 2 4 (fs(x1) Efs(x T +1)) ηT + 2(αη2 LK2D2 + βσ2) (97) Note that we want to use dt 2 are used as the metrics, where dt = λT t (Diag(FA )) and λt is calculated based on accumulated (stochastic) gradients t. Then we have s [S] λs t fs(xt) dt + 2 dt 2 . (98) With constant learning rate ηt = η and averaging from T communication rounds, we have t [T ] E dt 2 1 s [S] λs t fs(xt) dt t [T ] 2E dt 2 (99) t [T ] 2SE X λs t( fs(xt) t s) 2 + 1 t [T ] 2E dt 2 (100) 8 (fs(x1) Efs(x T +1)) ηT + (2S2 + 4)(αη2 LK2D2 + βσ2) (101) With constant learning rate η = 1 T , local learning rate ηL = O( 1 T 1/4 ) and β = η, t [T ] E dt 2 = O( 1 D Further Experiments and Additional Results In the following, we provide the detailed machine learning models for our experiments: 1) Multi MNIST Datasets and Learning Tasks: We test the convergence performance of our algorithms using the Multi MNIST dataset [53], which is a multi-task learning version of the MNIST dataset [57] from LIBSVM repository. Specifically, to convert the hand-written classification problem into a multi-task problem, we randomly chose 60000 images and divided them into M agents. Each agent has two tasks, where each task has n = 60000/(2 M) samples. Due to space limitations, we only present the convergence results for the case of non-i.i.d. data partition (i.e., data heterogeneity) and relegate the results of the i.i.d. data case to the appendix. For the non-i.i.d. data partition, we use the same data partition strategy as in [28], where each client can access data with at most two labels. In our experiments, a group of images is positioned in the top left corner, while another group of images is positioned in the bottom right. The two tasks are task L (to categorize the top-left digit) and task R (to classify the bottom-right digit). The overall problem is to classify the images of different tasks at different agents. All algorithms use the same randomly generated initial point. Here, we present experiments with M = 10 agents, where each agent has two tasks (i.e., A RM 2 is an all-one matrix). We set the local update rounds K = 10. Experiments with a larger number of agents (M = 5, 10, 30) are provided here. The learning rates are chosen as ηL = 0.1 and ηt = 0.1, t. 2): River Flow Dataset and Learning Tasks: We further test our algorithms on FMOL problems of larger sizes. In this experiment, we use the River Flow dataset[54], which is for flow prediction flow at eight locations within the Mississippi River network. Thus, there are eight tasks in this problem. In this experiment, we set ηL = 0.001, ηt = 0.1, M = 10, and keep the batch size = 256 while comparing K, and keep K = 30 while comparing the batch size. To better visualize 8 different tasks, we illustrate the normalized loss in radar charts in Fig. 2(b). We again verify that utilizing a larger training batch size and conducting additional local steps K results in accelerated convergence. 3): Celeb A Dataset and Learning Tasks: We utilize the Celeb A dataset [55], consisting of 200K facial images annotated with 40 attributes. We approach each attribute as a binary classification task, resulting in a 40-way multi-task learning (MTL) problem. To create a shared representation function, we implement Res Net-18 [58] without the final layer, attaching a linear layer to each attribute for classification. In this experiment, we set ηL = 0.0005, ηt = 0.1, M = 10, and K = 10. Figure 3 displays a radar chart depicting the loss value of each binary classification task. In Figure 3, we demonstrate the efficacy of our FMGDA and FSMGDA algorithms in both i.i.d. case and non-i.i.d. case. Experiments on i.i.d. data: First, we compare the convergence results with the same experimental settings in our Section. 5 but tested on the i.i.d data. As shown in Fig. 4, both FMGDA and FSMGDA (a) Non-i.i.d. case. (b) i.i.d case. Figure 3: Experiments on Celeb A dataset. 0 20 40 60 80 100 Iterations Batch size = 16 Batch size = 32 Batch size = 128 Batch size = 256 0 20 40 60 80 100 Iterations Batch size = 16 Batch size = 32 Batch size = 128 Batch size = 256 (a) Loss v.s. Communication rounds. 0 20 40 60 80 100 Iterations K = 1 K = 5 K = 10 K = 20 K = 50 0 20 40 60 80 100 Iterations K = 1 K = 5 K = 10 K = 20 K = 50 (b) Loss v.s. Local update rounds. Figure 4: Experiments on i.i.d. data. successfully converged in i.i.d. data, and the algorithm with a larger training batch size and more local updates K may converge faster. Impact of the number of clients: In this experiment, we choose the different number of clients from the discrete set {5, 10, 30} and fix learning rates at 0.1 and local update rounds at 10. As shown in Fig. 5, a larger number of workers leads to faster convergence rates of our proposed algorithms both in i.i.d. case and non-i.i.d. case; this is mainly because more samples have been used while training while having more workers. 0 20 40 60 80 100 Iterations M = 5 M = 10 M = 30 0 20 40 60 80 100 Iterations M = 5 M = 10 M = 30 (a) i.i.d. case. 0 20 40 60 80 100 Iterations M = 5 M = 10 M = 30 0 20 40 60 80 100 Iterations 0.0025 0.0050 0.0075 0.0100 0.0125 0.0150 0.0175 0.0200 M = 5 M = 10 M = 30 (b) Non-i.i.d. case. Figure 5: Loss value comparisons of algorithms on a different numbers of clients M. 0 20 40 60 80 100 Iterations 0 20 40 60 80 100 Iterations 0 20 40 60 80 Iterations 0.0050 0.0075 0.0100 0.0125 0.0150 0.0175 0.0200 0.0225 0 20 40 60 80 Iterations (b) FSGMDA. Figure 6: Comparisons of different step-sizes. Impact of the Step-size: In this experiment, we choose the value of the learning rate ηL from the discrete set {0.05, 0.01, 0.1} and fix worker number at 5, local update rounds at 10. As shown in Fig. 6, larger local step-sizes lead to faster convergence rates on both FMGDA algorithm and FSMGDA algorithm.