# order_matters_agentbyagent_policy_optimization__1791b45a.pdf Published as a conference paper at ICLR 2023 ORDER MATTERS: AGENT-BY-AGENT POLICY OPTIMIZATION Xihuai Wang1,2 , Zheng Tian3 , Ziyu Wan1,2, Ying Wen1, Jun Wang2,4, Weinan Zhang1 1 Shanghai Jiao Tong University, 2 Digital Brain Lab, 3 Shanghai Tech University, 4 University College London While multi-agent trust region algorithms have achieved great success empirically in solving coordination tasks, most of them, however, suffer from a nonstationarity problem since agents update their policies simultaneously. In contrast, a sequential scheme that updates policies agent-by-agent provides another perspective and shows strong performance. However, sample inefficiency and lack of monotonic improvement guarantees for each agent are still the two significant challenges for the sequential scheme. In this paper, we propose the Agent-byagent Policy Optimization (A2PO) algorithm to improve the sample efficiency and retain the guarantees of monotonic improvement for each agent during training. We justify the tightness of the monotonic improvement bound compared with other trust region algorithms. From the perspective of sequentially updating agents, we further consider the effect of agent updating order and extend the theory of non-stationarity into the sequential update scheme. To evaluate A2PO, we conduct a comprehensive empirical study on four benchmarks: Star Craft II, Multiagent Mu Jo Co, Multi-agent Particle Environment, and Google Research Football full game scenarios. A2PO consistently outperforms strong baselines. 1 INTRODUCTION Trust region learning methods in reinforcement learning (RL) (Kakade & Langford, 2002) have achieved great success in solving complex tasks, from single-agent control tasks (Andrychowicz et al., 2020) to multi-agent applications (Albrecht & Stone, 2018; Ye et al., 2020). The methods deliver superior and stable performances because of their theoretical guarantees of monotonic policy improvement. Recently, several works that adopt trust region learning in multi-agent reinforcement learning (MARL) have been proposed, including algorithms in which agents independently update their policies using trust region methods (de Witt et al., 2020; Yu et al., 2022) and algorithms that coordinate agents policies during the update process (Wu et al., 2021; Kuba et al., 2022). Most algorithms update the agents simultaneously, that is, all agents perform policy improvement at the same time and cannot observe the change of other agents, as shown in Fig. 1c. The simultaneous update scheme brings about the non-stationarity problem, i.e., the environment dynamic changes from one agent s perspective as other agents also change their policies (Hernandez-Leal et al., 2017). Rollout Rollout Rollout Multiple Rollouts at a Stage Single Rollout at Policy Update Simultaneous Policy Update A Stage: All Agents are Updated Update Update Update Update Update Figure 1: The taxonomy on the rollout scheme and the policy update scheme. In contrast to the simultaneous update scheme, algorithms that sequentially execute agent-byagent updates allow agents to perceive changes made by preceding agents, presenting another perspective for analyzing inter-agent interaction (Gemp et al., 2022). Bertsekas (2021) proposed a sequential update framework, named Rollout and Policy Iteration for a Single Agent (RPISA) in this paper, which performs a rollout every time an agent updates its policy (Fig. 1a). RPISA effectively turns non-stationary MARL problems into stationary single agent reinforcement learning (SARL) ones. It retains the theo- Work done at Digital Brain Lab. Correspondence to Zheng Tian and Weinan Zhang . Published as a conference paper at ICLR 2023 retical properties of the chosen SARL base algorithm, such as the monotonic improvement (Kakade & Langford, 2002). However, it is sample-inefficient since it only utilizes 1/n of the collected samples to update n agents policies. On the other hand, heterogeneous Proximal Policy Optimization (HAPPO) (Kuba et al., 2022) sequentially updates agents based on their local advantages estimated from the same rollout samples (Fig. 1b). Although it avoids the waste of collected samples and has a monotonic improvement on the joint policy, the policy improvement of a single agent is not theoretically guaranteed. Consequently, one agent s policy update may offset previous agents policy improvement, reducing the overall joint policy improvement. In this paper, we aim to combine the merits of the existing single rollout and sequential policy update schemes. Firstly, we show that naive sequential update algorithms with a single rollout can lose the monotonic improvement guarantee of PPO for a single agent s policy. To tackle this problem, we propose a surrogate objective with a novel off-policy correction method, preceding-agent offpolicy correction (Pre OPC), which retains the monotonic improvement guarantee on both the joint policy and each agent s policy. Then we further show that the joint monotonic bound built on the single agent bound is tighter than those of other simultaneous update algorithms and is tightened during updating the agents at a stage1. This leads to Agent-by-agent Policy Optimization (A2PO), a novel sequential update algorithm with single rollout scheme (Fig. 1b). Further, we study the significance of the agent update order and extend the theory of non-stationarity to the sequential update scheme. We test A2PO on four popular cooperative multi-agent benchmarks: Star Craft II, multi-agent Mu Jo Co, multi-agent particle environment, and Google Research Football full game scenarios. On all benchmark tasks, A2PO consistently outperforms strong baselines with a large margin in both performance and sample efficiency and shows an advantage in encouraging interagent coordination. To sum up, the main contributions of this work are as follows: 1. Monotonic improvement bound. We prove that the guarantees of monotonic improvement on each agent s policy could be retained under the single rollout scheme with the off-policy correction method Pre OPC we proposed. We further prove that the monotonic bound on the joint policy achieved given theoretical guarantees of each agent is the tightest among single rollout algorithms, yielding effective policy optimization. 2. A2PO algorithm. We propose A2PO, the first agent-by-agent sequential update algorithm that retains the monotonic policy improvement on both each agent s policy and the joint policy and does not require multiple rollouts when performing policy improvement. 3. Agent update order. We further investigate the connections between the sequential policy update scheme, the agent update order, and the non-stationarity problem, which motivates two novel methods: a semi-greedy agent selection rule for optimization acceleration and an adaptive clipping parameter method for alleviating the non-stationarity problem. 2 RELATED WORKS Trust Region Policy Optimization (TRPO) (Schulman et al., 2015) and Proximal Policy Optimization (PPO) (Schulman et al., 2017) are popular trust region algorithms with strong performances, benefiting from the guarantee of monotonic policy improvement (Kakade & Langford, 2002). Several recent works delve deeper into understanding these methods (Wang et al., 2019; Liu et al., 2019; Wang et al., 2020). In the multi-agent scenarios, de Witt et al. (2020) and Papoudakis et al. (2020) empirically studied the performance of Independent PPO in multi-agent tasks. Yu et al. (2022) conducted a comprehensive benchmark and analyzed the factor influential to the performance of Multi-agent PPO (MAPPO), a variant of PPO with centralized critics. Coordinate PPO (Co PPO) (Wu et al., 2021) integrates the value decomposition (Sunehag et al., 2017) and approximately performs a joint policy improvement with monotonic improvement. Several further trials to implement trust region methods are discussed in Wen et al. (2021); Li & He (2020); Sun et al. (2022); Ye et al. (2022). However, these MARL algorithms suffer from the non-stationarity problem as they update agents simultaneously. The environment dynamic changes from one agent s perspective as others also change their policies. Consequently, agents suffer from the high variance of gradients and require more samples for convergence (Hernandez-Leal et al., 2017). To alleviate the non-stationarity problem, Multi-Agent Mirror descent policy algorithm with Trust region decomposition (MAMT) (Li et al., 2022b) factorizes the trust regions of the joint policy and constructs the connections among the factorized trust regions, approximately constraining the diversity of joint policy. 1We define a stage as a period during which all the agents have been updated once (Fig. 1). Published as a conference paper at ICLR 2023 Rollout and Policy Iteration for a Single Agent (RPISA) (Bertsekas, 2021) and Heterogeneous PPO (HAPPO) (Kuba et al., 2022) consider the sequential update scheme. RPISA suffers from sample inefficiency as it requires n times of rollout for n agents to complete their policies update. Additionally, their work lacks a practical algorithm for complex tasks. In contrast, we propose a practical algorithm A2PO that updates all agents using the same samples from a single rollout. HAPPO is derived from the advantage decomposition lemma, proposed as Lemma 1 in Kuba et al. (2022). It does not consider the distribution shift caused by preceding agents, and has no monotonic policy improvement guarantee for each agent s policy. While A2PO is derived without decomposing the advantage, and has a guarantee of monotonic improvement for each agent s policy. We further discuss other MARL methods in Appx. C. 3 TRUST REGION METHOD IN SEQUENTIAL POLICY UPDATE SCHEME 3.1 MARL PROBLEM FORMULATION We consider formulating the sequential decision-making problem in multi-agent scenarios as a decentralized Markov decision process (DEC-MDP) (Bernstein et al., 2002). An n-agent DEC-MDP can be formalized as a tuple (S, {Ai}i N , r, T , γ), where N = {1, . . . , n} is the set of agents, S is the state space. Ai is the action space of agent i, and A = A1 An is the joint action space. r : S A 7 R is the reward function, and T : S A S 7 [0, 1] is the dynamics function denoting the transition probability. γ [0, 1) is a reward discount factor. At time step t, each agent i takes action ai t from its policy πi( |st), simultaneously according to the state st, forming the joint action at = {a1 t, . . . , an t } and the joint policy π( |st) = π1 . . . πn. The joint policy π of these n agents induces a normalized discounted state visitation distribution dπ, where dπ(s) = (1 γ) P t=0 γt Pr(st = s|π) and Pr( |π) : S 7 [0, 1] is the probability function under π. We then define the value function V π(s) = Eτ (T ,π)[P t=0 γtr(st, at)|s0 = s] and the advantage function Aπ(s, a) = r(s, a) + γEs T ( |s,a)[V π(s )] V π(s), where τ = {(s0, a0), (s1, a1), . . .} denotes one sampled trajectory. The agents maximize their expected return, denoted as: π = argmaxπ J (π) = argmaxπ Eτ (T ,π)[P t=0 γtr(st, at)] ,. 3.2 MONOTONIC IMPROVEMENT IN SEQUENTIAL POLICY UPDATE SCHEME We assume agents are updated in the order 1, 2, . . . , n, without loss of generality. We define π as the joint base policy from which the agents are updated at a stage, ei = {1, . . . , i 1} as the set of preceding agents updated before agent i, and πi as the updated policy of agent i. We denote the joint policy composed of updated policies of agents in the set ei, the updated policy of agent i and base policies of other agents as ˆπi = π1 . . . πi πi+1 . . . πn, and define ˆπ0 = π and ˆπn = π. A general sequential update scheme is shown as follows, where Lˆπi 1(ˆπi) is the surrogate objective for agent i: π = ˆπ0 maxπ1 Lπ(ˆπ1) Update π1 ˆπ1 ˆπn 1 maxπn Lˆ πn 1(ˆπn) Update πn ˆπn = π. We wish our sequential update scheme retains the desired monotonic improvement guarantee while improving the sample efficiency. Before going to our method, we first discuss why naively updating agents sequentially with the same rollout samples will fail in monotonic improvement for each agent. Since agent i updates its policy from ˆπi 1, an intuitive surrogate objective (Schulman et al., 2015) used by agent i could be formulated as LI ˆπi 1(ˆπi) = J (ˆπi 1) + Oπ(ˆπi), where Oπ(ˆπi) = 1 1 γ E(s,a) (dπ,ˆπi)[Aπ(s, a)] and the superscript I means Intuitive . The expected return, however, is not guaranteed to improve with such a surrogate objective, as elaborated in the following proposition. Proposition 1 For agent i, let ϵ = maxs,a |Aπ(s, a)|, αj = Dmax T V (πj πj) j (ei {i}), where DT V (p q) is the total variation distance between distributions p and q and we define Dmax T V (π π) = maxs DT V (π( |s) π( |s)), then we have: J (ˆπi) LI ˆπi 1(ˆπi) 2ϵαi 3 1 γ 2 1 γ(1 P j (ei {i}) αj) + Uncontrollable z }| { 2ϵ P 1 γ = βI i . (1) The proof can be found in Appx. A.3. Published as a conference paper at ICLR 2023 Remark. From Eq. (1) and the definition of LI ˆπi 1, we know J (ˆπi) J (ˆπi 1) > Oπ(ˆπi) βI i . Thus J (ˆπi) > J (ˆπi 1) when Oπ(ˆπi) > βI i , which can be satisfied by constraining βI i and optimizing Oπ(ˆπi). However, in βI i , the term 2ϵ P j ei αj/(1 γ), is uncontrollable by agent i. Consequently, the upper bound βI i may be large and the expected performance J (ˆπi) may not be improved after optimizing Oπ(ˆπi) when Oπ(ˆπi) < βI i even if αi is well constrained. Although one can still prove a monotonic guarantee for the joint policy by summing Eq. (1) for all the agents, we will show that the monotonic improvement on every single agent, if guaranteed, brings a tighter monotonic bound on the joint policy and incrementally tightens the monotonic bound on the joint policy when updating agents during a stage. Uncontrollable terms also appear when similarly analyzing HAPPO and cause the loss of monotonic improvement for a single agent2. 3.3 PRECEDING-AGENT OFF-POLICY CORRECTION The uncontrollable term in Prop. 1 is caused by one ignoring how the updating of its preceding agents policies influences its advantage function. We investigate reducing the uncontrollable term in policy evaluation. Since agent i is updated from ˆπi 1, the advantage function Aˆπi 1 should be used in agent i s surrogate objective rather than Aπ. However, Aˆπi 1 is impractical to estimate using samples collected under π due to the off-policyness (Munos et al., 2016) of these samples. Nevertheless, we can approximate Aˆπi 1 by correcting the discrepancy between ˆπi 1 and π at each time step (Harutyunyan et al., 2016). To retain the monotonic improvement properties, we propose preceding-agent off-policy correction (Pre OPC), which approximates Aˆπi 1 using samples collected under π by correcting the state probability at each step with truncated product weights: Aπ,ˆπi 1(st, at) = δt + X j=1 λ min 1.0, ˆπi 1(at+j|st+j) π(at+j|st+j) δt+k , (2) where δt = r(st, at) + γV (st+1) V (st) is the temporal difference for V (st), λ is a parameter controlling the bias and variance, as used in Schulman et al. (2016). min(1.0, ˆπi 1(at+j|st+j) π(at+j|st+j) ) j {1, . . . , k} are truncated importance sampling weights, approximating the probability of st+k at time step t + k under ˆπi 1. The derivation of Eq. (2) can be found in Appx. A.8. With Pre OPC, the surrogate objective of agent i becomes Lˆπi 1(ˆπi) = J (ˆπi 1)+ 1 1 γ E(s,a) (dπ,ˆπi)[Aπ,ˆπi 1(s, a)] , and we summarize the surrogate objective of updating all agents as follows: Gπ( π) = J (π) + 1 1 γ i=1 E(s,a) (dπ,ˆπi)[Aπ,ˆπi 1(s, a)] . (3) Note that Eq. (3) takes the sum of expectations of the global advantage function approximated under different joint policies, different from the advantage decomposition lemma in Kuba et al. (2022) which decomposes the global advantage function into local ones. We can now prove that the monotonic policy improvement guarantee of both updating one agent s policy and updating the joint policy is retained by using Eq. (3) as the surrogate objective. The detailed proofs can be found in Appx. A.4. Theorem 1 (Single Agent Monotonic Bound) For agent i, let ϵi = maxs,a |Aˆπi 1(s, a)|, ξi = maxs,a |Aπ,ˆπi 1(s, a) Aˆπi 1(s, a)|, αj = Dmax T V (πj πj) j (ei {i}), then we have: J (ˆπi) Lˆπi 1(ˆπi) 4ϵiαi 1 1 γ 1 1 γ(1 P j (ei {i}) αj) + ξi (1 γ)2 αi X j (ei {i}) αj + ξi The single agent monotonic bound depends on ϵi, ξi, and αi and the total variation distances of preceding agents. Unlike Eq. (1), we can effectively constrain the monotonic bound by controlling αi since ξi decreases as agent i updating its value function (Munos et al., 2016) and does not 2More discussions about why HAPPO fails to guarantee monotonic improvement for a single agent s policy can be found in Appx. A.6. Published as a conference paper at ICLR 2023 Table 1: Comparisons of trust region MARL algorithms. The proofs of the monotonic bounds can be found in Appx. A. Note that we also provide the monotonic bound of RPISA-PPO, which implements RPISA with PPO as the base algorithm. We separate RPISA-PPO from other methods as it has low sample efficiency and thus does not constitute a fair comparison. Algorithm Rollout Update Sample Efficiency Monotonic Bound RPISA-PPO Multiple Sequential Low 4ϵ Pn i=1 αi( 1 1 γ 1 1 γ(1 αi)) Single Agent: 4ϵαi( 1 1 γ 1 1 γ(1 αi)) MAPPO Single Simultaneous High 4ϵ Pn i=1 αi 1 γ Co PPO Single Simultaneous High 4ϵ Pn i=1 αi( 1 1 γ 1 1 γ(1 Pn j=1 αj)) HAPPO Single Sequential High 4ϵ Pn i=1 αi( 1 1 γ 1 1 γ(1 Pn j=1 αj)) Single Agent: No Guarantee A2PO (ours) Single Sequential High 4ϵ Pn i=1 αi( 1 1 γ 1 1 γ(1 P j (ei {i}) αj)) + 1 γ Single Agent: 4ϵiαi( 1 1 γ 1 1 γ(1 P j (ei {i}) αj)) + ξi lead to an unsatisfiable bound when αi is well constrained, providing the guarantee for monotonic improvement when updating a single agent. Given the above bound, we can prove the monotonic improvement of the joint policy. Theorem 2 (Joint Monotonic Bound) For each agent i N, let ϵi = maxs,a |Aˆπi 1(s, a)| , αi = Dmax T V (πi πi), ξi = maxs,a |Aπ,ˆπi 1(s, a) Aˆπi 1(s, a)|, and ϵ = maxi ϵi, then we have: |J ( π) Gπ( π)| 4ϵ i=1 αi 1 1 γ 1 1 γ(1 P j (ei {i}) αj) + Pn i=1 ξi j (ei {i}) αj + Pn i=1 ξi Eq. (5) suggests a condition for monotonic improvement of the joint policy, similar to that in the remark under Prop. 1. We further prove that the joint monotonic bound is incrementally tightened when performing the policy optimization agent-by-agent during a stage due to the single agent monotonic bound, i.e., the condition for improving J ( π) is relaxed and more likely to be satisfied. The details can be found in Appx. A.5. We present the monotonic bounds of other algorithms in Tab. 1. Since 1 1 γ(1 P j (ei {i}) αj) < 1 1 γ(1 Pn j=1 αj), Eq. (5) achieves the tightest bound compared to other single rollout algorithms, with ξi i N small enough. The assumption about ξi is valid since preceding-agent off-policy correction is a contraction operator, which is a corollary of Theorem 1 in Munos et al. (2016). A tighter bound improves expected performance by optimizing the surrogate objective more effectively (Li et al., 2022a). 4 AGENT-BY-AGENT POLICY OPTIMIZATION We first give a practical implementation for optimizing the surrogate objective Gπ( π). When updating agent i, the monotonic bound in Eq. (4) consists of the total variation distances related to the preceding agents and agent i, i.e., αi P j (ei {i}) αj. It suggests that we can control the monotonic bound by controlling total variation distances αj j (ei {i}), to effectively improve the expected performance. We consider applying the clipping mechanism to control the total variation distances αj j (ei {i}) (Queeney et al., 2021; Sun et al., 2022). In the surrogate objective of agent i, i.e., J (ˆπi 1)+ 1 1 γ E(s,a) (dπ,π)[ πi Q j ei πj Aπ,ˆπi 1(s, a)], J (ˆπi 1) has no dependence to agent i, while the joint policy ratio πi Q j ei πj in the advantage estimation is appropriate for applying the clipping mechanism. We further consider reducing the instability in estimating agent i s policy gradient by clipping the joint policy ratio of preceding agents first, with a narrower clipping range (Wu et al., 2021). Thus we apply the clipping mechanism on the joint policy ratio twice: once on the joint policy ratio of preceding agents and once on the policy ratio of agent i. Finally, the practical objective for updating agent i becomes: Lˆπi 1(ˆπi) = E(s,a) (dπ,π) min l(s, a)Aπ,ˆπi 1, clip l(s, a), 1 ϵi Aπ,ˆπi 1 , (6) Published as a conference paper at ICLR 2023 where l(s, a) = πi(ai|s) πi(ai|s)g(s, a), and g(s, a) = clip( j ei πj(aj|s) Q j ei πj(aj|s), 1 ϵi 2 ). The clipping parameter ϵi is selected as ϵi = C(ϵ, i), where ϵ is the base clipping parameter and C( , ) is the clipping parameter adapting function. We summarize our proposed Agent-by-agent Policy Optimization (A2PO) in Alg. 1. Note that in Line 6, the agent for the next update iteration is selected according to the agent selection rule R( ). Algorithm 1: Agent-by-agent Policy Optimization (A2PO) 1 Initialize the joint policy π0 = {π1 0, . . . , πn 0 }, and the global value function V . 2 for iteration m = 1, 2, . . . do 3 Collect data using πm 1 = {π1 m 1, . . . , πn m 1}. 4 for Order k = 1, . . . , n do 5 Select an agent according to the selection rule as i = R(k). 6 Policy πi m = πi m 1, preceding agents ei = {R(1), . . . , R(k 1)}. 7 Joint policy ˆπi = {πi m, πj ek m , πj N ek m 1 }. 8 Compute the advantage approximation as Aπ,ˆπi 1(s, a) via Eq. (2). 9 Compute the value target v(st) = Aπ,ˆπi 1(s, a) + V (s). 10 for P epochs do 11 πi m = arg maxπim Lˆπi 1(ˆπi) as in Eq. (6). 12 V = arg min V Es dπ v(s) V (s) 2. Eq. (6) approximates the surrogate objective of a single agent. We remark that the monotonic improvement guarantee of a single agent reveals how the update of a single agent affects the overall objective. We will further discuss R( ) and C( , ) from the perspective of how to benefit the optimization of the overall surrogate objective by coordinating the policy updates of each agent. Semi-greedy Agent Selection Rule. With the monotonic policy improvement guarantee on the joint policy, as shown in Thm. 2, we can effectively improve the expected performance J ( π) by optimizing the surrogate objective of all agents Gπ( π) = J (π) + Pn i=1 Lˆπi 1(ˆπi). Since the policies except πi are fixed when maximizing Lˆπi 1(ˆπi), we recognize maximizing Pn i=1 Lˆπi 1(ˆπi) as performing a block coordinate ascent, i.e., iteratively seeking to update a block of chosen coordinates (agents) while other blocks (agents) are fixed. As a special case of the coordinate selection rule, the agent selection rule becomes crucial for convergence. On the one hand, intuitively, updating agent with a bigger absolute value of the advantage function contributes more to optimizing Gπ( π). Inspired by the Gauss-Southwell rule (Gordon & Tibshirani, 2015), we propose the greedy agent selection rule, under which an agent with a bigger absolute value of the expected advantage function is updated with a higher priority. We will verify that the agents with small absolute values of the advantage function also benefit from the greedy selelction rule in Appx. B.2.5. On the other hand, purely greedy selection may lead to early convergence which harms the performance. Therefore, we introduce randomness into the agent selection rule to avoid converging too early (Lu et al., 2018). Combining the merits, we propose the semi-greedy agent selection rule as ( R(k) = arg maxi (N e) Es,ai[|Aπ,ˆπR(k 1)|], k2 = 0 R(k) U(N e), k2 = 1, where e = {R(1), . . . , R(k 1)} and U is a uniform distribution. We verify that the semi-greedy agent selection rule contributes to the performance of A2PO in Sec. 5.2. Adaptive Clipping Parameter. We improve the sample efficiency by updating all agents using the samples collected under the base joint policy π. However, when updating agent i by optimizing 1 1 γ E(s,a) (dπ,ˆπi)[Aπ,ˆπi 1(s, a)], the expectation of advantage function is estimated using the states sampled under π instead of ˆπi 1, which reintroduces the non-stationarity since agent i can not perceive the change of the preceding agents. With the non-stationarity modeled by the state transition shift (Sun et al., 2022), we define the state transition shift encountered when updating agent i as π1,..., πi 1,πi,...,πn π1,...,πn (s |s) = P a[T (s |s, a)(ˆπi 1(a|s) π(a|s))]. The state transition shift has the following property. Published as a conference paper at ICLR 2023 Proposition 2 The state transition shift π1,..., πi 1,πi,...,πn π1,...,πn (s |s) can be decomposed as follows. π1,..., πi 1,πi,...,πn π1,...,πn = π1,π2,...,πn π1,...,πn + π1, π2,π3,...,πn π1,π2,...,πn + + π1,..., πi 1,πi,...,πn π1,..., πi 2,πi 1,...,πn Prop. 2 shows that the total state transition shift encountered by agent i can be decomposed into the sum of state transition shift caused by each agent whose policy has been updated. Shifts caused by agents with higher priorities will be encountered by more following agents and thus contribute more to the non-stationarity problem. Recall that the state transition shift effectively measures the total variation distance between policies. Therefore, in order to reduce the non-stationarity brought by the agents policy updates, we can adaptively clip each agent s surrogate objective according to their update priorities. We propose a simple yet effective method, named adaptive clipping parameter, to adjust the clipping parameters according to the updating order: C(ϵ, k) = ϵ cϵ + ϵ (1 cϵ) k/n, where cϵ is a hyper-parameter. We demonstrate how the agents with higher priorities affect the following agents in Fig. 2. Under the clipping mechanism, the influence of the agents with higher priority could be reflected in the clipping ranges of the joint policy ratio. The policy changes of the preceding agents may constrain the following agents to optimize the surrogate objective within insufficient clipping ranges, as shown on the left side of Fig. 2. The right side of Fig. 2 demonstrates that the adaptive clipping parameter method leads to balanced and sufficient clipping ranges. Figure 2: The clipping ranges of three agents. The surface a1 + a2 + a3 = 1 demonstrates the policy space of three discrete actions. The agents are updated in the order of 2, 3, 1. The areas in gray/pink are the clipping ranges with/without considering the joint policy ratio of preceding agents. Left: The agents have the same clipping parameters. The clipping range of agent 1 is insufficient due to the large variation in the policies of agent 2 and agent 3. Right: The clipping ranges are more balanced and sufficient with the adaptive clipping parameter method. 5 EXPERIMENTS In this section, we empirically evaluate and analyze A2PO in the widely adopted cooperative multiagent benchmarks, including the Star Craft II Multi-agent Challenge (SMAC) (Samvelyan et al., 2019), Multi-agent Mu Jo Co (MA-Mu Jo Co) (de Witt et al., 2020), Multi-agent Particle Environment (MPE) (Lowe et al., 2017)3, and more challenging Google Research Football (GRF) full-game scenarios (Kurach et al., 2020). Experimental results demonstrate that 1) A2PO achieves performance and efficiency superior to those of state-of-the-art MARL Trust Region methods, 2) A2PO has strength in encouraging coordination behaviors to complete complex cooperative tasks, and 3) the Pre OPC, the semi-greedy agent selection rule, and the adaptive clipping parameter methods significantly contribute to the performance improvement. 4 We compare A2PO with advanced MARL trust-region methods: MAPPO (Yu et al., 2022), Co PPO (Wu et al., 2021) and HAPPO (Kuba et al., 2022). We implement all the algorithms as parameter sharing in SMAC and MPE, and as parameter-independent in MA-Mu Jo Co and GRF, according to the homogeneity and heterogeneity of agents. We divide the agents into blocks for tasks with numerous agents to control the training time of A2PO comparable to other algorithms. Full experimental details can be found in Appx. B. 5.1 PERFORMANCE AND EFFICIENCY We evaluate the algorithms in 9 maps of SMAC with various difficulties, 14 tasks of 6 scenarios in MA-Mu Jo Co, and the 5-vs-5 and 11-vs-11 full game scenarios in GRF. Results in Tab. 2, Fig. 3, and Fig. 4 show that A2PO consistently outperforms the baselines and achieves higher sample efficiency in all benchmarks. More results and the experimental setups can be found in Appx. B.2. 3We evaluate A2PO in fully cooperative and general-sum MPE tasks respectively, showing the potential of extending A2PO to general-sum games, see Appx. B.2.3 for full results. 4Code is available at https://anonymous.4open.science/r/A2PO. Published as a conference paper at ICLR 2023 Star Craft II Multi-agent Challenge (SMAC). As shown in Tab. 2, A2PO achieves (nearly) 100% win rates in 6 out of 9 maps and significantly outperforms other baselines in most maps. In Tab. 2, we additionally compare the performance with that of Qmix (Rashid et al., 2018), a well known baseline in SMAC. We also observe that Co PPO and A2PO have better stability as they consider clipping joint policy ratios. Table 2: Median win rates and standard deviations on SMAC tasks. Map Difficulty MAPPO w/ PS Co PPO w/ PS HAPPO w/ PS A2PO w/ PS Qmix w/ PS MMM Easy 96.9(0.988) 96.9(1.25) 95.3(2.48) 100(1.07) 95.3(5.2) 3s5z Hard 84.4(4.39) 92.2(2.35) 92.2(1.74) 98.4(1.04) 88.3(2.9) 5m vs 6m Hard 84.4(2.77) 84.4(2.12) 87.5(2.51) 90.6(3.06) 75.8(3.7) 8m vs 9m Hard 84.4(2.39) 84.4(2.04) 96.9(3.78) 100(1.04) 92.2(2.0) 10m vs 11m Hard 93.8(18.7) 96.9(2.6) 98.4(2.99) 100(0.521) 95.3(1.0) 6h vs 8z Super Hard 87.5(1.53) 90.6(0.765) 87.5(1.49) 90.6(1.32) 9.4(2.0) 3s5z vs 3s6z Super Hard 82.8(19.2) 84.4(2.9) 37.5(13.2) 93.8(19.8) 82.8(5.3) MMM2 Super Hard 90.6(8.89) 90.6(6.93) 51.6(9.01) 98.4(1.25) 87.5(2.6) 27m vs 30m Super Hard 93.8(3.75) 93.8(2.2) 90.6(4.77) 100(1.55) 39.1(9.8) Overall / 88.7(6.96) 90.5(2.57) 81.9(4.67) 96.9(3.41) 74.0(3.83) Multi-agent Mu Jo Co environment (MA-Mu Jo Co). We investigate whether A2PO can scale to more complex continuous control multi-agent tasks in MA-Mu Jo Co. We calculate the normalized score return minimum return maximum return minimum return over all the 14 tasks in the left of Fig. 3. We also present part of results in the right of Fig. 3, where the control complexity and observation dimension, depending on the number of the robot s joints, increases from left to right. We observe that A2PO generally shows an increasing advantage over the baselines with increasing task complexity. 0.0 0.2 0.4 0.6 0.8 1.0 Steps 107 Normalized Score Normalized Score in 14 Tasks MAPPO w/o PS Co PPO w/o PS HAPPO w/o PS A2PO w/o PS 0.00 0.25 0.50 0.75 1.00 Steps 107 Episode Return Ant-v2 2x4d 0.00 0.25 0.50 0.75 1.00 Steps 107 0 Episode Return Walker2d-v2 3x2 0.00 0.25 0.50 0.75 1.00 Steps 107 Episode Return Humanoid-v2 17x1 MAPPO w/o PS Co PPO w/o PS HAPPO w/o PS A2PO w/o PS Figure 3: Experiments in MA-Mu Jo Co. Left: Normalized scores on all the 14 tasks. Right: Comparisons of averaged return on selected tasks. The number of robot joints increases from left to right. Google Research Football (GRF). We evaluate A2PO in GRF full-game scenarios, where agents have difficulty discovering complex coordination behaviors. A2PO obtains nearly 100% win rate in the 5-vs-5 scenario. In both scenarios, we attribute the performance gain of A2PO to the learned coordination behavior. We analyze the experiments in GRF to verify that A2PO encourages agents to learn coordination behaviors in complex tasks. In Tab. 3, an Assist is attributed to the player who passes the ball to the teammate that makes a score, a Pass is counted when the passing-andreceiving process is finished, Pass Rate is the proportion of success passes over the pass attempts. A2PO have an advantage in passing-and-receiving coordination, leading to more assists and scores. 0 5000 10000 15000 Episodes Football 5 vs 5 0 25000 50000 75000 100000 Episodes Football 11 vs 11 MAPPO Co PPO HAPPO A2PO Figure 4: Averaged win rate on the Google Research Football full-game scenarios. Table 3: Learned behaviors on the Google Research Football 5-vs-5 scenario. Bigger values are better except fot the Lost metric. Metric MAPPO Co PPO HAPPO A2PO Assist 0.04(0.02) 0.19(0.08) 0.07(0.05) 0.56(0.20) Goal 1.95(1.17) 4.42(2.08) 2.68(0.86) 9.01(0.95) Lost 0.49(0.11) 0.74(0.33) 1.04(0.12) 0.78(0.15) Pass 1.52(0.13) 3.44(1.04) 4.03(1.97) 6.42(2.23) Pass Rate 19.3(10.0) 35.0(10.3) 48.9(25.7) 67.1(11.7) Published as a conference paper at ICLR 2023 5.2 ABLATION STUDY This section studies how Pre OPC, the semi-greedy agent selection rule, and the adaptive clipping parameter affect the performance. Full ablation details can be found in Appx. B.2.5 0.00 0.25 0.50 0.75 1.00 Steps 107 Episode Return RPISA-PPO Asymptotic Performance Humanoid-v2 9|8 0.00 0.25 0.50 0.75 1.00 Steps 107 RPISA-PPO Asymptotic Performance MAPPO MAPPO w/ V-trace Co PPO Co PPO w/ V-trace HAPPO HAPPO w/ Pre OPC A2PO w/o Pre OPC A2PO Figure 5: Ablation experiments on precedingagent off-policy correction. Pre OPC. Fig. 5 shows the effects of utilizing off-policy correction in two cases: 1) Correction on all agents policies for simultaneous update algorithms, i.e., MAPPO w/ V-trace (Espeholt et al., 2018) and Co PPO w/ V-trace, and 2) Correction on the preceding agents policies for sequential update algorithms, i.e., HAPPO w/ Pre OC and A2PO. V-trace brings no general improvement to MAPPO and Co PPO, while Pre OPC significantly improves the sequential update cases. Pre OPC improves the performance of HAPPO significantly, while A2PO still outperforms HAPPO w/ Pre OPC. The performance gap lies in that A2PO clips the joint policy ratios, which matches the monotonic bound in Thm. 1. The results verify that A2PO reaches or outperforms the asymptotic performance of RPISA-PPO using an approximated advantage function and updating all the agents with the same rollout samples. Additionally, preceding-agent off-policy correction does not increase the sensitivity of the hyper-parameter λ, as shown in Appx. B.2.5. Agent Selection Rule. We provide comparisons of different agent selection rules in Fig. 6. The Cyclic rule means select agents in the order 1, . . . , n, and other rules have been introduced in sec. 4. The semi-greedy rule considers the optimization acceleration and the performance balance among agents and thus performs the best in all tasks. Adaptive Clipping Parameter. We propose the adaptive clipping parameter method for balanced and sufficient clipping ranges of agents. As shown in Fig. 7, the adaptive clipping parameter contributes to the performance gain of A2PO. 0.00 0.25 0.50 0.75 1.00 Steps 107 Episode Return Humanoid-v2 9|8 0.00 0.25 0.50 0.75 1.00 Steps 107 HAPPO-Cyclic HAPPO-Random HAPPO-Greedy HAPPO-Semi-greedy A2PO-Cyclic A2PO-Random A2PO-Greedy A2PO-Semi-greedy Figure 6: Ablation experiments on the agent selection rules. 0.00 0.25 0.50 0.75 1.00 Steps 107 Episode Return Humanoid-v2 9|8 0.00 0.25 0.50 0.75 1.00 Steps 107 MAPPO MAPPO w/ Adapt Co PPO Co PPO w/ Adapt HAPPO HAPPO w/ Adapt A2PO w/o Adapt A2PO Figure 7: Ablation experiments on the adaptive clipping parameter method. 6 CONCLUSION In this paper, we investigate the potential of the sequential update scheme in coordination tasks. We introduce A2PO, a sequential algorithm using a single rollout at a stage, which guarantees monotonic improvement on both the joint policy and each agent s policy. We also justify that the monotonic bound achieved by A2PO is the tightest among existing trust region MARL algorithms under single rollout scheme. Furthermore, A2PO integrates the proposed semi-greedy agent selection rule and adaptive clipping parameter method. Experiments in various benchmarks demonstrate that A2PO consistently outperforms state-of-the-art methods in performance and sample efficiency and encourages coordination behaviors for completing complex tasks. For future work, we plan to analyze the theoretical underpinnings of the agent selection rules and study the learnable methods to select agents and clipping parameters. Published as a conference paper at ICLR 2023 Acknowledgements. The SJTU team is partially supported by New Generation of AI 2030 Major Project (2018AAA0100900), the Shanghai Municipal Science and Technology Major Project (2021SHZDZX0102), the Shanghai Sailing Program (21YF1421900), the National Natural Science Foundation of China (62076161, 62106141). Xihuai Wang and Ziyu Wan are supported by Wu Wen Jun Honorary Scholarship, AI Institute, Shanghai Jiao Tong University. We thank Yan Song and He Jiang for their help in the football experiments. Ethics Statement. Our method and algorithm do not involve any adversarial attack, and will not endanger human security. All our experiments are performed in the simulation environment, which does not involve ethical and fair issues. Reproducibility Statement. The source code of this paper is available at https:// anonymous.4open.science/r/A2PO. We provide proofs in appx. A, including the proofs of intuitive sequential update, monotonic policy improvement of A2PO, incrementally tightened bound of A2PO and monotonic policy improvement of MAPPO, Co PPO and HAPPO. We specify all the experiments implementation details, the experiments setup, and the additional results in the appx. B. The related works of coordinate descent are shown in appx. D. Stefano V. Albrecht and Peter Stone. Autonomous agents modelling other agents: A comprehensive survey and open problems. Artif. Intell., 258:66 95, 2018. doi: 10.1016/j.artint.2018.01.002. URL https://doi.org/10.1016/j.artint.2018.01.002. Agarwal Alekh, Jiang Nan, M. Kakade Sham, and Sun Wen. Reinforcement Learning: Theory and Algorithms, 2022. URL https://rltheorybook.github.io/. Marcin Andrychowicz, Anton Raichuk, Piotr Stanczyk, Manu Orsini, Sertan Girgin, Rapha el Marinier, L eonard Hussenot, Matthieu Geist, Olivier Pietquin, Marcin Michalski, Sylvain Gelly, and Olivier Bachem. What matters in on-policy reinforcement learning? A large-scale empirical study. Co RR, abs/2006.05990, 2020. URL https://arxiv.org/abs/2006.05990. Daniel S. Bernstein, Robert Givan, Neil Immerman, and Shlomo Zilberstein. The complexity of decentralized control of markov decision processes. Math. Oper. Res., 27(4):819 840, 2002. doi: 10.1287/moor.27.4.819.297. URL https://doi.org/10.1287/moor.27.4.819.297. Dimitri Bertsekas. Multiagent Reinforcement Learning: Rollout and Policy Iteration. IEEE/CAA Journal of Automatica Sinica, 8(2):249 272, February 2021. ISSN 2329-9266, 2329-9274. doi: 10.1109/JAS.2021.1003814. Christian Schr oder de Witt, Tarun Gupta, Denys Makoviichuk, Viktor Makoviychuk, Philip H. S. Torr, Mingfei Sun, and Shimon Whiteson. Is independent learning all you need in the starcraft multi-agent challenge? Co RR, abs/2011.09533, 2020. URL https://arxiv.org/abs/ 2011.09533. Lasse Espeholt, Hubert Soyer, R emi Munos, Karen Simonyan, Volodymyr Mnih, Tom Ward, Yotam Doron, Vlad Firoiu, Tim Harley, Iain Dunning, Shane Legg, and Koray Kavukcuoglu. IMPALA: scalable distributed deep-rl with importance weighted actor-learner architectures. In Jennifer G. Dy and Andreas Krause (eds.), Proceedings of the 35th International Conference on Machine Learning, ICML 2018, Stockholmsm assan, Stockholm, Sweden, July 10-15, 2018, volume 80 of Proceedings of Machine Learning Research, pp. 1406 1415. PMLR, 2018. URL http://proceedings.mlr.press/v80/espeholt18a.html. Ian M. Gemp, Charlie Chen, and Brian Mc Williams. The generalized eigenvalue problem as a nash equilibrium. Co RR, abs/2206.04993, 2022. doi: 10.48550/ar Xiv.2206.04993. URL https: //doi.org/10.48550/ar Xiv.2206.04993. Tobias Glasmachers and Ur un Dogan. Accelerated coordinate descent with adaptive coordinate frequencies. In Cheng Soon Ong and Tu Bao Ho (eds.), Asian Conference on Machine Learning, ACML 2013, Canberra, ACT, Australia, November 13-15, 2013, volume 29 of JMLR Workshop and Conference Proceedings, pp. 72 86. JMLR.org, 2013. URL http://proceedings. mlr.press/v29/Glasmachers13.html. Published as a conference paper at ICLR 2023 Geoff Gordon and Ryan Tibshirani. Coordinate descent. Optimization, 10(36):725, 2015. Tuomas Haarnoja, Aurick Zhou, Pieter Abbeel, and Sergey Levine. Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor. In International conference on machine learning, pp. 1861 1870. PMLR, 2018. Anna Harutyunyan, Marc G. Bellemare, Tom Stepleton, and R emi Munos. Q(λ) with off-policy corrections. In Ronald Ortner, Hans Ulrich Simon, and Sandra Zilles (eds.), Algorithmic Learning Theory - 27th International Conference, ALT 2016, Bari, Italy, October 19-21, 2016, Proceedings, volume 9925 of Lecture Notes in Computer Science, pp. 305 320, 2016. doi: 10.1007/ 978-3-319-46379-7\ 21. URL https://doi.org/10.1007/978-3-319-46379-7_ 21. Qiang He and Xinwen Hou. Wd3: Taming the estimation bias in deep reinforcement learning. In 2020 IEEE 32nd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 391 398, 2020. doi: 10.1109/ICTAI50040.2020.00068. Pablo Hernandez-Leal, Michael Kaisers, Tim Baarslag, and Enrique Munoz de Cote. A survey of learning in multiagent environments: Dealing with non-stationarity. Co RR, abs/1707.09183, 2017. URL http://arxiv.org/abs/1707.09183. Sham M. Kakade and John Langford. Approximately optimal approximate reinforcement learning. In Claude Sammut and Achim G. Hoffmann (eds.), Machine Learning, Proceedings of the Nineteenth International Conference (ICML 2002), University of New South Wales, Sydney, Australia, July 8-12, 2002, pp. 267 274. Morgan Kaufmann, 2002. Jakub Grudzien Kuba, Ruiqing Chen, Muning Wen, Ying Wen, Fanglei Sun, Jun Wang, and Yaodong Yang. Trust region policy optimisation in multi-agent reinforcement learning. In International Conference on Learning Representations, 2022. URL https://openreview.net/forum? id=Ec GGFk NTxd J. Karol Kurach, Anton Raichuk, Piotr Stanczyk, Michal Zajac, Olivier Bachem, Lasse Espeholt, Carlos Riquelme, Damien Vincent, Marcin Michalski, Olivier Bousquet, and Sylvain Gelly. Google research football: A novel reinforcement learning environment. In AAAI 2020, pp. 4501 4510. AAAI Press, 2020. URL https://ojs.aaai.org/index.php/AAAI/ article/view/5878. Chenghao Li, Tonghan Wang, Chengjie Wu, Qianchuan Zhao, Jun Yang, and Chongjie Zhang. Celebrating Diversity in Shared Multi-Agent Reinforcement Learning. In Advances in Neural Information Processing Systems, November 2021. Hepeng Li and Haibo He. Multi-agent trust region policy optimization. Co RR, abs/2010.07916, 2020. URL https://arxiv.org/abs/2010.07916. Hepeng Li, Nicholas Clavette, and Haibo He. An analytical update rule for general policy optimization. In Kamalika Chaudhuri, Stefanie Jegelka, Le Song, Csaba Szepesv ari, Gang Niu, and Sivan Sabato (eds.), International Conference on Machine Learning, ICML 2022, 17-23 July 2022, Baltimore, Maryland, USA, volume 162 of Proceedings of Machine Learning Research, pp. 12696 12716. PMLR, 2022a. URL https://proceedings.mlr.press/v162/li22d.html. Wenhao Li, Xiangfeng Wang, Bo Jin, Junjie Sheng, and Hongyuan Zha. Dealing with nonstationarity in MARL via trust-region decomposition. In International Conference on Learning Representations, 2022b. URL https://openreview.net/forum?id=XHUxf5a RB3s. Boyi Liu, Qi Cai, Zhuoran Yang, and Zhaoran Wang. Neural trust region/proximal policy optimization attains globally optimal policy. In Hanna M. Wallach, Hugo Larochelle, Alina Beygelzimer, Florence d Alch e-Buc, Emily B. Fox, and Roman Garnett (eds.), Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, Neur IPS 2019, December 8-14, 2019, Vancouver, BC, Canada, pp. 10564 10575, 2019. URL https://proceedings.neurips.cc/paper/2019/ hash/227e072d131ba77451d8f27ab9afdfb7-Abstract.html. Published as a conference paper at ICLR 2023 Ryan Lowe, Yi Wu, Aviv Tamar, Jean Harb, Pieter Abbeel, and Igor Mordatch. Multi-agent actorcritic for mixed cooperative-competitive environments. In Isabelle Guyon, Ulrike von Luxburg, Samy Bengio, Hanna M. Wallach, Rob Fergus, S. V. N. Vishwanathan, and Roman Garnett (eds.), Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4-9, 2017, Long Beach, CA, USA, pp. 6379 6390, 2017. URL https://proceedings.neurips.cc/paper/2017/hash/ 68a9750337a418a86fe06c1991a1d64c-Abstract.html. Haihao Lu, Robert M. Freund, and Vahab S. Mirrokni. Accelerating greedy coordinate descent methods. In Jennifer G. Dy and Andreas Krause (eds.), Proceedings of the 35th International Conference on Machine Learning, ICML 2018, Stockholmsm assan, Stockholm, Sweden, July 1015, 2018, volume 80 of Proceedings of Machine Learning Research, pp. 3263 3272. PMLR, 2018. URL http://proceedings.mlr.press/v80/lu18b.html. Remi Munos, Tom Stepleton, Anna Harutyunyan, and Marc Bellemare. Safe and Efficient Off Policy Reinforcement Learning. In Advances in Neural Information Processing Systems, volume 29. Curran Associates, Inc., 2016. Georgios Papoudakis, Filippos Christianos, Lukas Sch afer, and Stefano V. Albrecht. Comparative evaluation of multi-agent deep reinforcement learning algorithms. Co RR, abs/2006.07869, 2020. URL https://arxiv.org/abs/2006.07869. Bei Peng, Tabish Rashid, Christian Schroeder de Witt, Pierre-Alexandre Kamienny, Philip Torr, Wendelin Boehmer, and Shimon Whiteson. Facmac: Factored multi-agent centralised policy gradients. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan (eds.), Advances in Neural Information Processing Systems, volume 34, pp. 12208 12221. Curran Associates, Inc., 2021. URL https://proceedings.neurips.cc/paper/2021/file/ 65b9eea6e1cc6bb9f0cd2a47751a186f-Paper.pdf. James Queeney, Yannis Paschalidis, and Christos G Cassandras. Generalized proximal policy optimization with sample reuse. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan (eds.), Advances in Neural Information Processing Systems, volume 34, pp. 11909 11919. Curran Associates, Inc., 2021. URL https://proceedings.neurips. cc/paper/2021/file/63c4b1baf3b4460fa9936b1a20919bec-Paper.pdf. Tabish Rashid, Mikayel Samvelyan, Christian Schroeder, Gregory Farquhar, Jakob Foerster, and Shimon Whiteson. Qmix: Monotonic value function factorisation for deep multi-agent reinforcement learning. In International conference on machine learning, pp. 4295 4304. PMLR, 2018. Mikayel Samvelyan, Tabish Rashid, Christian Schr oder de Witt, Gregory Farquhar, Nantas Nardelli, Tim G. J. Rudner, Chia-Man Hung, Philip H. S. Torr, Jakob N. Foerster, and Shimon Whiteson. The starcraft multi-agent challenge. In Edith Elkind, Manuela Veloso, Noa Agmon, and Matthew E. Taylor (eds.), Proceedings of the 18th International Conference on Autonomous Agents and Multi Agent Systems, AAMAS 19, Montreal, QC, Canada, May 13-17, 2019, pp. 2186 2188. International Foundation for Autonomous Agents and Multiagent Systems, 2019. URL http://dl.acm.org/citation.cfm?id=3332052. John Schulman, Sergey Levine, Pieter Abbeel, Michael I. Jordan, and Philipp Moritz. Trust region policy optimization. In Francis R. Bach and David M. Blei (eds.), Proceedings of the 32nd International Conference on Machine Learning, ICML 2015, Lille, France, 6-11 July 2015, volume 37 of JMLR Workshop and Conference Proceedings, pp. 1889 1897. JMLR.org, 2015. URL http://proceedings.mlr.press/v37/schulman15.html. John Schulman, Philipp Moritz, Sergey Levine, Michael I. Jordan, and Pieter Abbeel. Highdimensional continuous control using generalized advantage estimation. In Yoshua Bengio and Yann Le Cun (eds.), 4th International Conference on Learning Representations, ICLR 2016, San Juan, Puerto Rico, May 2-4, 2016, Conference Track Proceedings, 2016. URL http: //arxiv.org/abs/1506.02438. John Schulman, Filip Wolski, Prafulla Dhariwal, Alec Radford, and Oleg Klimov. Proximal policy optimization algorithms. Co RR, abs/1707.06347, 2017. URL http://arxiv.org/abs/ 1707.06347. Published as a conference paper at ICLR 2023 Hao-Jun Michael Shi, Shenyinying Tu, Yangyang Xu, and Wotao Yin. A Primer on Coordinate Descent Algorithms. ar Xiv:1610.00040 [math, stat], January 2017. Mingfei Sun, Sam Devlin, Katja Hofmann, and Shimon Whiteson. Monotonic Improvement Guarantees under Non-stationarity for Decentralized PPO. ar Xiv:2202.00082 [cs], January 2022. Peter Sunehag, Guy Lever, Audrunas Gruslys, Wojciech Marian Czarnecki, Vin ıcius Flores Zambaldi, Max Jaderberg, Marc Lanctot, Nicolas Sonnerat, Joel Z. Leibo, Karl Tuyls, and Thore Graepel. Value-decomposition networks for cooperative multi-agent learning. Co RR, abs/1706.05296, 2017. URL http://arxiv.org/abs/1706.05296. Hossein Taheri and Christos Thrampoulidis. Decentralized Learning with Separable Data: Generalization and Fast Algorithms, September 2022. J Terry, Benjamin Black, Nathaniel Grammel, Mario Jayakumar, Ananth Hari, Ryan Sullivan, Luis S Santos, Clemens Dieffendahl, Caroline Horsch, Rodrigo Perez-Vicente, et al. Pettingzoo: Gym for multi-agent reinforcement learning. Advances in Neural Information Processing Systems, 34: 15032 15043, 2021. Emanuel Todorov, Tom Erez, and Yuval Tassa. Mujoco: A physics engine for model-based control. In 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 5026 5033. IEEE, 2012. doi: 10.1109/IROS.2012.6386109. Jianhao Wang, Zhizhou Ren, Terry Liu, Yang Yu, and Chongjie Zhang. QPLEX: duplex dueling multi-agent q-learning. In 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3-7, 2021. Open Review.net, 2021. URL https://openreview. net/forum?id=Rcmk0xx IQV. Lingxiao Wang, Qi Cai, Zhuoran Yang, and Zhaoran Wang. Neural policy gradient methods: Global optimality and rates of convergence. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020. Open Review.net, 2020. URL https://openreview.net/forum?id=BJg Qfk SYDS. Yuhui Wang, Hao He, and Xiaoyang Tan. Truly proximal policy optimization. In Amir Globerson and Ricardo Silva (eds.), Proceedings of the Thirty-Fifth Conference on Uncertainty in Artificial Intelligence, UAI 2019, Tel Aviv, Israel, July 22-25, 2019, volume 115 of Proceedings of Machine Learning Research, pp. 113 122. AUAI Press, 2019. URL http://proceedings.mlr. press/v115/wang20b.html. Muning Wen, Jakub Grudzien Kuba, Runji Lin, Weinan Zhang, Ying Wen, Jun Wang, and Yaodong Yang. Multi-Agent Reinforcement Learning is a Sequence Modeling Problem, May 2022. Ying Wen, Hui Chen, Yaodong Yang, Zheng Tian, Minne Li, Xu Chen, and Jun Wang. A gametheoretic approach to multi-agent trust region optimization, 2021. URL https://arxiv. org/abs/2106.06828. Zifan Wu, Chao Yu, Deheng Ye, Junge Zhang, Haiyin Piao, and Hankz Hankui Zhuo. Coordinated proximal policy optimization. In Marc Aurelio Ranzato, Alina Beygelzimer, Yann N. Dauphin, Percy Liang, and Jennifer Wortman Vaughan (eds.), Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, Neur IPS 2021, December 6-14, 2021, virtual, volume 34, pp. 26437 26448. Curran Associates, Inc., 2021. URL https://proceedings.neurips.cc/paper/2021/hash/ de73998802680548b916f1947ffbad76-Abstract.html. Deheng Ye, Zhao Liu, Mingfei Sun, Bei Shi, Peilin Zhao, Hao Wu, Hongsheng Yu, Shaojie Yang, Xipeng Wu, Qingwei Guo, Qiaobo Chen, Yinyuting Yin, Hao Zhang, Tengfei Shi, Liang Wang, Qiang Fu, Wei Yang, and Lanxiao Huang. Mastering complex control in MOBA games with deep reinforcement learning. In AAAI 2020, pp. 6672 6679. AAAI Press, 2020. URL https: //ojs.aaai.org/index.php/AAAI/article/view/6144. Jianing Ye, Chenghao Li, Jianhao Wang, and Chongjie Zhang. Towards Global Optimality in Cooperative MARL with Sequential Transformation, August 2022. Published as a conference paper at ICLR 2023 Chao Yu, Akash Velu, Eugene Vinitsky, Yu Wang, Alexandre Bayen, and Yi Wu. The Surprising Effectiveness of PPO in Cooperative, Multi-Agent Games, July 2022. Ming Zhou, Ziyu Wan, Hanjing Wang, Muning Wen, Runzhe Wu, Ying Wen, Yaodong Yang, Weinan Zhang, and Jun Wang. Malib: A parallel framework for population-based multi-agent reinforcement learning. Co RR, abs/2106.07551, 2021. URL https://arxiv.org/abs/ 2106.07551. Published as a conference paper at ICLR 2023 Supplementary Material Table of Contents A Proofs 16 A.1 Notations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16 A.2 Useful Lemmas . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16 A.3 Proofs of Intuitive Sequential Update . . . . . . . . . . . . . . . . . . . . . . . 19 A.4 Proofs of Monotonic Policy Improvement of A2PO . . . . . . . . . . . . . . . . 19 A.5 Proofs of Incrementally Tightened Bound of A2PO . . . . . . . . . . . . . . . . 20 A.6 Proofs of Monotonic Policy Improvement of MAPPO, Co PPO and HAPPO . . . 20 A.7 Comparisons on Monotonic Improvement Bounds . . . . . . . . . . . . . . . . 22 A.8 Preceding-agent Off-policy Correction . . . . . . . . . . . . . . . . . . . . . . 23 A.9 Why Off-policyness is More Serious in Sequential Update Scheme? . . . . . . . 23 B Experimental Details 24 B.1 Implementation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24 B.2 Experimental Setup and Additional Results . . . . . . . . . . . . . . . . . . . . 24 B.3 Wall Time Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32 B.4 Hyper-parameters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32 C The Related Work of Other MARL Methods 34 D The Related Work of Coordinate Descent 35 Published as a conference paper at ICLR 2023 A.1 NOTATIONS We list the main notations used in Tab. 4. Table 4: The notations and symbols used in this paper. Notation Definition S The state space N The set of agents n The number of agents i The agent index Ai The action space of agent i r The reward function T The transition function γ The discount factor t The time-step st The state at time-step t ai t The action of agent i at time-step t at The joint action at time-step t dπ The discounted state visitation distribution Pr The state probability function V The value function A The advantage function τ The trajectory of an episode e A set of preceding agents ei The set of preceding agents updated before agent i πi The policy of agent i πi The updated policy of agent i π The joint policy λ The bias and variance balance parameter π The joint target policy ˆπi The joint policy after updating agent i J (π) The expected return / performance of the joint policy π Lˆπi 1(ˆπi) The surrogate objective of agent i LI ˆπi 1(ˆπi) An intuitive surrogate objective of agent i Gπ( π) The surrogate objective of all agents ϵ The upper bound of an advantage function DT V The total variation distance function α The total variation distance between 2 policies ξi The off policy correction error of ˆπi 1 C The clipping parameter adaptation function R The agent selection function A.2 USEFUL LEMMAS Lemma 1 (Multi-agent Policy Performance Difference Lemma). Given any joint policies π and π, the difference between the performance of two joint policies can be expressed as: J ( π) J (π) = 1 1 γ E(s,a) (d π, π) [Aπ(s, a)] , where dπ = (1 γ) P t=0 γt Pr(st = s|π) is the normalized discounted state visitation distribution. Proof. A corollary of the Policy Performance Difference Lemma, see Lemma 1.16 in Alekh et al. (2022). For convenience, we give some properties and definitions of coupling5 and the definition of αcoupled policy pair (Schulman et al., 2015) here. 5The definition of coupling and the properties can be found in any textbook containing Markov Chains. Published as a conference paper at ICLR 2023 Definition 1 (Coupling) A coupling of two probability distributions µ and ν is a pair of random variables (X, Y ) such that the marginal distribution of X is µ and the marginal distribution of Y is ν. A coupling (X, Y ) satisfies the following constraints: Pr(X = x) = µ(x) and Pr(Y = y) = ν(y). Proposition 3 For any coupling (X, Y ) that DT V (µ ν) Pr(X = Y ). Proposition 4 There exists a coupling (X, Y ) that DT V (µ ν) = Pr(X = Y ). Corollary 1 For all s, there exists a coupling (π( |s), π( |s)), that Pr(a = a) 1 Dmax T V (π π), for a π( |s), a π( |s). Proof. By prop. 4 there exists a coupling (π( |s), π( |s)), s.t. 1 Pr(a = a) = Pr(a = a) = DT V (π, π) Dmax T V (π π) Corollary 2 For all s, DT V (π( |s) π( |s)) Pn i=1 DT V (πi( |s) πi( |s)). Proof. We denote π( |s) as π( ) for brevity. DT V (π( |s) π( |s)) a1,a2,...,an a1,a2,...,an i=1 πi(ai) π1(a1) i=2 πi(ai) + π1(a1) π1(a1) π1(a1) X π1(a1) π1(a1) ai |πi(ai) πi(ai)| i=1 DT V (πi( |s) πi( |s)) Definition 2 (α-coupled policy pair) If (π, π) is an α-coupled policy pair, then (a, a|s) satisfies Pr(a = a|s) α for all s, and a π( |s), a π( |s). From Corollaries 1 and 2, we know that given any joint policy pair π and π, select α = Dmax T V (π( |s) π( |s)), then (π, π) is an α-coupled policy pair that for all s, Pr(a = a|s) Dmax T V (π( |s) π( |s)) Pn i=1 αi, where αi = Dmax T V (πi πi). Lemma 2 Given any joint policies π and π, if ( π,π) is a coupled policy pair, the following inequality holds: |Ea π [Aπ(s, a)] | 2ϵ where αi = Dmax T V ( πi πi) and ϵ = maxs,a |Aπ(s, a)|. Published as a conference paper at ICLR 2023 Proof. Note that Ea π[Aπ(s, a)] = 0. We have |Ea π [Aπ(s, a)]| = |E a π [Aπ(s, a)] Ea π [Aπ(s, a)]| = E( a,a) ( π,π) [Aπ(s, a) Aπ(s, a)] = Pr( a = a|s)E( a,a) ( π,π) [Aπ(s, a) Aπ(s, a)] i=1 αi E( a,a) ( π,π) [|Aπ(s, a) Aπ(s, a)|] i=1 αi 2 max s,a |Aπ(s, a)| Lemma 3 (Multi-agent Advantage Discrepancy Lemma). Given any joint policies π1, π2 and π3, if (π1, π2) and (π2, π3) are coupled policy pairs, the following inequality holds: E(st,at) (P rπ2,π2) h Aπ1i E(st, at) (P rπ3,π2) h Aπ1i 4ϵπ1 Dmax T V (π1 π2) (1 (1 Dmax T V (π2 π3)) t) , where ϵπ1 = maxs,a Aπ1(s, a) and we denote A(s, a) as A for brevity. Proof. Let nt represent the times a = a (π1 disagrees with π3) before timestamp t. E(st,at) (P rπ2,π2) h Aπ1i E(st, at) (P rπ3,π2) h Aπ1i =Pr(nt > 0) E(st,at) (P rπ2,π2)|nt>0 h Aπ1i E(st, at) (P rπ3,π2)|nt>0 h Aπ1i (a) =(1 Pr(nt = 0)) E k=1 Pr(ak = ak|ak π2( |sk), ak π3( |sk))) E k=1 (1 Dmax T V (π2 π3))) E =(1 (1 Dmax T V (π2 π3)) t) E (1 (1 Dmax T V (π2 π3)) t) 2 2 Dmax T V (π1 π2) ϵπ1 =4ϵπ1 Dmax T V (π1 π2) (1 (1 Dmax T V (π2 π3)) t) In (a), we denote |E(st,at) (P rπ2,π2)|nt>0[Aπ1] E(st, at) (P rπ3,π2)|nt>0[Aπ1]| as E for brevity. (b) follows the definition of α-coupled policy pair. We provide a useful equation of the normalized discounted state visitation distribution here. Proposition 5 E(s,a) (dπ1,π2) [f(s, a)] = (1 γ) X t=0 γt Pr(st = s|π1) X a π2(a|s)f(s, a) s Pr(st = s|π1) X a π2(a|s)f(s, a) t=0 γt E(st,at) (P rπ1,π2)[f(st, at)] Published as a conference paper at ICLR 2023 A.3 PROOFS OF INTUITIVE SEQUENTIAL UPDATE J (ˆπi) J (ˆπi 1) 1 1 γ E(s,a) (dπ,ˆπi) [Aπ] E(s,a) (dˆ πi,ˆπi) h Aˆπi 1i E(s,a) (dπ,ˆπi) [Aπ] E(s,a) (dˆ πi,ˆπi) h Aˆπi 1i E(s,a) (dπ,ˆπi) h Aˆπi 1i E(s,a) (dπ,ˆπi) h Aˆπi 1i E(s,a) (dπ,ˆπi) [Aπ] 4ϵˆπi 1αi X t=0 γt(1 (1 X j (ei {i}) αj)t) + 1 1 γ E(s,a) (dπ,ˆπi) h Aˆπi 1 Aπ i 4ϵˆπi 1αi( 1 1 γ 1 1 γ(1 P j (ei {i}) αj)) + 1 1 γ 4αiϵˆπi 1 + 2 X A.4 PROOFS OF MONOTONIC POLICY IMPROVEMENT OF A2PO Theorem 1 (Single Agent Monotonic Bound) For agent i, let ϵi = maxs,a |Aˆπi 1(s, a)|, ξi = maxs,a |Aπ,ˆπi 1(s, a) Aˆπi 1(s, a)|, αj = Dmax T V (πj πj) j (ei {i}), then we have: J (ˆπi) Lˆπi 1(ˆπi) 4ϵiαi 1 1 γ 1 1 γ(1 P j (ei {i}) αj) + ξi (1 γ)2 αi X j (ei {i}) αj + ξi Proof. Using Lemma 3 and Prop. 5, we get J (ˆπi) J (ˆπi 1) 1 1 γ E(s,a) (dπ,ˆπi) h Aπ,ˆπi 1i E(s,a) (dˆ πi,ˆπi) h Aˆπi 1i E(s,a) (dπ,ˆπi) h Aπ,ˆπi 1i E(s,a) (dˆ πi,ˆπi) h Aˆπi 1i E(s,a) (dπ,ˆπi) h Aˆπi 1i E(s,a) (dπ,ˆπi) h Aˆπi 1i E(s,a) (dπ,ˆπi) h Aπ,ˆπi 1i 4ϵˆπi 1αi X t=0 γt(1 (1 X j (ei {i}) αj)t) + 1 1 γ E(s,a) (dπ,ˆπi) h Aˆπi 1 Aπ,ˆπi 1 i 4ϵˆπi 1αi( 1 1 γ 1 1 γ(1 P j (ei {i}) αj)) + 1 1 γ ξi Published as a conference paper at ICLR 2023 Theorem 2 (Joint Monotonic Bound) For each agent i N, let ϵi = maxs,a |Aˆπi 1(s, a)| , αi = Dmax T V (πi πi), ξi = maxs,a |Aπ,ˆπi 1(s, a) Aˆπi 1(s, a)|, and ϵ = maxi ϵi, then we have: |J ( π) Gπ( π)| 4ϵ i=1 αi 1 1 γ 1 1 γ(1 P j (ei {i}) αj) + Pn i=1 ξi j (ei {i}) αj + Pn i=1 ξi |J ( π) Gπ( π)| J ( π) J (π) i=1 E(s,a) (dπ,ˆπi) h Aπ,ˆπi 1(s, a) i J (ˆπn) J (ˆπn 1) + + J (ˆπ1) J (ˆπ0) 1 1 γ i=1 E(s,a) (dπ,ˆπi) h Aπ,ˆπi 1(s, a) i J (ˆπi) J (ˆπi 1) 1 1 γ E(s,a) (dπ,ˆπi) h Aπ,ˆπi 1(s, a) i i=1 αi 1 1 γ 1 1 γ(1 P j (ei {i}) αj) + Pn i=1 ξi j (ei {i}) αj + Pn i=1 ξi A.5 PROOFS OF INCREMENTALLY TIGHTENED BOUND OF A2PO Assume agent k is updated with order k in the sequence 1, . . . , n, since ˆπk 1 is known, we have |J ( π) Gπ( π)| J (ˆπi) Lˆπi 1(ˆπi) + 4ϵ i=k αi 1 1 γ 1 1 γ(1 P j (ei {i}) αj) + Pn i=k ξi J (ˆπi) Lˆπi 1(ˆπi) + 4ϵ i=k 1 αi 1 1 γ 1 1 γ(1 P j (ei {i}) αj) Pn i=k 1 ξi i=1 αi 1 1 γ 1 1 γ(1 P j (ei {i}) αj) + Pn i=1 ξi Thus the condition for improving J ( π) is relaxed during updating agents at a stage. A.6 PROOFS OF MONOTONIC POLICY IMPROVEMENT OF MAPPO, COPPO AND HAPPO In this section, we give proof of the monotonic policy improvement of MAPPO, and unify the formats of the monotonic bounds of Co PPO and HAPPO, without considering the parameter-sharing method. MAPPO. For MAPPO, Lπ( π) = Pn i=1 J (π) + 1 1 γ h E(s,a) (dπ,π) h πi πi Aπii . We first prove that for agent i, J ( π) J (π) 1 1 γ h E(s,a) (dπ,π) h πi πi Aπii is bounded. Published as a conference paper at ICLR 2023 J ( π) J (π) 1 1 γ E(s,a) (dπ,π) E(s,a) (d π, π) [Aπ] E(s,a) (dπ,π) t=0 γt E(st,at) (P r π, π)Aπ E(st,at) (P rπ,π) Sum the bounds for all agents and take the average, we get J ( π) J (π) 1 E(s,a) (dπ,π) Finally, the monotonic bound for MAPPO is J ( π) J (π) 1 1 γ E(s,a) (dπ,π) J ( π) J (π) 1 E(s,a) (dπ,π) E(s,a) (dπ,π) j=1 αj + n 1 1 1 γ αi 2ϵπ Co PPO. We prove the results of Co PPO in a unified and convenient form. For Co PPO, Lπ( π) = J (π) + 1 1 γ E(s,a) (dπ, π)[Aπ(s, a)], we prove the bound using Lemma 3. Published as a conference paper at ICLR 2023 J ( π) J (π) 1 1 γ E(s,a) (dπ, π)[Aπ] E(s,a) (d π, π)[Aπ] E(s,a) (dπ, π)[Aπ] t=0 γt E(s,a) (P r π, π) [Aπ] E(s,a) (P rπ, π) [Aπ] i=1 αi 1 (1 Dmax T V (π π))t i=1 αi 1 1 γ 1 1 γ(1 Pn j=1 αj) HAPPO. Following the proof of Lemma 2 in Kuba et al. (2022), we know that HAPPO has the same monotonic improvement bound as that of Co PPO. For the monotonic improvement of a single agent, we formulate the surrogate objective of agent i using HAPPO as J (ˆπi 1) + 1 1 γ E(s,a) (dπ,ˆπi)[Aπ(s, a)] 1 1 γ E(s,a) (dπ,ˆπi 1)[Aπ(s, a)], as shown in Proposition 3 of Kuba et al. (2022). Following the proof of Thm. 1, we get the following inequality. J (ˆπi) J (ˆπi 1) 1 1 γ E(s,a) (dπ,ˆπi) [Aπ] + 1 1 γ E(s,a) (dπ,ˆπi 1)[Aπ(s, a)] E(s,a) (dˆ πi,ˆπi) h Aˆπi 1i E(s,a) (dπ,ˆπi) [Aπ] + 1 1 γ E(s,a) (dπ,ˆπi 1)[Aπ(s, a)] E(s,a) (dˆ πi,ˆπi) h Aˆπi 1i 1 1 γ E(s,a) (dπ,ˆπi) h Aˆπi 1i E(s,a) (dπ,ˆπi) h Aˆπi 1i E(s,a) (dπ,ˆπi) [Aπ] + 2 1 1 γ 4ϵˆπi 1αi X t=0 γt(1 (1 X j (ei {i}) αj)t) + 1 1 γ E(s,a) (dπ,ˆπi) h Aˆπi 1 Aπ i + 2 1 1 γ 4ϵˆπi 1αi( 1 1 γ 1 1 γ(1 P j (ei {i}) αj)) + 1 1 γ 4αiϵˆπi 1 + 4 X The right side of the last inequality is not a monotonic improvement bound, or it does not provide a guarantee for improving the expected performance J (ˆπi) since the term P j ei αjϵπ is not controllable for agent i, whether through policy improvement or value learning. The uncontrollable term means the expected performance may not be improved even if the total variation distances of consecutive policies are well constrained. A.7 COMPARISONS ON MONOTONIC IMPROVEMENT BOUNDS Co PPO and HAPPO have the same monotonic bound that is tighter than that of MAPPO. A2PO achieves the tightest monotonic bound given mild assumptions about the errors of preceding-agent off-policy correction, which is valid and easy to achieve since preceding-agent off-policy correction is a contraction operator. A sufficient condition that A2PO has the tightest bound is that ξi < γ(1 γ) P j N ei {i} αj j ei {i} αj))(1 γ(1 Pn j=1 αj)), for all i N. Published as a conference paper at ICLR 2023 A.8 PRECEDING-AGENT OFF-POLICY CORRECTION In Retrace(λ) (Munos et al., 2016), consider the current policy as ˆπi=1 and base policy as π, we have the following definition: Rt = rt + γQt+1 + X j=1 λ min 1.0, ˆπi 1(at+j|st+j) π(at+j|st+j) (rt+k + γQt+k+1 Qt+k) , Following that same structure, we have: Rt = rt + γVt+1 + X j=1 λ min 1.0, ˆπi 1(at+j|st+j) π(at+j|st+j) (rt+k + γVt+k+1 Vt+k) , By subtracting Vt, we get the definition of Pre OPC. Or one can get γAπ,ˆπi 1 by substituting rt + γVt+1 for Qt and subtracting rt + γVt+1. A.9 WHY OFF-POLICYNESS IS MORE SERIOUS IN SEQUENTIAL UPDATE SCHEME? As shown in Fig. 13, the off policy correction in sequential update algorithms improves the performance significantly while similar performance gaps are not observed when used in simultaneous update algorithms. We attribute the difference to the influence of the clipping mechanism on the total variation distance. From Corollary 2, DT V (π π) < Pn i=1 DT V (πi πi). Although we can not prove exact relations, clipping the agents independently tends to larger total variation distances between the current and future policies of the agents, leading to more off-policyness in sequential update algorithms. Published as a conference paper at ICLR 2023 B EXPERIMENTAL DETAILS B.1 IMPLEMENTATION For a fair comparison, we (re)implement A2PO and the baselines based on the implementation of MAPPO. We keep the same structures for all the algorithms and tune all the algorithms following the same process, i.e., a grid search over a small collection of hyper-parameters, to avoid the influence of different implementation details on the results. The grid search is performed on three hyperparameters: the learning rate, λ and the agent block num in the tasks with numerous agents. The algorithms, including A2PO and baselines, are implemented into both parameter sharing and parameter independent versions. A2PO in the parameter sharing version is implemented as in Alg. 2. The main modifications are colored in blue. We rearrange the loops of agents and ppo epochs. The number of ppo epochs is divided by n for comparable updating times with the simultaneous algorithms. The approximated advantage is estimated by correcting the action probabilities of all the agents given such ei. Algorithm 2: Agent-by-agent Policy Optimization (Parameter Sharing) 1 Initialize the shared joint policy π0 = {π1 0, . . . , πn 0 } with π1 0 = = πn 0 , and the global value function V . 2 for iteration m = 1, 2, . . . do 3 Collect data using πm 1. 4 Policy πm = πm 1. n epochs do 6 for k = 1, . . . , n do 7 Agent i = R(k), preceding agents ei = {R(1), . . . , R(n 1)}. 8 Joint policy ˆπi = πm. 9 Compute the advantage approximation as Aπ,ˆπi 1(s, a) via eq. (2). 10 Compute the value target v(st) = Aπ,ˆπi 1(s, a) + V (s). 11 πi m = arg maxπim Lˆπi 1(ˆπi) as in eq. (6). 12 V = arg min V Es dπ v(s) V (s) 2. Practically, each agent is equipped with a value function, we generate the agent order at once to avoid estimating the advantage function n(n 1) 2 times. The order becomes [1, . . . , i, . . . , j, . . . , n] in which E|Ai| >= E|Aj|. B.2 EXPERIMENTAL SETUP AND ADDITIONAL RESULTS B.2.1 STARCRAFTII MULTI-AGENT CHALLENGE Star Craft II Multi-agent Challenge (SMAC) (Samvelyan et al., 2019) provides a wide range of multiagent tasks in the battle scenarios of Star Craft II. Algorithms adopting parameter sharing have shown superior performance in SMAC, so all the algorithms are implemented as parameter sharing. As shown in Tab. 5, we evaluate the algorithms in 12 maps of SMAC with various difficulties, in which the baselines can not achieve 100% win rates easily. We use the results of Qmix in Yu et al. (2022). The learning curves for episode return are summarized in Fig. 8. B.2.2 MULTI-AGENT MUJOCO Multi-agent Mu Jo Co (MA Mu Jo Co) (Peng et al., 2021) contains a range of multi-agent robot continuous control tasks, in which an agent controls the composition of robot joints. MA Mu Jo Co extends the high-dimensional single-agent locomotion tasks in Mu Jo Co (Todorov et al., 2012), a widely adopted benchmark for SARL algorithms (Haarnoja et al., 2018; He & Hou, 2020), into the multi-agent case. Agents must cooperate in their actions for robot locomotion, and different agents control different compositions of the robot joints. We use the reward settings of the original paper Published as a conference paper at ICLR 2023 0 2 4 Steps 106 0.00 0.25 0.50 0.75 1.00 Steps 107 0.00 0.25 0.50 0.75 1.00 Steps 107 0 2 4 Steps 106 0.00 0.25 0.50 0.75 1.00 Steps 107 0.00 0.25 0.50 0.75 1.00 Steps 107 0.00 0.25 0.50 0.75 1.00 Steps 107 0.00 0.25 0.50 0.75 1.00 Steps 107 0.00 0.25 0.50 0.75 1.00 Steps 107 3s5z vs 3s6z 0.00 0.25 0.50 0.75 1.00 Steps 107 0.0 0.5 1.0 1.5 Steps 107 0.00 0.25 0.50 0.75 1.00 Steps 107 MAPPO w/ PS Co PPO w/ PS HAPPO w/ PS A2PO w/ PS Figure 8: Comparisons of median win rate on SMAC. Published as a conference paper at ICLR 2023 Table 5: Median win rates and standard deviations on SMAC tasks. w/ PS means the algorithm is implemented as parameter sharing Map Difficulty MAPPO w/ PS Co PPO w/ PS HAPPO w/ PS A2PO w/ PS Qmix w/ PS MMM Easy 96.9(0.988) 96.9(1.25) 95.3(2.48) 100(1.07) 95.3(2.5) 3s vs 5z Hard 100(1.17) 100(2.08) 100(0.659) 100(0.534) 98.4(2.4) 2c vs 64zg Hard 98.4(1.74) 96.9(0.521) 96.9(0.521) 96.9(0.659) 92.2(4.0) 3s5z Hard 84.4(4.39) 92.2(2.35) 92.2(1.74) 98.4(1.04) 88.3(2.9) 5m vs 6m Hard 84.4(2.77) 84.4(2.12) 87.5(2.51) 90.6(3.06) 75.8(3.7) 8m vs 9m Hard 84.4(2.39) 84.4(2.04) 96.9(3.78) 100(1.04) 92.2(2.0) 10m vs 11m Hard 93.8(18.7) 96.9(2.6) 98.4(2.99) 100(0.521) 95.3(1.0) 6h vs 8z Super Hard 87.5(1.53) 90.6(0.765) 87.5(1.49) 90.6(1.32) 9.4(2.0) 3s5z vs 3s6z Super Hard 82.8(19.2) 84.4(2.9) 37.5(13.2) 93.8(19.8) 82.8(5.3) MMM2 Super Hard 90.6(8.89) 90.6(6.93) 51.6(9.01) 98.4(1.25) 87.5(2.6) 27m vs 30m Super Hard 93.8(3.75) 93.8(2.2) 90.6(4.77) 100(1.55) 39.1(9.8) corridor Super Hard 96.9(0) 100(0.659) 96.9(0.96) 100(0) 84.4(2.5) overall / 91.1(5.46) 92.6(2.2) 85.9(3.68) 97.4(2.65) 78.4(3.6) but set the environment to be fully observable6. The agents are heterogeneous and mostly asymmetric in MA-Mu Jo Co, so we implement the algorithms as parameter-independent. We test 14 tasks of 6 scenarios in MA Mu Jo Co, as illustrated in Fig. 9. B.2.3 MULTI-AGENT PARTICLE ENVIRONMENT We consider the Navigation task of the Multi-agent Particle Environment (MPE) (Lowe et al., 2017) implemented in Petting Zoo (Terry et al., 2021) which implements MPE with minor fixes and provides convenience for customizing the number of agents and landmarks, and customizing the global and local rewards., with 3 and 5 agents and corresponding numbers of landmarks. The agents are rewarded based on the minimum distance to the landmarks and penalized for colliding with each other, meaning that the reward is entirely up to the coordination behavior. We adopted two different reward settings: Fully Cooperative and General-sum. In the Fully Cooperative setting, the agents share the same reward, while in the General-sum setting, the agents are additionally rewarded based on the local collision detection. The results in Fig. 10 show that A2PO generally outperforms the baselines on the average return and the sample efficiency. Noted that A2PO is developed in fully cooperative games, the results in the General-sum setting reveal the potential of extending A2PO into general-sum games. Further, the performance gap between A2PO and the baselines enlarges with the increasing number of agents. B.2.4 GOOGLE RESEARCH FOOTBALL 0 5000 10000 15000 Episodes Football 5 vs 5 (Parameter Sharing) MAPPO Co PPO Figure 11: 5-vs-5 scenario with Parameter sharing. In the above experiments, we have evaluated A2PO in tasks where agents can learn both their micro-operations and coordination behaviors (SMAC and MA-Mu Jo Co) and tasks where agents can only learn coordination behaviors (the Navigation task). However, the coordination behaviors in the above tasks are relatively easy to discover, e.g., agents learn to concentrate their fire to shoot the enemies and cover each other in SMAC. Recent works (Wen et al., 2022; Yu et al., 2022) have conducted experiments on Google Research Football academic scenarios with a small number of players and easily accessible targets, making the coordination behavior also easy to discover. In contrast, we evaluate A2PO in the full-game scenarios, where the players of the left team, except for the goalkeeper, are controlled to play a football match against the right team controlled by the built-in AI provided by GRF. The agents in the full-game scenarios have high-dimensional observations, complex action spaces, and a long-span timescale (3000 steps). We reconstruct the observation space and 6Empirically, we find the fully observable setting does not make the tasks easier because of the information redundancy. Published as a conference paper at ICLR 2023 0.00 0.25 0.50 0.75 1.00 Steps 107 Episode Return 0.00 0.25 0.50 0.75 1.00 Steps 107 Episode Return Ant-v2 2x4d 0.00 0.25 0.50 0.75 1.00 Steps 107 Episode Return 0.00 0.25 0.50 0.75 1.00 Steps 107 Episode Return 0.00 0.25 0.50 0.75 1.00 Steps 107 0 Episode Return Walker2d-v2 2x3 0.00 0.25 0.50 0.75 1.00 Steps 107 0 Episode Return Walker2d-v2 3x2 0.00 0.25 0.50 0.75 1.00 Steps 107 0 Episode Return Walker2d-v2 6x1 0.00 0.25 0.50 0.75 1.00 Steps 107 Episode Return Hopper-v2 3x1 0.00 0.25 0.50 0.75 1.00 Steps 107 Episode Return Half Cheetah-v2 2x3 0.00 0.25 0.50 0.75 1.00 Steps 107 Episode Return Half Cheetah-v2 3x2 0.00 0.25 0.50 0.75 1.00 Steps 107 Episode Return Half Cheetah-v2 6x1 0.00 0.25 0.50 0.75 1.00 Steps 107 Episode Return Humanoid-v2 9|8 0.00 0.25 0.50 0.75 1.00 Steps 107 Episode Return Humanoid-v2 17x1 0.00 0.25 0.50 0.75 1.00 Steps 107 Episode Return Humanoid Standup-v2 9|8 MAPPO w/o PS Co PPO w/o PS HAPPO w/o PS A2PO w/o PS Figure 9: Comparisons of average episode return on MA-Mu Jo Co. Published as a conference paper at ICLR 2023 0 1 2 3 Steps 106 Episode Return 3 Agents (Fully Cooperative) 0 2 4 Steps 106 Episode Return 5 Agents (Fully Cooperative) 0.00 0.25 0.50 0.75 1.00 Steps 107 Episode Return 3 Agents (General Sum) 0.00 0.25 0.50 0.75 1.00 Steps 107 45.0 Episode Return 5 Agents (General Sum) MAPPO Co PPO HAPPO A2PO Figure 10: Comparisons of averaged return on the Multi-agent Particle Environment Navigation task. Left: The fully cooperative setting. Right: The general-sum setting. design a dense reward to facilitate training in these scenarios based on Football-paris. The observation is formed to be agent-specific. The reward function estimates the behaviors of the entire team, including scoring, and carrying the ball to the opponent s restricted area et al., but not the individual behaviors such as ball-passing (Li et al., 2021). We implement all the algorithms for the 5-vs-5 scenario as both parameter sharing and parameter-independent. The additional results with algorithms implemented as parameter sharing are shown in Fig. 11, in which A2PO gets free from the trouble that the controlled agents have similar behavior and compete for the ball (Li et al., 2021). We implement all the algorithms on the 11-vs-11 scenario as parameter sharing using MALib (Zhou et al., 2021) for acceleration and train the algorithms for 300M environment steps. We summarize the learned behaviors observed in the game videos: Basic Skills. The agents trained by MAPPO and Co PPO perform unsatisfactorily in basic skills such as dribbling, shooting, and the agents even run out of bounds frequently. In contrast, the agents trained by HAPPO and A2PO perform better in the basic skills. We attribute the problems to the non-stationarity issue that seriously influences the simultaneous updating algorithms. We also note that the agents trained by all the algorithms fail to understand the off-side mechanism and occasionally gather together on the opponent s bottom line. Passing and Receiving Coordination. We analyze the direct way for coordination: passing and receiving the ball. As illustrated in Tab. 3, the agents trained by MAPPO have the lowest number of successful passes and the lowest successful pass rate, and we can hardly observe the agents passing the ball. Agents trained by Co PPO perform better on passing the ball but suffer from poor basic skills, and get tackled after receiving the ball. Agents trained by HAPPO prefer passing the ball without considering the teammates situations, e.g., the receiver is marked by several opponents. Agents trained by A2PO can pass the ball to their teammates in a way that leads to a score. We attribute the performance gain to the preceding-agent off-policy correction, which means that agents estimate the teammates situations and intentions better. We further visualize the learned behaviors of A2PO in Fig. 12. In the top of Fig. 12, two players cooperatively break through the opponent s defense and complete a passing and receiving coordination for scoring. In the bottom of Fig. 12, three players make a fast thrust by two long passes: the goalkeeper passes the ball to the player at the edge, and the player at the edge passes the ball to the player behind the opponents. The complex coordination strategies are hardly observed in other baselines. B.2.5 ABLATION Preceding agent off-policy correction. More ablations on preceding-agent off-policy correction are shown in Fig. 13. The baselines are: MAPPO w/ V-trace, Co PPO w/ V-trace: Simultaneous update methods with advantage estimation as V-trace. HAPPO w/ Pre OPC: HAPPO with advantage estimation as Pre OPC. Published as a conference paper at ICLR 2023 The player Turing beat an opponent with the player Johnson marking. Turing beat another opponent. Johnson prepares to take the pass from Turing. Turing passes the ball to Johnson, then Johnson receives the pass and thrusts to shoot. Johnson breaks through the opponent's goalkeeper, shoots, and makes a goal. The goalkeeper makes a goal kick and plays a long pass to the player near the sideline. The player Turing receives the pass from the goalkeeper, then dribbles and passes the ball to the player in the midfield. The player Curie receives the pass from Turing, then plays a fast break. Curie shoots and makes a goal. location of the player Motion of the player Motion of the ball History location of the player Figure 12: Visualization of trained A2PO policies on the Google Research Football 11-vs-11 scenario, which shows that A2PO encourages complex cooperation behaviors to make a goal. Top: Player Turing and Johnson cooperate to beat multiple opponents to break through the defense and make a goal. Bottom: The goalkeeper, player Turing, and Curie achieve the pass and receive cooperation twice. A fast thrust is made by consecutively passing the ball. In this ablation study, the baselines are equipped with off-policy correction methods. The experiment yields the following three conclusions: The results firstly support the conclusion in Sec. 3.3 that applying Pre OPC to sequential update methods results in a greater performance improvement than applying V-trace to simultaneous update methods. Secondly, the primary distinction between A2PO and HAPPO with Pre OPC is the clipping objective. The results demonstrate that the clipping objective derived from the single-agent improvement bound contributes to the performance improvement. And thirdly, although we were unable to assess the error of Pre OPC, we compare A2PO with RPISA-PPO, which can be viewed as A2PO algorithms with error-free off-policy correction methods (the advantage estimation is error-free) at the expense of sample inefficiency. A2PO reaches or outperforms the asymptotic performance of RPISA-PPO. A2PO outperforms RPISA-PPO since RPISA-PPO suffers from performance degradation as a result of agents updating policies with separated data (Taheri & Thrampoulidis, 2022). We further analyze the sensitivity to the hyper-parameter λ. Results in Fig. 14 illustrate that preceding-agent off-policy correction does not introduce more sensitivity. 0.00 0.25 0.50 0.75 1.00 Steps 107 Episode Return RPISA-PPO Asymptotic Performance Humanoid-v2 9|8 0.00 0.25 0.50 0.75 1.00 Steps 107 Episode Return RPISA-PPO Asymptotic Performance 0.00 0.25 0.50 0.75 1.00 Steps 107 RPISA-PPO Asymptotic Performance 0.00 0.25 0.50 0.75 1.00 Steps 107 RPISA-PPO Asymptotic Performance MAPPO MAPPO w/ V-trace Co PPO Co PPO w/ V-trace HAPPO HAPPO w/ Pre OPC A2PO w/o Pre OPC A2PO Figure 13: Ablation experiments on preceding-agent off-policy correction. Agent Selection Rule. More ablations on the agent selection rules are shown in Fig. 13. We compare two additional rules: Reverse-greedy and Reverse-semi-greedy . Reverse means selecting the agent with the minimal advantage first. While we observe that the effect of the selection rule becomes less significant in tasks with homogeneous or symmetric agents. Going deeper into the effects of agent selection rules, we show that the agents with implicit guidance from the advantage estimation benefit from greedily selecting agents in Fig. 16 and 17. More Published as a conference paper at ICLR 2023 0.0 0.2 0.4 0.6 0.8 1.0 Steps 107 Episode Return 0.0 0.2 0.4 0.6 0.8 1.0 Steps 107 Episode Return MAPPO w/ MA-A-trace 0.0 0.2 0.4 0.6 0.8 1.0 Steps 107 Episode Return A2PO w/o MA-A-trace 0.0 0.2 0.4 0.6 0.8 1.0 Steps 107 Episode Return A2PO Humanoid-v2 9 8 λ=0.90 λ=0.93 λ=0.95 λ=0.97 λ=0.99 0.0 0.2 0.4 0.6 0.8 1.0 Steps 107 Episode Return 0.0 0.2 0.4 0.6 0.8 1.0 Steps 107 Episode Return MAPPO w/ MA-A-trace 0.0 0.2 0.4 0.6 0.8 1.0 Steps 107 Episode Return A2PO w/o MA-A-trace 0.0 0.2 0.4 0.6 0.8 1.0 Steps 107 Episode Return A2PO Ant-v2 4x2 λ=0.90 λ=0.93 λ=0.95 λ=0.97 λ=0.99 0.0 0.2 0.4 0.6 0.8 1.0 Steps 107 0.0 0.2 0.4 0.6 0.8 1.0 Steps 107 MAPPO w/ MA-A-trace 0.0 0.2 0.4 0.6 0.8 1.0 Steps 107 A2PO w/o MA-A-trace 0.0 0.2 0.4 0.6 0.8 1.0 Steps 107 λ=0.90 λ=0.93 λ=0.95 λ=0.97 λ=0.99 0.0 0.2 0.4 0.6 0.8 1.0 Steps 107 0.0 0.2 0.4 0.6 0.8 1.0 Steps 107 MAPPO w/ MA-A-trace 0.0 0.2 0.4 0.6 0.8 1.0 Steps 107 A2PO w/o MA-A-trace 0.0 0.2 0.4 0.6 0.8 1.0 Steps 107 λ=0.90 λ=0.93 λ=0.95 λ=0.97 λ=0.99 Figure 14: Sensitivity analysis of λ. Published as a conference paper at ICLR 2023 0.00 0.25 0.50 0.75 1.00 Steps 107 Episode Return Humanoid-v2 9|8 0.00 0.25 0.50 0.75 1.00 Steps 107 0.00 0.25 0.50 0.75 1.00 Steps 107 Episode Return 0 2 4 Steps 106 0.0 HAPPO-Cyclic HAPPO-Random HAPPO-Greedy HAPPO-Semi-greedy HAPPO-Reverse-greedy HAPPO-Reverse-semi-greedy A2PO-Cyclic A2PO-Random A2PO-Greedy A2PO-Semi-greedy A2PO-Reverse-greedy A2PO-Reverse-semi-greedy Figure 15: Ablation experiments on the agent selection rules. Left: Heterogeneous or asymmetric agents. Right: Homogeneous or symmetric agents. even bars appear in one fig means the agents are more balanced in terms of the guidance from the advantage estimation. Take the agent 10 in Fig. 16 for example, under Cyclic and Random rules, agent 10 perform the worst with high proportions, while it has higher proportions in prior ranks under Greedy rule. 1 2 3 4 5 6 7 8 9 10 Agent 1 2 3 4 5 6 7 8 9 10 Agent 1 2 3 4 5 6 7 8 9 10 Agent Order 1 Order 2 Order 3 Order 4 Order 5 Order 6 Order 7 Order 8 Order 9 Order 10 (a) The imbalance of the agents. The bar of agent i illustrates the proportion of its ranks in terms of Es,ai[|Aπ,ˆπi|]. Especially, agent 10 has implicit guidance, i.e., a small absolute value of advantage function when using Cyclic and Random selection rule, but is comparable with other agents with Greedy selection rule. Cyclic Random Greedy Selection Strategy Coefficient of Variation (b) The coefficient of variance of agent 10 s order proportions. Figure 16: Agents imbalance in terms of the estimated advantage. The experiment is conducted on the MMM2 task of SMAC. Adaptive Clipping Parameter. More ablations on the adaptive clipping parameter are shown in Fig. 18. Similarly, we observe that the effect of the adaptive clipping parameter becomes less significant in tasks with homogeneous or symmetric agents. 1 2 3 4 5 6 7 8 Agent 1 2 3 4 5 6 7 8 Agent 1 2 3 4 5 6 7 8 Agent 1 2 3 4 5 6 7 8 Agent Semi-greedy 1 2 3 4 5 6 7 8 Agent Reverse-greedy 1 2 3 4 5 6 7 8 Agent Reverse-semi-greedy Order 1 Order 2 Order 3 Order 4 Order 5 Order 6 Order 7 Order 8 1 2 3 4 5 6 7 8 9 10 Agent 1 2 3 4 5 6 7 8 9 10 Agent 1 2 3 4 5 6 7 8 9 10 Agent 1 2 3 4 5 6 7 8 9 10 Agent Semi-greedy 1 2 3 4 5 6 7 8 9 10 Agent Reverse-greedy 1 2 3 4 5 6 7 8 9 10 Agent Reverse-semi-greedy Order 1 Order 2 Order 3 Order 4 Order 5 Order 6 Order 7 Order 8 Order 9 Order 10 Figure 17: More experiments on the agents imbalance in terms of the estimated advantage. Top: 3s5z task. Bottom: MMM2 task. Published as a conference paper at ICLR 2023 0.00 0.25 0.50 0.75 1.00 Steps 107 Episode Return Humanoid-v2 9|8 0.00 0.25 0.50 0.75 1.00 Steps 107 0.00 0.25 0.50 0.75 1.00 Steps 107 Episode Return 0 2 4 Steps 106 MAPPO MAPPO w/ Adapt Co PPO Co PPO w/ Adapt HAPPO HAPPO w/ Adapt A2PO w/o Adapt A2PO Figure 18: More ablation experiments on the adaptive clipping parameter. Left: Heterogeneous or asymmetric agents. Right: Homogeneous or symmetric agents. 0.00 0.25 0.50 0.75 1.00 Steps 107 Episode Return Humanoid-v2 9|8 0 50 100 150 Minutes Episode Return Humanoid-v2 9|8 MAPPO w/o PS Co PPO w/o PS HAPPO w/o PS A2PO w/o PS (a) Comparison on Humanoid 9|8 over both environment steps and training time. 0.0 2.5 5.0 7.5 10.0 Hours Football 11 vs 11 MAPPO Co PPO HAPPO A2PO (b) Comparison on GRF 11-vs-11 scenario. Figure 19: Wall time Analysis. B.3 WALL TIME ANALYSIS Multiple updates in a stage may increase training time, and the need for more training time may impact the scalability of A2PO, which is a common concern regarding the sequential update scheme. Nevertheless, a sequential update scheme will increase training time less than might be expected. Before proceeding, we note that the majority of experiments in our work are synchronously implemented, and the training time consists of the time spent updating policies and collecting samples. We have proposed a simple yet effective method for controlling training time in order to reduce training time. As a trade-off between performance and training time, we divide the agents into blocks to reduce the number of update iterations. For example, the tasks with 10 agents can be divided into 3 blocks, with sizes 3, 3, 4, respectively, and only 3 updates will be performed in a policy update iteration. From the implementation perspective, since the number of samples used in a single update decreases, the sequential update scheme requires less memory and less updating time when update policies. Therefore, it is possible to control the training time as less than 1.5 times the training time of the simultaneous update methods. In addition, assuming a good implementation, fewer update iterations will be performed if mini-batches are used in a single policy update, as the size of a mini-batch can be greater in sequential update methods under limited memory resources. In such a case, fewer mini-batches will be used, further decreasing the training time. Moreover, sampling consumes the majority of the training time, and the increased updating time appears less significant when analysing the wall time for on-policy algorithms with synchronized implementations. The training time is depicted in Tab. 6. A2PO achieves significantly greater performance with only marginally more training time. In addition, we illustrate the Humanoid 9|8 comparisons regarding environment steps and training time in Fig. 19a, and the comparisons on the GRF 11-vs-11 scenario in Fig. 19b. A2PO maintains an advantage in terms of training time. B.4 HYPER-PARAMETERS We tune several hyper-parameters in all the benchmarks, other hyper-parameters refer to the settings used in MAPPO. cϵ are selected to be 0.5 in all the tasks. Published as a conference paper at ICLR 2023 Table 6: The comparison of training duration. The format of the first line in a cell is: Training time(Sampling time+Updating Time). The second line of a cell represents the time normalized. Task MAPPO Co PPO HAPPO A2PO 3s5z 3h29m(3h3m+0h26m) 3h33m(3h6m+0h27m) 3h49m(3h7m+0h42m) 4h32m(3h41m+0h51m) 1.00(0.87 + 0.13) 1.02(0.89 + 0.13) 1.10(0.89 + 0.20) 1.30(1.06 + 0.25) 27m vs 30m 13h23m(8h31m + 4h52m) 13h19m(8h24m + 4h55m) 16h2m(8h20m + 7h42m) 15h53m(8h7m + 7h46m) 1.00(0.64 + 0.36) 1.00(0.63 + 0.37) 1.20(0.62 + 0.58) 1.19(0.61 + 0.58) Humanoid 9|8 2h0m(1h45m + 0h15m) 1h58m(1h43m + 0h15m) 2h15m(1h45m + 0h30m) 2h31m(2h0m + 0h31m) 1.00(0.87 + 0.13) 0.99(0.86 + 0.13) 1.12(0.87 + 0.25) 1.26(1.00 + 0.26) Ant 4x2 6h42m(6h16m + 0h26m) 6h45m(6h19m + 0h26m) 7h29m(6h5m + 1h24m) 7h2m(5h34m + 1h28m) 1.00(0.93 + 0.07) 1.01(0.94 + 0.07) 1.12(0.91 + 0.21) 1.05(0.83 + 0.22) Humanoid 17x1 12h9m(10h6m + 2h3m) 17h7m(15h5m + 2h2m) 16h55m(11h2m + 5h53m) 19h25m(11h59m + 7h26m) 1.00(0.83 + 0.17) 1.41(1.24 + 0.17) 1.39(0.91 + 0.48) 1.60(0.99 + 0.61) Football 5vs5 34h46m(32h47m + 1h59m) 32h46m(30h49m + 1h57m) 39h26m(31h54m + 7h32m) 37h26m(30h2m + 7h24m) 1.00(0.94 + 0.06) 0.94(0.89 + 0.06) 1.13(0.92 + 0.22) 1.08(0.86 + 0.21) B.4.1 STARCRAFTII MULTI-AGENT CHALLENGE We list the hyper-parameters used for each task of SMAC in Tab. 7. Table 7: Hyper-parameters in SMAC. Hyperparameters agent block ppo epoch λ ϵ MMM 3 12 0.95 0.2 3s vs 5z 3 15 0.95 0.05 2c vs 64zg 2 5 0.95 0.2 3s5z 3 8 0.95 0.2 5m vs 6m 2 10 0.93 0.05 8m vs 9m 5 15 0.95 0.05 10m vs 11m 2 10 0.97 0.2 6h vs 8z 2 8 0.99 0.2 3s5z vs 3s6z 2 5 0.90 0.2 MMM2 2 5 0.95 0.2 27m vs 30m 3 5 0.95 0.2 corridor 2 5 0.95 0.2 B.4.2 MULTI-AGENT MUJOCO For the model structure in MA Mu Jo Co, the output from the last layer is processed by a Tanh layer and the action distribution is modeled as a Gaussian distribution initialized with mean as 0 and log std as -0.5. The probability output of different actions are averaged when computing the policy ratio. The common hyper-parameters used in MA Mu Jo Co are listed in Tab. 8. Table 8: Common hypermeters in MA Mu Jo Co. Hyperparameters Values entropy 0 gain 0.01 batch size 4000 B.4.3 MULTI-AGENT PARTICLE ENVIRONMENT We list the hyper-parameters used in MPE in Tab. 10. B.4.4 GOOGLE RESEARCH FOOTBALL We list the hyper-parameters used in the GRF 5-vs-5 scenario in Tab. 11. Published as a conference paper at ICLR 2023 Table 9: Hypermeters for the scenarios in MA Mu Jo Co. Hyperparameters Ant Half Cheetah Hopper Humanoid Humanoid Standup Walker2d agent block 8x1:4 / / 17x1:5 17x1:4 / ppo epoch 8 5 8 5 5 5 actor lr 3e-4 3e-4 1e-4 3e-4 3e-4 3e-4 critic lr 3e-4 3e-4 1e-4 3e-4 3e-4 3e-4 λ 0.93 0.93 0.95 0.9 0.93 0.93 ϵ 0.2 0.2 0.1 0.2 0.2 0.2 Table 10: Hypermeters for the scenarios in MPE. Hyperparameters Values ppo epoch 8 chunk length 5 entropy 0.05 actor lr 2e-4 critic lr 2e-4 λ 0.97 ϵ 0.2 C THE RELATED WORK OF OTHER MARL METHODS Value decomposition methods. The value decomposition methods such as VDN (Sunehag et al., 2017) and Qmix (Rashid et al., 2018), factorize the joint value function and adopt the centralized training and decentralized execution paradigm. The Individual-Global-MAX (IGM) principle is proposed to ensure consistency between the joint and local greedy action selections in the joint Q-value function Qtot(τ, a) and the individual Q-value function {Qi(τ i, ai}n i=1: τ T , arg maxa A Qtot(τ, a) = (arg maxa1 A1 Q1(τ 1, a1), . . . , arg maxan An Qn(τ n, an)). Two sufficient conditions, the additivity and the monotonicity, to satisfy IGM are proposed in Sunehag et al. (2017) and Rashid et al. (2018) respectively. In addition to the V function and Q function decomposition, QPLEX (Wang et al., 2021) considers implementing IGM in the dueling structure where Q = V + A. QPLEX only constrains the advantage functions to satisfy the IGM principle. The global advantage function is decomposed as Atot(τ, a) = Pn i=1 λi(τ, a)Ai(τ, ai), where λi(τ, a) > 0. We evaluate the performance of Qmix in Tab. 2 and Tab. 5. Integrating the IGM principle into A2PO without compromising the monotonic improvement guarantee is a desirable extension. Specifically, the advantage-based IGM establishes a connection between the global advantage function and the local advantage functions, and the advantage decomposition Atot(τ, a) = Pn i=1 λi(τ, a)Ai(τ, ai) will not jeopardize the derivation of the monotonic improvement guarantee. Convergence and optimality of MARL. T-PPO (Ye et al., 2022) firstly introduce a framework called Generalized Multi-Agent Actor-Critic with Policy Factorization (GPF-MAC), which consists of methods with factorized local policies and may become stuck in sub-optimality. To address this problems, T-PPO transforms a multi-agent MDP into a special single-agent MDP with a sequential structure. T-PPO transforms a multi-agent MDP into a single-agent MDP with a sequential structure to address this issue. T-PPO has been shown to produce an optimal policy if implemented properly. Theoretically, sequential update methods, such as A2PO and HAPPO, are also instances of GPF-MAC and may be stuck into sub-optimal policies. The main differences between A2PO and T-PPO include that A2PO updates the factorized policies sequentially and makes decisions simultaneously, while T-PPO makes decisions sequentially, and that A2PO does not introduce the virtual state and the sequential transformation framework network. And theoretically, T-PPO may compromise the monotonic improvement guarantee. In Tab. 12, we compare A2PO, MAPPO and T-PPO on SMAC tasks empirically. A2PO is superior to T-PPO in the majority of tasks. Published as a conference paper at ICLR 2023 Table 11: Hypermeters for the scenarios in MPE. Hyperparameters Values ppo epoch 10 chunk length 10 entropy 0.001 actor lr 5e-4 critic lr 5e-4 λ 0.95 ϵ 0.25 γ 0.995 Table 12: Comparisons of A2PO, MAPPO and T-PPO. Map Difficulty MAPPO w/ PS T-PPO w/ PS A2PO w/ PS 1c3s5z Easy 100(0.0) 99.8((0.0) 100(0.0) MMM2 Super Hard 90.6(8.9) 81.6(7.7) 98.4(1.3) 3s5z vs 3s6z Super Hard 82.8(19.2) 85.5(5.2) 93.8(19.8) 6h vs 8z Super Hard 87.5(1.5) 91.8(1.1) 90.6(1.3) corridor Super Hard 99.1(0.3) 96.9(0.0) 100(0.0) D THE RELATED WORK OF COORDINATE DESCENT Realizing the similarity between the sequential policy update scheme and the block coordinate descent algorithms, we borrow the optimization techniques in the coordinate descent algorithms to accelerate the optimization and amplify the convergence advantage over the simultaneous update scheme (Gordon & Tibshirani, 2015; Shi et al., 2017). One of the critical questions in the coordinate descent algorithms is selecting the coordinate for the next-step optimization. Glasmachers & Dogan (2013); Lu et al. (2018) provided analyses of the convergence rate advantage of the Gauss Southwell rule, i.e., greedily selecting the coordinate with the maximal gradient, over the random selection rule. We recognize the optimization of our surrogate objective (Schulman et al., 2017) agent-by-agent as a block coordinate descent problem. Therefore the agent selection rule plays a crucial role in accelerating the optimization. Inspired by the coordinate selection rules, we propose greedy and semi-greedy agent selection rules and empirically show that the underperforming agents benefit from the greedily selecting agents.