# adaptive_order_qlearning__b435fd81.pdf Adaptive Order Q-learning Tao Tan1 , Hong Xie2 , Defu Lian2 1College of Computer Science, Chongqing University 2University of Science and Technology of China tantao2020@foxmail.com, xiehong2018@foxmail.com, liandefu@ustc.edu.cn This paper revisits the estimation bias control problem of Q-learning, motivated by the fact that the estimation bias is not always evil, i.e., some environments benefit from overestimation bias or underestimation bias, while others suffer from these biases. Different from previous coarse-grained bias control methods, this paper proposes a fine-grained bias control algorithm called Order Q-learning. It uses the order statistic of multiple independent Qtables to control bias and flexibly meet the personalized bias needs of different environments, i.e., the bias can vary from underestimation bias to overestimation bias as one selects a higher order Qvalue. We derive the expected estimation bias and its lower bound and upper bound. They reveal that the expected estimation bias is inversely proportional to the number of Q-tables and proportional to the index of order statistic function. To show the versatility of Order Q-learning, we design an adaptive parameter adjustment strategy, leading to Ada Order (Adaptive Order) Q-learning. It adaptively selects the number of Q-tables and the index of order statistic function via the number of visits to state-action pair and the average Q-value. We extend Order Q-learning and Ada Order Q-learning to the large scale setting with function approximation, leading to Order DQN and Ada Order DQN, respectively. Finally, we consider two experiment settings: deep reinforcement learning experiments show that our method outperforms several SOTA baselines drastically; tabular MDP experiments reveal fundamental insights into why our method can achieve superior performance. Our supplementary file can be found in https://1drv.ms/f/s! Atddp1ia Dm L2gjv31Ca Gquw5Ww YI. 1 Introduction Q-learning [Watkins, 1989] is one of the most fundamental algorithms in reinforcement learning [Chu et al., 2022; Zhao et al., 2023; Zhang et al., 2023], and it has been Corresponding Author successfully applied to various real-world problems [Yang et al., 2020; Upadhyay, 2021; Mock and Muknahallipatna, 2023]. However, Q-learning has an overestimation bias problem [Thrun and Schwartz, 1993; Van Hasselt, 2013], and it can be exacerbated in deep reinforcement learning with nonlinear neural network approximation [Wu et al., 2022; Kondrup et al., 2023; Cetin and Celiktutan, 2023]. This overestimation bias originates from that in the updating rule of Q-learning, the maximum Q-value is an essential term and is obtained by maximizing the stochastic estimation of Q-value. Note that these stochastic estimations are caused by stochastic and unknown rewards and state transitions. A number of algorithms were proposed to mitigate this overestimation bias, but may lead to an underestimation bias [Hasselt, 2010; Peer et al., 2021; Chen et al., 2021]. The reason is that they are controlling estimation bias in a coarsegrained manner. It is shown in [Lan et al., 2020] and illustrated in Section 3.3 that different environments may have personalized bias needs for underestimation or overestimation. Unfortunately, the coarse-grained property has a fundamental limitation in satisfying such personalized bias needs. To address the above coarse-grained limitation, we propose a fine-grained bias control algorithm called Order Q-learning. It uses the order statistic of multiple independent Q-tables to control bias and flexibly adapt to different environments. More specifically, the number of Q-tables controls the boundary of bias, i.e., the larger the number of Q-tables, the larger the upper bound, and the smaller the lower bound; the index of order statistic function controls the specific size of bias in the above bounds, i.e., the bias varies from the lower bound to the upper bound as one selects a larger index. Two control parameters of Order Q-learning can provide a fine-grained bias to meet the personalized bias needs of different environments. To address the challenge in Order Q-learning, i.e., the suitable control parameters are unknown in advance, we propose an adaptive parameter adjustment algorithm called Ada Order Q-learning. Figure 2 shows the power of Order Q-learning and Ada Order Q-learning. In summary, our contributions are as follows: 1) We propose Order Q-learning, which uses the order statistic of multiple independent Q-tables to attain finegrained bias control, and aims to meet the personalized bias needs of different environments Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24) 2) We design an adaptive parameter adjustment strategy to Order Q-learning, leading to Ada Order Q-learning. We extend Order Q-learning and Ada Order Q-learning to Order DQN and Ada Order DQN, respectively. 3) We evaluate the efficiency of our method in two experiment settings, i.e., tabular MDP and deep reinforcement learning. Extensive experiment results show that our method outperforms SOTA baselines drastically in different settings. More specifically, Ada Order DQN improves the average score per episode over SOTA baselines by at least 19.49% in the Pixelcopte environments. 2 Related Work Remove overestimation bias. Double Q-learning [Hasselt, 2010] estimates the maximum Q-value via cross-validation [Van Hasselt, 2013], but it has a risk of an underestimation bias. EBQL [Peer et al., 2021] is a natural extension of Double Q-learning to ensembles. It splits the sample data into at least two disjoint sets: one set estimates the optimal action that attains maximum Q-value; the other sets estimate the mean of Q-values with previously estimated optimal action. However, with a finite number of disjoint sets, the estimation bias of maximum Q-value is always negative. Different from Maxmin Q-learning [Lan et al., 2020], REDQ [Chen et al., 2021] minimizes over random Q-tables, i.e., the default choice for the number of randomly Q-tables is two. The Qvalue of REDQ has a lower variance, but it still maintains a negative bias throughout most rounds of learning. Control estimation bias. Weighted Double Q-learning [Zhang et al., 2017] is a weighted combination of Q-learning and Double Q-learning, which controls the estimation bias through the weight parameter. Like Weighted Double Qlearning, Balanced Q-learning [Karimpanal et al., 2023] weights the optimistic bias and pessimistic bias via the balancing factor. SCQ [Zhu and Rigotti, 2021] uses the current step Q-value and the last step Q-value to build a new selfcorrecting estimator. This estimator balances the overestimation bias and the underestimation bias via the dependence hyperparameter. Softmax Q-learning [Song et al., 2019] finds that the estimation bias is proportional to the hyperparameter of softmax operation. Ensemble Q-learning [Anschel et al., 2017] averages multiple independent Q-tables to reduce the variance of Q-value. The larger the number of Q-tables, the smaller the variance, and the smaller the estimation bias. This estimation bias is inversely proportional to the number of Q-tables. However, with a finite number of Q-tables, the estimation bias is always positive. Maxmin Q-learning [Lan et al., 2020] controls the estimation bias through minimum operation on multiple independent Q-tables, and this estimation bias is inversely proportional to the number of Q-tables. To adaptively select the number of Q-tables for Maxmin Qlearning [Lan et al., 2020] in practice, Ada EQ [Wang et al., 2021] adjusts the ensemble size based on the approximation Q-value error, and this adjustment method relies on the discounted MC return [Jones and Qin, 2022]. AEQ [Gong et al., 2023] uses the uncertainty of Q-values and the familiarity of sampling trajectories to control the estimation bias naturally, and can adaptively change for specific state-action pairs. The above bias control methods are coarse-grained, and can not meet the personalized needs of different environments. Different from coarse-grained bias control, this paper proposes a fine-grained bias control algorithm to adapt to different environments. 3 Background & Model 3.1 Basics of an MDP We consider an infinite-horizon MDP [Chen et al., 2022], where each decision step is indexed by t N. Let S and A denote the state space and the action space, respectively. Let P : S A S [0, 1] denote the state transition probability function. Let R : S A R denote the reward function. Let s 7 π(s) denote the policy, where s S and π(s) A. Let Qπ(s, a) denote the expected cumulative discounted reward with discounting factor γ (0, 1) as: Qπ(s,a)=E[R(St,At)+γR(St+1,At+1)+ |St =s,At =a] , where Sκ is generated from P(Sκ|Sκ 1, Aκ 1) and Aκ = π(Sκ), κ t + 1. The learning objective is to find the optimal Q-value and the optimal policy. 3.2 Overestimation Bias & Underestimation Bias Q-learning [Watkins, 1989] maintains one Q-table denoted by Qt(s, a), and uses it to estimate the optimal Q-value. It starts with the initialization of Q-table and gets Q0(s, a). Then in each step t, from state s, the agent takes action a with εgreedy policy [Ghasemipour et al., 2021], i.e., ε (0, 1), gets reward R(s, a) and next state s . Q-learning uses the interaction data {s, a, R(s, a), s } to update the Q-table as: (Qt+1(s, a) = Qt(s, a) + α[Y (s, a) Qt(s, a)], Y (s, a) = R(s, a) + γ max a A Qt(s , a ), (1) where α is the learning rate, Y (s, a) is the target Q-value of Qt+1(s, a). Following Bellman expectation equation [Bouten et al., 2005], the expectation of target Q-value should be as E [R(s, a)] + γ maxa A E [Qt(s , a )] . However, according to Jensen s inequality [Williams et al., 2017], we have: E [Y (s, a)] = E R(s, a) + γ max a A Qt(s , a ) E [R(s, a)] + γ max a A E [Qt(s , a )] . It illustrates the overestimation bias of Q-learning. Double Q-learning [Hasselt, 2010] maintains two independent Q-tables denoted by QA t (s, a) and QB t (s, a). The updating rule, such as QA t (s, a), is as: (QA t+1(s, a) = QA t (s, a) + α[Y A(s, a) QA t (s, a)], Y A(s, a) = R(s, a) + γQB t (s , arg max a A QA t (s , a )). (2) Note that QA t (s, a) and QB t (s, a) are updated with the same probability. Following previous work [Van Hasselt, 2013], we have: E Y A(s, a) = E R(s, a)+γQB t (s , arg max a A QA t (s , a )) E [R(s, a)] + γ max a A E QA t (s , a ) . It shows the underestimation bias of Double Q-learning. Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24) Figure 1: Diagram of a simple MDP. 3.3 Effect of Estimation Bias Following Maxmin Q-learning [Lan et al., 2020], the effect of estimation bias is not always evil and depends on the environment, i.e., some environments benefit from overestimation bias or underestimation bias, while others suffer from these biases. Further, we use an example to discuss it from a more intuitive and comprehensive perspective. Example 1. Consider a simple MDP depicted in Figure 1, which is adapted from Figure 6.5 in [Sutton and Barto, 2018]. This simple episodic MDP task has three non-terminal states, i.e., state A, state B, and state C, and always starts with state A. State A has two actions, i.e., left and right: the right action transitions to the state C with a reward of zero; the left action transitions to the state B with a reward of zero. State B and state C include ten identical actions, and each action transitions to the terminal state with a reward of the distribution N(µ1, σ1) and N(µ2, σ2), respectively. We set µ1 < µ2, σ1 0, and σ2 0. Thus, the optimal action at state A is right. We discuss the effect of estimation bias as follows: 1) When σ1 = 0 and σ2 > 0, the estimation bias occurs in the high reward area. Overestimation bias can attract the agent to choose the right action at state A. Underestimation bias may make the agent choose the left action at state A. Thus, the overestimation bias is helpful, and the underestimation bias is hurtful. 2) When σ1 > 0 and σ2 = 0, the estimation bias occurs in the low reward area. Overestimation bias may induce the agent to choose the left action at state A. Underestimation bias can guide the agent to choose the right action at state A. Thus, the overestimation bias is hurtful, and the underestimation bias is helpful. 3) When σ1 > 0 and σ2 > 0, the estimation bias occurs both in the high reward area and the low reward area. Both overestimation bias and underestimation bias have a risk of making the agent choose the left action at state A. Thus, both overestimation bias and underestimation bias may be hurtful. 4 Order Q-learning & Order DQN 4.1 Order Q-learning Example 1 illustrates that the correct approach should be to control the estimation bias according to the environment, such as increasing the estimation bias in case one, reducing the estimation bias in case two, and removing the estimation Algorithm 1 Order Q-learning 1: Parameter: M, m 2: Initialize: Q1 0(s, a), Q2 0(s, a), , QM 0 (s, a) 3: Get the starting state s 4: for t = 0, 1, 2, do: 5: Choose action a at state s by ε-greedy policy 6: Take action a, get reward R(s, a) and next state s 7: Compute Qm:M t (s , a ) as Equation (3), a A 8: Randomly select one Q-table i M 9: Update Qi t(s, a) as Equation (4) 10: s s bias in case three. To meet the personalized bias needs of different environments, we propose a fine-grained bias control algorithm called Order Q-learning. We first maintain M N+ independent Q-tables denoted by Q1 t(s, a), Q2 t(s, a), , QM t (s, a), (s, a) S A. Then we reorder them from smallest to largest, and define order Q-value as: Qm:M t (s, a) = order m:M Q1 t(s, a), , QM t (s, a) , (3) where order m:M is the order statistic function and m [1, M]. Note that this order statistic function is used to extract the mth largest Q-value. We can write it more explicitly as: Q1:M t (s, a) = min i Q1 t(s, a), Q2 t(s, a), , QM t (s, a) , QM:M t (s, a) = max i Q1 t(s, a), Q2 t(s, a), , QM t (s, a) . Based on the above order Q-value, Order Q-learning uses εgreedy policy with order Q-value to interact with the environment and gets {s, a, R(s, a), s }. Following previous works [Hasselt, 2010; Lan et al., 2020; Peer et al., 2021], we update each Q-table with the same probability, such as Qi t(s, a), as: (Qi t+1(s, a) = Qi t(s, a) + α[Y i(s, a) Qi t(s, a)], Y i(s, a) = R(s, a) + γ max a A Qm:M t (s , a ), (4) where i [1, M] is the index of Q-table. Algorithm 1 outlines our Order Q-learning. Note that Order Q-learning turns to Q-learning [Watkins, 1989] when M = 1, and turns to Maxmin Q-learning [Lan et al., 2020] when m = 1. To analyze the estimation bias of Order Q-learning, following previous works [Thrun and Schwartz, 1993; Lan et al., 2020], we first make a common assumption as follows: Assumption 1. Each independent Q-table Qi t(s, a) has a uniform random error ei t(s, a), we have: Qi t(s, a) = E Qi t(s, a) + ei t(s, a), (5) where ei t(s, a) obeys the uniform distribution U( τ, τ) for some τ > 0. Due to all Q-tables being independent and updated with the same probability, we have: E Q1 t(s, a) = E Q2 t(s, a) = = E QM t (s, a) . Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24) Equation (4) illustrates the updated rule of our Order Qlearning, where Y i(s, a) is the target Q-value of Qi t+1(s, a). Thus, we define the estimation bias of Order Q-learning as: Z = Y i(s, a) E [R(s, a)] + γ max a A E Qi t(s , a ) = (R(s, a) E [R(s, a)]) + γ max a A Qm:M t (s , a ) max a A E Qi t(s , a ) . Theorem 1. Based on Assumption 1, the expected estimation bias of Order Q-learning is as: where Ci M 1 2τ M i is the ith term probability of binomial distribution B(M, p), p = 1 2 + x 2τ and x [ τ, τ]. For compactness, we write pi instead of Ci M 1 2τ M i, and rewrite E [Z] as: According to the binomial distribution B(M, p), we have PM i=1 pi = 1 and pi 0. Thus, the larger the m, the smaller the PM i=m pi, the larger the E [Z]. Similarly, the larger the M, the larger the PM i=m pi, the smaller the E [Z]. According to the above analysis, we find that the expected estimation bias of Order Q-learning is inversely proportional to M and proportional to m. Theorem 2. Based on Assumption 1, the boundary of expected estimation bias for Order Q-learning is as: τ 2τ |A| (|A| 1) 1 |A|+ 1 E[Z] γ τ 2τ 1 M|A|+1 Equation (7) illustrates that the boundary depends on M, i.e., the larger the M, the larger the upper bound, the smaller the lower bound. When M = 1, the lower bound equals the upper bound, and the expected estimation bias is γτ |A| 1 |A|+1. When M + , the lower bound and the upper bound are γτ and γτ, respectively. Combinate Theorem 1 and Theorem 2, Order Q-learning can provide a fine-grained bias via (M, m) to adapt to different environments, where M controls the boundary of bias and m controls the specific size of bias in the above bounds. More specifically, when overestimation bias is helpful, Order Q-learning can increase the estimation bias by increasing m and decreasing M; when underestimation bias is helpful, Order Q-learning can decrease the estimation bias by decreasing m and increasing M; when both overestimation bias and underestimation bias are hurtful, Order Q-learning can remove the estimation bias by adjusting (m, M). We prove Theorem 1 and Theorem 2 in our supplementary file. Algorithm 2 Order DQN 1: Parameter: M, m 2: Initialize: θ0 = {θ1 0, θ2 0, , θM 0 }, θ 0 = θ0 3: Get the starting state s 4: for t = 0, 1, 2, do: 5: Choose action a at state s by ε-greedy policy 6: Take action a, get reward r and next state s 7: Store (s, a, r, s ) in D 8: Randomly sample a set of (s B, a B, r B, s B) from D 9: for (sj, aj, rj, s j) (s B, a B, r B, s B) do: 10: Yj = rj + γmax a j AQm:M(s j, a j; θ t ) 11: Randomly select one Q-function i M 12: θi t+1 arg minθi t Ej [1,2,...,|B|] Yj Q(sj, aj; θi t) 2 13: Every V steps reset θ t+1 θt+1 14: s s 4.2 Order DQN Following previous deep reinforcement learning algorithms, such as DQN [Mnih et al., 2015], DDQN [Van Hasselt et al., 2016], Averaged DQN [Anschel et al., 2017], and Maxmin DQN [Lan et al., 2020], we represent the Q-function by a neural network for the high-dimensional environment. Based on experience replay and target network techniques [Mnih et al., 2015], we extend Order Q-learning to Order DQN as Algorithm 2, where θ is the network weight, θ is the target network weight, |D| is the buffer size, |B| is the batch size, and V N+ is the updated steps for target network weight. 5 Ada Order Q-learning & Ada Order DQN To address the challenge in Order Q-learning, i.e., the suitable parameters (M, m) are unknown in advance, we propose Ada Order Q-learning and extend it to Ada Order DQN. In particular, Ada Order Q-learning is just one of the adaptive adjustment algorithms for Order Q-learning, used to finegrained control bias close to zero. 5.1 Ada Order Q-learning When both overestimation bias and underestimation bias are hurtful, such as the case three in Example 1, Order Q-learning requires multiple experiments to find the suitable parameters (m, M) for removing the estimation bias, and it is timeconsuming. Thus, we need to design an adaptive parameter adjustment strategy. Theorem 2 illustrates that the range of bias is proportional to M. According to temporal difference [Tesauro and others, 1995], the larger the number of visits to each state-action pair, the more accurate the estimated Q-value. Thus, the range of bias should decrease as the number of visits to each state-action pair increases. According to the above analysis, the number of Q-tables for Order Q-learning should decrease with the increase of the number of visits to (s, a) as: [n(s, a) + 1]β Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24) Algorithm 3 Ada Order Q-learning 1: Parameter: C, β 2: Initialize: Q1 0(s, a), Q2 0(s, a), , QC 0 (s, a) 3: Set n(s, a) as zero, (s, a) (S, A) 4: Get the starting state s 5: for t = 0, 1, 2, do: 6: Choose action a at state s by ε-greedy policy 7: Take action a, get reward R(s, a) and next state s 8: n(s, a) = n(s, a) + 1 9: Compute M(s , a ) as Equation (8), a A 10: Compute m(s , a ) as Equation (9), a A 11: Compute Qm(s ,a ):M(s ,a ) t (s ,a ) as Equation (3) 12: Randomly select one Q-table i C 13: Y i(s, a)=R(s, a)+γmax a AQm(s ,a ):M(s ,a ) t (s , a ) 14: Qi t+1(s, a) = Qi t(s, a) + α[Y i(s, a) Qi t(s, a)] 15: s s where C N+ is the initial number of independent Q-tables, n(s, a) N is the number of visits to (s, a), and β (0, 1) controls the decreasing speed of the number of Q-tables. Theorem 1 shows that the specific bias of Order Q-learning is proportional to m, i.e., the bias varies from the lower bound to the upper bound as one selects a larger m. According to Ensemble Q-learning [Anschel et al., 2017], it uses the average operation to estimate the maximum expected Q-value. As a result, the estimation bias of Ensemble Q-learning is inversely proportional to the number of Q-tables, and it is always positive. We take Ensemble Q-learning as the baseline and select the mth largest Q-value, which is closest and smaller than the average Q-value. This way can help the estimation bias of Order Q-learning to be close to zero. The index of order statistic function for each state-action pair is as: 1 if Q1:M(s,a) t (s,a) PC i=1 Qi t(s,a) C , ( Qm:M(s,a) t (s,a)< PC i=1 Qi t(s,a) C , Qm+1:M(s,a) t (s,a) PC i=1 Qi t(s,a) C , M(s,a) if QM(s,a):M(s,a) t (s,a) PC i=1 Qi t(s,a) C . Note that for Qm:M(s,a) t (s, a), we randomly select M(s, a) Q-tables from C Q-tables as the input to Equation (3). Incorporating the above adaptive parameter adjustment strategy into Order Q-learning, we obtain Adaptive Order Qlearning as Algorithm 3, and call it Ada Order Q-learning. 5.2 Ada Order DQN During Equation (8), n(s, a) can not be calculated in the highdimensional or continuous state environment. Due to M(s, a) is inversely proportional to n(s, a) and n(s, a) is proportional to the training step t, we have that the number of Q-functions should be inversely proportional to t. Thus, we adjust the number of Q-functions as: M(t) = C 1 t T + 1 Algorithm 4 Ada Order DQN 1: Parameter: C 2: Initialize: θ0 = {θ1 0, θ2 0, , θC 0 }, θ 0 = θ0 3: Get the starting state s 4: for t = 0, 1, 2, do: 5: Compute M(t) as Equation (10) 6: Choose action a at state s by ε-greedy policy 7: Take action a, get reward r and next state s 8: Store (s, a, r, s ) in D 9: Randomly sample a set of (s B, a B, r B, s B) from D 10: for (sj, aj, rj, s j) (s B, a B, r B, s B) do 11: Compute m(s j, a j) as Equation (9) 12: Yj = rj + γmax a j AQm(s j,a j):M(t)(s j, a j; θ t ) 13: Randomly select one Q-function i C 14: θi t+1 arg minθi t Ej [1,2,...,|B|] Yj Q(sj, aj; θi t) 2 15: Every V steps reset θ t+1 θt+1 16: s s where T N+ is the total training steps. Like Order DQN in Section 4.2, we extend Ada Order Q-learning to Ada Order DQN as Algorithm 4. 6 Experiments We start with tabular MDP experiments, to reveal fundamental insights into why Order Q-learning and Ada Order Qlearning can achieve superior performance. We then evaluate the impact of Order DQN and Ada Order DQN in deep reinforcement learning settings. The code of all experiments can be found in link1. 6.1 Tabular MDP Experiments We introduce three tabular MDP environments: (1) Multiarmed bandit is adapted from [Mannor et al., 2007], which considers the single-state with ten action and the reward of each action obeys the distribution N(0, 1); (2) A simple MDP environment is depicted in Figure 1, where µ1 = 0.1, σ1 = 1.0, µ2 = 0.1, σ2 = 1.0; (3) Gridworld [Zhang et al., 2017] has four actions, i.e., up, down, left, and right for each state. The start and goal states are in the southwest and northeast positions of grid, respectively. The agent will return to the start state after it does any action at the goal state. If the agent selects an action and makes it out of grid, i.e., the grid size is 3 3, it will not move and stay in the previous state. The agent gets a stochastic reward of r1 = 12 or r2 = 10 with the same probability for any action at the non-goal state, and gets a stochastic reward of r3 = 40 or r4 = 50 with the same probability for any action at the goal state. Following [Hasselt, 2010; Zhu and Rigotti, 2021; Pentaliotis and Wiering, 2021], we set γ = 0.95, α = 1 n(s,a)0.8 , and ε = 1 n(s)0.5 for the Multiarmed bandit and Gridworld environments; and set γ = 1.0, α = 0.1, and ε = 0.1 for the MDP environment. 1https://1drv.ms/u/s!Atddp1ia Dm L2ghdc Hy YXNO785mo D Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24) Figure 2: The performance of Order Q-learning and Ada Order Q-learning in tabular MDP settings. For compactness, we write Q, DQ, EQ, MQ, SQ, WDQ, OQ, and Ada OQ as Q-learning, Double Q-learning, Ensemble Q-learning, Maxmin Q-learning, Softmax Q-learning, Weighted Double Q-learning, Order Q-learning and Ada Order Q-learning, respectively. We vary m = 1, 2, 4; M = 4, 8, 16 for OQ, C = 4, 8, 16; β = 0.05, 0.1, 0.2 for Ada OQ, and set (C = 8, β = 0.1) for default. For a fair comparison, we set the number of Q-tables of EQ, MQ, REDQ, EBQL, and Ada EQ as 8. The hyperparameters for SCQ, SQ, and WDQ are 2, 2, and 0.5, respectively, which are recommended and fine-tuned. Note that all experimental results are averaged over 1, 000 times. Figure 2 shows the maximum Q-value, the probability of taking action left at state A denoted by Pr[left|state = A], and the mean reward of Order Q-learning and Ada Order Qlearning in three tabular MDP environments. We discuss the performance of Order & Ada Order Q-learning under different bias preference environments, in our supplementary file. Order Q-learning. Figure 2(a) and Figure 2(b) show the maximum Q-value of Order Q-learning with different (M, m) in the Multi-armed bandit environment. One can observe that the estimation bias of Order Q-learning is proportional to m and inversely proportional to M, and it can be higher than that of Q-learning and lower than that of Double Q-learning. This implies that Order Q-learning can provide a more fine-grained bias for meeting the personalized needs of different environments than Q-learning and Double Q-learning. Ada Order Q-learning. Figure 2(c) and Figure 2(d) show the maximum Q-value of Ada Order Q-learning with different (C, β) in the Multi-armed bandit environment. One can observe that the estimation bias curves of Ada Order Q-learning with different (C, β) overlap, and they are close to zero. This implies that Ada Order Q-learning is not sensitive to (C, β), and can fine-grained control bias close to zero. Comparison with SOTA. Figure 2(e), Figure 2(f), and Figure 2(g) show the comparison results in the Multi-armed ban- dit, MDP, and Gridworld environments, respectively. We consider eight SOTA as Ensemble Q-learning [Anschel et al., 2017], Maxmin Q-learning [Lan et al., 2020], SCQ [Zhu and Rigotti, 2021], Softmax Q-learning [Song et al., 2019], Weighted Double Q-learning [Zhang et al., 2017], REDQ [Chen et al., 2021], EBQL [Peer et al., 2021], and Ada EQ [Wang et al., 2021]. One can observe that the maximum Qvalue curve of Ada Order Q-learning closes to zero, the probability curve of Ada Order Q-learning lies at the bottom, and the mean reward curve of Ada Order Q-learning lies at the top. They imply that Ada Order Q-learning can learn more accurate estimation bias than SOTA through fine-grained bias control, resulting in better policy and higher reward. Different adaptive methods. Figure 2(h) shows the maximum Q-value of different adaptive methods in the Multiarmed bandit environment, where Ada AQ (Adaptive Average Q-learning) and Ada MQ (Adaptive Minimum Q-learning) are obtained by replacing the order operation of Ada OQ with the average and minimum operations, respectively. Note that we keep the same (C, β) in above three methods. One can observe that the maximum Q-value curve of Ada Order Qlearning lies between that of Ada AQ and Ada MQ, and is close to zero. This implies that the fine-grained bias control, i.e., order operation, is more suitable for adaptively removing the estimation bias than the coarse-grained bias control, i.e., the average and minimum operations. Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24) Figure 3: The average score per episode of Ada Order DQN in deep reinforcement learning settings. For compactness, we write ADQN, MDQN, SDQN, WDDQN, and Ada ODQN as Averaged DQN, Maxmin DQN, Softmax DQN, Weighted Double DQN, and Ada Order DQN, respectively. We vary C = 4, 8, 16, 32 for Ada ODQN, and set C = 8 for default. Note that we keep the same hyperparameters of comparison baselines with our tabular MDP experiments, and all experimental results are averaged over 20 repeated runs. 6.2 Deep Reinforcement Learning Experiments To evaluate the impact of Order DQN and Ada Order DQN, we choose three common deep reinforcement learning games from Py Game Learning Environment [Urtans and Nikitenko, 2018] and Min Atar [Young and Tian, 2019]: Pixelcopter, Breakout, and Asterix. Following [Young and Tian, 2019; Lan et al., 2020], we reuse the settings of neural networks in [Lan et al., 2020], and set γ = 0.99, |B| = 32 for all environments. For the Pixelcopter environment, we set |D| = 10, 000, V = 200, and α = 0.001. ε decreases linearly from 1.0 to 0.01 in 1, 000 steps, and fixes to 0.01 after 1, 000 steps. For the Breakout and Asterix environments, we set |D| = 100, 000, V = 1, 000, and α = 0.01. ε decreases linearly from 1.0 to 0.1 in 100, 000 steps, and fixes to 0.1 after 100, 000 steps. Figure 3 shows the average score per episode of Ada Order DQN in deep reinforcement learning settings, where the score is averaged over the last 100 episodes and the shaded area represents one standard error. From Figure 3(a), Figure 3(b), and Figure 3(c), one can observe that the score curves of Ada Order DQN lie at the top. They imply that Ada Order DQN can get a higher score than SOTA baselines. From Figure 3(d), one can observe that the score curves of Ada Order DQN with different C overlap, and lie much higher than that of DQN and DDQN. This implies that Ada Order DQN is not sensitive to C, and outperforms DQN and DDQN drastically. We provide the experiment results of Order DQN with different (M, m), and discuss the performance of different adaptive methods, in our supplementary file. Table 1 shows the average score per episode of different algorithms, where the values are evaluated in the final round. We consider ten SOTA baselines as DQN [Mnih et al., 2015], DDQN [Van Hasselt et al., 2016], Averaged DQN [Anschel et al., 2017], Maxmin DQN [Lan et al., 2020], SCDQN [Zhu and Rigotti, 2021], Softmax DQN [Song et al., 2019], WDDQN [Zhang et al., 2017], REDQ [Chen et al., 2021], EBQL [Peer et al., 2021], and Ada EQ [Wang et al., 2021]. Our Ada Order DQN improves the average score per episode over SOTA baselines by at least 38.57 32.28 32.28 = 19.49%, Algorithm Average score per episode Pixelcopter Breakout Asterix DQN 14.09 5.06 0.65 DDQN 13.82 5.29 0.83 Averaged DQN 28.50 6.85 10.70 Maxmin DQN 13.31 4.66 1.93 SCDQN 12.23 0.89 0.56 Softmax DQN 11.22 8.90 2.73 WDDQN 17.57 5.39 0.92 REDQ 32.28 9.41 8.41 EBQL 28.06 7.42 13.33 Ada EQ 31.77 9.44 9.91 Ada Order DQN(Ours) 38.57 10.68 14.57 Table 1: The average score per episode of different algorithms. 9.44 = 13.14%, and 14.57 13.33 13.33 = 9.30%, in the Pixelcopter, Breakout, and Asterix environments, respectively. 7 Conclusion Different from previous coarse-grained bias control manner, this paper proposes Order Q-learning, which controls the estimation bias via the order statistic of multiple independent Q-tables and can provide a fine-grained bias to flexibly meet the personalized bias needs of different environments. Based on Order Q-learning, we design an adaptive parameter adjustment strategy, leading to Ada Order Q-learning. We extend Order Q-learning and Ada Order Q-learning to Order DQN and Ada Order DQN, respectively. Extensive experiment results show that our proposed method outperforms SOTA baselines drastically in different settings. We think future work should focus on the adaptive adjustment strategies, finding more flexible adaptive adjustment strategies to adapt to highly complex or changing environments, such as non-stationary environments and reward sparse environments. Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24) Acknowledgments The work of Hong Xie was supported in part by National Nature Science Foundation of China (61902042), Chongqing Natural Science Foundation(cstc2020jcyi-msxm X0652). References [Anschel et al., 2017] Oron Anschel, Nir Baram, and Nahum Shimkin. Averaged-dqn: Variance reduction and stabilization for deep reinforcement learning. In International conference on machine learning, pages 176 185. PMLR, 2017. [Bouten et al., 2005] Luc Bouten, Simon Edwards, and VP Belavkin. Bellman equations for optimal feedback control of qubit states. Journal of Physics B: Atomic, Molecular and Optical Physics, 38(3):151, 2005. [Cetin and Celiktutan, 2023] Edoardo Cetin and Oya Celiktutan. Learning pessimism for reinforcement learning. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 37, pages 6971 6979, 2023. [Chen et al., 2021] Xinyue Chen, Che Wang, Zijian Zhou, and Keith W. Ross. Randomized ensembled double qlearning: Learning fast without a model. In International Conference on Learning Representations, 2021. [Chen et al., 2022] Liyu Chen, Rahul Jain, and Haipeng Luo. Learning infinite-horizon average-reward markov decision process with constraints. In International Conference on Machine Learning, pages 3246 3270. PMLR, 2022. [Chu et al., 2022] Chen Chu, Yong Li, Jinzhuo Liu, Shuyue Hu, Xuelong Li, and Zhen Wang. A formal model for multiagent q-learning dynamics on regular graphs. In Luc De Raedt, editor, Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, IJCAI 2022, Vienna, Austria, 23-29 July 2022, pages 194 200. ijcai.org, 2022. [Ghasemipour et al., 2021] Seyed Kamyar Seyed Ghasemipour, Dale Schuurmans, and Shixiang Shane Gu. Emaq: Expected-max q-learning operator for simple yet effective offline and online rl. In International Conference on Machine Learning, pages 3682 3691. PMLR, 2021. [Gong et al., 2023] Xiaoyu Gong, Shuai L u, Jiayu Yu, Sheng Zhu, and Zongze Li. Adaptive estimation q-learning with uncertainty and familiarity. In Proceedings of the Thirty Second International Joint Conference on Artificial Intelligence, IJCAI 2023, 19th-25th August 2023, Macao, SAR, China, pages 3750 3758. ijcai.org, 2023. [Hasselt, 2010] Hado Hasselt. Double q-learning. Advances in neural information processing systems, 23:2613 2621, 2010. [Jones and Qin, 2022] Galin L Jones and Qian Qin. Markov chain monte carlo in practice. Annual Review of Statistics and Its Application, 9:557 578, 2022. [Karimpanal et al., 2023] Thommen George Karimpanal, Hung Le, Majid Abdolshah, Santu Rana, Sunil Gupta, Truyen Tran, and Svetha Venkatesh. Balanced q-learning: Combining the influence of optimistic and pessimistic targets. Artificial Intelligence, 325:104021, 2023. [Kondrup et al., 2023] Flemming Kondrup, Thomas Jiralerspong, Elaine Lau, Nathan de Lara, Jacob Shkrob, My Duc Tran, Doina Precup, and Sumana Basu. Towards safe mechanical ventilation treatment using deep offline reinforcement learning. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 37, pages 15696 15702, 2023. [Lan et al., 2020] Qingfeng Lan, Yangchen Pan, Alona Fyshe, and Martha White. Maxmin q-learning: Controlling the estimation bias of q-learning. In International Conference on Learning Representations, 2020. [Mannor et al., 2007] Shie Mannor, Duncan Simester, Peng Sun, and John N Tsitsiklis. Bias and variance approximation in value function estimates. Management Science, 53(2):308 322, 2007. [Mnih et al., 2015] Volodymyr Mnih, Koray Kavukcuoglu, David Silver, Andrei A Rusu, Joel Veness, Marc G Bellemare, Alex Graves, Martin Riedmiller, Andreas K Fidjeland, Georg Ostrovski, et al. Human-level control through deep reinforcement learning. nature, 518(7540):529 533, 2015. [Mock and Muknahallipatna, 2023] James W Mock and Suresh S Muknahallipatna. A comparison of ppo, td3 and sac reinforcement algorithms for quadruped walking gait generation. Journal of Intelligent Learning Systems and Applications, 15(1):36 56, 2023. [Peer et al., 2021] Oren Peer, Chen Tessler, Nadav Merlis, and Ron Meir. Ensemble bootstrapping for q-learning. In International Conference on Machine Learning, pages 8454 8463. PMLR, 2021. [Pentaliotis and Wiering, 2021] Andreas Pentaliotis and Marco A Wiering. Variation-resistant q-learning: Controlling and utilizing estimation bias in reinforcement learning for better performance. In ICAART (2), pages 17 28, 2021. [Song et al., 2019] Zhao Song, Ron Parr, and Lawrence Carin. Revisiting the softmax bellman operator: New benefits and new perspective. In International Conference on Machine Learning, pages 5916 5925. PMLR, 2019. [Sutton and Barto, 2018] Richard S Sutton and Andrew G Barto. Reinforcement learning: An introduction. 2018. [Tesauro and others, 1995] Gerald Tesauro et al. Temporal difference learning and td-gammon. Communications of the ACM, 38(3):58 68, 1995. [Thrun and Schwartz, 1993] Sebastian Thrun and Anton Schwartz. Issues in using function approximation for reinforcement learning. In Proceedings of the Fourth Connectionist Models Summer School, pages 255 263. Hillsdale, NJ, 1993. [Upadhyay, 2021] Indrima Upadhyay. ANALYSIS OF QLEARNING BASED GAME PLAYING AGENTS FOR ABSTRACT BOARD GAMES WITH INCREASING STATE- Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24) SPACE COMPLEXITY. Ph D thesis, Miami University, 2021. [Urtans and Nikitenko, 2018] Evalds Urtans and Agris Nikitenko. Survey of deep q-network variants in pygame learning environment. In Proceedings of the 2018 2nd International Conference on Deep Learning Technologies, pages 27 36, 2018. [Van Hasselt et al., 2016] Hado Van Hasselt, Arthur Guez, and David Silver. Deep reinforcement learning with double q-learning. In Proceedings of the AAAI conference on artificial intelligence, volume 30, 2016. [Van Hasselt, 2013] Hado Van Hasselt. Estimating the maximum expected value: an analysis of (nested) cross validation and the maximum sample average. ar Xiv preprint ar Xiv:1302.7175, 2013. [Wang et al., 2021] Hang Wang, Sen Lin, and Junshan Zhang. Adaptive ensemble q-learning: Minimizing estimation bias via error feedback. Advances in Neural Information Processing Systems, 34, 2021. [Watkins, 1989] Christopher John Cornish Hellaby Watkins. Learning from delayed rewards. 1989. [Williams et al., 2017] Grady Williams, Nolan Wagener, Brian Goldfain, Paul Drews, James M Rehg, Byron Boots, and Evangelos A Theodorou. Information theoretic mpc for model-based reinforcement learning. In 2017 IEEE International Conference on Robotics and Automation (ICRA), pages 1714 1721. IEEE, 2017. [Wu et al., 2022] Haolin Wu, Jianwei Zhang, Zhuang Wang, Yi Lin, and Hui Li. Sub-avg: Overestimation reduction for cooperative multi-agent reinforcement learning. Neurocomputing, 474:94 106, 2022. [Yang et al., 2020] Yongjian Yang, Xintao Wang, Yuanbo Xu, and Qiuyang Huang. Multiagent reinforcement learning-based taxi predispatching model to balance taxi supply and demand. Journal of Advanced Transportation, 2020, 2020. [Young and Tian, 2019] Kenny Young and Tian Tian. Minatar: An atari-inspired testbed for more efficient reinforcement learning experiments. ar Xiv preprint ar Xiv:1903.03176, 59:60, 2019. [Zhang et al., 2017] Zongzhang Zhang, Zhiyuan Pan, and Mykel J Kochenderfer. Weighted double q-learning. In IJCAI, pages 3455 3461, 2017. [Zhang et al., 2023] Zhe Zhang, Yukun Zou, Junjie Lai, and Qing Xu. M2DQN: A robust method for accelerating deep q-learning network. In Proceedings of the 15th International Conference on Machine Learning and Computing, ICMLC 2023, Zhuhai, China, February 17-20, 2023, pages 116 120. ACM, 2023. [Zhao et al., 2023] Fuqing Zhao, Qiaoyun Wang, and Ling Wang. An inverse reinforcement learning framework with the q-learning mechanism for the metaheuristic algorithm. Knowledge-Based Systems, 265:110368, 2023. [Zhu and Rigotti, 2021] Rong Zhu and Mattia Rigotti. Selfcorrecting q-learning. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pages 11185 11192, 2021. Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24)