# towards_decisionfriendly_auc_learning_multiclassifier_with_aucµ__4e941eb9.pdf Towards Decision-Friendly AUC: Learning Multi-Classifier with AUCµ Peifeng Gao 1, Qianqian Xu 2, *, Peisong Wen 1, 2 Huiyang Shao 1, 2, Yuan He 3, Qingming Huang 1, 2, 4, 5, 1 School of Computer Science and Tech., University of Chinese Academy of Sciences 2 Key Laboratory of Intelligent Information Processing, Institute of Computing Technology, Chinese Academy of Sciences 3 Alibaba Group 4 BDKM, University of Chinese Academy of Sciences 5 Peng Cheng Laboratory {gaopeifeng21, shaohuiyang21}@mails.ucas.ac.cn, {xuqianqian, wenpeisong20z}@ict.ac.cn heyuan.hy@alibaba-inc.com, qmhuang@ucas.ac.cn Area Under the ROC Curve (AUC) is a widely used ranking metric in imbalanced learning due to its insensitivity to label distributions. As a well-known multiclass extension of AUC, Multiclass AUC (MAUC, a.k.a. M-metric) measures the average AUC of multiple binary classifiers. In this paper, we argue that simply optimizing MAUC is far from enough for imbalanced multi-classification. More precisely, MAUC only focuses on learning scoring functions via ranking optimization, while leaving the decision process unconsidered. Therefore, scoring functions being able to make good decisions might suffer from low performance in terms of MAUC. To overcome this issue, we turn to explore AUCµ, another multiclass variant of AUC, which further takes the decision process into consideration. Motivated by this fact, we propose a surrogate risk optimization framework to improve model performance from the perspective of AUCµ. Practically, we propose a twostage training framework for multi-classification, where at the first stage a scoring function is learned maximizing AUCµ, and at the second stage we seek for a decision function to improve the F1-metric via our proposed soft F1. Theoretically, we first provide sufficient conditions that optimizing the surrogate losses could lead to the Bayes optimal scoring function. Afterward, we show that the proposed surrogate risk enjoys a generalization bound in order of O(1/ N). Experimental results on four benchmark datasets demonstrate the effectiveness of our proposed method in both AUCµ and F1metric. Introduction Over the past few decades, learning with imbalanced data has been attracting researchers attention in the machine learning community (Japkowicz and Stephen 2002; C ardenas and Baras 2006; He and Garcia 2009; Li, Chaudhuri, and Tewari 2016; Wang et al. 2021). In some real-world tasks like disease prediction tasks (Hao et al. 2020; Zhou et al. 2020) and rare event detection tasks (Liu et al. 2020a, 2018; Wu et al. 2020), the data distributions are largely skewed, i.e., a few categories occupy the vast majority of observations, whereas the rest account for a small fraction. *Corresponding author. Copyright 2023, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. In this case, commonly used classification metrics like accuracy are not ideal choices to evaluate classifiers since they might ignore the important minority categories. Hence, evaluating the model performance in the case of imbalanced data is critical for imbalanced classification problems. For binary classification tasks, Area Under the ROC Curve (AUC) has been widely used as one of the standard metrics since it is insensitivity to label distributions. Due to its importance, researchers have made efforts to direct AUC optimization (Rakotomamonjy 2004; Zhang, Saha, and Vishwanathan 2012; Ying, Wen, and Lyu 2016; Liu et al. 2020b). Since several real-world classification tasks involve more than two classes, it is natural to extend AUC to multiclass such that classifiers over imbalanced multiclass datasets can be measured and optimized similarly. More related work is presented in the Appendices. One simple extension is Multiclass AUC (MAUC), which takes the average AUC between classes in a one-vs-one or one-vs-all manner (Hand and Till 2001; Yang et al. 2021a). Despite the success of the MAUC in imbalanced learning, we argue that it is not an ideal metric for imbalanced multiclassification. Specifically, multi-classification can be solved with two stages: scoring and decision process, where the former is predicting continuous score for each category, and the latter is mapping scores to discrete categories. MAUC focuses on the scoring, but leaves the decision process unconsidered. This might lead to inconsistency of MAUC w.r.t. the same decision results, i.e., even if two scoring functions have same predictions, MAUC might be different, while another multiclass extension AUCµ(Kleiman and Page 2019) outputs consistent results. As shown in Fig.1, given two examples (x1, 1), (x2, 2) and a scoring function f with two components f 1, f 2, denote 1 = f 1(x1) f 1(x2), 2 = f 2(x1) f 2(x2). For any f with f(x1) = f(x1) + b1, f(x2) = f(x2) + b2, these two scoring functions output the same predictions, and 1,2 might be different with b1, b2 changing. However, MAUC measures 1, 2 independently, leading to different evaluation results for f and f. Since AUCµ measures margin = 1 2 invariable w.r.t. b1, b2, it is able to avoid this issue. To this end, we turn to explore AUCµ (Kleiman and Page 2019) which takes both scoring and The Thirty-Seventh AAAI Conference on Artificial Intelligence (AAAI-23) Figure 1: Illustration on inconsistency of MAUC w.r.t. same decisions. Left: f. Right: f with f(x1) = f(x1) + b1, f(x2) = f(x2) + b2. We have MAUC(f) = 1 = MAUC( f) = 0.5, AUCµ(f) = AUCµ( f) = 1. decision process into consideration. First of all, on top of Kleiman and Page s work, we further analyze AUCµ under the imbalanced multi-classification setting. Specifically, we show that AUCµ could efficiently avoid the imbalanced issue and the decision challenge by considering ranking pairs and sub-classifier pairs simultaneously. Therefore, AUCµ is more reasonable to measure performance of scoring functions for imbalanced classification problems, and it is necessary to study direct optimization of AUCµ. Based on the above considerations, we propose to train a multi-classifier with two-stages: 1) learn a scoring function with AUCµ and 2) search optimal thresholds. At the first stage, we propose to directly optimize AUCµ. The main challenge is that AUCµ is non-differentiable, which makes gradient-based optimization methods infeasible. To overcome this problem, we investigate replacing the non-differentiable 0-1 loss with differentiable surrogate losses. We provide sufficient conditions that the surrogate losses are Fisher consistent with AUCµ, i.e., optimizing the surrogate risk will lead to Bayes optimal scoring function under the AUCµ criterion. On top of this, we further propose an empirical surrogate risk of AUCµ. In this way, the AUCµ of models can be optimized without acknowledging the data prior. Additionally, we provide generalization error bounds to ensure the expected risk could be optimized. At the second stage, we target to find a set of thresholds such that the macro F1-metric is maximized. Since the argmax operation is non-differentiable, it is replaced by softmax, leading to a differentiable soft F1-metric. By optimizing the thresholds such that the soft F1-metric is minimized, we obtain a better decision function. In short, the main contribution of this paper is three-fold: From the view of decision process, we show that AUCµ is a more ideal metric compared to MAUC since both the imbalance issue and the decision process are taken into account. We propose a two-stage training framework for imbalanced multi-classification. The key components include: learning a scoring function by optimizing AUCµ, and searching a decision function with soft F1-metric. We propose an empirical surrogate risk minimization framework to directly optimize AUCµ. Theoretically, we provide consistency analysis and generalization error bounds. Learning Multi-Classifier with AUCµ In this section, we provide the formal definition of MAUC and AUCµ, and then analyze the properties of both in multiclassification problems. Preliminary Notations Denote the sample space as Z = X Y, where X Rd is the feature space with d dimensions and Y is the label space. Under the context of multi-classification, Y = {1, . . . , NC}, where NC is the number of categories. Given two examples z1 = (x1, y1) and z2 = (x2, y2), we denote the proposition that y1 = i, y2 = j or y1 = j, y2 = i as ε(i,j). Denote a scoring function for multiclassification as f 7 RNC, where the i-th component f i : X 7 R predicts the score of an instance being ith category, and the hypothesis space of f i is denoted as F = f i : X 7 R is a measurable function . I( ) is the indicator function, which returns 1 if argument ( ) is true and returns 0 otherwise. We define the 0-1 loss function I(x) = I(x > 0) + 1 Multiclass extensions of AUC The main idea of MAUC is to decompose the multi-class problem into several binary problems in a one-vs-one (ovo) or one-vs-all (ova) manner. As suggested in (Yang et al. 2021a), ovo is more comprehensive, thus we only discuss MAUC decomposed in an ovo manner. Formally, MAUC formulates multiclass AUC metric as the average AUC of each class pair (i, j): MAUC(f) = 1 NC(NC 1) j =i AUC(i,j)(f), (1) AUC(i,j)(f) = E z1,z2 h I(f i(x1) f i(x2)) ε(i,j)i . Intuitively, AUC(i,j)(f) equals the probability that scores of positive examples are higher than negative ones, thus it is insensitive to the distribution skew. Similarly, AUCµ decomposes a multi-class problem in an ovo manner. The main difference from MAUC is the definition of binary AUC: AUCµ(f) = 1 NC(NC 1) j =i AUC(i,j) µ (f), (2) AUC(i,j) µ (f) = E z1,z2 h I( margin(i, j, x1, x2, f)) ε(i,j)i , margin(i, j, x1, x2, f) = f i(x1) f i(x2) f j(x1) f j(x2) . More information about how it is defined could be found in (Kleiman and Page 2019). Multi-classification Given a sample space Z = X Y, a classifier could be viewed as a compound of a scoring function f : X 7 RNC and a decision process g : RNC 7 Y. Typically, g could be implemented as an argmax operation: g(f(x)) = arg max i [NC] f i(x). AUCµ in Decision Process In this subsection, we provide detailed explanations that compared with MAUC, showing that both AUCµ and MAUC could measure the ability of ranking. However in classification problems, AUCµ is a more ideal metric considering the decision process. To begin with, we show the necessity of considering the decision process by the following example. Example 1 (Decision bias of MAUC & AUCµ). Given a dataset {(x1, 1), (x2, 2)} and a scoring function f( ) with f(x1) = [0.9, 0.1], f(x2) = [0.8, 0.2]. Obviously we have MAUC(f) = AUCµ(f) = 1. However, if implementing the decision process g as an argmax operation, then the predicted categories are g(f(x1)) = g(f(x2)) = 1. The above phenomenon stems from the fact that AUC only constrains the relative scores for example pairs. Therefore, even if we have an optimal scoring function with MAUC = AUCµ(f) = 1, the classifier might still fail when using the argmax. To avoid this issue, it is necessary to introduce thresholds into the decision process as follows. Definition 1 (Decision process with thresholds). Given an NC-class scoring function f and a set of thresholds τ = {τ i R}NC i=1, the decision function g is defined as g(f(x); τ) = arg max i [NC] f i(x) + τ i . By introducing thresholds into the decision process, the decision bias can be addressed. For example, by setting τ = {0, 0.7} in Example 1, we have g(f(x1)) = 1, g(f(x2)) = 2, where both x1, x2 are correctly classified. Based on the above observation, we argue that the evaluation of scoring functions should take the above defined decision process into consideration. For the sake of the presentation, we propose two concepts to measure scoring functions as follows. Definition 2 (Equivalent scoring function). Given two NCclass scoring functions f, f, and a feature space X, if for any thresholds τ RNC and x X, the following condition holds: g(f(x); τ) = g( f(x); τ), then it is said that f is equivalent to f. Definition 3 (Optimal scoring function). Given an NC-class scoring functions f, and a sample space Z, if there exists a set of thresholds τ RNC, such that the predictions are correct for any (x, y) Z: g(f(x); τ) = y, then it is said that f is an optimal scoring function. Intuitively, if two scoring functions are equivalent, they always lead to same predictions no matter how the decision function is chosen. This motivates us to propose two principles for evaluating scoring functions from the decision perspective: Claim 1. An appropriate evaluation metric should: Output same results for equivalent scoring functions; Achieve score 1 for an optimal scoring function. However, it is easy to give an example that MAUC violates the above principles, while AUCµ satisfies them. Example 2 (Failure case of MAUC). Consider a dataset {(x1, 1), (x2, 2), (x3, 3)} and two scoring functions f, f with f(x1) = [0.30, 0.45, 0.25], f(x1) = [0.10, 0.25, 0.05], f(x2) = [0.20, 0.50, 0.30], f(x2) = [0.20, 0.50, 0.30], f(x3) = [0.31, 0.40, 0.29], f(x3) = [0.31, 0.40, 0.29]. Obviously, f is an optimal scoring function (by setting τ = [0.17, 0.01, 0.2]), and f is equivalent to f. According to the definition we have MAUC(f) = 2/3, MAUC( f) = 0.5. As opposed to MAUC, we have AUCµ(f) = AUCµ( f) = 1. Besides the above intuitive explanation, we further provide a formal presentation that AUCµ satisfies our principle in the following propositions. See Appendices for proofs. Proposition 1. Given two equivalent scoring function f, f, we have AUCµ(f) = AUCµ( f). Proposition 2. Given an optimal scoring function f, we have AUCµ(f) = 1. In a nutshell, AUCµ is more consistent with the prediction of a classifier. Methodology As analyzed in the last section, we can learn a classifier in a two-stage manner: 1) train a scoring function f such that AUCµ(f) is maximized; 2) optimize the thresholds τ, such that the F1-metric of g f is maximized. In this section, we first propose a training framework for maximizing AUCµ. Then we optimize Soft F1 to derive thresholds τ. By our proposed methods, one could get a classifier that has great ranking and classification performances simultaneously. Maximization of AUCµ In this subsection, we focus on the minimization of expected risk AUC µ = 1 AUCµ: R(f) = AUC µ(f) = 1 NC(NC 1) h I( margin(y1, y2, x1, x2, f)) ε(i,j)i . Surrogate Risk of AUC µ Since the 0-1 loss I( ) is nondifferentiable, gradient-based methods are infeasible to optimize the above objective function. To overcome this problem, we replace I( ) with a differentiable surrogate loss ℓ( ), and construct the following surrogate risk: Rℓ(f) = 1 NC(NC 1) h ℓ( margin(i, j, x1, x2, f)) ε(i,j)i . However, this brings up another problem: can R(f) be minimized by minimizing Rℓ(f)? In search of an answer to this problem, we introduce the concern of Fisher consistency. Definition 4 (Consistency of Rℓ(f)). Surrogate loss ℓ( ) is consistent with AUC µ if for any distributions over sample space and any sequence {ft}t N+, the following condition holds: R(ft) inf f FNCR(f) , if Rℓ(ft) inf f FNCRℓ(f). This definition gives conditions that a surrogate loss is helpful to original optimization problem. To investigate what kinds of surrogate loss are consistent, we first focus on the Bayes optimal scoring functions what kinds of scoring function f can minimize R(f). Specifically, a scoring function f is a Bayes optimal scoring function of R(f) if f arg inf f FNC R(f). And the following theorem provides sufficient conditions of Bayes optimal scoring functions: Theorem 1 (Bayes Optimal Scoring Functions). A scoring function f is Bayes optimal if i = j Y, x1, x2 X s.t. ηi(x1)ηj(x2) = ηi(x2)ηj(x1), we have [ηi(x1)ηj(x2) ηj(x1)ηi(x2)] margin > 0, ηi(x) := P z(y = i|x), margin := margin(i, j, x1, x2, f). Based on the above theorem, we provide a sufficient condition for consistency of surrogate losses as follows: Theorem 2 (Consistent Surrogate Loss of AUC µ). A surrogate loss ℓis consistent with AUC µ if it is convex, differentiable at 0 and ℓ (0) < 0. Remark 1. Note that the derivation of the above theorem is non-trivial. Previous work (Gao and Zhou 2015; Yang et al. 2021a) provides proofs of binary AUC s and MAUC s consistency by contradiction. However, since AUCµ involves relative scores between different components of the score function, we find it hard to follow existing techniques. Instead, we propose a new proof. See Appendices for details. From the above theorem, we find that most of widely-used surrogate loss functions are consistent with AUC µ. Corollary 1. These surrogate losses are consistent with AUC µ according to Thm.2. Exp loss: ℓ(x) = exp( ax); Square Loss: ℓ(x) = (1 x/a)2; Generalized Hinge Loss: 1 x if x 1 ϵ (x 1 ϵ)2/4ϵ if 1 ϵ x 1 0 otherwise Remark 2. Note that Hinge Loss max(1 x, 0) equals to Generalized Hinge Loss ℓϵ by taking ϵ 0, which means Hinge Loss is an asymptotically consistent surrogate loss. Empirical Risk Minimization So far, we have found many proper surrogate losses for surrogate risk minimization. Unfortunately, even with a proper surrogate loss, there is still a challenge to optimize Rℓ(f). Concretely, the minimization of Rℓ(f) is intractable due to the sample distribution is unavailable. This requires us to estimate Rℓ(f) over a finite training set. The following proposition gives an unbiased estimation of Rℓ(f) over a sample set S. The proof could be found in Appendices. Proposition 3 (Unbiased Empirical Estimation of Rℓ(f)). Consider a sample set S = {(xi, yi)}N i=1, where N is the number of samples. We denote {xj|yj = i, (xj, yj) S} as Si and |Si| as Ni. The unbiased estimation of Rℓ(f) on S is given by ˆRℓ,S(f) = 1 NC(NC 1) ℓ( margin(i, j, x1, x2, f)) This allows gradient-based optimization technologies to come into play. By plugging the above empirical risk into the standard optimization framework, we can obtain an approximately optimal scoring function in the sense of AUCµ. Decision Thresholds Learning After learning a scoring function with the above framework, we still need to determine a set of decision thresholds τ to form a classifier. The decision thresholds are expected to maximize the F1-metric corresponding to the decision results. Formally, given a fixed scoring function f and a training set S, we target to solve the following problem: max τ RNC F(f, τ) = [Preci(f, τ)] 1 + [Reci(f, τ)] 1 + α τ 2 2, where α is a hyperparameter to control the norm of τ, and Preci(f, τ) = P z S I (g(f(x); τ) = i) I (y = i) P z S I (g(f(x); τ) = i) , Reci(f, τ) = P z S I (g(f(x); τ) = i) I (y = i) P z S I (y = i) . However, optimizing the above target involves complicated discrete programming problems. To avoid this, we replace the non-differentiable argmax operation with softmax, and the term I (g(f(x); τ) = i) is softened to the i-th component of the predicted score f(x) + τ after softmax: exp λ (f i(x) + τ i) P j exp (λ (f j(x) + τ j)), where λ is a tunable hyperparameter. In this way, we can optimize τ with gradient-based methods. Generalization Analysis In the previous section, we have proposed a training framework to minimize the empirical risk of AUCµ. In this section, we explore whether the expected risk could be optimized by minimizing the empirical surrogate risk with sufficient examples. Formally, we target to find out an upper bound of the gap R(f) Rℓ(f) ˆRℓ.S(f) + ϵ(N), such that the upper bound ϵ(N) 0 when N . By choosing surrogate loss ℓ I, the first inequality is obvious, thus we focus on the second inequality in the rest of this section. All details of proof could be found in Appendices. Following the standard analysis framework of Probably Approximately Correct (PAC) learning, given fixed labels Y , we denote h ˆRℓ,S|Y(S) = Y i ˆRℓ,S i , then we have the following conclusion: Lemma 1. Assume dom(ℓ) = [0, B], then f FNC, δ (0, 1), we have Rℓ,S(f) ˆRℓ,S(f) + E S h Φ(S ) Y(S ) = Y i + 2B NC ψ(Y ) where ψ(Y ) = q P i N Ni , Ni = P y Y I(y = i). The above lemma shows that the generalization error could be upper bounded by bounding h Φ(S ) Y(S ) = Y i . To this end, we introduce the concept of Rademacher Complexity (Mohri, Rostamizadeh, and Talwalkar 2018), which is a kind of complexity measure of hypothesis space. Unfortunately, the standard Rademacher Complexity is infeasible here since it required the risk to be expressed as a summation of independent terms, while the risk of AUCµ is a pair-wise form which couples many examples. To handle this problem, similar to the pair-wise Rademacher Complexity of MAUC (Yang et al. 2021a), we propose the Rademacher Complexity of AUCµ as follows: Definition 5 (Rademacher Complexity of AUC µ). The empirical Rademacher Complexity of AUC µ over the dataset S is ˆRS ℓ FNC = 1 NC(NC 1) xn Sj T i,j,m,n T i,j,m,n = σm i + σn j 2 ℓ( margin(i, j, xm, xn, f)) i [NC], {σm i }i,m are i.i.d. Rademacher random variables taking 1 or 1 uniformly. Here xm Si represents that xm is the m-th example of category i in dataset S. With the above definition, ES h Φ(S ) Y(S ) = Y i can be bounded by the expected Rademacher Complexity of AUCµ as shown in Lem.2. The proof follows techniques of (Usunier, Amini, and Gallinari 2005; Agarwal et al. 2005; Yang et al. 2021a). Lemma 2 (Symmetrization of Rademacher Complexity for AUC µ). For sake of the presentation, we denote Y(S) as labels of dataset S. We have the the following condition holds: E S h Φ(S) Y(S) = Y i 4RY ℓ FNC , where RY ℓ FNC is the conditional expected Rademacher Complexity when the labels Y of dataset S is fixed, i.e., RY ℓ FNC = E S|Y ˆRS ℓ FNC . By combining the Lem.1 and Lem.2, we get a close form to the final result. The last two steps are: 1). bound expected Rademacher Complexity by the empirical form. 2). replace the conditional expectation with the general expectation. The final generalization error bound is given by the following theorem: Theorem 3 (Generalization Error Bound of AUC µ). Assume dom(ℓ) = [0, B], then f FNC, δ (0, 1), we have Rℓ,S(f) ˆRℓ,S(f)+ 4 ˆRS[ℓ FNC] + 10B Remark 3. According to the literature (Yang et al. 2021a; Golowich, Rakhlin, and Shamir 2018; Long and Sedghi 2019), the Rademacher Complexity could be bounded in order of O(1/ N) for a wide range of models including deep neural networks. Therefore, we can draw the conclusion that the following fact holds with high probability: Rℓ(f) ˆRℓ,S(f) = O(1/ N). Remark 4. From the above result, it can be seen that AUCµ enjoys an imbalance-aware generalization bound like MUAC, showing both AUCµ and MAUC metric are imbalance-friendly. Type Method CIFAR-10 CIFAR-100 Tiny Image Net Image Net 50 100 200 50 100 200 100 50 100 200 Baseline CE 93.90 93.01 92.44 93.20 92.32 91.42 89.54 96.40 95.52 94.97 Imbalanced methods Focal 94.82 94.90 90.86 93.10 92.43 92.30 88.79 96.26 95.53 94.16 CBFocal 95.48 94.40 91.16 92.85 92.71 89.88 88.72 96.04 94.90 94.46 CBCE 94.44 93.99 92.25 92.48 91.94 91.63 89.87 96.68 95.91 94.78 LDAM 94.49 94.60 92.36 92.38 92.75 91.88 90.68 96.66 95.85 94.58 TL 94.99 94.45 92.83 93.40 91.15 92.20 89.49 96.69 96.00 94.92 IHT 95.31 95.13 91.31 93.16 91.52 92.76 89.88 96.71 95.53 95.08 NM 93.62 92.58 92.55 92.96 92.04 89.75 89.40 96.31 95.99 95.18 AUC-based losses MAUC + Square 95.15 95.05 90.05 93.28 92.15 92.12 89.71 96.67 95.88 95.08 MAUC + Exp 94.59 94.15 91.18 93.26 92.41 92.49 89.76 96.51 95.93 95.04 MAUC + Hinge 93.82 93.96 91.94 93.06 92.11 91.70 89.36 96.42 95.71 95.18 Ours + Square 95.68 94.54 91.81 94.51 93.54 93.53 90.90 96.75 95.99 95.13 Ours + Exp 94.87 94.55 93.28 92.96 92.53 92.44 89.12 96.67 96.02 95.30 Ours + Hinge 95.68 95.61 92.20 93.73 92.56 92.41 89.88 96.57 95.79 95.14 Table 1: Performance comparison on AUCµ. The best results and the best competitors of each type are marked as bold and underline. 62.5 60.0 57.5 55.0 52.5 50.0 47.5 45.0 65.0 62.5 60.0 57.5 55.0 52.5 50.0 47.5 1 5 10 15 20 25 0.8 1.0 1.25 1.66 2.0 0.5 1.0 1.5 2.0 2.5 64 62 60 58 56 54 52 50 1 5 10 15 20 25 30 (a)Square Loss; . (b)Square Loss; . (d)Exp Loss; . (c)Exp Loss; . Figure 2: Sensitivity analysis of our method on CIFAR-10 with an imbalance ratio of 50. (a) and (c): sensitivity of the warm-up epochs K. (b) and (d): sensitivity of the hyperparameter a in surrogate losses. More results could be found in Appendices. Experiments To demonstrate the effectiveness of our proposed framework, we conduct a series of experiments in four benchmark datasets for imbalanced multi-classification: CIFAR10, CIFAR100 (Krizhevsky 2012), Tiny Image Net (Russakovsky et al. 2015) and Image Net (Deng et al. 2009). These datasets are resampled with imbalance ratios of 50, 100, 200. Experimental Settings Network Architecture For all experiments, we implement the scoring function as Res Net-18 trained from scratch. To normalize the scores into [0, 1], we add a softmax function after the last linear layer of the model. Competitors We compare popular methods for the imbalanced classification task. We set up standard Cross Entropy loss (CE) as a baseline. Other competitors are divided into two technical routes: 1) AUC-based loss functions including MAUC (Yang et al. 2021a) and ours using square loss, exponential loss and hinge loss respectively; 2) imbalanced methods including, Focal loss (Focal) (Lin et al. 2017), Class-Balanced CE loss (CB-CE), Class-Balanced Focal loss (Cui et al. 2019, CB-Focal), LDAM (Cao et al. 2019). Besides these loss functions, we also compare samplingbased methods including Instance Hardness Threshold sampling (Smith, Martinez, and Giraud-Carrier 2014, IHT), Near Miss sampling (Mani and Zhang 2003, NM) and Tomek Links sampling (Tomek 1976, TL). All samplingbased methods are coupled with the CE loss. Training Strategy Limited by the space, we only present a part of key training settings, and more details are provided in Appendices. We utilize the Adam optimizer (Kingma and Ba 2017) for all methods. The initial learning rates are searched in [10 4, 10 3], and decays by 0.99 per epoch. We keep the models with the highest AUCµ in the validation set and report the corresponding AUCµ and F1-metric on the test set. The training epochs are set to 25 for Image Net and 80 for other datasets. Experimental Results In our experiments, evaluation metrics including AUCµ, MAUC, precision, recall, and F1-metric are reported. Due to the limitation of the space, only test AUCµ and F1-metric of all methods are shown in Tab.1 and Tab.2 respectively, and more results of other metrics are provided in Appendices. Type Method CIFAR-10 CIFAR-100 Tiny Image Net Image Net 50 100 200 50 100 200 100 50 100 200 Baseline CE 56.21 49.37 49.14 22.21 25.24 21.04 9.50 11.23 10.28 9.15 Imbalanced methods Focal 60.39 58.26 50.23 26.92 23.40 16.91 10.72 9.40 7.85 6.41 CBFocal 61.07 56.92 46.31 23.28 20.87 21.67 9.24 9.48 6.94 6.47 CBCE 59.38 52.23 46.24 24.89 22.21 16.96 12.88 13.29 11.19 7.85 LDAM 58.34 54.31 45.91 24.11 19.64 21.07 12.88 10.76 9.35 8.58 TL 60.49 53.85 48.07 25.16 22.71 22.69 9.33 12.50 11.11 9.22 IHT 60.25 58.70 44.68 22.83 22.59 19.85 11.61 13.85 10.20 10.12 NM 47.32 42.31 35.11 23.25 25.86 19.18 9.91 10.54 10.55 8.62 AUC-based losses MAUC + Square 61.03 57.79 46.01 27.28 22.96 22.40 11.69 14.43 12.51 11.19 MAUC + Exp 58.32 51.29 48.06 24.92 25.56 22.85 13.94 13.88 12.62 11.02 MAUC + Hinge 56.15 53.46 43.89 24.91 23.12 23.72 11.66 13.51 11.76 11.30 Ours + Square 61.12 55.23 45.89 28.18 28.04 23.92 13.04 15.27 12.85 10.69 Ours + Exp 59.66 52.55 50.73 29.42 26.92 21.78 11.71 15.04 12.89 11.96 Ours + Hinge 60.68 59.42 45.82 29.45 23.34 18.48 14.35 14.94 13.14 11.64 Ours + Square 61.24 55.64 46.17 28.26 28.08 23.95 13.34 15.39 12.90 10.73 Ours + Exp 60.28 53.05 50.73 29.42 27.26 22.01 12.01 15.15 13.02 12.07 Ours + Hinge 61.98 59.61 46.41 29.53 23.53 18.64 14.42 15.08 13.26 11.74 Table 2: Performance comparison on F1-metric. For our method, we report the F1-metric with and without learning decision thresholds, denoted as Ours and Ours, respectively. We have the following observations from Tab.1 and 2: 1). Our method shows great superiority in terms of both F1-metric and AUCµ, which validates the effectiveness of the proposed framework in promoting AUCµ. 2). Compared with MAUC in the perspective of F1, ours outperforms MAUC with a large margin in most cases. This supports our claim that AUCµ taking scoring and decision thresholds into consideration simultaneously, thus learning with AUCµ could significantly promote the classification performance. 3). The AUC-based methods are usually outperforms other competitors. We attribute this to insensitivity of AUC to label distributions and argue AUC-based methods are more suitable for imbalance learning. Sensitivity Analysis To study sensitivity of our method to hyperparameters, we conduct a series of empirical studies. Specifically, we study two key hyperparameters: the hyperparameter a controlling surrogate losses and the number of warm-up epochs K. We conduct experiments by grid searching, where K is set in {1, 5, 10, 15, 20, 25, 30, 35, 50, 45}, a is set in {0.8, 1.0, 1.25, 1.66, 2.0, 2.5} for the Square loss and {0.5, 1.0, 1.5, 2.0, 2.5, 3.0}. Afterward, we study the effect of a hyperparameter by fixing the other one. The results are shown in Fig.2, where the horizontal and vertical axis represent the hyperparameters, the corresponding F1-Metric, respectively. The length of each box shows the variation of the corresponding parameter. Effect of a For a of Square Loss, one can observe a clear trend from Fig.2 (b) that the classifier achieves optimal per- formance when a is around 1 and 1.25. For a of Exp Loss, the trend is not so clear. As Fig.2 (d) shown, when we tune a in [0.5, 3.0], F1-metric do not show any obvious change. The difference comes from the gradient of Square Loss and Exp Loss. The gradient of Square Loss at 0 is very sensitive to a, so choosing the appropriate a is crucial for training. While it does not affect gradient of Exp Loss at 0 that much. Effect of K According to Fig.2 (a) and (c), the number of warm-up epochs K has a huge impact on training. If K is too small, the classifier cannot be trained well. For both surrogate losses, increasing K from 1 to 30 makes a great improvement in terms of F1-metric. This demonstrates the importance of the warm-up phase for AUC training. In this work, we argue that MAUC metric only considers the ranking of a classifier, leaving the classification decision process unconsidered. This leads to the phenomenon that a classifier enjoying a high MAUC may have poor classification ability. We turn to study another form of multiclass AUC, AUCµ. Through simple analysis, we find that AUCµ is more consistent with the prediction of a classifier. Inspired by this, we propose an empirical surrogate risk minimization framework to optimize AUCµ directly. The main challenge of optimizing AUCµ is the 0-1 loss is not differentiable. To overcome this difficulty, we replace 0-1 loss with surrogate loss. Then, we find optimizing surrogate risk of AUCµ using some widely-used surrogate losses can lead to Bayes optimal scorer. Moreover, we provide generalization analysis of our training framework. Finally, experiments on four datasets consistently show advantages of our methods. Acknowledgments This work was supported in part by the National Key R&D Program of China under Grant 2018AAA0102000, in part by National Natural Science Foundation of China: 62236008, U21B2038, 61931008, 6212200758 and 61976202, in part by the Fundamental Research Funds for the Central Universities, in part by Youth Innovation Promotion Association CAS, in part by the Strategic Priority Research Program of Chinese Academy of Sciences, Grant No. XDB28000000. References Agarwal, S.; Graepel, T.; Herbrich, R.; Har-Peled, S.; and Roth, D. 2005. Generalization Bounds for the Area Under the ROC Curve. Journal of Machine Learning Research, 6(14): 393 425. Cao, K.; Wei, C.; Gaidon, A.; Arechiga, N.; and Ma, T. 2019. Learning Imbalanced Datasets with Label Distribution-Aware Margin Loss. In Advances in Neural Information Processing Systems, volume 32, 1567 1578. Curran Associates, Inc. C ardenas, A. A.; and Baras, J. S. 2006. B-ROC Curves for the Assessment of Classifiers over Imbalanced Data Sets. In AAAI Conference on Artificial Intelligence, 1581 1584. Cortes, C.; and Mohri, M. 2003. AUC Optimization vs. Error Rate Minimization. In Advances in Neural Information Processing Systems, volume 16, 313 320. MIT Press. Cui, Y.; Jia, M.; Lin, T.-Y.; Song, Y.; and Belongie, S. 2019. Class-Balanced Loss Based on Effective Number of Samples. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 9268 9277. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; and Fei Fei, L. 2009. Imagenet: A Large-scale Hierarchical Image Database. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 248 255. Ieee. Ferri, C.; Hern andez-Orallo, J.; and Salido, M. A. 2003. Volume under the ROC Surface for Multi-Mlass Problems. In European Conference on Machine Learning, 108 120. Springer. Gao, W.; and Zhou, Z.-H. 2015. On the Consistency of AUC Pairwise Optimization. In AAAI Conference on Artificial Intelligence. Golowich, N.; Rakhlin, A.; and Shamir, O. 2018. Sizeindependent Sample Complexity of Neural Networks. In Conference On Learning Theory, 297 299. PMLR. Hand, D. J.; and Till, R. J. 2001. A Simple Generalisation of the Area Under the ROC Curve for Multiple Class Classification Problems. In Machine Learning, volume 45, 171 186. Springer. Hao, H.; Fu, H.; Xu, Y.; Yang, J.; Li, F.; Zhang, X.; Liu, J.; and Zhao, Y. 2020. Open-Appositional-Synechial Anterior Chamber Angle Classification in AS-OCT Sequences. In Medical Image Computing and Computer Assisted Intervention, 715 724. ISBN 978-3-030-59721-4. He, H.; and Garcia, E. A. 2009. Learning from Imbalanced Data. In IEEE Transactions on Knowledge and Data Engineering, volume 21, 1263 1284. Ieee. Herschtal, A.; and Raskutti, B. 2004. Optimising Area under the ROC Curve using Gradient Descent. In International Conference on Machine learning, 49. Honz ık, P.; Kuˇcera, P.; Hynˇcica, O.; and Jirs ık, V. 2009. Novel Method for Evaluation of Multi-Class Area under Receiver Operating Characteristic. In International Conference on Soft Computing, Computing with Words and Perceptions in System Analysis, Decision and Control, 1 4. IEEE. Japkowicz, N.; and Stephen, S. 2002. The Class Imbalance Problem: A Systematic Study. In Intelligent Data Analysis, volume 6, 429 449. IOS Press. Kingma, D. P.; and Ba, J. 2017. Adam: A method for stochastic optimization. In International Conference on Learning Representations. Kleiman, R.; and Page, D. 2019. Aucµ: A Performance Metric for Multi-Class Machine Learning Models. In International Conference on Machine Learning, 3439 3447. PMLR. Krizhevsky, A. 2012. Learning Multiple Layers of Features from Tiny Images. University of Toronto. Lane, T. 2000. Extensions of ROC Analysis to Multi Class Domains. In ICML-2000 Workshop on Cost-Sensitive Learning, Standford. Li, B.; Chaudhuri, S.; and Tewari, A. 2016. Handling Class Imbalance in Link Prediction Using Learning to Rank Techniques. In AAAI Conference on Artificial Intelligence, volume 30. Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; and Doll ar, P. 2017. Focal Loss for Dense Object Detection. In IEEE/CVF International Conference on Computer Vision, 2980 2988. Ling, C. X.; Huang, J.; Zhang, H.; et al. 2003. AUC: A Statistically Consistent and More Discriminating Measure than Accuracy. In International Joint Conference on Artificial Intelligence, volume 3, 519 524. Liu, C.; Zhong, Q.; Ao, X.; Sun, L.; Lin, W.; Feng, J.; He, Q.; and Tang, J. 2020a. Fraud Transactions Detection via Behavior Tree with Local Intention Calibration. In ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, 3035 3043. Liu, M.; Yuan, Z.; Ying, Y.; and Yang, T. 2020b. Stochastic AUC Maximization with Deep Neural Networks. In International Conference on Learning Representations. Liu, W.; Luo, W.; Lian, D.; and Gao, S. 2018. Future Frame Prediction for Anomaly Detection A New Baseline. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 6536 6545. Long, P. M.; and Sedghi, H. 2019. Generalization Bounds for Deep Convolutional Neural Networks. In International Conference on Learning Representations. Mani, I.; and Zhang, I. 2003. KNN Approach to Unbalanced Data Distributions: A Case Study Involving Information Extraction. In Proceedings of Workshop on Learning from Imbalanced Datasets, volume 126, 1 7. International Conference on Machine Learning. Mc Clish, D. K. 1989. Analyzing a Portion of the ROC Curve. In Medical decision making, volume 9, 190 195. Sage Publications Sage CA: Thousand Oaks, CA. Mohri, M.; Rostamizadeh, A.; and Talwalkar, A. 2018. Foundations of Machine Learning. MIT press. Mossman, D. 1999. Three-way Rocs. In Medical Decision Making, volume 19, 78 89. Sage Publications Sage CA: Thousand Oaks, CA. Narasimhan, H.; and Agarwal, S. 2013. A Structural SVM Based Approach for Optimizing Partial AUC. In International Conference on Machine Learning, 516 524. PMLR. Natole, M.; Ying, Y.; and Lyu, S. 2018. Stochastic Proximal Algorithms for AUC Maximization. In International Conference on Machine Learning, 3710 3719. PMLR. Pepe, M. S.; and Thompson, M. L. 2000. Combining Diagnostic Test Results to Increase Accuracy. In Biostatistics, volume 1, 123 140. Oxford University Press. Provost, F.; and Domingos, P. 2000. Well-Trained PETs: Improving Probability Estimation Trees. In Raport instytutowy IS-00-04, Stern School of Business, New York University. Rakotomamonjy, A. 2004. Support Vector Machines and Area under ROC Curve. In PSI-INSA de Rouen: Technical Report. Citeseer. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. 2015. Imagenet Large Scale Visual Recognition Challenge. 115(3): 211 252. Smith, M. R.; Martinez, T.; and Giraud-Carrier, C. 2014. An Instance Level Analysis of Data Complexity. In Machine learning, volume 95, 225 256. Springer. Tomek, I. 1976. Two Modifications of CNN. SMC-6(11): 769 772. Usunier, N.; Amini, M.-R.; and Gallinari, P. 2005. A Data Dependent Generalisation Error Bound for the AUC. In ICML Workshop on ROC Analysis in Machine Learning. Wang, L.; Xu, S.; Wang, X.; and Zhu, Q. 2021. Addressing Class Imbalance in Federated Learning. In AAAI Conference on Artificial Intelligence, volume 35, 10165 10173. Wang, Z.; and Chang, Y.-C. I. 2011. Marker Selection via Maximizing the Partial Area under the ROC Curve of Linear Risk Scores. In Biostatistics, volume 12, 369 385. Oxford University Press. Wu, H.; Hu, Z.; Jia, J.; Bu, Y.; He, X.; and Chua, T.-S. 2020. Mining Unfollow Behavior in Large-scale Online Social Networks via Spatial-Temporal Interaction. In AAAI Conference on Artificial Intelligence, volume 34, 254 261. Yang, B. 2009. The Extension of the Area under the Receiver Operating Characteristic Curve to Multi-Class Problems. In ISECS International Colloquium on Computing, Communication, Control, and Management, volume 2, 463 466. IEEE. Yang, Z.; Xu, Q.; Bao, S.; Cao, X.; and Huang, Q. 2021a. Learning with Multiclass AUC: theory and algorithms. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(11): 7747 7763. Yang, Z.; Xu, Q.; Bao, S.; He, Y.; Cao, X.; and Huang, Q. 2021b. When All We Need is a Piece of the Pie: A Generic Framework for Optimizing Two-way Partial AUC. In International Conference on Machine Learning, 11820 11829. PMLR. Yang, Z.; Xu, Q.; Bao, S.; He, Y.; Cao, X.; and Huang, Q. 2022. Optimizing Two-way Partial AUC with an End-to-end Framework. In IEEE Transactions on Pattern Analysis and Machine Intelligence. IEEE. Ying, Y.; Wen, L.; and Lyu, S. 2016. Stochastic Online AUC Maximization. In Advances in Neural Information Processing Systems, volume 29. Yuan, Z.; Yan, Y.; Sonka, M.; and Yang, T. 2021. Largescale Robust Deep AUC Maximization: A New Surrogate Loss and Empirical Studies on Medical Image Classification. In IEEE/CVF International Conference on Computer Vision, 3040 3049. Zhang, X.; Saha, A.; and Vishwanathan, S. 2012. Smoothing Multivariate Performance Measures. In Journal of Machine Learning Research, volume 13, 3623 3680. JMLR. org. Zhou, K.; Gao, S.; Cheng, J.; Gu, Z.; Fu, H.; Tu, Z.; Yang, J.; Zhao, Y.; and Liu, J. 2020. Sparse-GAN: Sparsityconstrained Generative Adversarial Network for Anomaly Detection in Retinal OCT Image. In IEEE International Symposium on Biomedical Imaging (ISBI), 1227 1231. IEEE.