# adaptivegrained_label_distribution_learning__239115ce.pdf Adaptive-Grained Label Distribution Learning Yunan Lu1, Weiwei Li2, Dun Liu3, Huaxiong Li4, Xiuyi Jia1* 1 School of Computer Science and Engineering, Nanjing University of Science and Technology, Nanjing, China 2 College of Computer Science and Technology, Nanjing University of Aeronautics and Astronautics, Nanjing, China 3 School of Economics and Management, Southwest Jiaotong University, Chengdu, China 4 Department of Control Science and Intelligence Engineering, Nanjing University, Nanjing, China {luyn, jiaxy}@njust.edu.cn Label polysemy, where an instance can be associated with multiple labels, is common in real-world tasks. LDL (label distribution learning) is an effective learning paradigm for handling label polysemy, where each instance is associated with a label distribution. Although numerous LDL algorithms have been proposed and achieved satisfactory performance on most existing datasets, they are typically trained directly on the collected label distributions which often lack quality guarantees in real-world tasks due to annotator subjectivity and algorithm assumptions. Consequently, direct learning from such uncertain label distributions can lead to unpredictable generalization performance. To address this problem, we propose an adaptive-grained label distribution learning framework whose main idea is to extract relatively reliable supervision information from unreliable label distributions, and thus the label distribution learning task can be decomposed into three subtasks: coarsening label distributions, learning coarse-grained labels and refining coarse-grained labels. In this framework, we design an adaptive label coarsening algorithm to extract an optimal coarsen-grained labels and a label refining function to enhance the coarse-grained label into the final label distributions. Finally, we conduct extensive experiments on real-world datasets to demonstrate the advantages of our proposal. 1 Introduction Label polysemy, where an instance is relevant to multiple labels, is ubiquitous in the real world. The most direct method for addressing label polysemy is to assign a vector of logical values (either 0 or 1) to each instance, where each logical value indicates whether the instance is relevant to the corresponding label or not; the process of learning a mapping from features to a vector of logical values is called MLL (multi-label learning (Tsoumakas and Katakis 2006)). However, MLL only gives which labels can describe the instance but cannot answer a question with more polysemy, i.e., how much does each label describe the instance. Therefore, label distribution (Geng 2016) is introduced to answer this question. Label distribution is a real-valued vector analogous to a probability distribution, where each element, called LDD (label description degree), represents the degree to which *Corresponding author Copyright 2025, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. Figure 1: An example of why CGL mitigates noise. the corresponding label describes the instance. The process of learning a mapping from features to a label distribution is called LDL (label distribution learning). So far, LDL has been applied in many practical tasks, such as sentiment analysis (Jia et al. 2019b; Machajdik and Hanbury 2010; Yang, She, and Sun 2017) and facial age estimation (Gao et al. 2018; Geng, Smith-Miles, and Zhou 2013; Wen et al. 2020). Until now, a large number of algorithms have been proposed to handle the LDL task, such as (Geng 2016; Jia et al. 2023; Liu et al. 2021; Wang and Geng 2023). We refer the readers to Section 2 for more details. Although these algorithms have achieved satisfactory performance on various existing datasets, they typically learn a predictive model directly from the collected label distributions. However, the quality of the collected label distributions is unguaranteed in real-world tasks. On the one hand, the label distributions annotated by humans are susceptible to various factors, such as the subjectivity and expertise of the annotators, as well as the incentives for the annotation task. On the other hand, the label distributions generated by label enhancement algorithms (Xu, Tao, and Geng 2018), which aims to recover label distributions from logical labels by mining potential information, are susceptible to the algorithm assumptions or data distributions. These factors collectively contribute to the challenge of ensuring the quality of collected label distributions. Consequently, direct learning such collected label distributions can yield unreliable predictive performance. Therefore, in this paper, we propose an Adaptive-Grained Label Distribution Learning framework called AGLDL. The main idea is to adaptively extract the coarse-grained label (abbr. CGL) information from unreliable label distributions. Since CGL contains a range of potential fine-grained label distributions, it is possible that the noise-disturbed label distributions and the ground truth label distribution correspond to the same CGL. As shown in Figure 1, if the value The Thirty-Ninth AAAI Conference on Artificial Intelligence (AAAI-25) Figure 2: Overall diagram of AGLDL. In the training process, features are fed into two branches. The first branch trains an unreliable LDD predictor f using the collected (unreliable) label distribution through an off-the-shelf LDL algorithm; The second branch learns a partition of the set of training instances by FDGI, which transforming the LDD into ordinal discrete CGL; a CGL predictor f can be trained on this CGL using any off-the-shelf classification algorithm. In the predicting process, the prediction from f is mapped to a distribution of LDD bijectively, which is then refined by the prediction from f . of the ground-truth LDD is d and the CGL corresponds to d δ < LDD < d + δ, where δ is a positive real number, then both the ground-truth LDD d and any LDD disturbed by noise with an amplitude smaller than δ will be mapped to the same CGL, which means that the noise with an amplitude smaller than δ can be mitigated. Therefore, learning on the coarse-grained labels to some extent can alleviate the impact of noise. Following the above idea, the LDL task can be decomposed into three sub-tasks: coarsening, learning and refining, whose workflow is shown in Figure 2. 1) In the coarsening process, the objective is to learn a set of adjacent intervals that partition the range of LDD; the ordinal number of the interval to which each LDD value belongs is then taken as its corresponding CGL value. To achieve this, we propose an adaptive label coarsening algorithm which is inspired by a basic idea from fuzzy rough feature selection, i.e., finding the smallest feature subset that maintains the fuzzy dependency (or fuzzy positive region) of the instance set. In the adaptive label coarsening algorithm, we design a metric called FDGI (fuzzy dependency gain reweighted by inclusion probability) based on the original fuzzy dependency metric, and aim to maximize FDGI in the coarsening process. 2) In the learning process, we employ an off-the-shelf classification algorithm to train a CGL predictor. 3) In the refining process, we first train an unreliable LDD predictor f directly on the collected (unreliable) label distributions using an off-the-shelf LDL algorithm. Then, within the intervals specified by CGL, we identify the label distribution that is closest to the output of f , which serves as the finegrained prediction of AGLDL. Finally, we validate AGLDL through extensive experiments on real-world datasets. The merits of AGLDL can be summarized as follows. More stable performance: Our framework is able to work stably in the environments where the quality of label distributions is uncertain or unreliable. More informative predictions: Our framework is able to output both coarse-grained and fine-grained predictions and their respective uncertainties. More compatible architecture: Our framework can seamlessly integrate with any off-the-shelf classification algorithm for learning CGL; it can also serve as a postprocessor to enhance the existing LDL algorithms. 2 Related Work Our work is mainly related to LDL. Existing works on LDL algorithms can be broadly divided into two topics, i.e., model representation and learning criterion. The focus of model representation is how to represent the mapping from features to a label distribution. For example, Geng (2016) derive a maximum entropy model for LDL. AAk NN (Geng 2016), GMM-k LDL (Zhai and Dai 2019), and LDL-LCR (Xu et al. 2020) are the instance-based models which use the training instances to represent the mapping from the features to label distribution. These works focus on how to find the neighbors of an instance and calculate the similarities. LDLogit Boost (Xing, Geng, and Xue 2016), LDLFs (Shen et al. 2017), ENN-LDL (Zhai, Dai, and Shi 2018), Struct RF (Chen et al. 2018), and BC-LDL (Wang and Geng 2018) extend the standard ensemble models for LDL. The focus of learning criterion is how to design a loss function to improve the generalization ability of the LDL algorithm. The most popular idea is to mine the label relation underlying the label distribution and design regularization terms to improve the generalization performance (Jia et al. 2018, 2024, 2023, 2019b; Ren et al. 2019a,b; Wang and Geng 2023; Zhao and Zhou 2018; Zheng, Jia, and Li 2018). Besides, some works focus on designing loss functions for special scenarios (such as incomplete and unbalanced labels). For example, Incom LDL (Xu and Zhou 2017) complements the incomplete label distributions by the low rank assumption. WSLDL-MCSC (Jia et al. 2019a) is a weakly supervised LDL method by matrix completion. Incom LDLLR (Zeng et al. 2019) is an incomplete LDL method with local reconstruction. 3 Preliminaries 3.1 Fuzzy Dependency The main idea of fuzzy rough feature selection is to find a smallest subset of features that can maintain the fuzzy dependency (or fuzzy positive region) of the original feature set. Some concepts are introduced below. Definition 1 (Triangular norm) A function T : [0, 1]2 [0, 1] is called a triangular norm if the following three conditions hold for any u, v, w [0, 1]: 1. Exchangeability, i.e., T (u, v) = T (u, v). 2. Associativity, i.e., T (u, T (v, w)) = T (T (u, v), w). 3. Monotonicity, i.e., u v, T (w, u) T (w, v). Definition 2 (Fuzzy similarity relation) Given a finite nonempty set of objects U, a function R : U 2 [0, 1] is called a fuzzy similarity relation if the following two conditions hold for any u, v U: 1) Reflexibity, i.e., R(u, u) = 1. 2) Symmetry, i.e., R(u, v) = R(v, u). Definition 3 (Fuzzy T -equivalence relation) Given a finite nonempty set of objects U, a fuzzy similarity relation R is called a fuzzy T -equivalence relation if T -transitivity is additionally imposed, i.e., T (R(u, v), R(v, w)) R(u, w) holds for any u, v, w U, where T is a triangular norm. Definition 4 (Fuzzy implicator) A mapping I : [0, 1]2 [0, 1] is called fuzzy implicator if it satisfies that I(0, 0) = 1, x [0, 1], I(1, x) = x, and I is decreasing in its first component, and increasing in its second component. Definition 5 (Fuzzy dependency) Given a finite nonempty set of objects U and a fuzzy subset V , and π is a partition on U. The fuzzy dependency γR π are defined as γR π = 1 |U| u U sup V π inf v U I R(u, v), V (v) . (1) 3.2 Problem Formulation We cope with the training datasets that appear as pairs {(xi, di)}N i=1, where xi = [xi1, xi2, , xi C] and di = [di1, di2, , di M] denote the i-th feature vector and the ith label distribution (i.e., the vector of LDD), respectively. We denote the CGL vector corresponding to the label distribution di as yi, and its element yij is an ordinal number. X = [x1; ; x N] RN C, D = [d1; ; d N] RN M and Y = [y1; ; y N] denote the matrices of features, LDD and CGL, respectively. We denote the i-th column of X and D as x i and d i, respectively, i.e., x i = [x1i, x2i, , x Ni] and d i = [d1i, d2i, , d Ni] . Our goal is to learn a mapping from a feature vector to a label distribution according to training set (X, D). 4 Methodology 4.1 General Framework of AGLDL Here we show the general framework of AGLDL in detail. The computational flow of AGLDL includes three phases. The first phase is to obtain the CGL Y and multiple partitions of training set π by the label coarsening function f , i.e., Y , π f (X, D); each CGL is a discrete value which corresponds to an interval of LDD; each partition is derived from the description degree of a certain label. The details will be illustrated in Section 4.2. The second phase is to train an unreliable LDD predictor and a CGL predictor with the training sets (X, Y ) and (X, D), respectively: f arg minf LLDD([f(xi)]N i=1, D), f arg minf LCGL([f(xi)]N i=1, Y ), (2) where LLDD( ) and LCGL( ) are the loss functions of LDD predictor and CGL predictor, respectively. These two processes correspond to two existing learning paradigms (i.e., LDL tasks and classification tasks), so they can be implemented with the help of off-the-shelf techniques. f is a mapping from a vector of features to a label distribution; f (x) = {f 1 (x), f 2 (x), , f M(x)} is a set of mappings from features to a discrete probability distribution whose components represent the probabilities that the instance belongs to the corresponding CGL. The third phase is to integrate the outputs of the CGL predictor and unstable LDD predictor by the label refining function f to obtain the final label distribution, i.e., { ˆdi} ˆ N i=1 {f (f ( ˆxi), f ( ˆxi))} ˆ N i=1, and the details is illusrated in Section 4.3. The algorithm of AGLDL is shown in Appendix. The above mechanism converts the unreliable LDD into reliable CGL. In this way, the CGL prediction can provide a reliable range for the LDD prediction. There are two questions of interest in this process: how well CGL approximates the true LDD and how probable is the LDD range specified by CGL to include the true LDD. We answer them by two theorems. Theorem 1 Given an unreliable LDD d, its corresponding true value d is subject to a uniform distribution on the interval (d ϵl, d + ϵr), and d is mapped to the interval (d δl, d + δr) by a coarsening process. ϵl (0, d), ϵr (0, 1 d), δl (0, d), δr (0, 1 d) are four independent real values. Then, the expected error of approximating 2 4 6 8 10 12 14 0.00 ε = 0.05 ε = 0.1 ε = 0.15 (a) Theorem 1 2 4 6 8 10 12 14 0.2 Inclusion probability ε = 0.05 ε = 0.1 ε = 0.15 (b) Theorem 2 Figure 3: Visualization of theorems. The horizontal axis denotes the number of identical-width sub-intervals. the true LDD by any value ˆd uniformly generated from the interval (d δl, d + δr) is given by: E[( d ˆd)2] = 3 1(ϵl ϵr)2 + 3 1(δl δr)2 +3 1(ϵlϵr + δlδr) 2 1(ϵl ϵr)(δl δr). (3) Theorem 1 quantifies how well CGL approximates the true LDD with the help of the EAE (expected approximation error) defined in (Lu and Jia 2022) and (Lu et al. 2023). The expected value E[( d ˆd)2] can be easily obtained by the integration of d and ˆd. Theorem 2 Given an unreliable LDD d, its corresponding true value d is subject to a uniform distribution on the interval (d ϵl, d + ϵr), and d is mapped to the interval (d δl, d + δr) by a coarsening process. ϵl (0, d), ϵr (0, 1 d), δl (0, d), δr (0, 1 d) are four independent real values. Then we can obtain the inclusion probability: p( d (d δl, d + δr)) = min ϵl + δr ϵl + ϵr , ϵr + δl ϵl + ϵr , δl + δr ϵl + ϵr , 1 . Theorem 2 quantifies how probable is the LDD range specified by CGL to incorporate the true LDD, which can be easily measured by the proportion of the intersection of (d δl, d + δr) and the interval of d in (d δl, d + δr). In Figure 3, we show how the granularity of CGL affects the expected approximation error and the inclusion probability. For simplicity, we in Figure 3 assume that δl = δr, ϵl = ϵr, and the interval (0, 1) is divided into K intervals of identical width, i.e., δl = δr = (2K) 1. As can be seen from Figure 3(a), if K exceeds 6, then refining the granularity of CGL (i.e., increasing K) cannot significantly reduce the expected approximation error. Figure 3(b) shows that refining the granularity of CGL significantly reduces the inclusion probability if ϵ is large, i.e., the collected LDD is unreliable. 4.2 Coarsening Label Description Degree In fuzzy rough feature selection, fuzzy dependency is an important metric to quantify the association strength between features and labels. Therefore, in the process of coarsening the label description degree, we desire the coarse-grained labels and features to have a large fuzzy dependency. Nevertheless, it should be noted that a coarser label can easily lead to a large value of fuzzy dependency. For example, if all instances belong to the same coarse-grained label (i.e., the elements of Y are identical), the fuzzy dependency will be 1. Obviously, this situation is meaningless since it leads to a large expected approximation error to the true LDD and provides an uninformative range of LDD, i.e., LDD (0, 1). Therefore, we design a metric called Fuzzy Dependency Gain reweighted by Inclusion probability (abbr. FDGI) to strike a balance between the fuzzy dependency and expected approximation error. FDGI is composed of the fuzzy dependency reweighted by error probability and expected error of approximating true LDD. Fuzzy Dependency Reweighted by Inclusion Probability Given a training set (X, D) and a CGL-based instance partition π, we first calculate the fuzzy dependency. According to Eq. (1), the fuzzy dependency relies on the fuzzy T - equivalence relation and fuzzy similarity relation. We first calculate the fuzzy similarity relation. The fuzzy similarity relation between the i-the instance and the j-th instance based on the c-th feature can be defined as: RS c (i, j) = 1 | max(x c) min(x c)| 1|xic xjc|. (5) Since it has been shown that the fuzzy intersection of all fuzzy similarity relations based on individual feature can preserve T -transitivity (Wallace, Avrithis, and Kollias 2006), we derive the fuzzy T -equivalence relation between the i-th instance and the j-th instance: RT (i, j) = \ c [C]{RS c (i, j)}, [C] = {1, 2, . . . , C}. (6) Besides, we take the fuzzy intersection as u, v 7 min(u, v), and Lukasiewicz implicator I : u, v 7 min{1, 1 u+v} is used as the fuzzy implicator in this paper. According to Theorem 2, it can be seen that the CGL of each instance probably does not contain the true LDD. Hence, we use Eq. (4) to reweight each instance, and then we can obtain the fuzzy dependency reweighted by inclusion probability: F(X, πi, d i) = 1 |U| 1 ϵl ui + ϵr ui min ϵl ui + dui dui, ϵr ui + dui dui, dui dui, ϵl ui + ϵr ui sup V πi inf v U min 1 + I(v V ) min c [C] 1 |xuc xvc| | max(x c) min(x c)| where I( ) is the indicator function, U = {1, 2, . . . , N}, dui = maxj πu i dji, dui = minj πu i dji, and πu i denotes the group (in πi) containing the instance u. ϵl ui and ϵr ui measure the unreliability of the training LDD, which are set as ϵl ui = min{dui, ζ} and ϵr ui = min{1 dui, ζ} where ζ is set to 0.05 in this paper. Fuzzy Dependency Gain Reweighted by Inclusion Probability To define FDGI, it is necessary to obtain the expected error of approximating LDD based on Theorem 1: E(πi, d i) = 1 |U| ϵl uiϵr ui ui ui ϵr ui ϵl ui 2 +( ui + ui)2 3 + (ϵl ui ϵr ui)( ui + ui) Algorithm 1: Label coarsening function f Input: training set (X, D), a set distance function dset( , ); Output: coarse-grained labels Y and partitions of training set π; 1: Y 1 Initilize a matrix of coarse-grained labels.; 2: for i = 1, 2, , M do 3: πi {{u}}N u=1; Initialize a partition of training set. 4: π i πi; Initialize the optimal partition. 5: γ 0; Initialize the maximum FDGI. 6: while |πi| > 1 do 7: v 1, v 2 arg minv1 =v2 πi πi dset(v1, v2); Find two groups closest to each other in terms of d i. 8: πi πi\v1\v2 (v1 v2); Merge the two closest groups to update the trainint set partition. 9: if γ(X, πi, d i) > γ then 10: γ γ(X, πi, d i); Update FDGI. 11: π i πi; Update the optimal partition. 12: for k = 1, 2, , |π i | do 13: π ik the group in π i which ranks the k-th position in ascending order w.r.t. the average LDD; 14: for u π ik do 15: yui k; Assign the coarse-grained label. return Y , π. where ui dui dui, ui dui dui, other symbols are defined as in Eq. (7). Finally, we define the FDGI as the quotient of the reweighted fuzzy dependency by the expected approximating error, i.e., the quotient of Eq. (7) by Eq. (8): γ(X, πi, d i) = E(πi, d i) 1F(X, πi, d i). (9) The process of coarsening the label description degree can be summarized in Algorithm 1. 4.3 Refining Coarse-Grained Label Here we investigate how to refine the CGL predictions by the unreliable LDD predictions. The main idea is to find the label distribution closest to the predicted LDD within the range of LDD specified by CGL as the result, i.e., f f (ˆx), f (ˆx), π =E ˆy p ( ˆy|ˆx) h arg mind i B(i,ˆyi,π),i [M] d f (ˆx) i , (10) where the operation arg mind i B(i,ˆyi,π),i [M] d f (ˆx) returns the label distribution closest to f (ˆx) while satisfying that d i B(i, ˆyi, π) holds for any i {1, 2, , M}; the interval B, which provides the LDD range corresponding to the CGL, and the probability distribution of coarsegrained labels p (ˆy|ˆx) are defined as follows: B(m, k, π) = (min{dim|i πmk}, max{dim|i πmk}), p (ˆy|ˆx) = m=1 Cat(ˆym|f m(ˆx)), (11) where dim is the description degree of the m-th label to the i-th training instance, πmk is the group in πm, which ranks the k-th position in ascending order w.r.t. the mean LDD value of the instance group. Cat(ˆym|f m(ˆx)) is a categorical distribution with the parameter of f m(ˆx). It should be noted that the time complexity of computing the exact solution of the expectation in Eq. (10) is at least O(KM), where K is the number of possible values for CGL, so we use Monte Carlo to approximate it: f f (ˆx), f (ˆx), π 1 l=1 arg min d i B(i,ˆy(l) i ,π),i [M] d f (ˆx) , (12) where ˆy(l) is a Monte Carlo sample generated from the categorical distribution Cat(ˆym|f m(ˆx)), and the number of Monte Carlo samples L is set to 20 in this paper. 5 Experiments 5.1 Datasets and Evaluation Measures We adopt six datasets from several representative real-world tasks, including JAFFE (Lyons et al. 1998) from a facial emotion recognition task, Movie (Geng 2016) from a movie rating prediction task, Emotion6 (Peng et al. 2015) and Painting (Machajdik and Hanbury 2010) from image sentiment recognition tasks, M2B (Nguyen et al. 2012) and FBP5500 (Liang et al. 2018) from facial beauty perception tasks. More details of the datasets can be found in Appendix. We use five LDL measures, which are Cheb (Chebyshev distance), Clark (Clark distance), KL (Kullback-Leibler divergence), Cosine (cosine coefficient), and Intersec (intersection similarity), to evaluate the performance. The formula of these measures can be found in paper (Geng 2016). The lower value of distance-based measures (i.e., Cheb, Clark, KL) or the higher value of similarity-based measures (i.e., Cosine and Intersec) indicates the better performance, which are denoted by and , respectively. 5.2 Predictive Experiments Experimental Procedure Given a dataset, it is randomly divided into two chunks (70% for training and 30% for testing). In the noise-free scenario, we train an LDL model directly through the training set and evaluate the performance on the test set; in the noisy scenario, we first inject noise into the label distribution of the training instance according to the following equation: d αϵ + (1 α)d, (13) where ϵ is a random label distribution, α is the proportion of noise in supervision information, then train an LDL model on such a noise-disturbed training set and evaluate the predictive performance on the test set. We repeat the above process ten times and record the mean and standard deviation. Comparison Methods We use SABFGS (Geng 2016), AAk NN (Geng 2016), BD-LDL (Liu et al. 2021), LDLLRR (Jia et al. 2023), and LDL-LDM (Wang and Geng 2023) for comparison. The hyperparameter K in AAk NN is selected from {5, 6, , 10}; the hyperparameters λ1 and λ2 in BD-LDL are both selected from {10 3, 10 2, , 103}; the hyperparameters λ and β in LDL-LRR are selected from {10 6, 10 5, , 10 1} Method Cheb ( ) Clark ( ) KL ( ) Cosine ( ) Intersec ( ) AGLDL (1) 0.090 0.005 (3) 0.346 0.011 (3) 0.049 0.003 (3) 0.955 0.003 (3) 0.881 0.004 AAk NN (5) 0.102 0.006 (4) 0.358 0.013 (5) 0.053 0.004 (5) 0.949 0.004 (5) 0.871 0.006 SABFGS (6) 0.117 0.007 (6) 0.463 0.028 (6) 0.086 0.011 (6) 0.926 0.009 (6) 0.841 0.009 BD-LDL (4) 0.097 0.005 (5) 0.362 0.011 (4) 0.050 0.003 (4) 0.953 0.003 (4) 0.874 0.004 LDL-LDM (1) 0.090 0.004 (1) 0.332 0.009 (1) 0.044 0.003 (1) 0.958 0.003 (1) 0.884 0.005 LDL-LRR (1) 0.090 0.004 (1) 0.332 0.009 (1) 0.044 0.003 (1) 0.958 0.003 (1) 0.884 0.005 AGLDL (3) 0.117 0.001 (3) 0.521 0.005 (3) 0.101 0.002 (3) 0.934 0.001 (3) 0.835 0.002 AAk NN (5) 0.123 0.001 (5) 0.552 0.003 (4) 0.113 0.003 (5) 0.925 0.002 (5) 0.823 0.001 SABFGS (6) 0.137 0.002 (6) 0.592 0.008 (6) 0.143 0.005 (6) 0.911 0.003 (6) 0.808 0.003 BD-LDL (3) 0.117 0.001 (4) 0.539 0.004 (5) 0.117 0.007 (3) 0.934 0.001 (4) 0.833 0.002 LDL-LDM (1) 0.115 0.001 (1) 0.515 0.004 (1) 0.098 0.002 (1) 0.936 0.001 (1) 0.837 0.002 LDL-LRR (1) 0.115 0.001 (1) 0.515 0.004 (1) 0.098 0.002 (1) 0.936 0.001 (1) 0.837 0.002 AGLDL (1) 0.252 0.006 (1) 1.622 0.013 (1) 0.427 0.012 (1) 0.805 0.006 (1) 0.659 0.005 AAk NN (2) 0.260 0.007 (4) 1.654 0.013 (3) 0.482 0.022 (2) 0.786 0.008 (3) 0.649 0.007 SABFGS (5) 0.300 0.009 (6) 1.762 0.017 (5) 0.621 0.030 (5) 0.733 0.010 (5) 0.612 0.009 BD-LDL (4) 0.282 0.005 (2) 1.638 0.015 (4) 0.517 0.019 (3) 0.775 0.006 (4) 0.620 0.005 LDL-LDM (6) 0.336 0.013 (3) 1.651 0.031 (6) 0.647 0.287 (6) 0.715 0.016 (6) 0.601 0.057 LDL-LRR (3) 0.281 0.006 (5) 1.678 0.032 (2) 0.476 0.067 (4) 0.764 0.013 (2) 0.655 0.011 AGLDL (1) 0.255 0.007 (3) 1.718 0.034 (1) 0.553 0.010 (1) 0.727 0.005 (1) 0.599 0.006 AAk NN (2) 0.258 0.011 (5) 1.754 0.035 (4) 0.590 0.035 (5) 0.707 0.014 (5) 0.587 0.013 SABFGS (6) 0.321 0.014 (6) 1.911 0.037 (6) 1.164 0.319 (6) 0.634 0.022 (6) 0.535 0.016 BD-LDL (4) 0.261 0.010 (3) 1.718 0.034 (5) 0.594 0.100 (3) 0.719 0.008 (3) 0.592 0.009 LDL-LDM (5) 0.264 0.011 (1) 1.714 0.037 (3) 0.570 0.026 (4) 0.716 0.011 (4) 0.588 0.012 LDL-LRR (3) 0.260 0.008 (1) 1.714 0.034 (2) 0.557 0.016 (2) 0.723 0.007 (2) 0.594 0.008 AGLDL (1) 0.356 0.007 (1) 1.205 0.011 (1) 0.561 0.019 (1) 0.709 0.006 (1) 0.581 0.007 AAk NN (4) 0.371 0.008 (2) 1.216 0.014 (5) 0.664 0.051 (5) 0.678 0.015 (3) 0.564 0.009 SABFGS (6) 0.392 0.008 (3) 1.254 0.015 (6) 0.815 0.048 (6) 0.645 0.010 (6) 0.551 0.008 BD-LDL (5) 0.373 0.008 (6) 1.310 0.005 (4) 0.636 0.206 (4) 0.705 0.006 (5) 0.558 0.008 LDL-LDM (3) 0.370 0.009 (5) 1.298 0.033 (3) 0.572 0.019 (3) 0.706 0.009 (4) 0.559 0.010 LDL-LRR (2) 0.365 0.007 (4) 1.273 0.034 (2) 0.562 0.055 (2) 0.708 0.022 (2) 0.570 0.006 AGLDL (2) 0.144 0.004 (1) 1.259 0.007 (2) 0.134 0.010 (2) 0.953 0.002 (2) 0.832 0.006 AAk NN (2) 0.151 0.003 (4) 1.315 0.008 (4) 0.170 0.009 (5) 0.942 0.003 (2) 0.832 0.004 SABFGS (6) 0.161 0.002 (6) 1.334 0.009 (5) 0.221 0.004 (6) 0.935 0.001 (3) 0.829 0.002 BD-LDL (5) 0.160 0.002 (5) 1.331 0.006 (6) 0.226 0.015 (4) 0.943 0.002 (6) 0.822 0.003 LDL-LDM (4) 0.155 0.002 (3) 1.294 0.005 (3) 0.139 0.004 (2) 0.945 0.002 (5) 0.823 0.003 LDL-LRR (1) 0.136 0.002 (2) 1.278 0.005 (1) 0.107 0.004 (1) 0.954 0.002 (1) 0.850 0.002 Table 1: Prediction performance formatted as (rank) mean std statistical significance. and {10 3, 10 2, , 102}, respectively; in LDL-LDM, the hyperparameters λ1, λ2, and λ3 are selected from {10 3, 10 2, , 103}, the hyperparameter g is selected from {1, 2, , 14}. In our framework, we use the ordinal logistic model (All-Threshold variant) proposed in (Rennie and Srebro 2005), abbreviated as Logistic AT, as the CGL predictor whose L2 regularization weight is selected from {1, (2n) 1, 2n}5 n=1. To prevent the performance of AGLDL from benefiting from any specific LDL algorithm, we do not use any LDD predictor to refine the predictions of the CGL predictor in experiments. The label refining function can be formalized as f f (ˆx), f (ˆx), π 1 L PL l=1 d , where d i Uni(d i|B(i, ˆy(l) i , π)). To accelerate the convergence, 0.1 0.2 0.3 0.4 0.05 0.1 0.2 0.3 0.4 0.10 0.1 0.2 0.3 0.4 0.1 0.2 0.3 0.4 0.1 0.2 0.3 0.4 0.1 0.2 0.3 0.4 Figure 4: KL performance under varying extent of noise. we preprocess the feature by min-max normalization. Results and Discussions Table 1 shows the predictive performance of each comparison method on seven real-world datasets. Each experimental result is formatted as (rank) mean std t-test ; (rank) is the rank of each comparison method among all methods; / denotes whether AGLDL is statistically superior/inferior to the corresponding comparison method under the pairwise two-tailed t-test at 0.05 significance level; if neither nor is shown, there is no significant difference between the corresponding method and AGLDL; It can be observed that our method AGLDL performs weakly on the JAFFE and Movie datasets, but performs well on the Emotion6 , Painting , and M2B datasets; the performance on the FBP5500 dataset is moderate. Furthermore, we compute average performance ranking of AGLDL and the fuzzy dependency of each dataset; the results are shown in the following table: Emotion6 M2B Painting FBP5500 JAFFE Movie 0.34 / 1 0.38 / 1 0.23 / 1.4 0.41 / 1.8 0.58 / 2.6 0.86 / 3 where each entry is formatted as fuzzy dependency / average performance ranking . It can be seen that AGLDL performs weakly on the datasets with fuzzy dependency over FBP5500, and AGLDL performs well on the datasets with fuzzy dependency below FBP5500. We believe the reason is that the fuzzy dependency is positively correlated with the quality of label information. Specifically, a lower fuzzy dependency indicates that the label information may contain more noise, and our proposed AGLDL specializes in noisy environments, which is why our method performs well. Nevertheless, we find that our method performs comparably to the top-performing method on JAFFE and Movie datasets, which indicates that our method can achieve good performance even under high-quality label information. Furthermore, we test the performance of comparison algorithms under conditions where the labels contain varying degrees of noise. As shown in Figure 4, where different styles of lines represent different comparison algorithms, the horizontal axis represents the proportion of noise. It is evi- 3 5 7 9 11 13 15 JAFFE ( = 0:0) 3 5 7 9 11 13 15 0.100 Movie ( = 0:0) 3 5 7 9 11 13 15 0.550 M2B ( = 0:0) 3 5 7 9 11 13 15 JAFFE ( = 0:3) 3 5 7 9 11 13 15 Movie ( = 0:3) 3 5 7 9 11 13 15 M2B ( = 0:3) Figure 5: KL performance of varying label coarsening methods. The horizontal axis denotes the number of possible values of CGL obtained by the label coarsening method, and the vertical axis denotes the KL performance. dent that our algorithm performs the best in most cases and is relatively insensitive to the increasing noise. 5.3 Experiments on Label Coarsening Function As shown in Figure 5, we use uniform discretization, quantile discretization and k-means discretization as alternatives to our proposed label coarsening function. Uniform discretization divides the interval of label description degree into multiple groups with identical width; quantile discretization divides the interval of label description degree into groups containing identical number of elements; kmeans discretization divides the interval of label description degree into multiple groups by the k-means clustering algorithm. It can be found that in most cases the predictive performance of our proposal is better than that of uniform discretization and quantile discretization. In addition, we observe that the predictive performance obtained by the comparison discretization methods gets better first and then worse as the label gets finer. This phenomenon is consistent with our intuition: too coarse a label can lead to a large error in approximating the true label description degree, and too fine a label can easily lead to a noise-sensitive model. 6 Conclusion This paper proposes an adaptive-grained label distribution learning framework to cope with the ubiquitous situation that the quality of the collected label distributions cannot be guaranteed. On the one hand, we novelly propose a label coarsening algorithm to transform unreliable label distributions into relatively reliable ordinal discrete labels, which can be mined by any off-the-shelf ordinal regression model. In the label coarsening algorithm, we design FDGI to obtain the best coarse-grained label. On the other hand, we design a label refining algorithm to enhance the coarse-grained labels into label distributions by any existing LDL algorithm. The experimental results on multiple real-world datasets show that our framework can significantly improve the performance of LDL in the cases where the quality of the collected label distributions is not guaranteed. Acknowledgements This work was partially supported by the Natural Science Foundation of Jiangsu Province (BK20242045), and the National Natural Science Foundation of China (62176123, 62476130, 62276217, 62176116). References Chen, M.; Wang, X.; Feng, B.; and Liu, W. 2018. Structured random forest for label distribution learning. Neurocomputing, 320: 171 182. Gao, B.; Zhou, H.; Wu, J.; and Geng, X. 2018. Age Estimation Using Expectation of Label Distribution Learning. In International Joint Conference on Artificial Intelligence, 712 718. Geng, X. 2016. Label Distribution Learning. IEEE Transactions on Knowledge and Data Engineering, 28(7): 1734 1748. Geng, X.; Smith-Miles, K.; and Zhou, Z.-H. 2013. Facial Age Estimation by Learning from Label Distributions. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35: 2401 2412. Jia, X.; Li, W.; Liu, J.; and Zhang, Y. 2018. Label Distribution Learning by Exploiting Label Correlations. In AAAI Conference on Artificial Intelligence, 3310 3317. Jia, X.; Qin, T.; Lu, Y.; and Li, W. 2024. Adaptive Weighted Ranking-Oriented Label Distribution Learning. IEEE Transactions on Neural Networks and Learning Systems, 35(8): 11302 11316. Jia, X.; Ren, T.; Chen, L.; Wang, J.; Zhu, J.; and Long, X. 2019a. Weakly supervised label distribution learning based on transductive matrix completion with sample correlations. Pattern Recognition Letters, 125: 453 462. Jia, X.; Shen, X.; Li, W.; Lu, Y.; and Zhu, J. 2023. Label Distribution Learning by Maintaining Label Ranking Relation. IEEE Transactions on Knowledge and Data Engineering, 35(2): 1695 1707. Jia, X.; Zheng, X.; Li, W.; Zhang, C.; and Li, Z. 2019b. Facial Emotion Distribution Learning by Exploiting Low-Rank Label Correlations Locally. In IEEE Conference on Computer Vision and Pattern Recognition, 9833 9842. Liang, L.; Lin, L.; Jin, L.; Xie, D.; and Li, M. 2018. SCUT-FBP5500: A Diverse Benchmark Dataset for Multi Paradigm Facial Beauty Prediction. In International Conference on Pattern Recognition, 1598 1603. Liu, X.; Zhu, J.; Zheng, Q.; Li, Z.; Liu, R.; and Wang, J. 2021. Bidirectional Loss Function for Label Enhancement and Distribution Learning. Knowledge-Based System, 213: 106690. Lu, Y.; and Jia, X. 2022. Predicting Label Distribution from Multi-Label Ranking. In Advances in Neural Information Processing Systems, 36931 36943. Lu, Y.; Li, W.; Li, H.; and Jia, X. 2023. Predicting Label Distribution From Tie-Allowed Multi-Label Ranking. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45(12): 15364 15379. Lyons, M.; Akamatsu, S.; Kamachi, M.; and Gyoba, J. 1998. Coding Facial Expressions with Gabor Wavelets. In IEEE International Conference on Automatic Face and Gesture Recognition, 200 205. Machajdik, J.; and Hanbury, A. 2010. Affective Image Classification Using Features Inspired by Psychology and Art Theory. In ACM International Conference on Multimedia, 83 92. Nguyen, T. V.; Liu, S.; Ni, B.; Tan, J.; Rui, Y.; and Yan, S. 2012. Sense Beauty via Face, Dressing, and/or Voice. In ACM International Conference on Multimedia, 239 248. Peng, K.-C.; Chen, T.; Sadovnik, A.; and Gallagher, A. 2015. A Mixed Bag of Emotions: Model, Predict, and Transfer Emotion Distributions. In IEEE Conference on Computer Vision and Pattern Recognition, 860 868. Ren, T.; Jia, X.; Li, W.; Chen, L.; and Li, Z. 2019a. Label Distribution Learning with Label-specific Features. In International Joint Conference on Artificial Intelligence, 3318 3324. Ren, T.; Jia, X.; Li, W.; and Zhao, S. 2019b. Label Distribution Learning with Label Correlations via Low-Rank Approximation. In International Joint Conference on Artificial Intelligence, 3325 3331. Rennie, J. D.; and Srebro, N. 2005. Loss Functions for Preference Levels: Regression with Discrete Ordered Labels. In IJCAI Multidisciplinary Workshop on Advances in Preference Handling, 180 186. Shen, W.; Zhao, K.; Guo, Y.; and Yuille, A. 2017. Label Distribution Learning Forests. In Advances in Neural Information Processing Systems, 834 843. Tsoumakas, G.; and Katakis, I. 2006. Multi-Label Classification: An Overview. International Journal of Data Warehousing and Mining, 3(3): 1 13. Wallace, M.; Avrithis, Y.; and Kollias, S. 2006. Computationally Efficient Sup-T Transitive Closure for Sparse Fuzzy Binary Relations. Fuzzy Sets and Systems, 157(3): 341 372. Wang, J.; and Geng, X. 2023. Label Distribution Learning by Exploiting Label Distribution Manifold. IEEE Transactions on Neural Networks and Learning Systems, 34(2): 839 852. Wang, K.; and Geng, X. 2018. Binary Coding based Label Distribution Learning. In International Joint Conference on Artificial Intelligence, 2783 2789. Wen, X.; Li, B.; Guo, H.; Liu, Z.; Hu, G.; Tang, M.; and Wang, J. 2020. Adaptive Variance Based Label Distribution Learning for Facial Age Estimation. In European Conference on Computer Vision, 379 395. Xing, C.; Geng, X.; and Xue, H. 2016. Logistic Boosting Regression for Label Distribution Learning. In IEEE Conference on Computer Vision and Pattern Recognition, 4489 4497. Xu, M.; and Zhou, Z.-H. 2017. Incomplete Label Distribution Learning. In International Joint Conference on Artificial Intelligence, 3175 3181. Xu, N.; Tao, A.; and Geng, X. 2018. Label Enhancement for Label Distribution Learning. In International Joint Conference on Artificial Intelligence, 1632 1643. Xu, S.; Ju, H.; Shang, L.; Pedrycz, W.; Yang, X.; and Li, C. 2020. Label distribution learning: A local collaborative mechanism. International Journal of Approximate Reasoning, 121: 59 84. Yang, J.; She, D.; and Sun, M. 2017. Joint Image Emotion Classification and Distribution Learning via Deep Convolutional Neural Network. In International Joint Conference on Artificial Intelligence, 3266 3272. Zeng, X.-Q.; Chen, S.-F.; Xiang, R.; Wu, S.-X.; and Wan, Z.- Y. 2019. Filling missing values by local reconstruction for incomplete label distribution learning. International Journal of Wireless and Mobile Computing, 16(4): 314 321. Zhai, Y.; and Dai, J. 2019. Geometric Mean Metric Learning for Label Distribution Learning. In International Conference on Neural Information Processing, 260 272. Zhai, Y.; Dai, J.; and Shi, H. 2018. Label Distribution Learning Based on Ensemble Neural Networks. In International Conference on Neural Information Processing, 593 602. Zhao, P.; and Zhou, Z.-H. 2018. Label Distribution Learning by Optimal Transport. In AAAI Conference on Artificial Intelligence, 4506 4513. Zheng, X.; Jia, X.; and Li, W. 2018. Label Distribution Learning by Exploiting Sample Correlations Locally. In AAAI Conference on Artificial Intelligence, 4556 4563.