# predicting_label_distribution_from_multilabel_ranking__7b16542d.pdf Predicting Label Distribution from Multi-label Ranking Yunan Lu, Xiuyi Jia School of Computer Science and Engineering Nanjing University of Science and Technology, Nanjing 210094, China {luyn, jiaxy}@njust.edu.cn Label distribution can provide richer information about label polysemy than logical labels in multi-label learning. There are currently two strategies including LDL (label distribution learning) and LE (label enhancement) to predict label distributions. LDL requires experts to annotate instances with label distributions and learn a predictive mapping on such a training set. LE requires experts to annotate instances with logical labels and generates label distributions from them. However, LDL requires costly annotation, and the performance of the LE is unstable. In this paper, we study the problem of predicting label distribution from multi-label ranking which is a compromise w.r.t. annotation cost but has good guarantees for performance. On the one hand, we theoretically investigate the relation between multi-label ranking and label distribution. We define the notion of EAE (expected approximation error) to quantify the quality of an annotation, give the bounds of EAE for multi-label ranking, and derive the optimal range of label distribution corresponding to a particular multi-label ranking. On the other hand, we propose a framework of label distribution predicting from multi-label ranking via conditional Dirichlet mixtures. This framework integrates the processes of recovering and learning label distributions end-to-end and allows us to easily encode our knowledge about current tasks by a scoring function. Finally, we implement extensive experiments to validate our proposal. 1 Introduction The label polysemy problem has been a popular research topic in machine learning area, in which an instance is described by multiple labels simultaneously. MLL (multi-label learning) [23] deals with label polysemy by assigning a vector with logical values to the instance, in which each logical value indicates whether the corresponding label is associated with the instance. However, MLL only gives which labels can describe the instance, but cannot directly answer a question with more polysemy, i.e., how much does each label describe the instance. Hence, label distribution [4], a real-valued vector that explicitly gives the description degrees of labels to an instance, is introduced to answer this question. Obviously, label distribution provides richer information about label polysemy than logical labels, and it has been applied in many practical application scenarios, such as sentiment analysis [14, 35, 39], facial age estimation [3, 6, 30], and so on. There are two main methods for obtaining label distribution to characterize label polysemy. The first type is LDL (label distribution learning) [4], that is, learning a predictive mapping from a feature vector to a label distribution. LDL requires experts to directly annotate the instances with label distributions as a training set. Such methods focus on how to design a well performed LDL algorithm. Typical works include some algorithms [10, 20, 25, 37, 38] that improve LDL performance by mining Corresponding author 36th Conference on Neural Information Processing Systems (Neur IPS 2022). label correlations, and some [21, 31] that improve the ability to fit complex label distributions by introducing more flexible models. The second type is LE (label enhancement) [34], that is, generating label distributions from a vector of logical labels. LE requires experts to annotate the instances with logical labels first, and then recover the label distribution by analyzing the features and labels of the training instances, finally train a predictive model by these recovered label distributions. In other words, LE can be regarded as the pre-processing of LDL to obtain the label distribution. Such methods focus on how to recover the true label distribution accurately from the given logical labels. For example, most LE algorithms [12, 22, 29, 34, 40] consider recovering more accurate label distributions by mining sample and label correlations. Since LDL methods are trained directly on instances with true label distributions, they usually produce better performance. However, annotating instances with label distributions is costly and even impractical in some cases [34]. In contrast, LE methods only require experts to annotate the instance with logical labels, thus reducing the annotation cost. However, there is no reliable theory to guarantee that the label distribution recovered from logical labels converges to the true label distribution. In terms of the annotation form, two labels taking the same logical value are indistinguishable in the MLL case, while the label distribution usually characterizes them as labels with different description degrees. In terms of solution, logical labels provide a large solution space for the label distribution (e.g., the label distribution component corresponding to a label with a logical value of 1 can take any real value from 0 to 1), making the solution unstable and inaccurate. Therefore, we propose a hypothesis that the label distribution recovered from logical labels is not guaranteed to have the same label ranking as the true label distribution, while accurate label ranking is the key to recovering and predicting an accurate label distribution [13, 26, 27]. Fortunately, multi-label ranking [1, 2, 7] is a good annotation form to address the above problems. Multi-label ranking2 requires experts to give which labels are relevant to the instance and further to give the ranking (strict order) of these relevant labels. Although the annotation cost of multi-label ranking is slightly higer than that of logical labels, it guarantees a ranking consistent with the true label distribution and constrains the approximated label distribution in a narrow solution space. Hence, in this paper, we investigate the problem of predicting label distribution from multi-label ranking. On the one hand, we theoretically investigate the relation between multi-label rankings and label description vectors3 (unnormalized label distributions). We define the notion of EAE (expected approximation error) to quantify the quality of annotation w.r.t. recovering the true label description vector; we derive the EAE for multi-label ranking and logical labels to clarify the advantage of multi-label ranking; we give the bounds of EAE for multi-label ranking, and derive the optimal range of label description vector corresponding to a particular multi-label ranking. On the other hand, we propose a generic framework named DRAM (label Distribution predicting from multi-label RAnking via conditional Dirichlet Mixtures). This framework first forms a semi-adaptive prior p (d) for the label distribution via a scoring function with a predefined functional form and adaptive parameters, then models the predictive distribution p(d|x) by conditional Dirichlet mixtures, finally learns the model parameters by minimizing the cross-entropy of p(d|x) relative to p (d). This framework has two merits: 1) It allows us to flexibly encode our prior knowledge about the tasks by a scoring function, and 2) it integrates the processes of recovering and learning label distributions end-to-end. Besides, we design a comparison method whose main idea is to transform the dataset with multi-label rankings into the dataset with logical labels such that any existing LE method can be borrowed. Finally, to validate our proposal, we conduct experiments on reduced LDL datasets and a new real-world dataset that we create directly according to the task. The experimental results show that our method significantly outperforms the comparison methods and also outperforms the LDL methods directly trained on the examples with true label distribution in most cases. 2In some literature, multi-label ranking is a learning task; in this paper, it only denotes an annotation form. 3In general, label distribution d satisfies that each element di [0, 1] and P di = 1. For simplicity, we sometimes do not consider P di = 1. We call such an unnormalized label distribution a label description vector. 2 Theoretical analysis 2.1 Preliminary Let x denote the feature vector of the instance and Y = {yi}M i=1 denote the label set. The label description vector z is the expert s internal view of how much does each label describe the instance; zi [0, 1] indicates the description degree of yi to x. If the expert is asked to annotate the instance with logical labels, the internal label description vector will be degenerately expressed as a logical label vector l {0, 1}M; the element li = 1 (or li = 0) in l means that the label yi is the relevant (or irrelevant) to the instance x. Let m denote the number of relevant labels. If the expert is asked to annotate the instance with a multi-label ranking, the internal label description vector will be degenerately expressed as a permutation σ (which represents a total strict order) on the relevant labels; σi indicates that the label yσi is at the i-th position in ascending order of the description degree; for j [M]\σ, the label yj is an irrelevant label for x, where [M] {1, 2, , M}. Since both logical labels and multi-label ranking are reduced versions of the internal label description vector, there is consistency between them, which can be described in the following two assumptions: Assumption 1 If the expert s internal label description vector z is expressed as a logical label vector l, and δ > 0 is the minimum margin4 of label description degrees, then we have z Sδ l , where Sδ l = {z | ( li = 0, zi = 0) ( lj = 1, δ zj 1)}. Assumption 2 If the expert s internal label description vector z is expressed as a multi-label ranking σ, and δ > 0 is the minimum margin of label description degrees, then we have z Sδ σ, where Sδ σ = z | ( i σ, δ zi 1) ( i [m 1], zσi zσi+1 δ) ( i [M]\σ, zi = 0) . Note that the margin δ is an implicit variable in the annotation process, which is not explicitly indicated by the annotation results. Therefore, the range of internal label description vector (e.g. Sδ σ and Sδ l ) is also implicit, so the range determined by the implicit interval δ is called the implicit range. Although δ is implicit, we can predefine an explicit margin ˆδ; the range determined by the explicit margin (e.g. S ˆδ σ and S ˆδ l ) is called the explicit range. We are then able to generate a label description vector from the explicit range to approximate the internal label description vector. 2.2 Theoretical results We first define EAE to measure the quality of a certain annotation form w.r.t. approximating the internal label description vector. Definition 1 Suppose that an instance is annotated with an annotation r; δ and ˆδ are implicit and explicit margins, respectively; Sδ r and S ˆδ r are the implicit and explicit ranges, respectively. Then the expected approximation error of r to the internal label description vector is ϵδ,ˆδ r = Z V ˆδ r V δ r z ˆz 2 2dzdˆz, V δ r = Z z Sδr dz, V ˆδ r = Z z S ˆδr dˆz. (1) Eq. (1) is essentially derived from the expectation of the squared Euclidean distance between z and ˆz, i.e., Ez,ˆz z ˆz 2 2 , which measures the average distance between the estimated label description vector and the internal one given the ranges Sδ r and S ˆδ r. Since z and ˆz are independent, p(z, ˆz) = p(z)p(ˆz). Besides, we do not consider additional assumptions to reduce the uncertainty of z and ˆz; we assume that z and ˆz follow the uniform distributions on Sδ r and S ˆδ r, i.e., p(z)p(ˆz) = (V δ r V ˆδ r ) 1, and derive the Eq. (1). Next, we give the EAE of multi-label ranking. Theorem 1 If an instance is annotated by a multi-label ranking σ, m is the number of relevant labels, δ and ˆδ are the implicit and explicit margins, respectively, then the EAE of σ is ϵδ,ˆδ σ = m 6(m + 1) (m + 1)2(δ2 + ˆδ2) 2m(δ + ˆδ) (4m + 2)δˆδ + 2 . (2) 4Note that it is hard for an expert to rank a set of labels with close description degrees. We hence introduce the minimum margin δ to represent the smallest difference of description degrees required for the expert to distinguish and rank these labels. 5 10 15 0.00 upper bound mean value (a) Upper bound and mean value of ϵδ,ˆδ σ /m. (b) Lower bound of ϵδ,ˆδ l ϵδ,ˆδ σ . Figure 1: Visualization of corollaries. The horizontal coordinates of the two figures above indicate the number of relevant labels m. Theorem 1 can be proved by mathematical induction, and the details can be found in the appendix. Before giving more corollaries, we need to specify the range of the margins, i.e., δ and ˆδ. Lemma 1 If an instance is annotated by a multi-label ranking σ, then the margins δ and ˆδ satisfy that 0 δ m 1 and 0 ˆδ m 1. Next we give some interesting corollaries to understand Theorem 1. Corollary 1 If an instance is annotated by a multi-label ranking σ, m is the number of relevant labels, the explicit margin ˆδ minimizing the EAE of σ is ˆδ = ((2m + 1)δ + m)(m + 1) 2. It is clear that the optimal explicit margin ˆδ depends on the implicit margin δ; hence we cannot obtain an exact optimum for ˆδ . Nevertheless, Corollary 1 helps us to narrow down the range of the optimal explicit margin considerably, i.e., m(m + 1) 2 ˆδ m 1. Corollary 2 If an instance is annotated by a multi-label ranking σ, m is the number of relevant labels, 0 δ m 1, m(m + 1) 2 ˆδ m 1, then the EAE of σ is bounded by: 0 ϵδ,ˆδ σ m(m2 + 4m + 2) 6(m + 1)3 < 1 Corollary 2 gives m-dependent bounds on the EAE of the multi-label ranking. See the appendix for details of the proof. Corollary 3 If an instance is annotated by a multi-label ranking σ, m is the number of relevant labels, δ and ˆδ are uniform over 0, m 1 and m(m + 1) 2, m 1 , respectively, then we have: h ϵδ,ˆδ σ i = 2m4 + 8m3 + 8m2 + 4m + 1 36m(m + 1)3 . (4) Corollary 3 can be obtained by a simple integral calculation, as detailed in the appendix. In Fig. 1(a), we visualize the expected value in Corollary 3 and the upper bound in Corollary 2. It is obvious that we can obtain a significant performance gain by ranking a small number of labels. Theorem 2 If an instance is annotated by a logical label vector l, m is the number of relevant labels, δ and ˆδ are the implicit and explicit margins, respectively, then the EAE of l is ϵδ,ˆδ l = m 6 (2δ2 + 2ˆδ2 δ ˆδ 3δˆδ + 1). (5) Theorem 2 gives the EAE of the logical label vector, and the proof is detailed in the appendix. Corollary 4 Suppose that ϵδ,ˆδl l and ϵδ,ˆδσ σ are the EAE of the logical label vector l and the EAE of the multi-label ranking σ, respectively, we have the following inequality holds for m 3: ϵδ,ˆδl l ϵδ,ˆδσ σ 7m 48 (δ2 2δ) + m(m 1)(7m2 + 20m + 9) > 7m5 m4 46m3 30m2 + 7m + 7 48m(m + 1)3 > 0. (6) Corollary 4 shows the advantage of the multi-label ranking over the logical labels w.r.t. approximating the true label description vector. It is obvious that as the number of relevant labels m increases, ϵδ,ˆδl l ϵδ,ˆδσ σ increases at least at the rate of O(m), which is visualized in Fig. 1(b). 3 Algorithms In this section, we consider how to train a model on the dataset {(xn, σn)}N n=1 for predicting label distributions. We propose a framework DRAM to deal with this problem. Besides, we also design a comparison method. 3.1 DRAM framework First, we describe how to enhance multi-label rankings into label distributions. Then, we formally give the predictive model. Finally, we derive a generic EM algorithm for our framework that works for any instantiation of the basic models and consider a concrete instantiation. 3.1.1 Recovering label distributions from multi-label rankings In order to learn a mapping from instance features to label distributions, we consider enhancing multi-label rankings to label distributions. The process of enhancing the simple label (e.g., multi-label ranking and logical label) into the label distribution can be viewed as selecting label distributions that satisfy a predefined prior assumption from those consistent with simple labels. For example, some algorithms [9, 34] select those label distributions that satisfy the smoothness assumption [24]; In fact, according to the no-free-lunch axiom, no prior assumption can work for all tasks; hence we do not consider a concrete assumption in our framework. We provide a semi-adaptive scoring function φ(d; θ) such that any assumption can be easily encoded. The scoring function allows us to build a prior distribution of label distribution p (d): p (d) = 1 Zp φ (d; θ) Z 0 I(td S ˆδ σ)dt, (7) where S ˆδ σ is the explicit range of σ, Zp is a normalization constant, and I( ) denotes the indicator function. The integral term in Eq. (7) intuitively indicates how many label description vectors can be normalized to d. According to Corollary 1, we set ˆδ corresponding to the example (xn, σn) as |σn|(|σn| + 1) 2. The functional form of φ(d; θ) is predefined and the parameters can be learned adaptively. For example, we can predefine φ(d; θ) as a Gaussian likelihood function, and leave its mean and variance to be learned. 3.1.2 Predictive model: conditional Dirichlet mixtures Here we need to determine the distribution form of d conditioned on x. We use Dirichlet distribution to model d|x. Since φ is any non-negative real-valued function, the prior distribution of d is usually multimodal. Therefore, we model p(d|x) with the mixture of Dirichlet distributions: k=1 fk(x; α)Dir(d|f(x; βk)), (8) where α, β1, , βK are learnable parameters; f(x; α) outputs a K-dimensional positive realvalued vector with a sum of 1, and fk(x; α) is its k-th value; f(x; βk) outputs a M-dimensional positive real-valued vector; Dir( ) denotes Dirichlet distribution whose details are in the appendix. Algorithm 1 Generic DRAM Require: training set {(xn, σn)}N n=1, testing instance x , score function φ, number of mixture components K, number of Monte Carlo samples L; 1: α, θ, β1, , βK Initialize model parameters; 2: while the likelihood is not converged do 3: for n = 1, 2, , N; i = 1, 2, , L do 4: ˆδn |σn|(|σn| + 1) 2; 5: z(i) n Generate a sample uniformly from S ˆδn σn; 6: q(c(i) n ) Infer the posterior of latent variable c(i) n as in Eq. (9); 7: α, θ, β1, , βK Update the model parameters as in Eq. (10); 8: d Predict the label distribution for instance x as in Eq. (11); 9: return the label distribution d for instance x ; 3.1.3 Learning algorithm We consider minimizing the cross-entropy of p(d|x) relative to p (d), i.e., maximizing Ep (d) [ln p(d|x)]. Since p (d) is usually a complex distribution and Ep (d) [ln p(d|x)] involves the integration of p (d), it is often intractable. Therefore, we approximate it by the importance sampling method (whose detailed derivation can be found in the appendix): E p (d) [ln p(d|x)] φ(d(i); θ) ln p(d(i)|x) PL j=1 φ(d(j); θ) , d(i) = 1 Z(i) z(i), z(i) Uni(z|S ˆδ σ), where z(i) Uni(z|S ˆδ σ) denotes that z(i) is sampled uniformly from S ˆδ σ, and Z(i) equals to the sum of all elements in z(i). Since our model contains discrete latent variables, we use the EM algorithm [8] to train the model. We introduce the variational distribution q(c(i) n ) and obtain ln p(d(i) n |xn) = E q(c(i) n ) ln p(d(i) n |c(i) n , xn)p(c(i) n |xn) | {z } ELBO (Evidence Lower Bound) + KL q(c(i) n ) p(c(i) n |d(i) n , xn) | {z } c-posterior error We then alternate between M-step (maximizing ELBO) and E-step (minimizing the c-posterior error). E-step: Infer the posterior of latent variables to minimize the c-posterior error: γ(i) nk q(c(i) n = k) = fk(xn; α)Dir(d(i) n |f(xn; βk)) PK j=1 fj(xn; α)Dir(d(i) n |f(xn; βj)) , k [K]. (9) M-step: Update the model parameters to maximize the ELBO: arg max α,θ,β1, ,βK φ(d(i) n ; θ) PL j=1 φ(d(j) n ; θ) k=1 γ(i) nk ln fk(xn; α)Dir(d(i) n |f(xn; βk)) Once the model parameters are learned, we can predict the label distribution d for the test instance x according to Eq. (8). To evaluate DRAM, we take the expectation of the label distribution based on p(d|x ) as a deterministic output: 1 Zk fk(x; α)f(x , βk), Zk = i=1 fi(x , βk). (11) The overall learning process is shown in the Algorithm 1. 3.1.4 Instantiation: DRAM with linear learner and noninformative scoring function Here we consider a concrete instantiation for our framework. For simplicity, we use the noninformative scoring function, i.e., φ(d; θ) = 1. f(x; α) and f(x; βk) are modelled as linear models with the softmax and softplus activation functions, respectively. To avoid over-fitting, we regularize the parameters {βk}K k=1 by L2 norm. Then Eq. (10) can be rewritten as: arg min α,β1, ,βK λN k=1 γ(i) nk ln fk(xn; α)Dir(d(i) n |f(xn; βk)) , fk(xn; α) = exp(α k x n) PK j=1 exp(α j x n) , fj(xn; βk) = ln(1 + exp(βk j x n)), x n = [xn; 1], where αk is the k-th parameter vector in α, βk j is the j-th parameter vector in βk, and λ is a trade-off parameter. We use L-BFGS [17] to optimize Eq. (12). We denote this instantiation as DRAM-LN. 3.2 Comparison method: data transformation Since there is no existing method that can directly predict label distribution from multi-label ranking, we propose a comparison method called DT (dataset transformation). The main idea of DT is to transform multi-label rankings into logical labels. Then, we can recover label distributions by any existing LE algorithm and learn a label distribution predictor by any existing LDL algorithm. Specifically, the multi-label ranking dataset {(xn, σn)}N n=1 will be transformed into SN n=1 xn, P j σn,:i vj, |σn| 1 |σn| i=1 , where σn,:i is the set of the 1st, 2nd, , and i-th elements in σn vector, vj is an |σn|-dimensional one-hot vector with the j-th element being 1, and the instance weight |σn| 1 is to avoid that the model learning favours instances with more relevant labels. 4 Related work Our research is mainly related to LE [34] and LDL [4]. LE is a process of recovering label distributions from logical labels in the dataset. Most of the existing LE algorithms follow some basic assumptions. For example, some LE algorithms [9, 15, 22, 34] assume that instances with similar features have similar label distributions; some LE algorithms [12, 32] assume that semantically related labels also have close values in the label distribution; some LE algorithms [33, 16] assume that the label distribution is the low-dimensional representation of feature and logical label vectors. LDL is the learning process on the instances annotated by label distributions. The LDL problem can be addressed in two ways. The first is to directly design algorithms that match the prerequisites of the LDL problem. Some prominent examples include the LDL algorithms [11, 19, 20, 25, 37] that mine label correlation and the LDL algorithms that maintain label ranking [13] or classification accuracy [26, 27, 28]. Another way is to extend existing learning algorithms. For example, LDSVR [5] fits each component of the label distribution by a support vector machine; LDLLogit Boost [31] extends the boosting method by additive weighted regressors; LDLF [21] designs a normalization layer to model the multi-modal distribution by extending differentiable decision trees. 5 Experiments 5.1 Datasets and evaluation measures We adopt several widely used label distribution datasets, including Movie [4], Emotion6 [18], Twitter LDL, and Flickr-LDL [36].5 We manually reduce the label distributions in these datasets to multi-label rankings, train the model on these multi-label rankings, and then evaluate the model using the original label distributions in these datasets.6 In addition, we create a dataset called NSRD (Natural Scene with multi-label Rankings and label Distributions). The instances and label sets in NSRD are the same as in the Natural-Scene [7]. Three experts are requested to annotate the instances with multilabel rankings and label distributions. Then we can directly train and evaluate the model using the multi-label rankings and label distributions, respectively. Details of these datasets can be found in the 5Although there are many label distribution datasets, we only adopt those whose label distributions are generated by expert annotation rather than algorithms or experimental instruments. 6These datasets contain some label distributions with identical values, e.g., [0.3, 0.3, 0.4, 0] , which does not satisfy the prerequisites of this paper, so we remove them and their corresponding feature vectors beforehand. Table 1: Experimental results ((rank) mean std t-test) evaluated by four measures. Dataset Method Cheb ( ) Canber ( ) Cosine ( ) Rho ( ) DRAM-LN (2) 0.124 0.001 (2) 1.058 0.008 (2) 0.932 0.001 (1) 0.720 0.006 DT+VI+SA (6) 0.163 0.001 (6) 1.337 0.007 (6) 0.888 0.002 (5) 0.685 0.020 DT+VI+DM (7) 0.166 0.003 (7) 1.355 0.017 (7) 0.884 0.004 (4) 0.712 0.006 DT+GL+SA (4) 0.143 0.001 (5) 1.207 0.003 (4) 0.912 0.001 (6) 0.669 0.010 DT+GL+DM (3) 0.142 0.001 (4) 1.204 0.004 (3) 0.915 0.001 (2) 0.717 0.006 GT+SA (5) 0.144 0.003 (3) 1.201 0.021 (5) 0.903 0.003 (7) 0.625 0.010 GT+DM (1) 0.114 0.001 (1) 0.990 0.007 (1) 0.936 0.001 (3) 0.715 0.006 DRAM-LN (1) 0.282 0.004 (1) 3.953 0.050 (1) 0.785 0.006 (1) 0.588 0.009 DT+VI+SA (2) 0.316 0.007 (3) 3.984 0.073 (3) 0.743 0.015 (3) 0.506 0.035 DT+VI+DM (3) 0.319 0.006 (2) 3.966 0.065 (2) 0.748 0.012 (2) 0.564 0.051 DT+GL+SA (5) 0.336 0.007 (5) 4.121 0.041 (5) 0.685 0.014 (7) 0.319 0.041 DT+GL+DM (4) 0.332 0.007 (4) 4.047 0.037 (4) 0.708 0.010 (6) 0.386 0.035 GT+SA (7) 0.567 0.011 (7) 5.789 0.045 (7) 0.494 0.016 (5) 0.406 0.034 GT+DM (6) 0.380 0.009 (6) 4.769 0.043 (6) 0.643 0.014 (4) 0.473 0.029 Twitter-LDL DRAM-LN (1) 0.355 0.009 (1) 6.526 0.018 (1) 0.828 0.009 (1) 0.604 0.005 DT+VI+SA (6) 0.546 0.003 (4) 6.604 0.020 (6) 0.621 0.005 (7) 0.506 0.019 DT+VI+DM (7) 0.580 0.010 (6) 6.638 0.023 (7) 0.549 0.027 (4) 0.559 0.011 DT+GL+SA (4) 0.520 0.003 (2) 6.545 0.018 (3) 0.682 0.001 (3) 0.578 0.005 DT+GL+DM (5) 0.534 0.003 (3) 6.559 0.018 (4) 0.656 0.001 (2) 0.592 0.006 GT+SA (3) 0.436 0.017 (7) 6.937 0.026 (5) 0.653 0.020 (6) 0.516 0.009 GT+DM (2) 0.372 0.004 (5) 6.626 0.017 (2) 0.763 0.006 (5) 0.554 0.006 DRAM-LN (1) 0.324 0.005 (1) 6.013 0.017 (1) 0.815 0.004 (1) 0.627 0.006 DT+VI+SA (6) 0.456 0.005 (4) 6.116 0.021 (5) 0.657 0.006 (6) 0.542 0.021 DT+VI+DM (7) 0.472 0.011 (5) 6.146 0.040 (7) 0.629 0.021 (2) 0.592 0.010 DT+GL+SA (4) 0.440 0.005 (2) 6.076 0.021 (3) 0.690 0.003 (4) 0.573 0.005 DT+GL+DM (5) 0.450 0.005 (3) 6.090 0.020 (4) 0.674 0.003 (3) 0.592 0.006 GT+SA (3) 0.439 0.008 (7) 6.730 0.020 (6) 0.634 0.010 (7) 0.524 0.007 GT+DM (2) 0.363 0.004 (6) 6.360 0.014 (2) 0.720 0.005 (5) 0.554 0.005 DRAM-LN (3) 0.509 0.006 (1) 7.649 0.017 (3) 0.599 0.009 (2) 0.459 0.013 DT+VI+SA (5) 0.576 0.008 (7) 7.835 0.028 (6) 0.462 0.013 (7) 0.187 0.017 DT+VI+DM (7) 0.595 0.007 (6) 7.813 0.029 (7) 0.459 0.008 (6) 0.240 0.034 DT+GL+SA (4) 0.574 0.006 (4) 7.741 0.028 (4) 0.523 0.004 (4) 0.442 0.012 DT+GL+DM (6) 0.579 0.006 (5) 7.755 0.027 (5) 0.509 0.007 (1) 0.462 0.013 GT+SA (1) 0.468 0.008 (3) 7.700 0.021 (1) 0.610 0.012 (3) 0.447 0.008 GT+DM (2) 0.488 0.010 (2) 7.659 0.035 (2) 0.604 0.014 (5) 0.438 0.013 DRAM-LN (3) 0.509 0.006 (1) 7.649 0.017 (3) 0.599 0.009 (2) 0.459 0.013 DT+VI+SA (5) 0.570 0.006 (7) 7.811 0.024 (6) 0.469 0.018 (7) 0.198 0.028 DT+VI+DM (7) 0.588 0.009 (6) 7.793 0.030 (7) 0.465 0.012 (6) 0.251 0.026 DT+GL+SA (4) 0.568 0.005 (4) 7.725 0.025 (4) 0.525 0.004 (4) 0.444 0.016 DT+GL+DM (6) 0.574 0.006 (5) 7.739 0.028 (5) 0.511 0.008 (1) 0.462 0.014 GT+SA (1) 0.461 0.011 (3) 7.680 0.028 (1) 0.617 0.018 (3) 0.450 0.012 GT+DM (2) 0.485 0.012 (2) 7.664 0.035 (2) 0.605 0.014 (5) 0.438 0.014 DRAM-LN (3) 0.554 0.011 (1) 7.699 0.023 (3) 0.577 0.013 (2) 0.455 0.012 DT+VI+SA (4) 0.615 0.009 (7) 7.845 0.031 (6) 0.456 0.018 (7) 0.204 0.023 DT+VI+DM (7) 0.638 0.005 (6) 7.817 0.013 (7) 0.446 0.013 (6) 0.226 0.049 DT+GL+SA (5) 0.619 0.004 (4) 7.760 0.021 (4) 0.504 0.004 (5) 0.437 0.015 DT+GL+DM (6) 0.624 0.005 (5) 7.771 0.019 (5) 0.493 0.010 (1) 0.455 0.016 GT+SA (1) 0.490 0.012 (3) 7.748 0.028 (1) 0.601 0.019 (3) 0.443 0.011 GT+DM (2) 0.520 0.009 (2) 7.701 0.030 (2) 0.599 0.016 (4) 0.440 0.012 appendix. We used the six distance-based measures suggested in the paper [4] and a ranking-based measure suggested in the paper [13] to evaluate the performance of the model, which are Cheb (Chebyshev distance), Clark (Clark distance), Canber (Canberra distance), KL (Kullback-Leibler divergence), Cosine (cosine coefficient), Intersec (intersection similarity), and Rho (Spearman s rho coefficient). Due to page limitations, we only show the results on Cheb, Canber, Cosine, and Rho. Results on other measures are similar. 5.2 Comparison methods On the one hand, we compare DRAM with the baseline method DT proposed in Section 3.2. GL (Graph Laplacian LE) [34] and SA (specialized LDL algorithm with BFGS optimizer) [4] are the classical LE and LDL algorithms respectively. VI (LE with variational inference) [33] and DM (LDL with label distribution manifold) [25] are the state-of-the-art LE and LDL algorithms, respectively. We combine them in pairs to construct four comparison methods, i.e., DT+GL+SA, DT+GL+DM, DT+VI+SA, and DT+VI+DM. The hyperparameter configuration of GL, VI and DM Table 2: Average ranks of methods. Method Cheb Canber Cosine Rho DRAM-LN 2.00 1.14 2.00 1.43 DT+GL+DM 5.00 4.14 4.29 2.29 DT+GL+SA 4.29 3.71 3.86 4.71 DT+VI+DM 6.43 5.43 6.29 4.29 DT+VI+SA 4.86 5.43 5.43 6.00 GT+DM 2.43 3.43 2.43 4.43 GT+SA 3.00 4.71 3.71 4.86 0.56 Cheb ( ) K (number of mixture components) K (number of mixture components) 0.46 Rho ( ) Figure 2: Performance with varying K on NSRD. follows their respective literature. For our method, we set K = 3 and L = 20, and λ is selected from {10 5, 5 10 5, 10 4, 5 10 4, , 101, 5 101} by five-fold cross-validation. For the above comparison methods, since the label distributions are unavailable during training, the hyperparameter configuration that gives the highest Rho on the validation set will be used. On the other hand, we directly train DM and SA on the ground-truth label distributions for comparison. We refer to these two as GT+DM and GT+SA for short, respectively. For these two comparison methods, the hyperparameter configuration that gives the best Cheb, Canber, Cosine, and Rho on the validation set will be used. Each method is run for ten times on random dataset partitions (70% for training and 30% for test); the average values and standard derivations are recorded. 5.3 Results and discussions Table 1 shows the four performance measures of each method on four reduced datasets and NSRD dataset. Since the NSRD is annotated by three experts, we can obtain three corresponding datasets, denoted by the -e1 , -e2 and -e3 suffixes, respectively; these three datasets have the same feature vectors and different annotation values. Each experimental result is formatted as (rank) mean std t-test ; (rank) denotes the rank of each method among the seven comparison methods; / indicates whether DRAM-LN is statistically superior/inferior to the corresponding methods (pairwise two-tailed t-test at 0.05 significance level); if neither nor is shown, it means that there is no significant difference between the corresponding method and DRAM-LN; denotes the lower the better , and denotes the higher the better . Table 2 shows the average rank of each comparison method on each measure. Overall, our method achieves significant advantages. On the Emotion6, Twitter-LDL, and Flickr-LDL datasets, our method significantly outperforms almost any comparison methods. On the NSRD and Movie datasets, our method is only inferior to the GT-based methods on Cheb and Cosine measures. It is worth noting that DRAM-LN and DT-based methods outperform GT-based methods in many cases, such as the performance on Emotion6, Twitter-LDL, and Flickr-LDL datasets. We believe this is because some datasets are difficult to annotate; thus, the label distributions given by experts are noisy; then, fitting such label distributions exactly may lead to overfitting. This argument can be further supported by the fact that GT-based methods outperform our method on the NSRD dataset (where the label distributions are carefully annotated and less noisy). Figure 2 shows how the number of mixture components K affects the performance of our method on NSRD dataset. To save space, we show the average performance on NSRD-e1, NSRD-e2 and NSRD-e3 rather than showing them separately. It is obvious that the mixture model (K > 1) always outperform the single model (K = 1). In addition, it can be seen that appropriately increasing the Dirichlet components in the mixture can improve the model capacity and thus improve the predictive performance, but too many Dirichlet components may lead to overfitting and thus degrade the predictive performance. 6 Limitations and conclusion Limitations. 1) EAE is defined for the label description vector and does not directly reflect the approximation error to the label distribution; this limitation arises because normalizing the label description vector to a label distribution will lead to an extremely complex closed form of EAE. We do not believe that this limitation have a significant impact on our main results since the approximation error to the label distribution does not exceed that to the label description vector. For example, if the true and estimated label description vectors are z and tz (where t is a scaler), respectively, they will produce non-zero errors w.r.t. the label description vector; but z and tz are the same after normalization, i.e., they do not produce errors w.r.t. the label distribution. 2) If several labels actually describe the instance to the same degree, then requiring experts to give the strict order of these labels may lead to errors and invalidate Theorem 1; we plan to extend the theoretical results to this case in the future. Fortunately, DRAM framework can suit this case by a minor modification on the explicit range, i.e., allowing the description degree of these labels to be identical in line 5 of Algorithm 1. Conclusion. We derive some theorems and corollaries to reveal the relation between multi-label ranking and label distribution, and propose a generic framework, DRAM, for predicting label distribution from multi-label ranking. DRAM is cost-effective: It is trained on the examples with multi-label rankings and achieves performance comparable to that of LDL methods which require expensive label distribution annotations; DRAM is flexible: It allows users to easily encode their prior knowledge by a scoring function; DRAM is end-to-end: It integrates the processes of recovering and learning label distributions into one learning criterion, rather than performing them separately. Experimental results show the superiority of our proposal. 7 Acknowledgments This work was partially supported by the National Key Research and Development Program of China under Grant 2019YFB1706900, the National Natural Science Foundation of China (62176123, U20B2064), and the Fundamental Research Funds for the Central Universities (30920021131). [1] Klaus Brinker, Johannes Fürnkranz, and Eyke Hüllermeier. A unified model for multilabel classification and ranking. In European Conference on Artificial Intelligence, pages 489 493, 2006. [2] Klaus Brinker and Eyke Hüllermeier. Case-based multilabel ranking. In International Joint Conference on Artificial Intelligence, pages 702 707, 2007. [3] Binbin Gao, Hongyu Zhou, Jianxin Wu, and Xin Geng. Age estimation using expectation of label distribution learning. In International Joint Conference on Artificial Intelligence, pages 712 718, 2018. [4] Xin Geng. Label distribution learning. IEEE Transactions on Knowledge and Data Engineering, 28(7):1734 1748, 2016. [5] Xin Geng and Peng Hou. Pre-release prediction of crowd opinion on movies by label distribution learning. In International Joint Conference on Artificial Intelligence, pages 3511 3517, 2015. [6] Xin Geng, Kate Smith-Miles, and Zhi-Hua Zhou. Facial age estimation by learning from label distributions. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35:2401 2412, 2013. [7] Xin Geng, Renyi Zheng, Jiaqi Lv, and Yu Zhang. Multilabel ranking with inconsistent rankers. IEEE Transactions on Pattern Analysis and Machine Intelligence, pages 1 1, 2021. [8] Maya R. Gupta and Yihua Chen. Theory and use of the em algorithm. Foundations and Trends in Signal Processing, 4(3):223 296, 2011. [9] Peng Hou, Xin Geng, and Minling Zhang. Multi-label manifold learning. In AAAI Conference on Artificial Intelligence, pages 1680 1686, 2016. [10] Xiuyi Jia, Weiwei Li, Junyu Liu, and Yu Zhang. Label distribution learning by exploiting label correlations. In AAAI Conference on Artificial Intelligence, pages 3310 3317, 2018. [11] Xiuyi Jia, Weiwei Li, Junyu Liu, and Yu Zhang. Label distribution learning by exploiting label correlations. In AAAI Conference on Artificial Intelligence, pages 3310 3317, 2018. [12] Xiuyi Jia, Yunan Lu, and Fangwen Zhang. Label enhancement by maintaining positive and negative label relation. IEEE Transactions on Knowledge and Data Engineering, 2021. [13] Xiuyi Jia, Xiaoxia Shen, Weiwei Li, Yunan Lu, and Jihua Zhu. Label distribution learning by maintaining label ranking relation. IEEE Transactions on Knowledge and Data Engineering, pages 1 1, 2021. [14] Xiuyi Jia, Xiang Zheng, Weiwei Li, Changqing Zhang, and Zechao Li. Facial emotion distribution learning by exploiting low-rank label correlations locally. In IEEE Conference on Computer Vision and Pattern Recognition, pages 9833 9842, 2019. [15] Yukun Li, Minling Zhang, and Xin Geng. Leveraging implicit relative labeling-importance information for effective multi-label learning. In IEEE International Conference on Data Mining, pages 251 260, 2015. [16] Xinyuan Liu, Jihua Zhu, Qinghai Zheng, Zhongyu Li, Ruixin Liu, and Jun Wang. Bidirectional loss function for label enhancement and distribution learning. Knowledge-Based System, 213:106690, 2021. [17] Jorge Nocedal and Stephen Wright. Numerical Optimization. Springer, 2006. [18] Kuan-Chuan Peng, Tsuhan Chen, Amir Sadovnik, and Andrew Gallagher. A mixed bag of emotions: Model, predict, and transfer emotion distributions. In IEEE Conference on Computer Vision and Pattern Recognition, pages 860 868, 2015. [19] Tingting Ren, Xiuyi Jia, Weiwei Li, Lei Chen, and Zechao Li. Label distribution learning with label-specific features. In International Joint Conference on Artificial Intelligence, pages 3318 3324, 2019. [20] Tingting Ren, Xiuyi Jia, Weiwei Li, and Shu Zhao. Label distribution learning with label correlations via low-rank approximation. In International Joint Conference on Artificial Intelligence, pages 3325 3331, 2019. [21] Wei Shen, Kai Zhao, Yilu Guo, and Alan Yuille. Label distribution learning forests. In Advances in Neural Information Processing Systems, 2017. [22] Haoyu Tang, Jihua Zhu, Qinghai Zheng, Jun Wang, Shanmin Pang, and Zhongyu Li. Label enhancement with sample correlations via low-rank representation. In AAAI Conference on Artificial Intelligence, pages 5932 5939, 2020. [23] Grigorios Tsoumakas and Ioannis Katakis. Multi-label classification: An overview. International Journal of Data Warehousing and Mining, 3(3):1 13, 2006. [24] Jesper E. van Engelen and Holger H. Hoos. A survey on semi-supervised learning. Machine Learning, 109:373 440, 2020. [25] Jing Wang and Xin Geng. Label distribution learning by exploiting label distribution manifold. IEEE Transactions on Neural Networks and Learning Systems, pages 1 14, 2021. [26] Jing Wang and Xin Geng. Label distribution learning machine. In International Conference on Machine Learning, pages 10749 10759, 2021. [27] Jing Wang and Xin Geng. Learn the highest label and rest label description degrees. In International Joint Conference on Artificial Intelligence, pages 3097 3103, 2021. [28] Jing Wang, Xin Geng, and Hui Xue. Re-weighting large margin label distribution learning for classification. IEEE Transactions on Pattern Analysis and Machine Intelligence, pages 1 1, 2021. [29] Ke Wang, Ning Xu, Miaogen Ling, and Xin Geng. Fast label enhancement for label distribution learning. IEEE Transactions on Knowledge and Data Engineering, pages 1 1, 2021. [30] Xin Wen, Biying Li, Haiyun Guo, Zhiwei Liu, Guosheng Hu, Ming Tang, and Jinqiao Wang. Adaptive variance based label distribution learning for facial age estimation. In European Conference on Computer Vision, pages 379 395, 2020. [31] Chao Xing, Xin Geng, and Hui Xue. Logistic boosting regression for label distribution learning. In IEEE Conference on Computer Vision and Pattern Recognition, pages 4489 4497, 2016. [32] Ning Xu, Yun-Peng Liu, and Xin Geng. Label enhancement for label distribution learning. IEEE Transactions on Knowledge and Data Engineering, 33(4):1632 1643, 2021. [33] Ning Xu, Jun Shu, Yun-Peng Liu, and Xin Geng. Variational label enhancement. In International Conference on Machine Learning, pages 10597 10606, 2020. [34] Ning Xu, An Tao, and Xin Geng. Label enhancement for label distribution learning. In International Joint Conference on Artificial Intelligence, pages 2926 2932, 2018. [35] Jufeng Yang, Dongyu She, and Ming Sun. Joint image emotion classification and distribution learning via deep convolutional neural network. In International Joint Conference on Artificial Intelligence, pages 3266 3272, 2017. [36] Jufeng Yang, Ming Sun, and Xiaoxiao Sun. Learning visual sentiment distributions via augmented conditional probability neural network. In AAAI Conference on Artificial Intelligence, pages 224 230, 2017. [37] Peng Zhao and Zhi-Hua Zhou. Label distribution learning by optimal transport. In AAAI Conference on Artificial Intelligence, pages 4506 4513, 2018. [38] Xiang Zheng, Xiuyi Jia, and Weiwei Li. Label distribution learning by exploiting sample correlations locally. In AAAI Conference on Artificial Intelligence, pages 4556 4563, 2018. [39] Deyu Zhou, Xuan Zhang, Yin Zhou, Quan Zhao, and Xin Geng. Emotion distribution learning from texts. In Conference on Empirical Methods in Natural Language Processing, pages 638 647, 2016. [40] Wenfang Zhu, Xiuyi Jia, and Weiwei Li. Privileged label enhancement with multi-label learning. In International Joint Conference on Artificial Intelligence, pages 2376 2382, 2020. 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] (b) Did you describe the limitations of your work? [Yes] (c) Did you discuss any potential negative societal impacts of your work? [No] (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [Yes] (b) Did you include complete proofs of all theoretical results? [Yes] All proofs are detailed in the appendix. 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] The supplemental material provides a detailed instruction for reproducing the main results. (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [Yes] (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [No] 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [Yes] (b) Did you mention the license of the assets? [No] (c) Did you include any new assets either in the supplemental material or as a URL? [Yes] We include the new dataset we created in the supplemental material. (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [No] (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [Yes] The data we are using doesn t contain personally identifiable information or offensive content 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]