# instancedependent_partial_label_learning__549cb9b3.pdf Instance-Dependent Partial Label Learning Ning Xu, Congyu Qiao, Xin Geng , and Min-Ling Zhang School of Computer Science and Engineering, Southeast University, Nanjing 210096, China MOE Key Laboratory of Computer Network and Information Integration, Ministry of Education, China {xning, qiaocy, xgeng, zhangml}@seu.edu.cn Partial label learning (PLL) is a typical weakly supervised learning problem, where each training example is associated with a set of candidate labels among which only one is true. Most existing PLL approaches assume that the incorrect labels in each training example are randomly picked as the candidate labels. However, this assumption is not realistic since the candidate labels are always instancedependent. In this paper, we consider instance-dependent PLL and assume that each example is associated with a latent label distribution constituted by the real number of each label, representing the degree to each label describing the feature. The incorrect label with a high degree is more likely to be annotated as the candidate label. Therefore, the latent label distribution is the essential labeling information in partially labeled examples and worth being leveraged for predictive model training. Motivated by this consideration, we propose a novel PLL method that recovers the label distribution as a label enhancement (LE) process and trains the predictive model iteratively in every epoch. Specifically, we assume the true posterior density of the latent label distribution takes on the variational approximate Dirichlet density parameterized by an inference model. Then the evidence lower bound is deduced for optimizing the inference model and the label distributions generated from the variational posterior are utilized for training the predictive model. Experiments on benchmark and real-world datasets validate the effectiveness of the proposed method. Source code is available at https://github.com/palm-ml/valen. 1 Introduction Partial label learning (PLL) deals with the problem where each training example is associated with a set of candidate labels, among which only one label is valid [7, 5, 37]. Due to the difficulty in collecting exactly labeled data in many real-world scenarios, PLL leverages inexact supervision instead of exact labels. The need to learn from the inexact supervision leads to a wide range of applications for PLL techniques, such as web mining [25], multimedia content analysis [38, 4], ecoinformatics [24, 31], etc. To accomplish the task of learning from partial label data, many approaches have been proposed. Identification-based PLL approaches [17, 28, 24, 5, 37] regard the ground-truth label as a latent variable and try to identify it. Average-based approaches [15, 7, 40] treat all the candidate labels equally and average the modeling outputs as the prediction. For confidence-based approaches [8, 34, 42], the confidence of each label is estimated instead of identifying the ground-truth label. These approaches always adopt the randomly picked candidate labels to corrupt benchmark data into partially labeled version despite having no explicit generation process of candidate label sets. To depict the instance-independent generation process of candidate label sets, Feng [9] proposes a statistical model and deduces a risk-consistent method and a classifier-consistent method. Under the Corresponding author 35th Conference on Neural Information Processing Systems (Neur IPS 2021). 0 1 2 3 4 5 6 description degree (a) Handwritten digits in MNIST [22] Ship Bird Cat Airplane Deer Frog description degree Bird Airplane (b) Color image in CIFAR-10 [21] Figure 1: The examples about the latent label distributions for partial label learning. The candidate labels are in the box and the red one is valid. same generation process, another classifier-consistent risk estimator is proposed for deep model and stochastic optimizers [26]. The previous methods assume that the candidate labels are randomly sampled with the uniform generating procedure [26, 9], which is commonly adopted to corrupt benchmark datasets into partially labeled versions in their experiments. However, the candidate labels are always instance-dependent (feature-dependent) in practice as the incorrect labels related to the feature are more likely to be picked as candidate label set for each instance. These methods usually do not perform as well as expected due to the unrealistic assumption on the generating procedure of candidate label sets. In this paper, we consider instance-dependent PLL and assume that each instance in PLL is associated with a latent label distribution [33, 35, 11] constituted by the real number of each label, representing the degree to each label describing the feature. Then, the incorrect label with a high degree in the latent label distribution is more likely to be annotated as the candidate label. For example, the candidate label set of the handwritten digits in Figure 1(a) contains 1 , 3 and 5 , where 1 and 3 are not ground-truth but selected as candidate labels due to their high degrees in the latent label distribution of the instance. The object in Figure 1(b) is annotated with bird and airplane as the degrees of these two labels are much higher than others in the label distribution. The intrinsical ambiguity increases the difficulty of annotating, which leads to the result that annotators pick the candidate labels with high degrees in the latent label distribution of each instance instead of annotating the ground-truth label directly in PLL. Therefore, the latent label distribution is the essential labeling information in partially labeled examples and worth being leveraged for predictive model training. Motivated by the above consideration, we deal with the PLL problem from two aspects. First, we enhance the labeling information by recovering the latent label distribution for each training example as a label enhancement process [33, 35]. Second, we run label enhancement and train the predictive model with recovered label distributions iteratively. The proposed method named VALEN, i.e., VAriational Label ENhancement for instance-dependent partial label learning, uses the candidate labels to initialize the predictive model in the warm-up training stage, then recovers the latent label distributions via inferring the variational posterior density parameterized by an inference model with the deduced evidence lower bound, and trains the predictive model with a risk estimator by leveraging the candidate labels as well as the label distributions. Our contributions can be summarized as follows: We for the first time consider the instance-dependent PLL and assume that each partially labeled example is associated with a latent label distribution, which is the essential labeling information and worth being recovered for predictive model training. We infer the posterior density of the latent label distribution via taking on the approximate Dirichlet density parameterized by an inference model and deduce the evidence lower bound for optimization, in which the topological information and the features extracted from the predictive model are leveraged. We train predictive model with a proposed empirical risk estimator by leveraging the candidate labels as well as the label distributions. We iteratively recover the latent label distributions and train the predictive model in every epoch. After the network has been fully trained, the predictive model can perform predictions for future test examples alone. Experiments on the corrupted benchmark datasets and real-world PLL datasets validate the effectiveness of the proposed method. 2 Proposed Method First of all, we briefly introduce some necessary notations. Let X = Rq be the q-dimensional instance space and Y = {y1, y2, ..., yc} be the label space with c class labels. Given the PLL training set D = {(xi, Si)|1 i n} where xi denotes the q-dimensional instance and Si Y denotes the candidate label set associated with xi. Note that Si contains the correct label of xi and the task of PLL is to induce a multi-class classifier f : X 7 Y from D. For each PLL training example (xi, Si), we use the logical label vector li = [ly1 i , ly2 i , . . . , lyc i ] {0, 1}c to represent whether yj is the candidate label, i.e., lyj i = 1 if yj Si, otherwise lyj i = 0. The label distribution of xi is denoted by di = [dy1 i , dy2 i , . . . , dyc i ] [0, 1]c where Pc j=1 dyj i = 1. Then L = [l1, l2, . . . , ln] and D = [d1, d2, . . . , dn] represent the logical label matrix and label distribution matrix, respectively. 2.1 Overview To deal with PLL problem, we iteratively recover the latent label distribution for each example x and train the predictive model by leveraging the recovered label distribution. We start with a warm-up period, in which we train the predictive model with the PLL minimal loss [26]. This allows us to attain a reasonable predictive model before it starts fitting incorrect labels. After the warm-up period, the features extracted from the predictive model can help for recovering the latent label distribution. Benefited from the essential labeling information in the recovered label distribution, the performance of the predictive model could be further improved. VALEN implements label enhancement and classifier training iteratively in every epoch. In label enhancement, we assume the true posterior density of the latent label distribution takes on the variational approximate Dirichlet density parameterized by an inference model. Then the evidence lower bound is deduced for optimizing the inference model and the label distributions can be generated from the variational posterior. In classifier training, the predictive model is trained by leveraging the recovered label distributions and candidate labels with an empirical risk estimator. After the models has been fully trained, the predictive model can perform prediction for future test instances alone. 2.2 Warm-up Training The predictive model θ is trained on partially labeled examples by minimizing the following PLL minimal loss function [26]: i=1 min yj Si ℓ(f(xi), eyj), (1) where ℓis cross-entropy loss and e Y = {eyj : yj Y} denotes the standard canonical vector in Rc, i.e., the j-element in eyj equals 1 and others equal 0. Similar to [26], the min operator in Eq. (1) is replaced by using the current predictions for slightly weighting on the possible labels in warm-up training. Then we could extract the feature φ of each x via using the predictive model. 2.3 Label Enhancement We assume that the prior density p(d) is a Dirichlet with ˆα, i.e., p (d) = Dir (d | ˆα) where ˆα = [ε, ε, . . . , ε] is a c-dimensional vector with a minor value ε. Then we let the prior density p(D) be the product of each Dirichlet i=1 Dir(di| ˆα). (2) We consider the topological information of the feature space, which is represented by the affinity graph G = (V, E, A). Here, the feature vector φi of each example could be extracted from the predictive model θ in current epoch, V = {φi | 1 i n} corresponds to the vertex set consisting of feature vectors, E = {(φi, φj) | 1 i = j n} corresponds to the edge set, and a sparse adjacency matrix A = [aij]n n can be obtained by aij = 1 if φi N(φj) 0 otherwise , (3) where N(φj) is the set for k-nearest neighbors of φj and the diagonal elements of A are set to 1. Let features matrix Φ = [φ1, φ2, . . . , φn], adjacency matrix A and logical labels L be observed matrix, VALEN aims to infer the posterior density p(D|L, Φ, A). As the computation of the exact posterior density p(D|L, Φ, A) is intractable, a fixed-form density q(D|L, Φ, A) is employed to approximate the true posterior. We let the approximate posterior be the product of each Dirichlet parameterized by a vector αi = [α1 i , α2 i , . . . , αc i] : qw(D | L, Φ, A) = i=1 Dir (di|αi) . (4) Here, the parameters = [α1, α2, . . . , αn] are outputs of the inference model parameterized by w, which is defined as a two-layer GCN [20] by GCN(L, Φ, A) = A Re LU AZW0 W1, with Z = [Φ; L] and weight W0, W1. Here A = ˆA 1 2 is the symmetrically normalized weight matrix where ˆA is the degree matrix of A. By following the Variational Bayes techniques, a lower bound on the marginal likelihood of the model is derived which ensures that qw(D|L, Φ, A) is as close as possible to p(D|L, Φ, A). For logical label matrix L, feature matrix Φ, and the corresponding A, the log marginal probability can be decomposed as follows 1: log p(L, Φ, A) = LELBO + KL[qw(D|L, Φ, A)||p(D|L, Φ, A)]. (5) LELBO = Eqw(D|L,Φ,A)[log p(L, Φ, A|D)] KL[qw(D|L, Φ, A)||p(D)]. (6) Due to the non-negative property of KL divergence, the first term LELBO constitutes a lower bound of log p(L, Φ, A), which is often called as evidence lower bound (ELBO), i.e., log p(L, Φ, A) LELBO. According to Eq. (2) and Eq. (4) , the KL divergence in Eq. (6) can be analytically calculated as follows: KL (qw(D|L, Φ, A) p(D)) = Pn i=1 log Γ Pc j=1 αj i Pc j=1 log Γ αj i log Γ (c ε) + c log Γ (ε) + Pc j=1 αj i ε ψ αj i ψ Pc j=1 αj i . (7) where Γ( ) and ψ( ) are Gamma function and Digamma function, respectively. As the first part of Eq. (6) is intractable, we employ the implicit reparameterization trick [10] to approximate it by Monte Carlo (MC) estimation. Inspired by [20], we simply drop the dependence on Φ: p(L | A, D) = i=1 p (li | A, D) , j=1 p (aij | di, dj) , with p (aij = 1 | di, dj) = s d i dj . Here, s( ) is the logistic sigmoid function. We further assume that p (li|A, D) is a multivariate Bernoulli with probabilities τi. In order to simplify the observation model, T(m) = [τ (m) 1 , τ (m) 2 , . . . , τ (m) n ] is computed from m-th sampling D(m) with a three-layer MLP parameterized by η. Then the first part of Eq. (6) can be tractable: Eqw(D|L,Φ,A)[log pη(L, Φ, A|D)] = 1 tr (I L) log I T(m) + tr L log T(m) A S D(m)D(m) 2 F 1More detailed calculations can be seen in Appendix A.1. Algorithm 1 VALEN Algorithm Input: The PLL training set D = {(xi, Si)}n i=1, epoch T and iteration I; 1: Initialize the predictive model θ by warm-up training, the reference model w and observation model η; 2: Extract the features Φ from predictive model θ and calculate the adjacency matrix A; 3: for t = 1, . . . , T do 4: Shuffle training set D = {(xi, Si)}n i=1 into I mini-batches; 5: for k = 1, . . . , I do 6: Obtain label distribution di for each example xi by Eq. (4); 7: Update θ, w and η by forward computation and back-propagation by fusing Eq. (12) and Eq. (13); 8: end for 9: end for Output: The predictive model θ. Note that we can use only one MC sample in Eq. (9) during the training process as suggested in [19, 35]. In addition, VALEN improves the label enhancement by employing the compatibility loss, which enforces that the recovered label distributions should not be completely different from the confidence ζ(xi) [9, 26] estimated by current prediction f(xi; θ): j=1 ζj(xi) log dyj i (10) ζj(xi) = fj (xi; θ) / P yk Si fk (xi; θ) if yj Si 0 otherwise (11) Now we can easily get the objective of label enhancement LLE as follows: LLE = λLo LELBO (12) where λ is a hyper-parameter. The label distribution matrix D is sampled from q(D|L, Φ, A), i.e., di Dir(αi). Note that the implicit reparameterization gradient [10] is employed, which avoids the inversion of the standardization function, which makes the gradients can be computed analytically in backward pass. 2.4 Classifier Training To train the predictive model, we minimize the following empirical risk estimator by levering the recovered label distributions: b RV (f) = 1 yj Si dyj i ℓ(f(xi), eyj) Here we adopt the average value of di sampled by di Dir(αi). We can use any deep neural network as the predictive model, and then equip it with the VALEN framework to deal with PLL. Note that we could train the predictive model and update the label distributions in a principled end-to-end manner by fusing the objective Eq. (12) and Eq. (13). The algorithmic description of the VALEN is shown in Algorithm 1. Let bf V = minf F b RV (f) be the empirical risk minimizer and f = minf F RV (f) be the optimal risk minimizer where RV (f) is the risk estimator. Besides, we define the function space Hyj for the label yj Y as h : x 7 fyj(x) | f F . Let Rn Hyj be the expected Rademacher complexity [2] of Hyj with sample size n, then we have the following theorem. Theorem 1 Assume the loss function ℓ(f(x), eyj) is L-Lipschitz with respect to f(x)(0 < L < ) for all yj Y and upper-bounded by M, i.e., M = supx X,f F,yj Y ℓ(f(x), eyj). Then, for any δ > 0, with probability at least 1 δ, R bf V R (f ) 4 j=1 Rn Hyj + M The proof of Theorem 1 is provided in Appendix A.2. Theorem 1 shows that the empirical risk minimizer f V converges to the optimal risk minimizer f as n and Rn Hyj 0 for all parametric models with a bounded norm. 3 Related Work As shown in Section 1, supervision information conveyed by partially labeled training examples is implicit as the ground-truth label is hidden within the candidate label set. Therefore, partial label learning can be regarded as a weak supervision learning framework [18] with implicit labeling information. Intuitively, the basic strategy for handling partial label learning is disambiguation, i.e., trying to identify the ground-truth label from the candidate label set associated with each training example, where existing strategies include disambiguation by identification or disambiguation by averaging. For identification-based disambiguation, the ground-truth label is regarded as a latent variable and identified [17, 28, 24, 5, 37]. For averaging-based disambiguation, all the candidate labels are treated equally and the prediction is made by averaging their modeling outputs [15, 7, 40]. Most existing algorithms aim to fulfill the learning task by fitting widely-used learning techniques to partial label data. For maximum likelihood techniques, the likelihood of observing each partially labeled training example is defined over its candidate label set instead of the unknown ground-truth label [17, 24]. K-nearest neighbor techniques determine the class label of unseen instances via voting among the candidate labels of its neighboring examples [15, 40]. For maximum margin techniques, the classification margins over the partially labeled training examples are defined by discriminating modeling outputs from candidate labels and non-candidate labels [28, 37]. For boosting techniques, the weight over each partially labeled training example and the confidence over the candidate labels are updated in each boosting round [31]. For disambiguation-free strategies, the generalized description degree is estimated by using a graph Laplacian and induce a multi-output regression [34]. The confidence of each candidate label is estimated by using the manifold structure of feature space [42]. However, these methods just estimate the soft labeling information and train the predictive models in separate stages without considering the feedback of the predictive models. The above-mentioned works were solved in specific low-efficiency manners and incompatible with high-efficient stochastic optimization. To handle large-scale datasets, the deep networks are employed with an entropy-based regularizer to maximize the margin between the potentially correct label and the unlikely ones [36]. [26] proposes a classifier-consistent risk estimator and a progressive identification, which is compatible with deep models and stochastic optimizers. [9] proposes a statistical model to depict the generation process of candidate label sets, which deduces a risk-consistent method and a classifier-consistent method. The previous methods assume that the candidate labels are randomly sampled with the uniform generating procedure. However, the candidate labels are always instance-dependent (feature-dependent) in practice as the incorrect labels related to the feature are more likely to be picked as candidate label set for each instance. In this paper, we consider instance-dependent PLL and assume that each instance in PLL is associated with a latent label distribution [33, 35, 11] constituted by the real number of each label, representing the degree to each label describing the feature. Label enhancement (LE) [12, 33, 35] recovers the latent label distribution from the observed logical labels, in which the recovered label distribution is a kind of pseudo label [29, 23], actually. Note that the generation process of soft labels in label smoothing [30, 27] and distillation [14, 39] could also be regarded as a label enhancement process. Table 1: Classification accuracy (mean std) of each comparing approach on benchmark datasets corrupted by the instance-dependent generating procedure. MNIST Kuzushiji-MNIST Fashion-MNIST CIFAR-10 VALEN 97.85 0.05% 86.19 0.14% 86.17 0.19% 80.38 0.52% PRODEN 97.69 0.04% 85.71 0.12% 85.54 0.09% 79.80 0.28% RC 97.60 0.05% 84.86 0.11% 85.51 0.10% 79.46 0.25% CC 97.44 0.03% 82.67 1.82% 85.19 0.04% 78.98 0.60% D2CNN 94.63 0.16% 83.03 0.78% 82.42 0.21% 73.11 0.11% GA 95.25 0.07% 82.45 0.63% 80.41 0.24% 77.57 0.76% Table 2: Classification accuracy (mean std) of each comparing approach on benchmark datasets corrupted by the uniform generating procedure. MNIST Kuzushiji-MNIST Fashion-MNIST CIFAR-10 VALEN 97.93 0.05% 88.76 0.26% 88.98 0.16% 81.93 1.01% PRODEN 97.97 0.03% 88.55 0.10% 88.94 0.12% 81.53 0.53% RC 97.86 0.03% 86.65 0.10% 88.59 0.08% 81.30 1.30% CC 97.73 0.02% 87.99 0.03% 88.93 0.06% 80.17 1.09% D2CNN 95.12 0.16% 84.03 0.78% 80.42 0.21% 75.11 0.11% GA 96.29 0.19% 82.36 0.98% 81.81 0.99% 60.14 1.35% 4 Experiments 4.1 Datasets We adopt four widely used benchmark datasets including MNIST [22], Fashion-MNIST [32], Kuzushiji-MNIST [6], and CIFAR-10 [21], and five datasets from the UCI Machine Learning Repository [1], including Yeast, Texture, Dermatology, Synthetic Control, and 20Newgroups. We manually corrupt these datasets into partially labeled versions 2 by using a flipping probability ξyj i = P(lyj i = 1|ˆlyj i = 0, xi) , where ˆlyj i is the original clean label. To synthesize the instancedependent candidate labels, we set the flipping probability of each incorrect label corresponding to an example xi by using the confidence prediction of a clean neural network ˆθ (trained with the original clean labels) [43] with ξyj i = fj(xi; ˆθ) maxyj Yi fj(xi; ˆθ), where Yi is the incorrect label set of xi. The uniform corrupted version adopts the uniform generating procedure [26, 9] to flip the incorrect label into candidate label, where ξyj i = 1 In addition, five real-world PLL datasets are adopted, which are collected from several application domains including Lost [7], Soccer Player [38] and Yahoo!News [13] for automatic face naming from images or videos, MSRCv2 [24] for object classification, and Bird Song [3] for bird song classification. The detailed descriptions of these datasets are provided in Appendix A.3. We run 5 trials on the four benchmark datasets and perform five-fold cross-validation on UCI datasets and real-world PLL datasets. The mean accuracy as well as standard deviation are recorded for all comparing approaches. 4.2 Baselines The performance of VALEN is compared against five DNN based approaches: 1) PRODEN [26]: A progressive identification partial label learning approach which approximately minimizes a risk estimator and identifies the true labels in a seamless manner; 2) RC [9]: A risk-consistent partial label learning approach which employs the importance reweighting strategy to converges the true risk 2The datasets corrupted by the instance-dependent generating procedure are available at https://drive. google.com/drive/folders/1J_68Eq Or LN6t A56Rcy Tgcr1kom JB31Y1?usp=sharing. Table 3: Classification accuracy (mean std) of each comparing approach on UCI datasets corrupted by the instance-dependent generating procedure. Yeast Texture Synthetic Control Dermatology 20Newsgroup VALEN 57.57 1.08% 94.76 0.93% 82.86 0.76% 89.86 1.31% 81.88 0.47% PRODEN 54.78 1.28% 89.87 2.14% 71.16 6.19% 88.53 3.87% 78.06 0.74% RC 54.77 1.27% 89.57 2.37% 65.99 2.72% 88.53 3.49% 78.02 0.79% CC 54.98 0.91% 88.92 7.56% 66.99 2.47% 88.26 4.11% 77.88 0.39% D2CNN 44.94 1.87% 69.52 5.79% 62.66 8.92% 81.95 6.18% 73.55 0.92% GA 25.86 3.17% 74.84 2.87% 56.43 1.29% 84.85 1.43% 49.49 3.42% CLPL 54.92 2.38% 81.27 9.09% 66.33 3.25% 92.07 3.42% 77.62 0.23% PL-SVM 41.85 5.92% 39.03 4.35% 50.33 5.73% 84.98 4.56% 72.89 0.41% PL-KNN 47.44 2.69% 70.05 0.70% 80.50 1.26% 83.61 3.15% 33.28 1.09% IPAL 56.40 2.07% 93.49 0.89% 77.66 3.60% 78.94 8.34% 67.38 0.95% PLLE 55.53 1.74% 84.45 1.07% 66.16 7.96% 93.16 2.58% 75.54 0.66% Table 4: Classification accuracy (mean std) of each comparing approach on UCI datasets corrupted by the uniform generating procedure. Yeast Texture Synthetic Control Dermatology 20Newsgroup VALEN 58.18 1.46% 97.30 0.57% 97.17 0.47% 97.07 0.41% 71.75 3.02% PRODEN 56.32 1.98% 97.75 0.53% 95.83 1.95% 95.07 1.84% 68.28 0.91% RC 56.39 1.85% 97.77 0.55% 95.99 1.80% 95.62 1.51% 68.44 1.09% CC 56.25 1.89% 97.79 0.57% 96.33 1.39% 95.90 1.69% 67.95 0.95% D2CNN 54.04 1.90% 97.23 0.72% 81.16 8.11% 90.43 2.38% 65.88 2.56% GA 22.98 2.57% 95.09 1.07% 56.87 1.53% 51.95 3.89% 58.29 1.74% CLPL 56.54 3.35% 98.14 0.59% 94.66 6.41% 96.72 0.76% 70.45 0.91% PL-SVM 46.23 7.21% 39.74 2.11% 76.50 5.31% 92.37 5.08% 70.44 0.37% PL-KNN 44.40 2.50% 95.31 0.85% 95.33 2.98% 92.91 2.92% 27.10 0.49% IPAL 43.86 3.39% 98.71 0.37% 96.83 1.90% 95.35 2.08% 65.39 1.21% PLLE 53.58 2.86% 98.40 0.40% 89.66 1.91% 90.98 1.85% 53.88 0.59% minimizer; 3) CC [9]: A classifier-consistent partial label learning approach which uses a transition matrix to form an empirical risk estimator; 4) D2CNN [36]: A deep partial label learning approach which design an entropy-based regularizer to maximize the margin between the potentially correct label and the unlikely ones; 5) GA [16]: An unbiased risk estimator approach which can be applied for partial label learning. For all the DNN based approaches, we adopt the same predictive model for fair comparisons. Specifically, the 32-layer Res Net is trained on CIFAR-10 in which the learning rate, weight decay and mini-batch size are set to 0.05, 10 3 and 256, respectively. The three-layer MLP is trained on MNIST, Fashion-MNIST and Kuzushiji-MNIST where the learning rate, weight decay and mini-batch size are set to 10 2, 10 4 and 256, respectively. The linear model is trained on UCI and real-world PLL datasets where the learning rate, weight decay and mini-batch size are set to 10 2, 10 4 and 100, respectively. We implement the comparing methods with Py Torch. The number of epochs is set to 500, in which the first 10 epochs are warm-up training. In addition, we also compare with five classical partial label learning approaches, each configured with parameters suggested in respective literatures: 1) CLPL [7]: A convex partial label learning approach which uses averaging-based disambiguation; 2) PL-KNN [15]: An instance-based partial label learning approach which works by k-nearest neighbor weighted voting; 3) PL-SVM [28]: A maximum margin partial label learning approach which works by identification-based disambiguation; 4) IPAL [41]: A non-parametric method that applies the label propagation strategy to iteratively update the confidence of each candidate label; 5) PLLE [34]: A two-stage partial label learning approach which estimates the generalized description degree of each class label values via graph Laplacian and induces a multi-label predictive model with the generalized description degree in separate stages. Table 5: Classification accuracy (mean std) of each comparing approach on the real-world datasets. Lost MSRCv2 Bird Song Soccer Player Yahoo!News VALEN 70.28 2.29% 47.61 1.79% 72.02 0.37% 55.90 0.58% 67.52 0.19% PRODEN 68.62 4.86% 44.47 2.33% 71.68 0.83% 54.40 0.85% 67.12 0.97% RC 68.89 5.02% 44.59 2.65% 71.56 0.88% 54.23 0.89% 67.04 0.88% CC 62.21 1.77% 47.49 2.31% 68.42 0.99% 53.50 0.96% 61.92 0.96% D2CNN 68.56 6.68% 43.27 2.98% 65.48 2.57% 48.16 0.62% 52.46 1.71% GA 50.21 3.62% 30.91 4.31% 34.57 3.41% 50.65 0.94% 45.72 1.75% CLPL 74.15 3.03% 44.47 2.58% 65.76 1.19% 50.01 1.03% 53.25 1.12% PL-SVM 71.56 2.71% 38.25 3.89% 50.66 4.23% 36.39 1.03% 51.24 0.72% PL-KNN 33.87 2.48% 43.28 2.35% 64.34 0.75% 49.24 1.23% 40.38 0.37% IPAL 72.10 2.75% 52.96 1.36% 70.32 0.91% 54.41 0.68% 66.04 0.85% PLLE 72.55 3.55% 47.54 1.96% 70.63 1.24% 53.38 1.03% 59.45 0.43% 4.3 Experimental Results Table 1 reports the classification accuracy of each DNN-based method on benchmark datasets corrupted by the instance-dependent generating procedure. The best results are highlighted in bold. In addition, / indicates whether VALEN is statistically superior/inferior to the comparing approach on each dataset (pairwise t-test at 0.05 significance level). From the table, we can observe that VALEN always achieves the best performance and significantly outperforms other compared methods in most cases. In addition, we also validate the effectiveness of our approach on uniform corrupted versions that is commonly adopted in previous works. From Table 2, we can observe that VALEN achieves superior or at least comparable performance to other approaches on uniform corrupted versions. Table 3 and Table 4 report the classification accuracy of each method on UCI datasets corrupted by the instance-dependent generating procedure and the uniform generating procedure, respectively. VALEN always achieves the best performance and significantly outperforms other DNN-based methods in most cases on instance-dependent corrupted versions while achieves superior or at least comparable performance to other approaches on uniform corrupted versions. We further compare VALEN with five classical PLL methods that can hardly be implemented by DNNs on large-scale datasets. Despite the small scale of most UCI datasets, VALEN always achieve the best performance in most cases against the classical PLL methods as VALEN can deal with the high average number of candidate labels (can be seen in Appendix A.3) in the corrupted UCI datasets. Table 5 reports the experimental results on real-world PLL datasets. We can find that VALEN achieves best performance against other DNN-based methods on the real-world PLL datasets. Note that VALEN achieves best performance against classical methods on all datasets except Lost and MSRCv2 as these datasets are small-scale and the average number of candidate labels in each dataset is low (can be seen in Appendix A.3), which leads to the result that DNN-based methods cannot take full advantage. Figure 2(a) and Figure 2(b) illustrate the performance of VALEN on KMNIST corrupted by the instance-dependent generating procedure and the uniform generating procedure under different flipping probability, respectively. Besides, the performance of the ablation version that removes the label enhancement and trains the predictive model with PLL minimal loss (denoted by VALEN-NON) is recorded. These results clearly validate the usefulness of recovered label distributions for improving predictive performance. Figure 2(c) illustrates the recovered label distribution matrix over all training examples converges as the number of epoch (after warm-up training) on Kuzushiji-MNIST. We can see that the recovered label distributions converge fast with the increasing number of epoch. 5 Conclusion In this paper, the problem of partial label learning is studied where a novel approach VALEN is proposed. We for the first time consider the instance-dependent PLL and assume that each partially labeled example is associated with a latent label distribution, which is the essential labeling information and worth being recovered for predictive model training. VALEN recovers the latent label distribution via inferring the true posterior density of the latent label distribution by Dirichlet density 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 Average Flipping Probability Test Accuracy(%) VALEN VALEN-NON (a) Instance-dependent version 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 Average Flipping Probability Test Accuracy(%) VALEN VALEN-NON (b) Uniform version 0 100 200 300 400 500 Epoch D(t) D(t 1) F (c) Convergence curves of D Figure 2: Further analysis of VALEN on KMNIST. parameterized with an inference model and deduce the evidence lower bound for optimization. In addition, VALEN iteratively recovers latent label distributions and trains the predictive model in every epoch. The effectiveness of the proposed approach is validated via comprehensive experiments on both synthesis datasets and real-world PLL datasets. 6 Acknowledgments This research was supported by the National Key Research & Development Plan of China (No. 2018AAA0100104, No. 2018AAA0100100), the National Science Foundation of China (62125602, 62076063), China Postdoctoral Science Foundation (2021M700023), Jiangsu Province Science Foundation for Youths (BK20210220). [1] Arthur Asuncion and David Newman. UCI machine learning repository, 2007. [2] Peter L Bartlett and Shahar Mendelson. Rademacher and gaussian complexities: Risk bounds and structural results. Journal of Machine Learning Research, 3(Nov):463 482, 2002. [3] Forrest Briggs, Xiaoli Z Fern, and Raviv Raich. Rank-loss support instance machines for MIML instance annotation. In Proceedings of the 18th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pages 534 542, Beijing, China, 2012. [4] Ching-Hui Chen, Vishal M Patel, and Rama Chellappa. Learning from ambiguously labeled face images. IEEE Transactions on Pattern Analysis and Machine Intelligence, in press. [5] Yi-Chen Chen, Vishal M Patel, Rama Chellappa, and P Jonathon Phillips. Ambiguously labeled learning using dictionaries. IEEE Transactions on Information Forensics and Security, 9(12):2076 2088, 2014. [6] Tarin Clanuwat, Mikel Bober-Irizar, Asanobu Kitamoto, Alex Lamb, Kazuaki Yamamoto, and David Ha. Deep learning for classical japanese literature. ar Xiv preprint ar Xiv:1812.01718, 2018. [7] Timothee Cour, Ben Sapp, and Ben Taskar. Learning from partial labels. Journal of Machine Learning Research, 12(May):1501 1536, 2011. [8] Lei Feng and Bo An. Leveraging latent label distributions for partial label learning. In IJCAI, pages 2107 2113, 2018. [9] Lei Feng, Jiaqi Lv, Bo Han, Miao Xu, Gang Niu, Xin Geng, Bo An, and Masashi Sugiyama. Provably consistent partial-label learning. Advances in Neural Information Processing Systems, 2020. [10] Michael Figurnov, Shakir Mohamed, and Andriy Mnih. Implicit reparameterization gradients. Advances in Neural Information Processing Systems, 2018. [11] Xin Geng. Label distribution learning. IEEE Transactions on Knowledge and Data Engineering, 28(7):1734 1748, 2016. [12] Xin Geng, Chao Yin, and Zhi-Hua Zhou. Facial age estimation by learning from label distributions. IEEE transactions on pattern analysis and machine intelligence, 35(10):2401 2412, 2013. [13] Matthieu Guillaumin, Jakob Verbeek, and Cordelia Schmid. Multiple instance metric learning from automatically labeled bags of faces. In Lecture Notes in Computer Science 6311, pages 634 647. Springer, Berlin, 2010. [14] Geoffrey Hinton, Oriol Vinyals, and Jeff Dean. Distilling the knowledge in a neural network. ar Xiv preprint ar Xiv:1503.02531, 2015. [15] Eyke Hüllermeier and Jürgen Beringer. Learning from ambiguously labeled examples. Intelligent Data Analysis, 10(5):419 439, 2006. [16] Takashi Ishida, Gang Niu, Aditya Menon, and Masashi Sugiyama. Complementary-label learning for arbitrary losses and models. In International Conference on Machine Learning, pages 2971 2980. PMLR, 2019. [17] Rong Jin and Zoubin Ghahramani. Learning with multiple labels. In Advances in Neural Information Processing Systems 15, pages 897 904, Cambridge, MA, 2003. [18] Julian Katz-Samuels, Gilles Blanchard, and Clayton Scott. Decontamination of mutual contamination models. Journal of machine learning research, 20(41), 2019. [19] Diederik P Kingma and Max Welling. Auto-encoding variational bayes. In International Conference on Learning Representations, Banff, AB, Canada, 2014. [20] Thomas N Kipf and Max Welling. Variational graph auto-encoders. ar Xiv preprint ar Xiv:1611.07308, 2016. [21] Alex Krizhevsky, Geoffrey Hinton, et al. Learning multiple layers of features from tiny images. 2009. [22] Yann Le Cun, Léon Bottou, Yoshua Bengio, and Patrick Haffner. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278 2324, 1998. [23] Dong-Hyun Lee et al. Pseudo-label: The simple and efficient semi-supervised learning method for deep neural networks. In Workshop on challenges in representation learning, ICML, volume 3, page 896, 2013. [24] Liping Liu and Thomas G Dietterich. A conditional multinomial mixture model for superset label learning. In Advances in Neural Information Processing Systems 25, pages 557 565, Cambridge, MA, 2012. [25] Jie Luo and Francesco Orabona. Learning from candidate labeling sets. In Advances in Neural Information Processing Systems 23, pages 1504 1512. Cambridge, MA, 2010. [26] Jiaqi Lv, Miao Xu, Lei Feng, Gang Niu, Xin Geng, and Masashi Sugiyama. Progressive identification of true labels for partial-label learning. In International Conference on Machine Learning, pages 6500 6510. PMLR, 2020. [27] Rafael Müller, Simon Kornblith, and Geoffrey Hinton. When does label smoothing help? ar Xiv preprint ar Xiv:1906.02629, 2019. [28] Nam Nguyen and Rich Caruana. Classification with partial labels. In Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pages 381 389, Las Vegas, NV, 2008. [29] Hieu Pham, Zihang Dai, Qizhe Xie, and Quoc V Le. Meta pseudo labels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 11557 11568, 2021. [30] Christian Szegedy, Vincent Vanhoucke, Sergey Ioffe, Jon Shlens, and Zbigniew Wojna. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 2818 2826, 2016. [31] Cai-Zhi Tang and Min-Ling Zhang. Confidence-rated discriminative partial label learning. In Proceedings of the 31st AAAI Conference on Artificial Intelligence, pages 2611 2617, San Francisco, CA, 2017. [32] Han Xiao, Kashif Rasul, and Roland Vollgraf. Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms. ar Xiv preprint ar Xiv:1708.07747, 2017. [33] Ning Xu, Yun-Peng Liu, and Xin Geng. Label enhancement for label distribution learning. IEEE Transactions on Knowledge and Data Engineering, 33(4):1632 1643, 2021. [34] Ning Xu, Jiaqi Lv, and Xin Geng. Partial label learning via label enhancement. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, pages 5557 5564, 2019. [35] Ning Xu, Jun Shu, Yun-Peng Liu, and Xin Geng. Variational label enhancement. In Proceedings of the International Conference on Machine Learning, pages 10597 10606, Vienna, Austria, 2020. [36] Yao Yao, Jiehui Deng, Xiuhua Chen, Chen Gong, Jianxin Wu, and Jian Yang. Deep discriminative cnn with temporal ensembling for ambiguously-labeled image classification. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pages 12669 12676, 2020. [37] Fei Yu and Min-Ling Zhang. Maximum margin partial label learning. Machine Learning, 106(4):573 593, 2017. [38] Zinan Zeng, Shijie Xiao, Kui Jia, Tsung-Han Chan, Shenghua Gao, Dong Xu, and Yi Ma. Learning by associating ambiguously labeled images. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pages 708 715, Portland, OR, 2013. [39] Linfeng Zhang, Jiebo Song, Anni Gao, Jingwei Chen, Chenglong Bao, and Kaisheng Ma. Be your own teacher: Improve the performance of convolutional neural networks via self distillation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 3713 3722, 2019. [40] Min-Ling Zhang and Fei Yu. Solving the partial label learning problem: An instance-based approach. In Proceedings of the 24th International Joint Conference on Artificial Intelligence, pages 4048 4054, Buenos Aires, Argentina, 2015. [41] Min-Ling Zhang, Fei Yu, and Cai-Zhi Tang. Disambiguation-free partial label learning. IEEE Transactions on Knowledge and Data Engineering, 29(10):2155 2167, 2017. [42] Min-Ling Zhang, Bin-Bin Zhou, and Xu-Ying Liu. Partial label learning via feature-aware disambiguation. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pages 1335 1344, 2016. [43] Yikai Zhang, Songzhu Zheng, Pengxiang Wu, Mayank Goswami, and Chao Chen. Learning with feature dependent label noise: a progressive approach. ar Xiv preprint ar Xiv:2103.07756, 2021.