# classdistributionaware_pseudolabeling_for_semisupervised_multilabel_learning__7f31e8bd.pdf Class-Distribution-Aware Pseudo-Labeling for Semi-Supervised Multi-Label Learning Ming-Kun Xie1,2, Jia-Hao Xiao1,2, Hao-Zhe Liu1,2, Gang Niu3, Masashi Sugiyama3,4, Sheng-Jun Huang1,2 1Nanjing University of Aeronautics and Astronautics 2MIIT Key Laboratory of Pattern Analysis and Machine Intelligence, Nanjing, China 3RIKEN Center for Advanced Intelligence Project 4The University of Tokyo, Tokyo, Japan {mkxie, jiahaoxiao, haozheliu, huangsj}@nuaa.edu.cn gang.niu.ml@gmail.com sugi@k.u-tokyo.ac.jp Pseudo-labeling has emerged as a popular and effective approach for utilizing unlabeled data. However, in the context of semi-supervised multi-label learning (SSMLL), conventional pseudo-labeling methods encounter difficulties when dealing with instances associated with multiple labels and an unknown label count. These limitations often result in the introduction of false positive labels or the neglect of true positive ones. To overcome these challenges, this paper proposes a novel solution called Class-Aware Pseudo-Labeling (CAP) that performs pseudolabeling in a class-aware manner. The proposed approach introduces a regularized learning framework incorporating class-aware thresholds, which effectively control the assignment of positive and negative pseudo-labels for each class. Notably, even with a small proportion of labeled examples, our observations demonstrate that the estimated class distribution serves as a reliable approximation. Motivated by this finding, we develop a class-distribution-aware thresholding strategy to ensure the alignment of pseudo-label distribution with the true distribution. The correctness of the estimated class distribution is theoretically verified, and a generalization error bound is provided for our proposed method. Extensive experiments on multiple benchmark datasets confirm the efficacy of CAP in addressing the challenges of SSMLL problems. The implementation is available at https://github.com/milkxie/SSMLL-CAP. 1 Introduction In single-label supervised learning, each instance is assumed to be associated with only one class label, while many realistic scenarios may be multi-labeled, where each instance consists of multiple semantics. For example, an image of a nature landscape often contains the objects of sky, cloud, and mountain. Multi-label learning (MLL) is a practical and effective paradigm for handling examples with multiple labels. It trains a classifier that can predict all the relevant labels for unseen instances based on the given training examples. A large number of recent works have witnessed the great progress that MLL has made in many practical applications [27], e.g., image annotation [21], protein subcellular localization [28], and visual attribute recognition [32]. Thanks to its powerful capacity, the deep neural network (DNN) has become a prevalent learning model for handling MLL examples [34]. Unfortunately, it requires a large number of precisely labeled Correspondence to: Sheng-Jun Huang (huangsj@nuaa.edu.cn). 37th Conference on Neural Information Processing Systems (Neur IPS 2023). examples to achieve favorable performance. This leads to a high cost of manual annotation, especially when the dataset is large and the labeling task must be carried out by an expert. Given that it is hard to train an effective DNN based on a small subset of training examples, it is rather important to exploit the information from unlabeled instances. The problem has been formalized as a learning framework called semi-supervised multi-label learning (SSMLL), which aims to train a classifier based on a small set of labeled MLL examples and a large set of unlabeled ones. Compared to semi-supervised learning (SSL) that has made great progress [4, 38], SSMLL has received relatively less attention in the context of deep learning. Generally, there are still three main challenges towards the development of SSMLL. Firstly, since each instance is associated with multiple labels and the number is unknown, the commonly used pseudo-labeling strategy that selects the most probable label or the top-k probable labels cannot be applied to the SSMLL problems. It would face the dilemma of either introducing false positive labels or neglecting true positive ones. Secondly, due to the intrinsic class-imbalance property of MLL data, it is hard to achieve favorable performance by using a fixed threshold for each instance. Thirdly, recent studies have mainly focused on multi-label learning with missing labels (MLML) [11, 17, 2] scenarios, where each training instance is assumed to be assigned with a subset of true labels. Unfortunately, these methods often fail to achieve favorable performance, or cannot even be applied to the SSMLL scenarios, since most of them were designed under the assumption of MLML. To solve these challenges, in this paper, we propose a novel Class-Aware Pseudo-labeling (CAP) method for handling the SSMLL problems. Unlike the existing methods, we perform pseudo-labeling in a class-aware manner to avoid estimating the number of true labels for each instance, which can be very hard in practice. Specifically, a regularized learning framework is proposed to determine the numbers of positive and negative pseudo-labels for each class based on the class-aware thresholds. Given that the true class distribution is unknown, we alternatively determine the thresholds based on the estimated class distribution of labeled data, which can be a tight approximation according to our observation. Our theoretical results show the correctness of estimated class distribution and provide a generalization error bound for CAP. Extensive experimental results on multiple benchmark datasets with a variety of comparing methods validate that the proposed method can achieve state-of-the-art performance. 2 Related Work Thanks to the powerful learning capacity of DNNs, MLL has made great advances in the context of deep learning. Some methods designed architectures [6] or training strategies [22, 49] to exploit the label correlations. Some other methods designed sophisticated loss functions to improve the performance of MLL [34]. The last group of methods designed specific architectures to capture the objects related to semantic labels. Global-average-pooling (GAP) based models [33] and attentionbased models [22, 26] are two groups of representative methods. There are relatively few works that study how to improve the performance of deep models in SSMLL scenarios. Instead of end-to-end training, the only deep SSMLL method [43] performed the twostage training, which first used a DNN to extract features, and then used a linear model to perform classification. [37] proposed a deep sequential generative model to handle the noisy labels collected by crowdsourcing and unlabeled data simultaneously. [20] focused on the transductive and non-deep scenario, and thus cannot be applied to our setting. The method proposed by [39] utilized the graph neural network (GNN) to deal with SSMLL data with graph structures. In contrast, there are many works that trained linear models to solve the SSMLL problems [5, 15, 42, 52, 50, 40]. M3DN was proposed to deal with SSMLL data in multi-modal multi-instance learning scenarios by adopting optimal transport technique [48]. Pseudo-labeling has become a popular method in semi-supervised learning (SSL). The idea was firstly applied to semi-supervised training of deep neural networks [23]. Subsequently, a great deal of works have been devoted to improving the quality of pseudo-labels either by adopting consistency regularization [4, 38], or by using distribution alignment [30, 3]. The contrastive learning technique has been applied to improve the performance of SSL [24]. To improve the reliability of pseudolabels, an uncertainty-aware pseudo-labeling method proposed in [35] selected reliable pseudo-labels based on the prediction uncertainty. Recent studies have also paid attention to dealing with the class-imbalance problem of pseudo-labeling in SSL scenarios [18, 45, 14]. Unlike Fix Match that selects unlabeled examples with a fixed threshold, Dash [47] selected unlabeled examples with a dynamic threshold, with the goal of achieving better pseudo-labeling performance. Flex Match [51] was proposed to select unlabeled examples for every class according to the current learning status of the model. Several works have been explored the idea of selecting different thresholds for different classes to improve the performance of SLL [13, 14, 44].However, these methods are designed for the multi-class single-label scenario, and cannot be directly applied to the multi-label scenario. In order to reduce the annotation cost, a cost-effective strategy is to assign a subset of true labels to each instance. For example, [11] designed a partial binary cross entropy (BCE) loss that re-weights the losses of known labels. As an extreme case of MLML, single positive multi-label learning (SPML) [8, 53, 41] assumes that only one of multiple true labels can be observed during the training stage. The pioneering work [8] trains DNNs by simply treating unobserved labels as negative ones and utilizes the regularization to alleviate the harmfulness of false negative labels. [53] propose asymmetric pseudo labeling technique to recover true labels. 3 The Method In the SSMLL problem, let x X be a feature vector and y Y be its corresponding label vector, where X = Rd is the feature space and Y = {0, 1}q is the label space with q possible class labels. Here, yk = 1 indicates the k-th label is relevant to the instance, while yk = 0, otherwise. Suppose that we are given a labeled dataset with n training examples Dl = {(xi, yi)}n i=1 and an unlabeled dataset with m training instances Du = {xj}m j=1. Our goal is to train a DNN f(x; θ) based on the labeled dataset Dl and unlabeled dataset Du, where θ is the parameter of the network. For notational simplicity, we omit the notation θ and let f(x) be the predicted probability distribution over classes and fk(x) be the predicted probability of the k-th class for input x. Typical multi-label learning methods usually train a DNN with the commonly used binary cross entropy (BCE) loss, which decomposes the original task into multiple binary classification problems. Unfortunately, BCE loss often suffers from positive-negative imbalance issue. To mitigate this problem, we adopt the asymmetric loss (ASL) [34], which is a variant of focal loss with different focusing parameters for positive and negative instances. In our experiment, we found it works better than BCE loss. Formally, given the predicted probabilities f(x) on instance x, the ASL loss is defined as L(f(x), y) = Xq k=1 ykℓ1(fk(x)) + (1 yk)ℓ0(fk(x)), (1) Here, ℓ1(fk) = (1 fk)λ1 log(fk) and ℓ0(fk) = (fk)λ0 log(1 fk) represent the losses calculated on positive and negative labels, where λ1 and λ0 are positive and negative focusing parameters. 3.1 Instance-Aware Pseudo-Labeling The loss function may not be the best choice to solve the SSMLL problem, since besides the labeled training examples, there still exist a large number of unlabeled training examples. To exploit the information of unlabeled data, inspired by recent SSL works [4, 38], an intuitive strategy is assigning the unlabeled instances with pseudo-labels based on the model outputs. Formally, we define the unlabeled loss Lu as Lu(f(x), ˆy) = Xq k=1 ˆykℓ1(fk(x)) + (1 ˆyk)ℓ0(fk(x)), where ˆy = [ˆy1, , ˆyj] represents the pseudo-label vector for instance x. In the above formulation, the most significant element is how to obtain the pseudo-labels ˆy that significantly affects the final performance of SSMLL. Most of existing pseudo-labeling methods are performed in an instance-aware manner by assigning pseudo-labels to each unlabeled instance based on its probability distribution. Below, we briefly review three instance-aware pseudo-labeling strategies that can be applied to the SSMLL problems. The most commonly used strategy adopted by the SSL method called Fix Match [38] is to select the most probable label as the ground-truth one: (1 if k = arg max c [q] fc(x), 0 otherwise. (2) One advantage of the strategy is that it is likely to safely identify a true label for each unlabeled training instance. Unfortunately, it is obvious that the strategy would neglect multiple true labels. Generally, it transforms the unlabeled dataset into another learning scenario called single positive multi-label learning (SPML) [8], where only one of multiple positive labels is available for each instance. A straightforward strategy is to simply treat unobserved labels as negative ones. Although this strategy enables us to train a classifier based on SPML data, it would introduce a large number of false negative labels, leading to unfavorable performance. The second choice is an improved version of the above strategy, which selects the top l probable labels as the true ones: ˆyk = 1 if fk(x) τ l, 0 otherwise, (3) where τ l is the l-th predicted probability in a descending order. The strategy conducts a competition among labels, and selects top l winners. The optimal solution is to set l as the true number of positive labels for each unlabeled instance. Unfortunately, since the true number is unknown in practice, as a compromise, we set l as the average number of positive labels per instance. Given that the true number does not always equal to the average number, it would be caught in a dilemma of either introducing false positive labels or neglecting true positive ones. The last choice is to adopt an instance-aware threshold τj that separates positive and negative labels for each unlabeled instance. ˆyjk = 1 if fk(xj) τj, 0 otherwise. (4) Compared to the above methods, this strategy achieves a strong flexibility that allows it to assign different numbers of positive labels to different instances. A potential limitation is that it is hard to find the optimal thresholds for different instances. In practice, a feasible solution is to adopt a global threshold, that is j [m], τj = τ. Obviously, it is impossible to adopt a global threshold that is optimal for all instances, especially considering the class-imbalance property of MLL data. In general, a large threshold often leads to a small recall score of tail classes, which indicates that less positive labels would be identified. While a small threshold often results in a small precision score of head classes, which indicates besides positive labels, a great deal of negative ones would be treated as positive ones. The dilemma prevents the model from obtaining favorable performance. 3.2 Class-Aware Pseudo-Labeling As discussed above, in many real-world scenarios, it is really difficult to acquire the true number of positive labels for each instance. This leads the instance-aware pseudo-labeling methods to be caught in the dilemma of either mislabeling false positive labels or neglecting true positive labels, resulting in a noticeable decrease of the model performance. To solve this issue, we propose a regularized learning framework to assign pseudo-labels in a class-aware manner. Formally, we reformulate the optimization problem of SSMLL as k=1 yikℓ1(fk(xi)) + (1 yik)ℓ0(fk(xi)) k=1 ˆyjkℓ1(fk(xj)) + (1 ˆyjk)ℓ0(fk(xj)) k=1 αkˆyjk + βk(1 ˆyjk), s.t. j [m], ˆyj = [yj1, , yjq] {0, 1}q, k [q], αk > 0, βk > 0, where αk and βk are class-aware regularized parameters to control how many positive and negative labels would be included into model training for class k. Below, we primarily provide a solution of the optimization problem Eq.(5), and then discuss how to set parameters αk and βk to capture the true class distribution of unlabeled examples. Alternative Search It is hard to directly solve the optimization problem Eq.(5), since there are two sets of variables. A feasible solution is to adopt the alternative convex search [1, 54] strategy that optimizes a group of variables by fixing the other group of variables. Instance-aware thresholding Class-distribution-aware thresholding (a) Different Pseudo-Labling Strategies (b) Class Proportions Figure 1: (a) An illustration of the comparison between instance-aware and class-aware pseudolabeling methods. (b) The curves of the estimated and true class proportions on COCO and VOC. By using the CAT strategy, CAP can provide high-quality pseudo-labels by approximating the true class distribution. This can be validate by the results in (b), where the empirical and true class proportions of positive labels show high-level consistency. Suppose that pseudo labels ˆy are given, then the optimization problem Eq.(5) can be transformed into an ordinary loss by treating the pseudo labels as the true ones: i=1 L(f(xi), yi) + 1 j=1 Lu(f(xj), ˆyj), (6) which can be solved by applying the stochastic gradient decent (SGD) method. With the parameters θ fixed, we reformulate the optimization problem with respect to ˆy as k=1 ˆyjkℓ1(fk(xj)) + (1 ˆyjk)ℓ0(fk(xj)) k=1 αkˆyjk + βk(1 ˆyjk). (7) Consider that ˆyk is assume to be one or zero, we can obtain the following solution: 1 if fk(x) τ(αk), 0 if fk(x) τ(βk), 1 otherwise, (8) where τ(αk) = exp( αk) and τ(βk) = 1 exp( βk) are two class-aware thresholds, and ˆyk = 1 means that the label ˆyk would not be used for model training. 3.3 Class-Distribution-Aware Thresholding An important problem is how to set the thresholds τ(αk) and τ(βk), which determine the numbers of positive and negative pseudo-labels for every class k. In order to capture the true class distribution, we propose the Class-distribution-Aware Thresholding (CAT) strategy to determine τ(αk) and τ(βk). Suppose that we are given yj, j [m], i.e., the true label vectors of unlabeled training instances. By solving the following equation, we can obtain τ(αk) and τ(βk) that capture the true class distribution of unlabeled data. Pm j=1 I(fk(xj) τ(αk)) Pm j=1 I(fk(xj) τ(βk)) where γ k = Pm j=1 I(yjk=1) m and ρ k = Pm j=1 I(yjk=0) m are respectively the proportions of positive and negative labels in unlabeled data for class k. Although during the training process, the true labels of unlabeled instances are inaccessible, our observation shows that the estimated class distribution, i.e., the class proportions of positive and negative labels in labeled examples, can tightly approximate the true class distribution. As shown Figure 1 (b), we illustrate the proportions of positive labels in labeled examples and unlabeled examples for every class k on two benchmark datasets COCO and VOC. The proportions of labeled examples are respectively p = 0.05 and p = 0.1 for COCO and VOC. From the figures, it can be observed that even with a small proportion of labeled examples (p = 0.05), it achieves a nearly complete overlap between the estimated and true curves, which validates that the estimated class distribution can be a tight approximation of the true one. This motivate us to alternatively utilize the estimated class distribution to solve the solutions for τ(αk) and τ(βk): Pm j=1 I(fk(xj) τ(αk)) Pm j=1 I(fk(xj) τ(βk)) m = ˆρk, (9) where ˆγk = Pn i=1 I(yik=1) n and ˆρk = Pn i=1 I(yik=0) n are respectively the proportions of positive and negative labels in labeled data for class k. Figure 1 provides an illustration of the comparison between three instance-aware pseudo-labeling methods and the CAP method. By utilizing CAT strategy, CAP is expected to assign pseudo-labels with the class distribution that approximates the true one. In practice, to further improve the performance of CAP, one feasible solution is to discard a fraction of unreliable pseudo-labels with relatively low confidences, which may have a negative impact on the model training. Specifically, for any class k [q], we select top η1 ˆγk and η0 ˆρk proportion probable pseudo-labels, where η1, η0 [0, 1] are two parameters to control the reliable intervals of pseudo-labels. By substituting the two terms into the right sides of Eq.(9), we can obtain the thresholds correspondingly. In Section 5.4, we perform ablation experiments to study the influence of reliable intervals on the model performance. 4 Theoretical Analysis In this section, we perform theoretical analyses for the proposed method. In general, the performance of pseudo-labeling depends mainly on two factors, i.e., the quality of the model predictions and the correctness of estimated class distribution. Our work focuses on the latter. Consider an extreme case, where the model predictions are perfect, i.e., the confidences of positive labels are always greater than that of negative labels. In such a case, we still need an appropriate threshold to precisely separate the positive and negative labels. This implies that we need to capture the true class distribution of unlabeled data in order to achieve desirable pseudo-labeling performance. 4.1 Correctness of the Estimated Class Distribution To study the correctness of estimated class distribution, we provide the following theorem, which gives an upper bound on the difference between the estimated class proportion ˆγk and the true class proportion γ k (its proof is given in Appendix A). A similar result can be derived for ˆρk. Theorem 1. Assume the estimated class proportion ˆγk = 1 n Pn i=1 I(yik = 1), and the true class proportion γ k = 1 m Pm j=1 I(yjk = 1) for any k [q], where n and m are the numbers of labeled and unlabeled examples that satisfy m >> n. Then, with the probability larger than 1 2n 1 2m 1, we have, k [q], |ˆγk γ k| log n Theorem 1 tells us that the correctness of the estimated class distribution mainly depends on the number of labeled and unlabeled data. In general, the bound is dominated by the first term, since it always satisfies m >> n. By neglecting the second term, we can see that k [q], γ k ˆγk in the parametric rate Op(1/ n), where Op denotes the order in probability. Obviously, as the number of training examples increase, the estimated class distribution would quickly converge to the true one. 4.2 Generalization Bound Moreover, we study the generalization performance of CAP. Before providing the main results, we first define the true risk with respect to the classification model f(x; θ): R(f) = E(x,y) [L(f(x), y)] . Our goal is to learn a good classification model by minimizing the empirical risk b R(f) = b Rl(f) + b Ru(f), where b Rl(f) and b Ru(f) are respectively the empirical risk of the labeled loss Ll(f(x), y) and unlabeled loss Lu(f(x), y): b Rl(f) = 1 i=1 L(f(xi), yi), b Ru(f) = 1 j=1 Lu(f(xj), yj). Note that during the training, we cannot train a model directly by optimizing b Ru(f), since the labels of unlabeled data are inaccessible. Instead, we train the model with b R u(f) = 1 m Pm j=1 Lu(f(xj), ˆyj), where ˆyj represents the pseudo-label vector of the instance xj. Let ℓ(fk(x)) = ykℓ1(fk(x)) + (1 yk)ℓ0(fk(x)) be the loss for the class k, and Lℓbe any (not necessarily the best) Lipschitz constant of ℓ. Let RN(F) be the expected Rademacher complexity [31] of F with N = m + n training points. Let ˆf = arg minf F b R(f) be the empirical risk minimizer, where F is a function class, and f = arg minf F R(f) be the true minimizer. We derive the following theorem, which provides a generalization error bound for the proposed method (its proof is given in Appendix B). Theorem 2. Suppose that ℓ( ) is bounded by B. For some ϵ > 0, if Pm j=1 |I(fk(xj) τ(αk)) I(yjk = 1)|/m ϵ for any k [q], for any δ > 0, with probability at least 1 δ, we have R( ˆf) R(f ) 2q Bϵ + 4q LℓRN(F) + 2q B From Theorem 4, it can be observed that the generalization performance of ˆf mainly depends on two factors, i.e., the pseudo-labeling error ϵ and the number of training examples N. Apparently, a smaller pseudo-labeling error ϵ often leads to better generalization performance. Thanks to its ability to capture the true class distribution, CAP can achieve a much smaller pseudo-labeling error ϵ than existing instance-aware pseudo-labeling methods, which is beneficial for obtaining better classification performance. This can be further validated by our empirical results in Section 5.3. The second factor is the number of training examples. As N and ϵ 0, Theorem 4 shows that the empirical risk minimizer ˆf converges to the true risk minimizer f . 5 Experiments In this section, we first perform experiments to validate the effectiveness of the proposed method; then, we perform ablation studies to analyze the mechanism behind CAP. 5.1 Experimental Settings Datasets To evaluate the performance of the propose method, we conduct experiments on three benchmark image datasets, including Pascal VOC-2012 (VOC for short) 2 [12], MS-COCO-2014 (MS-COCO for short) 3 [25], and NUS-WIDE (NUS for short) 4 [7]. The detailed information of these datasets can be found in the appendix. For each dataset, we randomly sample a proportion p {0.05, 0.1, 0.15, 0.2} of examples with full labels while the others without any supervised information. Following the previous works [8, 53], we report the mean average precision (m AP) on the test set for each method. Comparing methods To validate the effectiveness of the proposed method, we compare it with five groups of methods: 1) three instance-aware pseudo-labeling methods: Top-1 (Eq.(2)), Top-k (Eq.(3)), IAT (Eq.(4)); 2) two state-of-the-art MLML methods: LL [19] (includes three variants LL-R, LL-Ct, and LL-Cp), PLC [46]; 3) Two state-of-the-art SSL methods: Adsh [14], Free Match [44]; 4) One state-of-the-art SSMLL method: DRML [43]; 5) Two baseline methods, BCE, ASL [34]. DRML is the only deep SSMLL method whose source code could be found on the Internet. 2http://host.robots.ox.ac.uk/pascal/VOC/ 3https://cocodataset.org 4https://lms.comp.nus.edu.sg/wp-content/uploads/2019/research/nuswide/NUS-WIDE. html Table 1: Comparison results on VOC and COCO in terms of m AP (%). The best performance is highlighted in bold. Method VOC COCO p = 0.05 p = 0.1 p = 0.15 p = 0.2 p = 0.05 p = 0.1 p = 0.15 p = 0.2 BCE 67.95 75.35 78.19 79.38 58.90 63.75 65.91 67.33 ASL 71.46 78.00 79.69 80.77 59.12 63.82 66.10 67.51 LL-R 75.69 80.96 82.31 83.55 59.31 64.25 66.61 68.01 LL-Ct 75.77 81.04 82.31 83.50 59.33 64.23 66.69 68.11 LL-Cp 75.79 81.03 82.36 83.68 59.27 64.19 66.68 68.12 PLC 74.49 80.35 82.35 83.39 59.85 65.03 67.62 69.14 Top-1 75.77 80.78 82.65 83.72 57.62 62.84 65.50 66.96 Top-k 75.07 80.20 81.99 83.16 58.25 63.52 66.11 67.49 IAT 73.24 80.27 82.39 83.55 60.34 65.54 67.88 69.25 ADSH 75.37 80.34 82.80 83.93 60.75 65.37 67.70 69.01 Free Match 75.11 80.66 82.63 83.60 59.94 64.46 66.79 68.04 DRML 61.77 71.01 72.98 74.49 53.60 57.06 58.53 59.24 Ours 76.16 82.16 83.48 84.41 62.43 67.36 69.11 70.41 Table 2: Comparison results on NUS in terms of m AP (%). The best performance is highlighted in bold. Method ASL LL-R PLC Top-1 Top-k IAT ADSH Free Match DRML Ours p = 0.05 42.87 40.20 43.55 40.99 40.89 42.58 43.94 43.12 30.61 44.82 p = 0.10 46.50 44.95 47.51 45.07 45.04 46.60 47.28 46.65 35.09 48.24 p = 0.15 48.42 47.32 49.75 47.43 47.22 48.76 49.22 48.74 37.91 49.90 p = 0.20 49.65 48.31 50.71 48.49 48.37 49.62 49.93 49.59 39.98 51.06 Furthermore, most MLML methods cannot be applied to the SSMLL scenario, since they assume that a subset of labels have been annotated for each training instance. The detailed information of these methods can be found in the appendix. Implementation We employ Res Net-50 [16] pre-trained on Image Net [36] for training the classification model. We adopt Rand Augment [9] and Cutout [10] for data augmentation. We employ Adam W [29] optimizer and one-cycle policy scheduler [10] to train the model with maximal learning rate of 0.0001. The number of warm-up epochs is set as 12 for all datasets. The batch size is set as 32, 64, and 64 for VOC, MS-COCO, and NUS. Furthermore, we perform exponential moving average (EMA) for the model parameter θ with a decay of 0.9997. For all methods, we use the ASL loss as the base loss function, since it shows superiority to BCE loss [34]. We perform all experiments on Ge Force RTX 3090 GPUs. The random seed is set to 1 for all experiments. 5.2 Comparison Results Table 1 and Table 2 report the comparison results between CAP and the comparing methods in terms of m AP on VOC, COCO, and NUS. From the tables, we can see that: 1) DRML obtains unfavorable performance, even worse than baselines BCE and ASL, since it performs two-stage training, which may destroy its representation learning. The original paper did not report the results on these three datasets. Therefore, it is rather important to design an effective SSMLL method in deep learning paradigm. 2) CAP outperforms three instance-aware pseudo-labeling methods, which demonstrates that by utilizing CAT strategy, CAP can precisely estimate the class distribution of unlabeled data and thus obtain desirable pseudo-labeling performance. 3) The performance of CAP is better than that of two state-of-the-art SSL methods. To achieve better performance, we have made several modifications for these two methods not limited to the following: a) use the ASL loss ; b) adopt stronger data augmentations; c) change the training scheme to make them more suitable for the multi-label scenario. 4) CAP achieves the best performance in all cases and significantly outperforms CAP Top-1 Top-k IAT 1 3 5 7 9 11 13 15 Epochs 45 50 55 60 65 70 CF1 Score (%) 1 3 5 7 9 11 13 15 Epochs 60 62 64 66 68 70 72 74 76 CF1 Score (%) 1 3 5 7 9 11 13 15 Epochs 64 66 68 70 72 74 76 CF1 Score (%) 1 3 5 7 9 11 13 15 Epochs 64 66 68 70 72 74 76 78 CF1 Score (%) 1 3 5 7 9 11 13 15 Epochs 35 40 45 50 55 60 CF1 Score (%) (a) p = 0.05 1 3 5 7 9 11 13 15 Epochs 35 40 45 50 55 60 65 CF1 Score (%) (b) p = 0.1 1 3 5 7 9 11 13 15 Epochs 35 40 45 50 55 60 65 70 CF1 Score (%) (c) p = 0.15 1 3 5 7 9 11 13 15 Epochs 35 40 45 50 55 60 65 70 CF1 Score (%) (d) p = 0.2 Figure 2: Pseudo-labeling performance in terms of CF1 score on VOC, COCO. Each row corresponds one dataset. 0.80 0.85 0.90 0.95 1.00 76 78 80 82 84 86 MAP Score (%) p = 0.05 p = 0.1 (a) VOC, η0 = 1 0.95 0.96 0.97 0.98 0.99 1.00 74 76 78 80 82 84 MAP Score (%) (b) VOC, η1 = 1 0.80 0.85 0.90 0.95 1.00 MAP Score (%) (c) COCO, η0 = 1 0.95 0.96 0.97 0.98 0.99 1.00 60 62 64 66 68 70 72 MAP Score (%) (d) COCO, η1 = 1 Figure 3: Performance of CAP on VOC and COCO in terms of m AP (%) with the increase of η1 and η0. the comparing methods, especially when the number of labeled examples is small. These results convincingly validate the effectiveness of the proposed method. 5.3 Study on the Performance of Pseudo-Labeling In this section, we explain why CAP is better than the conventional instance-aware pseudo-labeling methods. Figure 2 illustrates the performance of different pseudo-labeling methods in terms of CF1 score on VOC and COCO. From the figures, we can see that CAP achieves the best performance in all cases. As discussed above, the pseudo-labeling performance mainly depends on two factors, i.e., the quality of model predictions and the correctness of estimated class proportions. CAP improves the pseudo-labeling performance by precisely estimating the class distribution. An interesting observation is that at the first epoch, when the model predictions are the same for four methods, our method significantly outperforms the comparing methods, since it is able to capture the true class proportions. These results validate that CAP can achieve better pseudo-labeling performance. 5.4 Study on the Influence of Reliable Interval As mentioned above, to improve the performance, instead of using all pseudo-labels, we can train the model with only reliable pseudo-labels within the reliable intervals that are controlled by η1, η0. Figure 3 illustrates the performance of CAP as η1 and η0 change in the ranges of [0.8, 0.85, 0.9, 0.95, 1] and [0.95, 0.96, 0.97, 0.98, 0.99, 1]. From the figures, it can be observed that discarding the unreliable positive pseudo-labels would improve the performance, but discarding the unreliable negative pseudo-labels would degrade the performance. One possible reason behinds the phenomenon is due to the significant positive-negative imbalance in MLL data, i.e., the number of negative labels is often much greater than that of positive labels. This leads the model to be sensitive to false positive labels, while be robust to false negative labels. In our main experiments (Table 1 and Table 2), we set η1 = 1, η0 = 1. In practice, we are expected to achieve better performance by tuning the parameter η1. 6 Conclusion The paper studies the problem of semi-supervised multi-label learning, which aims to train a multilabel classifier by leveraging the information of unlabeled data. Different from the conventional instance-aware pseudo-labeling methods, we propose to assign pseudo-labels to unlabeled instances in a class-aware manner, with the aim of capturing the true class distribution of unlabeled data. Towards this goal, we propose the CAT strategy to obtain an estimated class distribution, which has been proven to be a desirable estimation of the true class distribution based on our observations. Theoretically, we first perform an analysis on the correctness of estimated class distribution; then, we provide the generalization error bound for CAP and show its dependence to the pseudo-labeling performance. Extensive experimental results on multiple benchmark datasets validate that CAP can achieve state-of-the-art performance. In general, the performance of pseudo-labeling depends mainly on two factors, i.e., the quality of the model predictions and the correctness of the estimated class distribution. This work focuses on the latter. In future, we plan to boost the performance of SSMLL by improving the quality of model predictions. Acknowledgments and Disclosure of Funding Sheng-Jun Huang was supported by Natural Science Foundation of Jiangsu Province of China (BK20222012, BK20211517), the National Key R&D Program of China (2020AAA0107000), and NSFC (62222605). Masashi Sugiyama was supported by JST CREST Grant Number JPMJCR18A2. [1] Mokhtar S Bazaraa, Hanif D Sherali, and Chitharanjan M Shetty. Nonlinear programming: theory and algorithms. John Wiley & Sons, 2013. [2] Emanuel Ben-Baruch, Tal Ridnik, Itamar Friedman, Avi Ben-Cohen, Nadav Zamir, Asaf Noy, and Lihi Zelnik-Manor. Multi-label classification with partial annotations using class-aware selective loss. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 4764 4772, 2022. [3] David Berthelot, Nicholas Carlini, Ekin D Cubuk, Alex Kurakin, Kihyuk Sohn, Han Zhang, and Colin Raffel. Remixmatch: Semi-supervised learning with distribution alignment and augmentation anchoring. ar Xiv preprint ar Xiv:1911.09785, 2019. [4] David Berthelot, Nicholas Carlini, Ian Goodfellow, Nicolas Papernot, Avital Oliver, and Colin A Raffel. Mixmatch: A holistic approach to semi-supervised learning. Advances in neural information processing systems, 32, 2019. [5] Gang Chen, Yangqiu Song, Fei Wang, and Changshui Zhang. Semi-supervised multi-label learning by solving a sylvester equation. In Proceedings of the 2008 SIAM international conference on data mining, pages 410 419. SIAM, 2008. [6] Zhao-Min Chen, Xiu-Shen Wei, Peng Wang, and Yanwen Guo. Multi-label image recognition with graph convolutional networks. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 5177 5186, 2019. [7] Tat-Seng Chua, Jinhui Tang, Richang Hong, Haojie Li, Zhiping Luo, and Yantao Zheng. Nuswide: a real-world web image database from national university of singapore. In Proceedings of the ACM international conference on image and video retrieval, pages 1 9, 2009. [8] Elijah Cole, Oisin Mac Aodha, Titouan Lorieul, Pietro Perona, Dan Morris, and Nebojsa Jojic. Multi-label learning from single positive labels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 933 942, 2021. [9] Ekin Dogus Cubuk, Barret Zoph, Jonathon Shlens, and Quoc Le. Randaugment: Practical automated data augmentation with a reduced search space. In Hugo Larochelle, Marc Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin, editors, Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems, 2020. [10] Terrance De Vries and Graham W Taylor. Improved regularization of convolutional neural networks with cutout. ar Xiv preprint ar Xiv:1708.04552, 2017. [11] Thibaut Durand, Nazanin Mehrasa, and Greg Mori. Learning a deep convnet for multi-label classification with partial labels. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 647 657, 2019. [12] Mark Everingham, SM Eslami, Luc Van Gool, Christopher KI Williams, John Winn, and Andrew Zisserman. The pascal visual object classes challenge: A retrospective. International journal of computer vision, 111(1):98 136, 2015. [13] Vasilii Feofanov, Emilie Devijver, and Massih-Reza Amini. Transductive bounds for the multiclass majority vote classifier. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, pages 3566 3573, 2019. [14] Lan-Zhe Guo and Yu-Feng Li. Class-imbalanced semi-supervised learning with adaptive thresholding. In International Conference on Machine Learning, pages 8082 8094, 2022. [15] Yuhong Guo and Dale Schuurmans. Semi-supervised multi-label classification: a simultaneous large-margin, subspace learning approach. In Machine Learning and Knowledge Discovery in Databases: European Conference, ECML PKDD 2012, Bristol, UK, September 24-28, 2012. Proceedings, Part II 23, pages 355 370. Springer, 2012. [16] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770 778, 2016. [17] Dat Huynh and Ehsan Elhamifar. Interactive multi-label cnn learning with partial labels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 9423 9432, 2020. [18] Jaehyung Kim, Youngbum Hur, Sejun Park, Eunho Yang, Sung Ju Hwang, and Jinwoo Shin. Distribution aligning refinery of pseudo-label for imbalanced semi-supervised learning. Advances in neural information processing systems, 33:14567 14579, 2020. [19] Youngwook Kim, Jae Myung Kim, Zeynep Akata, and Jungwoo Lee. Large loss matters in weakly supervised multi-label classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 14156 14165, 2022. [20] Xiangnan Kong, Michael K Ng, and Zhi-Hua Zhou. Transductive multilabel learning via label set propagation. IEEE Transactions on Knowledge and Data Engineering, 25(3):704 719, 2011. [21] Alina Kuznetsova, Hassan Rom, Neil Alldrin, Jasper Uijlings, Ivan Krasin, Jordi Pont-Tuset, Shahab Kamali, Stefan Popov, Matteo Malloci, Alexander Kolesnikov, et al. The open images dataset v4. International Journal of Computer Vision, 128(7):1956 1981, 2020. [22] Jack Lanchantin, Tianlu Wang, Vicente Ordonez, and Yanjun Qi. General multi-label image classification with transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 16478 16488, 2021. [23] Dong-Hyun Lee et al. Pseudo-label: The simple and efficient semi-supervised learning method for deep neural networks. In Workshop on challenges in representation learning, ICML, volume 3, page 896, 2013. [24] Junnan Li, Caiming Xiong, and Steven CH Hoi. Comatch: Semi-supervised learning with contrastive graph regularization. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 9475 9484, 2021. [25] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C Lawrence Zitnick. Microsoft coco: Common objects in context. In European conference on computer vision, pages 740 755, 2014. [26] Shilong Liu, Lei Zhang, Xiao Yang, Hang Su, and Jun Zhu. Query2label: A simple transformer way to multi-label classification. ar Xiv preprint ar Xiv:2107.10834, 2021. [27] Weiwei Liu, Haobo Wang, Xiaobo Shen, and Ivor W Tsang. The emerging trends of multi-label learning. IEEE transactions on pattern analysis and machine intelligence, 44(11):7955 7974, 2021. [28] Ziyi Liu, Zengmao Wang, and Bo Du. Multi-marginal contrastive learning for multi-label subcellular protein localization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 20626 20635, June 2022. [29] Ilya Loshchilov and Frank Hutter. Decoupled weight decay regularization. In 7th International Conference on Learning Representations, 2019. [30] Gideon S Mann and Andrew Mc Callum. Simple, robust, scalable semi-supervised learning via expectation regularization. In Proceedings of the 24th international conference on Machine learning, pages 593 600, 2007. [31] Mehryar Mohri, Afshin Rostamizadeh, and Ameet Talwalkar. Foundations of machine learning. MIT press, 2018. [32] Khoi Pham, Kushal Kafle, Zhe Lin, Zhihong Ding, Scott Cohen, Quan Tran, and Abhinav Shrivastava. Learning to predict visual attributes in the wild. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 13018 13028, 2021. [33] Kirill Prokofiev and Vladislav Sovrasov. Combining Metric Learning and Attention Heads For Accurate and Efficient Multilabel Image Classification. ar Xiv e-prints, 2022. [34] Tal Ridnik, Emanuel Ben-Baruch, Nadav Zamir, Asaf Noy, Itamar Friedman, Matan Protter, and Lihi Zelnik-Manor. Asymmetric loss for multi-label classification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 82 91, 2021. [35] Mamshad Nayeem Rizve, Kevin Duarte, Yogesh S Rawat, and Mubarak Shah. In defense of pseudo-labeling: An uncertainty-aware pseudo-label selection framework for semi-supervised learning. In International Conference on Learning Representations, 2020. [36] Olga Russakovsky, Jia Deng, Hao Su, Jonathan Krause, Sanjeev Satheesh, Sean Ma, Zhiheng Huang, Andrej Karpathy, Aditya Khosla, Michael Bernstein, Alexander C. Berg, and Li Fei-Fei. Image Net Large Scale Visual Recognition Challenge. International Journal of Computer Vision (IJCV), 115(3):211 252, 2015. [37] Wanli Shi, Victor S Sheng, Xiang Li, and Bin Gu. Semi-supervised multi-label learning from crowds via deep sequential generative model. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pages 1141 1149, 2020. [38] Kihyuk Sohn, David Berthelot, Nicholas Carlini, Zizhao Zhang, Han Zhang, Colin A Raffel, Ekin Dogus Cubuk, Alexey Kurakin, and Chun-Liang Li. Fixmatch: Simplifying semisupervised learning with consistency and confidence. Advances in neural information processing systems, 33:596 608, 2020. [39] Zixing Song, Ziqiao Meng, Yifei Zhang, and Irwin King. Semi-supervised multi-label learning for graph-structured data. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management, pages 1723 1733, 2021. [40] Qiaoyu Tan, Yanming Yu, Guoxian Yu, and Jun Wang. Semi-supervised multi-label classification using incomplete label information. Neurocomputing, 260:192 202, 2017. [41] Thomas Verelst, Paul K Rubenstein, Marcin Eichner, Tinne Tuytelaars, and Maxim Berman. Spatial consistency loss for training multi-label classifiers from single-label annotations. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pages 3879 3889, 2023. [42] Bo Wang, Zhuowen Tu, and John K Tsotsos. Dynamic label propagation for semi-supervised multi-class multi-label classification. In Proceedings of the IEEE international conference on computer vision, pages 425 432, 2013. [43] Lichen Wang, Yunyu Liu, Can Qin, Gan Sun, and Yun Fu. Dual relation semi-supervised multilabel learning. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pages 6227 6234, 2020. [44] Yidong Wang, Hao Chen, Qiang Heng, Wenxin Hou, Yue Fan, Zhen Wu, Jindong Wang, Marios Savvides, Takahiro Shinozaki, Bhiksha Raj, Bernt Schiele, and Xing Xie. Freematch: Selfadaptive thresholding for semi-supervised learning. In The Eleventh International Conference on Learning Representations, 2023. [45] Chen Wei, Kihyuk Sohn, Clayton Mellina, Alan Yuille, and Fan Yang. Crest: A classrebalancing self-training framework for imbalanced semi-supervised learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 10857 10866, 2021. [46] Ming-Kun Xie, Jia-Hao Xiao, and Sheng-Jun Huang. Label-aware global consistency for multi-label learning with single positive labels. In Advances in Neural Information Processing Systems, 2022. [47] Yi Xu, Lei Shang, Jinxing Ye, Qi Qian, Yu-Feng Li, Baigui Sun, Hao Li, and Rong Jin. Dash: Semi-supervised learning with dynamic thresholding. In International Conference on Machine Learning, pages 11525 11536. PMLR, 2021. [48] Yang Yang, Zhao-Yang Fu, De-Chuan Zhan, Zhi-Bin Liu, and Yuan Jiang. Semi-supervised multi-modal multi-instance multi-label deep network with optimal transport. IEEE Transactions on Knowledge and Data Engineering, 33(2):696 709, 2019. [49] Yang Yang, Yi-Feng Wu, De-Chuan Zhan, Zhi-Bin Liu, and Yuan Jiang. Complex object classification: A multi-modal multi-instance multi-label deep network with optimal transport. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pages 2594 2603, 2018. [50] Wang Zhan and Min-Ling Zhang. Inductive semi-supervised multi-label learning with cotraining. In Proceedings of the 23rd ACM SIGKDD international conference on knowledge discovery and data mining, pages 1305 1314, 2017. [51] Bowen Zhang, Yidong Wang, Wenxin Hou, Hao Wu, Jindong Wang, Manabu Okumura, and Takahiro Shinozaki. Flexmatch: Boosting semi-supervised learning with curriculum pseudo labeling. Advances in Neural Information Processing Systems, 34:18408 18419, 2021. [52] Feipeng Zhao and Yuhong Guo. Semi-supervised multi-label learning with incomplete labels. In Twenty-fourth international joint conference on artificial intelligence, 2015. [53] Donghao Zhou, Pengfei Chen, Qiong Wang, Guangyong Chen, and Pheng-Ann Heng. Acknowledging the unknown for multi-label learning with single positive labels, 2022. [54] Yang Zou, Zhiding Yu, Xiaofeng Liu, BVK Kumar, and Jinsong Wang. Confidence regularized self-training. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 5982 5991, 2019. A Proof of Theorem 1 Theorem 3. Assume the estimated class proportion ˆγk = 1 n Pn i=1 I(yik = 1), and the true class proportion γ k = 1 m Pm j=1 I(yjk = 1) for any k [q], where n and m are the numbers of labeled and unlabeled examples that satisfy m >> n. Then, with the probability larger than 1 2n 1 2m 1, we have, k [q], |ˆγk γ k| log n Proof. The proof is mainly based on Hoeffding s inequality that can be defined as follows. Lemma 1. (Hoeffding s inequality). Let z1, ..., z N be independent random variables bounded by [ai, bi]. Then ˆz = 1 N PN i=1 zi obeys for any ν > 0 Pr(|ˆz E[ˆz] ν|) 2 exp( 2N 2ν2 PN i=1(bi ai)2 ). Let γk = p(yk = 1) represents the expected class proportion. According to Hoeffding s inequality, for any k [q], we have Pr(|ˆγk γk| log n 2n ) 1 2n 1 or equivalently, with the probability at least 1 2n 1, we have |ˆγk γk| log n/ 2n. Similarly, with the probability at least 1 2m 1, for any k [q], we have |γ k γk| log m/ 2m. By applying the triangle inequality, with the probability at least 1 2n 1 2m 1, for any k [q], we have |γ k ˆγk| log n which completes the proof. B Proof of Theorem 2 Theorem 4. Suppose that ℓ( ) is bounded by B. For some ϵ > 0, if Pm j=1 |I(fk(xj) τ(αk)) I(yjk = 1)|/m ϵ for any k [q], for any δ > 0, with probability at least 1 δ, we have R( ˆf) R(f ) 2q Bϵ + 4q LℓRN(F) + 2q B Proof. Before proving the theorem, we first provide two useful lemmas as follows. We primarily derive the uniform deviation bound between R(f) and b R(f), which is a simple extension of the result in the binary setting [31]. Lemma 2. Suppose that the loss function ℓis Lℓ-Lipschitz continuous w.r.t. θ. For any δ > 0, with probability at least 1 δ, we have |R(f) b R(f)| 2q LℓRn+m(F) + q B δ 2(n + m) (1) Proof. In order to prove this lemma, we define the Rademacher complexity of L and F with m + n training examples as follows: i=1 σi L(f(xi), yi) + j=1 σj L(f(xj), yj) Considering that L(f(x), y) = Pq k=1 ℓ(fk, yk), we have Rn+m(L F) q Rn+m(ℓ F) q LℓRn+m(F) (10) where the second line is due to the Lipschitz continuity of the loss function ℓ. Then, we proceed the proof by showing that the one direction supf F R(f) b R(f) is bounded with probability at least 1 δ/2, and the other direction can be proved similarly. Note that replacing an example (xj, yj) with another (x j, y j) leads to a change of supf F R(f) b R(f) at most q B n+m due to the fact that ℓis bounded by B. According to Mc Diarmid s inequality [31], for any δ > 0, with probability at least 1 δ/2, we have sup f F R(f) b R(f) E sup f F R(f) b R(f) δ 2(n + m) (11) According to the result in [31] (Theorem 3.3) that shows E[supf F R(f) b R(f)] 2Rm(F), by further considering the other direction supf F b R(f) R(f), with probability at least 1 δ, we have R(f) b R(f) 2q LℓRm(F) + q B δ 2n + m (12) which completes the proof. Then, we can bound the difference between b R(f) and b R (f) as follows Lemma 3. Suppose that ℓ( ) is bounded by B. For some ϵ > 0, if Pm j=1 |I(fk(xj) τ(αk)) I(yjk = 1)|/m ϵ for any k [q], for any f F, we have: b R u(f) b Ru(f) q Bϵ (13) Proof. Without loss of generality, assume that ϵ is the largest pseudo-labeling error among q classes, i.e., ϵ = maxk [q] Pm j=1 |I(fk(xj) τ(αk)) I(yjk = 1)|/m. Obviously, ϵ consists of exactly two types of pseudo-labeling error: Pm j=1 I(fk(xj) < τ(αk), yjk = 1) Pm j=1 I(fk(xj) τ(αk), yjk = 0) where ϵ1 calculates the proportion of positive labels being treated as negative ones, and ϵ0 calculates the proportion of negative labels being treated as positive ones. Then, we prove the following two sides, which provide the bounds for b R u(f). Firstly, we prove its upper bound: b R u(f) = 1 k=1 I(fk(xj) τ(αk))ℓ1(fk(xj)) + I(fk(xj) < τ(αk))ℓ0(fk(xj)) k=1 I(yjk = 1)ℓ1(fk(xj)) + I(yjk = 0)ℓ0(fk(xj)) + I(yjk = 0, fk(xj) τ(αk))ℓ1(fk(xj)) + I(yjk = 1, fk(xj) < τ(αk))ℓ0(fk(xj)) j=1 L(f(xj), yj) + ϵ0 k=1 ℓ1(fk(xj)) + ϵ1 k=1 ℓ0(fk(xj)) b Ru(f) + q Bϵ where the second line holds based on Eq.(14). Then, we prove its low bound: b R u(f) = 1 k=1 I(fk(xj) τ(αk))ℓ1(fk(xj)) + I(fk(xj) < τ(αk))ℓ0(fk(xj)) k=1 I(yjk = 1)ℓ1(fk(xj)) + I(yjk = 0)ℓ0(fk(xj)) I(yjk = 1, fk(xj) < τ(αk))ℓ1(fk(xj)) I(yjk = 0, fk(xj) τ(αk))ℓ0(fk(xj)) j=1 L(f(xj), yj) ϵ1 k=1 ℓ1(fk(xj)) ϵ0 k=1 ℓ0(fk(xj)) b Ru(f) q Bϵ By combining these two sides, we can obtain the following result: b R u(f) b Ru(f) q Bϵ (15) which concludes the proof. For any δ > 0, with probability at least 1 δ, we have: b Rl( ˆf) + b Ru( ˆf) + 2q LℓRn+m(F) + q B b Rl( ˆf) + b R u( ˆf) + q Bϵ + 2q LℓRN(F) + q B b Rl(f) + b R u(f) + q Bϵ + 2q LℓRN(F) + q B b Rl(f) + b Ru(f) + 2q Bϵ + 2q LℓRN(F) + q B R(f) + 2q Bϵ + 4q LℓRN(F) + 2q B where the first and fifth lines are based on Eq.(1), and second and fourth lines are due to Lemma 2. The third line is by the definition of ˆf. C Details of Experimental Settings Table 3 reports the detailed characteristics of four benchmark datasets. VOC is a popular multi-label dataset that has been divided into a trainval set containing 5,011 examples and a test set containing 4,952 examples from 20 object categories. MS-COCO is another widely used multi-label dataset, which consists of 82,081 training examples and 40,504 validation examples belonging to 80 different categories. In our experiments, the validation set is used for testing. NUS is incomplete online due to many invalid URLs. Our collection consists of 126,034 training images and 84,226 testing images from 81 classes. Comparing methods To validate the effectiveness of the proposed method, we compare it with four groups of methods: 1) Three instance-aware pseudo labeling methods: Top-1, which is similar to Table 3: The detailed characteristics of benchmark datasets. Dataset # Training # Testing # Classes VOC 5,011 4,952 20 COCO 82,081 16,416 80 NUS 126,034 84,226 81 Fix Match [38] that selects the most probable label as the ground-truth one; Top-k, which selects the top k rank labels according to the predicted probabilities as the true labels; IAT, which utilizes the instance-aware thresholding strategy to separates positive and negative labels. 2) Two state-of-the-art MLML methods: LL [19] (includes three variants LL-R, LL-Ct, and LL-Cp), which rejects or corrects the large-loss examples to prevent model from memorizing noisy labels; PLC [46], which performs the pseudo-labeling consistency regularization for recovering the labeling information of potential labels. 3) Two state-of-the-art SSL methods: Adsh [14], which solves the class-imbalance issue in SSL by performing adaptive thresholding; Free Match [44], which improves the performance of Fix Match by introducing a self-adaptive class fairness regularization. 4) One state-of-the-art SSMLL method: DRML [43], which performs pseudo labeling by exploring the feature distribution and the label correlation simultaneously. 5) Two baseline methods, BCE, which is the most commonly used loss function for multi-label classification; ASL [34], which improves BCE by utilizing the dynamic down-weighting and hard-thresholding techniques. DRML is the only deep SSMLL method whose source code could be found on the Internet. Furthermore, most MLML methods cannot be applied to the SSMLL scenario, since they assume that a subset of labels have been annotated for each training instance.