# debiased_selftraining_for_semisupervised_learning__418ce4c1.pdf Debiased Self-Training for Semi-Supervised Learning Baixu Chen , Junguang Jiang , Ximei Wang, Pengfei Wan , Jianmin Wang, Mingsheng Long B School of Software, BNRist, Tsinghua University, China Y-tech, Kuaishou Technology {chenbx18,jjg20}@mails.tsinghua.edu.cn, {jimwang,mingsheng}@tsinghua.edu.cn Deep neural networks achieve remarkable performances on a wide range of tasks with the aid of large-scale labeled datasets. Yet these datasets are time-consuming and labor-exhaustive to obtain on realistic tasks. To mitigate the requirement for labeled data, self-training is widely used in semi-supervised learning by iteratively assigning pseudo labels to unlabeled samples. Despite its popularity, self-training is well-believed to be unreliable and often leads to training instability. Our experimental studies further reveal that the bias in semi-supervised learning arises from both the problem itself and the inappropriate training with potentially incorrect pseudo labels, which accumulates the error in the iterative self-training process. To reduce the above bias, we propose Debiased Self-Training (DST). First, the generation and utilization of pseudo labels are decoupled by two parameter-independent classifier heads to avoid direct error accumulation. Second, we estimate the worst case of self-training bias, where the pseudo labeling function is accurate on labeled samples, yet makes as many mistakes as possible on unlabeled samples. We then adversarially optimize the representations to improve the quality of pseudo labels by avoiding the worst case. Extensive experiments justify that DST achieves an average improvement of 6.3% against state-of-the-art methods on standard semisupervised learning benchmark datasets and 18.9% against Fix Match on 13 diverse tasks. Furthermore, DST can be seamlessly adapted to other self-training methods and help stabilize their training and balance performance across classes in both cases of training from scratch and finetuning from pre-trained models. 1 Introduction Deep learning has achieved great success in many machine learning problems in the past decades, especially where large-scale labeled datasets are present. In real-world applications, however, manually labeling sufficient data is time-consuming and labor-exhaustive. To reduce the requirement for labeled data, semi-supervised learning (SSL) improves the data efficiency of deep models by learning from a few labeled samples and a large number of unlabeled samples [20, 30, 51, 7]. Among them, self-training is an effective approach to deal with the lack of labeled data. Typical self-training methods [30, 47] assign pseudo labels to unlabeled samples with the model s predictions and then iteratively train the model with these pseudo labeled samples as if they were labeled examples. Although self-training has achieved great advances in benchmark datasets, they still exhibit large training instability and extreme performance imbalance across classes. For instance, the accuracy of Fix Match [47], one of the state-of-the-art self-training methods, fluctuates greatly when trained from scratch (see Figure 7). Though its performance will gradually recover after a sudden sharp drop, this is still not expected, since pre-trained models are more often adopted [14, 7, 24] are improve data efficiency, and the performance of pre-trained models is difficult to recover after a drastic decline due Equal contribution. 36th Conference on Neural Information Processing Systems (Neur IPS 2022). to catastrophic forgetting [25]. Besides, although Fix Match improves the average accuracy, it also leads to the Matthew effect, i.e., the accuracy of well-behaved categories is further increased while that of poorly-behaved ones is decreased to nearly zero (see Figure 4). This is also not expected, since most machine learning models prefer performance balance across categories, even when the class imbalance exists in the training data [65]. The above findings are caused by the bias between the pseudo labeling function with the unknown target labeling function. Training with biased and unreliable pseudo labels has the chance to accumulate errors and ultimately lead to performance fluctuations. And for those poorly-behaved categories, the bias of the pseudo labels gets worse and will be further enhanced as self-training progresses, ultimately leading to the Matthew effect. We delved into the bias issues arising from the self-training process and found that they can be briefly grouped into two kinds: (1) Data bias, the bias inherent in the SSL tasks; (2) Training bias, the bias increment brought by self-training with incorrect pseudo labels. In this regard, we present Debiased Self-Training (DST), a novel approach to decrease the undesirable bias in self-training. Specifically, to reduce the training bias, the classifier head is only trained with clean labeled samples and no longer trained with unreliable pseudo-labeled samples. In other words, the generation and utilization of pseudo labels are decoupled to mitigate bias accumulation and boost the model s tolerance to biased pseudo labels. Further, to decrease the data bias which cannot be calculated directly, we turn to estimate the worst case of training bias that implicitly reflects the data bias. Then we optimize the representations to decrease the worst-case bias and thereby improve the quality of pseudo labels. The contributions of this work are summarized as follows: (1) We systematically identify the problem and analyze the causes of self-training bias in semi-supervised learning. (2) We propose DST, a novel approach to mitigate the self-training bias and boost the stability and performance balance across classes, which can be used as a universal add-on for mainstream self-training methods. (3) We conduct extensive experiments and validate that DST achieves an average boost of 6.3% against state-of-the-art methods on standard datasets and 18.9% against Fix Match on 13 diverse tasks. 2 Related Work 2.1 Self-training for semi-supervised learning Self-training [60, 43, 20, 30] is a widely-used approach to utilize unlabeled data. Pseudo Label [30], one popular self-training method, iteratively generates pseudo labels and utilizes them with the same model. However, this paradigm suffers from the problem of confirmation bias [1], where the learner struggles to correct its own mistakes when learning from inaccurate pseudo labels. The bias issue is also mentioned in Debias Match [54] where they define the bias as the quantity imbalance for each category. Note that the bias in our paper refers to the deviation between the pseudo labeling function and the ground truth labeling function, which is a more essential problem existing in most self-training methods. Recent works mainly tackle this bias issue from the following two aspects. Generate higher-quality pseudo labels. Mix Match [4] averages predictions from multiple augmentations as pseudo labels. Re Mix Match [3], UDA [57], and Fix Match [47] adopt confidence thresholds to generate pseudo labels on weakly augmented samples and utilize these pseudo-labels as annotations for strongly augmented samples. Dash [59] and Flex Match [62] dynamically adjust the thresholds in a curriculum learning manner. Label Propagation methods [46, 23] assign pseudo labels with the density of the local neighborhood. DASO [38] blends the confidence-based pseudo labels and density-based pseudo labels differently for each class. Meta Pseudo Labels [41] proposes to generate pseudo labels with a meta learner. Different from the above methods that manually design specific criteria to improve the quality of pseudo labels, we estimate the worst case of self-training bias and adversarially optimize the representations to improve the quality of pseudo labels automatically. Improve tolerance to inaccurate pseudo labels. To mitigate the confirmation bias, existing methods maintain a mismatch between the generation and utilization of pseudo labels. Temporal Ensembling [29] and Mean Teacher [51] generate pseudo labels from the average of previous predictions or an exponential moving average of the model, respectively. Noisy Student [58] assigns pseudo labels by a fixed teacher from the previous round. Co-training [5], MMT [17], Divide Mix [31] and Multi-head Tri-training [44] introduce multiple models or classifier heads and learn in an online mutual-teaching manner. In these methods, each classifier head is still trained with potentially incorrect pseudo labels generated by other heads. In contrast, in our method, the classifier head that generates pseudo labels is never trained with pseudo labels, leading to better tolerance to inaccurate pseudo labels (Table 3). 2.2 Self-supervised learning for semi-supervised learning Self-supervised methods [14, 21] are also used on unlabeled data to improve the model with few labeled samples, either in the pre-training stage [7, 2] or in the downstream tasks [53, 32]. However, the training of self-supervision usually relies on big data and heavy computation, which is not feasible in most applications. Besides, although these methods avoid the use of unreliable pseudo labels, it is difficult for them to learn task-specific information from unlabeled data for better performance. 2.3 Adversarial training for semi-supervised learning Some works introduce adversarial training [18] into semi-supervised learning. A line of works [37, 45, 12, 15] exploit fake samples from the generator by labeling them with a new generated class and forcing the discriminator to output class labels. Another line of works use adversarial training to construct adversarial samples [19], e.g., VAT [34] injects additive noise into input, VAd D [39] introduces adversarial Dropout [48] layers and RAT [50] expands the noise in VAT into a set of input transformations. These methods aim to impose a local smoothness on the model and do not involve training with pseudo labels. In contrast, in our method, the goal of the adversarial training is to estimate the worst case of pseudo labeling and then avoid such cases (Section 4.2). 3 Analysis of Bias in Self-Training In this section, we provide some analysis of where the bias in self-training comes from. Let P denote a distribution over input space X. For classification with K classes, let P k denote the class-conditional distribution of x conditioned on ground truth f (x) = k. Assume that pseudolabeler fpl is obtained via training a classifier on n labeled samples b Pn. Let M(fpl) {x : fpl(x) = f (x)} denote the mistaken pseudolabeled samples. The bias in the self-training refers to the deviation between the learned decision hyperplanes and the true decision hyperplanes, which can be measured by the fraction of incorrectly pseudolabeled samples in any classes B(fpl) = {P k(M(fpl))}K k=1 [55]. By analyzing self-training bias under different training conditions, we have several nontrivial findings. Sample 2 Sample 1 Sample 3 Sample 4 Figure 1: Effect of data sampling. Top-1 accuracy of 7 randomly selected categories when trained with different labeled data sampled from CIFAR-100. The same category (such as cattle) may have completely different accuracy in different samples. Following Fix Match [47], 4 labeled data are sampled for each category by default in our analysis. (a) Supervised Pre-Train (b) Unsupervised Pre-Train Figure 2: Effect of pre-trained representations. Accuracy of 7 randomly selected categories with different pre-trained models on CIFAR-100. Different pre-trained models show different category preferences. (a) Baseline (b) Fix Match Figure 3: Effect of self-training algorithm. Accuracy of 7 randomly selected categories with different training methods on CIFAR-100. Fix Match largely increases the bias of poorlybehaved categories (Matthew effect). The sampling of labeled data will largely influence the self-training bias. As shown in Figure 1, when the data sampling is different, the accuracy of the same category may vary dramatically. The reason is that the distances between different data points and the true decision hyperplanes are not the same, with some supporting data points closer and others far away. When there are few labeled data, there may be a big difference in the distances between supporting data of each category and the true decision hyperplanes, hence the learned decision hyperplanes will be biased towards some categories. The pre-trained representations also affect the self-training bias. Figure 2 shows that different pre-trained representations lead to different category bias, even if the pre-trained dataset and the downstream labeled dataset are both identical. One possible reason is that the representations learned by different pre-trained models focus on different aspects of the data [64]. Therefore, the same data could also have different distances to the decision hyperplanes in the representation level with different pre-trained models. Training with pseudo labels aggressively in turn enlarges the self-training bias on some categories. Figure 3 shows that after training with pseudo labels (e.g., using Fix Match), the performance gap for different categories greatly enlarges, with the accuracy of some categories increasing from 60% to 80% and that of some categories dropping from 15% to 0%. The reason is that for well-behaved categories, the pseudo labels are almost accurate, hence using them for training could further reduce the bias. Yet for many poorly-behaved categories, the pseudo labels are not reliable, and the common self-training mechanism that uses these incorrect pseudo labels to train the model will further increase the bias, and fail to correct it back in the follow-up training. This results in the Matthew effect. Figure 4: Error rate of pseudo labels in any classes on CIFAR-100 (Res Net50, 4 labels per category). Fix Match decreases the bias on well-behaved categories while increasing that of poorly-behaved categories. In contrast, DST effectively balances the performance between different categories. Based on the above observations, we divide the bias caused by self-training into two categories. Data bias: the bias inherent in semi-supervised learning tasks, such as the bias of sampling and pretrained representations on unlabeled data. Formally, data bias is defined as B(fpl( b Pn, ψ0)) B(f ) (blue area in Figure 4), where the pseudolabeler fpl( b Pn, ψ0) is obtained from a biased sampling b Pn with a biased parameter initialization ψ0. Training bias: the bias increment brought by some unreasonable training strategies. Formally, training bias is B(fpl( b Pn, ψ0, S)) B(fpl( b Pn, ψ0)) (yellow area in Figure 4) where fpl( b Pn, ψ0, S) is a pseudolabeler obtained with self-training strategy S. Next we will introduce how to reduce training bias and data bias in self-training (red area in Figure 4). 4 Debiased Self-Training In semi-supervised learning (SSL), we have a labeled dataset L = {(xl i, yl i}nl i=1 of nl labeled samples and an unlabeled dataset U = {(xu j )}nu j=1 of nu unlabeled samples, where the size of the labeled dataset is usually much smaller than that of the unlabeled dataset, i.e., nl nu. Denote ψ the feature generator, and h the task-specific head. The standard cross-entropy loss on weakly augmented labeled examples is LL(ψ, h) = 1 i=1 LCE (h ψ α)(xl i), yl i , (1) where α is the weak augmentation function. Since there are few labeled samples, the feature generator and the task-specific head will easily over-fit, and typical SSL methods use these pseudo labels on plenty of unlabeled data to decrease the generalization error. Different SSL methods design different pseudo labeling function bf [30, 59, 42]. Take Fix Match [47] for an instance. Fix Match first generates predictions bp = (h ψ α)(x) on a weakly augmented version of given unlabeled images, and adopts a confidence threshold τ to filter out unreliable pseudo labels, bfψ,h(x) = arg max bp, max bp τ, 1, otherwise, (2) where bfψ,h refers to the pseudo labeling by model h ψ, hyperparameter τ specifies the threshold above which a pseudo label is retained and 1 indicates that this pseudo label is ignored in training. Then Fix Match utilizes selected pseudo labels to train on strongly augmented unlabeled images, LU(ψ, h, bf) = 1 j=1 LCE (h ψ A)(xu j ), bf(xu j ) , (3) where bf is a notation of general pseudo labeling function and A is the strong augmentation function. As shown in Figure 5(a), the optimization objective for Fix Match is min ψ,h LL(ψ, h) + λLU(ψ, h, bfψ,h), (4) where λ is the trade-off between the loss on labeled data and that on unlabeled data. Fix Match filters out low-confidence samples during the pseudo labeling process, yet two issues remain: (1) The pseudo labels are generated and utilized by the same head, which leads to the training bias, i.e., the errors of the model might be amplified as the self-training progresses. (2) When trained with extreme few labeled samples, the problem of unreliable pseudo labeling caused by data bias cannot be ignored anymore even with the confidence threshold mechanism. To tackle the above issues, we propose two important designs to decrease training bias and data bias in Section 4.1 and 4.2 respectively. 4.1 Generate and utilize pseudo labels independently The training bias of Fix Match stems from the way of training on the pseudo labels generated by itself. To alleviate this bias, some methods turn to generate pseudo labels from a better teacher model, such as the moving average of the original model [51] in Figure 5(b) or the model obtained from the previous round of training [58] in Figure 5(c), and then utilize these pseudo labels to train both the feature generator ψ and the task-specific head h. However, there is still a tight relationship between the teacher model that generates pseudo labels and the student model that utilizes pseudo labels in the above methods, and the decision hyperplanes of the student model h ψ strongly depend on the biased pseudo labeling bf. As a result, training bias is still large in the self-training process. Forward Propagation (a) Pseudo Labeling / Fix Match (b) Mean Teacher (c) Noisy Student (d) Ours (DST) generator generator pseudo head Backward Propagation Labeled Data Pseudo Label Parameter Replacement Independent Model Copied Model Unlabeled Data Figure 5: Comparisons on how different self-training methods generate and utilize pseudo labels. (a) Pseudo Labeling and Fix Match generate and utilize pseudo labels on the same model. (b) Mean Teacher generates pseudo labels from the Exponential Moving Average (EMA) of the current model. (c) Noisy Student generates pseudo labels from the teacher model which is obtained from the previous round of training. (d) DST generates pseudo labels from head h and utilizes pseudo labels on a parameter independent pseudo head hpseudo. To further decrease the training bias when utilizing the pseudo labels, we optimize the task-specific head h, only with the clean labels on L and without any unreliable pseudo labels from U. To prevent the deep models from over-fitting to the few labeled samples, we still use pseudo labels, but only for learning a better representation. As shown in Figure 5(d), we introduce a pseudo head hpseudo, which is connected to the feature generator ψ and only optimized with pseudo labels from U. Then the training objective is min ψ,h,hpseudo LL(ψ, h) + λLU(ψ, hpseudo, bfψ,h), (5) where the pseudo labels are generated by head h and utilized by a completely parameter independent pseudo head hpseudo. Although h and hpseudo are fed with features from the same backbone network, their parameters are independent, thus training the pseudo head hpseudo with some wrong pseudo labels will not accumulate the bias of head h directly in the iterative self-training process. Note that the pseudo head hpseudo is only responsible for gradient backpropagation to the feature generator ψ during training and will be discarded during inference, and thus will introduce no inference cost. 4.2 Reduce generation of erroneous pseudo labels Section 4.1 presents a solution to reduce the training bias, yet the data bias still exists in the pseudo labeling bf. As shown in Figure 6(a), due to the data bias, labeled samples of each class have different distances to the decision hyperplanes in the representation space, which leads to a deviation between the learned hyperplanes and the real decision hyperplanes, especially when the size of labeled samples is very small. As a result, pseudo labeling bf is very likely to generate incorrect pseudo labels on unlabeled data points that are close to these biased decision hyperplanes. And our objective now is to optimize the feature representations to reduce the data bias, and finally improve the quality of pseudo labels. Since we have no labels for U, we cannot directly measure the data bias and thereby reduce it. Yet training bias has some correlations with data bias. Recall in Section 4.1, the task-specific head h is only optimized with clean labeled data, since optimization with incorrect pseudo labels will push the learned hyperplanes in a more biased direction and lead to the training bias. Therefore, training bias can be considered as the accumulation of data bias with inappropriate utilization of pseudo labels, which is training algorithm dependent. And the worst training bias that can be achieved by some self-training methods is a good measure of data bias. Specifically, the worst training bias corresponds to the worst possible head h learned by pseudo labeling, such that h predicts correctly on all the labeled samples L while making as many mistakes as possible on unlabeled data U, hworst(ψ) = arg max h LU(ψ, h , bfψ,h) LL(ψ, h ), (6) where the mistakes of h on unlabeled data are estimated by its discrepancy with the current pseudo labeling function bf. Equation 6 aims to find the worst-case of task-specific head h that might be learned in the future when trained with pseudo labeling on the current feature generator ψ and the Different Classes Unlabeled Data True Hyperplane Learnt Hyperplane Worst Hyperplane (a) (b) (c) Figure 6: Concept explanations. (a) Shift between the hyperplanes learned on limited labeled data and the true hyperplanes. (b) The worst hyperplanes are hyperplanes that correctly distinguish labeled samples while making as many mistakes as possible on unlabeled samples. (c) Feature representations are optimized to improve the performance of the worst hyperplanes. current data sampling. It is also the worst hyperplanes as shown in Figure 6(b), which deviates as much as possible from the currently learned hyperplanes while ensuring that all labeled samples are correctly distinguished. Note that Equation 6 measures the degree of data bias, which depends on the feature representations generated by ψ, thus we can adversarially optimize feature generator ψ to indirectly decrease the data bias, min ψ LU(ψ, hworst(ψ), bfψ,h) LL(ψ, hworst(ψ)). (7) As shown in Figure 6(c), Equation 7 encourages the feature of unlabeled samples to be distinguished correctly even by the worst hyperplanes, i.e., be generated far away from the current hyperplanes, thereby reducing the data bias in feature representations. Overall loss. The final objective of the Debiased Self-Training (DST) approach is to reduce both training bias and data bias. The overall loss function simultaneously decouples the generation and utilization of pseudo-labels and avoids the worst-case hyperplanes. This is achieved by unifying Equations 5 7 into a minimax game: min ψ,h,hpseudo max h LL(ψ, h) + LU(ψ, hpseudo, bfψ,h) + LU(ψ, h , bfψ,h) LL(ψ, h ) . (8) 5 Experiments Following [47, 59], we evaluate Debiased Self-Training (DST) with random initialization on common SSL datasets, including CIFAR-10 [28], CIFAR-100 [28], SVHN [35] and STL-10 [10]. Following [53], we also evaluate DST with both supervised pre-trained models and unsupervised pre-trained models on 11 downstream tasks, including (1) superordinate-level object classification: CIFAR-10 [28], CIFAR-100 [28], Caltech-101 [16]; (2) fine-grained object classification: Food-101 [6], CUB200-2011 [52], Stanford Cars [27], FGVC Aircraft [33], Oxford IIIT Pets [40], Oxford Flowers [36]; (3) texture classification: DTD [9]; (4) scene classification: SUN397 [56]. The complete training dataset size ranges from 2, 040 to 75, 750 and the number of classes ranges from 10 to 397. Following [26], we report mean accuracy per-class on Caltech-101, FGVC Aircraft, Oxford IIIT Pets, Oxford Flowers, and top-1 accuracy for other datasets. Following [47], we construct a labeled subset with 4 labels per category to verify the effectiveness of DST in extremely label-scarce settings. To make a fair comparison, we keep the labeled subset for each dataset the same throughout our experiments. For experiments with random initialization, we follow [47] and adopt Wide Res Net variants [61]. For experiments with pre-trained models, we adopt Res Net50 [22] with an input size of 224 224 and pre-trained on Image Net [13]. We adopt Mo Co v2 [8] as unsupervised pre-trained models. We compare our method with many state-of-the-art SSL methods, including Pseudo Label [30], Π-Model [29], Mean Teacher [51], VAT [34], ALI [15], RAT [50], UDA [57], Mix Match [4], Re Mix Match [3], Fix Match [47], Dash [59], Self-Tuning [53], Flex Match [62] and Debias Match [54]. When training from scratch, we adopt the same hyperparameters as Fix Match [47], with learning rate of 0.03, mini-batch size of 512. For other experiments, we use SGD with momentum 0.9 and learning rates in {0.001, 0.003, 0.01, 0.03}. The mini-batch size is set to 64 following [49]. For each image, we first apply random-resize-crop and then use Rand Augment [11] for strong augmentation A and random-horizontal-flip for weak augmentation α. More details on hyperparameter selection can be found in Appendix A.2. Each experiment is repeated three times with different random seeds. We have released a benchmark containing both the code for our method and that for all the baselines at https://github.com/thuml/Debiased-Self-Training. 5.1 Main results Table 1 shows that DST yields consistent improvement on all tasks. On the challenging CIFAR100 and STL-10 tasks, DST boosts the accuracy of Fix Match and Flex Match by 8.3% and 10.7%, respectively. Figure 7 depicts the top-1 accuracy during the training procedure on CIFAR-100. We observe that the performance of Fix Match suffers from significant fluctuations during training. In contrast, the accuracy of DST (Fix Match) increases steadily and surpasses the best accuracy of Fix Match by 10.9%, relatively. Note that the accuracy of Flex Match also drops by over 6% in the final stages of training while DST (Flex Match) suffers from a much smaller drop by reducing erroneous pseudo labels during the self-training process. Besides, DST also improves the performance balance across categories (see Appendix B.2). Table 1: Top-1 accuracy on standard SSL benchmarks (train from scratch, 4 labels per category). Method CIFAR-10 CIFAR-100 SVHN STL-10 Avg Psuedo Label [30] 25.4 12.6 25.3 25.3 22.2 VAT [34] 25.3 15.1 26.1 25.5 23.0 ALI [15] 25.9 12.4 28.5 24.1 22.7 RAT [50] 33.2 20.5 52.6 30.7 34.2 Mix Match [4] 52.6 32.4 57.5 45.1 46.9 UDA [57] 71.0 40.7 47.4 62.6 55.4 Re Mix Match [3] 80.9 55.7 96.6 64.0 74.3 Dash [59] 86.8 55.2 97.0 64.5 75.9 Fix Match [47] 87.2 50.6 96.5 67.1 75.4 DST (Fix Match) 89.3 56.1 96.7 71.0 78.3 Flex Match [62] 94.7 59.5 89.6 71.3 78.8 DST (Flex Match) 95.0 65.4 94.2 79.6 83.6 Figure 7: Top-1 accuracy on CIFAR-100 (train from scratch, 4 labels per category). 0 2x105 4x105 6x105 8x105 106 Accuracy (%) Fix Match DST (Fix Match) Flex Match DST (Flex Match) 5.2 Transfer from a pre-trained model Supervisied pre-training. Table 2 reveals that typical self-training methods, e.g. Fix Match, lead to relatively mild improvements with supervised pre-trained models, which is consistent with previous findings [49, 53]. In contrast, incorporating DST into Fix Match significantly boosts the performance and surpasses Fix Match by 19.9% on all datasets. With a pre-trained model, self-training has better training stability. Yet once the performance degradation occurs, the process is also irreversible (Appendix B.1), partly due to the catastrophic forgetting of pre-trained representation. Also, selftraining suffers from a more severe performance imbalance across classes (Appendix B.1). DST effectively tackles these issues, indicating the importance of reducing bias. Table 2: Comparison between DST and various baselines (Res Net50, supervised and unsupervised pre-trained, 4 labels per category). indicates a performance degradation compared with the baseline. Baseline 81.4 65.2 48.2 39.9 47.7 25.4 46.5 85.2 78.1 33.3 33.8 53.2 Pseudo Label [30] 86.3 83.3 54.7 41.0 50.2 27.2 54.3 92.3 87.8 41.4 38.0 59.7 Π-Model [29] 83.5 73.1 49.2 39.7 50.3 24.3 47.1 90.7 82.2 30.9 33.9 55.0 Mean Teacher [51] 83.7 82.1 56.0 37.9 51.6 30.7 49.6 91.0 82.8 39.1 40.3 58.6 VAT [34] 84.1 72.2 48.8 39.5 50.6 25.9 48.1 89.4 81.8 32.4 36.7 55.4 ALI [15] 82.2 69.5 46.3 36.4 50.5 21.3 42.5 82.9 77.4 29.8 31.7 51.9 RAT [50] 84.0 81.8 55.4 39.0 49.1 31.6 50.0 89.9 84.1 37.9 38.4 58.3 Mix Match [4] 85.4 82.8 53.5 41.8 50.1 24.7 51.7 91.5 83.3 42.5 38.2 58.7 UDA [57] 85.8 83.6 54.7 41.3 49.0 27.1 52.1 92.0 83.1 45.6 41.7 59.6 Fix Match [47] 86.3 84.6 53.1 41.3 48.6 25.2 52.3 93.2 83.7 46.4 37.1 59.3 Self-Tuning [53] 87.2 76.0 57.1 41.8 50.7 35.2 58.9 92.6 86.6 58.3 41.9 62.4 Flex Match [62] 87.1 89.0 63.4 48.3 52.5 34.0 54.9 94.5 88.3 57.5 49.5 65.4 Debias Match [54] 88.6 91.0 65.7 46.6 52.4 37.5 58.6 95.6 86.4 60.5 53.5 66.9 DST (Fix Match) 89.6 94.9 70.4 48.1 53.5 43.2 68.7 94.8 89.8 71.0 58.5 71.1 DST (Flex Match) 90.6 95.9 71.2 49.8 56.2 44.5 70.5 95.8 90.4 72.7 57.1 72.2 Unsupervised Baseline 79.5 66.6 46.5 38.1 47.9 28.7 37.5 87.7 60.0 38.1 32.9 51.2 Pseudo Label [30] 86.2 70.8 49.8 38.6 50.0 26.6 41.8 93.0 68.4 37.3 32.8 54.1 Π-Model [29] 80.1 76.2 44.8 37.8 50.0 23.5 31.6 93.1 62.8 25.6 30.4 50.5 Mean Teacher [51] 80.4 80.8 51.3 34.2 48.8 33.8 41.6 92.9 67.0 50.5 39.1 56.4 VAT [34] 79.9 73.8 45.1 38.3 49.2 24.2 36.4 92.4 61.7 29.9 33.1 51.3 ALI [15] 76.4 69.2 44.4 34.9 50.1 22.2 33.8 84.9 59.6 33.1 31.0 49.1 RAT [50] 80.9 79.5 52.4 37.0 50.4 30.1 40.7 91.8 70.5 47.9 35.6 56.1 Mix Match [4] 84.1 81.5 51.7 38.4 47.0 31.7 39.8 93.5 66.4 47.1 34.6 56.0 UDA [57] 85.0 87.4 53.6 42.3 46.2 35.7 41.4 94.1 69.3 51.5 39.3 58.7 Fix Match [47] 83.1 82.2 51.4 39.2 43.9 30.1 36.8 94.3 65.7 48.6 36.8 55.6 Self-Tuning [53] 81.6 63.6 47.8 38.8 45.5 31.4 41.6 91.0 66.9 52.0 34.0 54.0 Flex Match [62] 86.4 96.7 60.2 45.3 53.9 42.0 49.2 95.8 72.9 69.0 37.5 64.4 Debias Match [54] 86.4 96.3 66.3 44.5 53.9 44.8 51.2 95.4 70.9 72.5 53.6 66.9 DST (Fix Match) 90.1 95.0 68.2 46.8 54.2 47.7 53.6 95.6 75.4 72.0 57.1 68.7 DST (Flex Match) 90.4 96.9 68.9 48.8 55.9 47.3 55.2 96.4 75.1 74.6 56.9 69.7 Unsupervised pre-training. Table 2 shows that with unsupervised pre-trained models, more methods suffer from performance degradation after self-training on the unlabeled data. The difficulty comes from that the unsupervised pre-training task has a larger task discrepancy with the downstream classification tasks than the supervised pre-training task. Thus, the representations learned by unsupervised pre-trained models usually exhibit stronger data bias, and inappropriate usage of pseudo labels will lead to rapid accumulation errors and increase the training bias. By eliminating training bias and reducing data bias, DST brings improvement on all datasets and relatively outperforms Fix Match by 23.5% on average, superior to Flex Match and Debias Match in 9 and 10 tasks, respectively. 5.3 Ablation studies We examine the design of our method on CIFAR-100 in Table 3 and have the following findings. (1) Compared with Mutual Learning [63, 17], where two heads provide pseudo labels to each other, the independent mechanism in our method where one head is only responsible for generating pseudo labels and the other head only uses them for self-training can better reduce the training bias. (2) A nonlinear pseudo head is always better than a linear pseudo head. We conjecture that nonlinear projection can reduce the degeneration of representation with biased pseudo labels. (3) The worst-case estimation of pseudo labeling improves the performance by large margins. Table 3: Ablation study on CIFAR-100 with different pre-trained models (4 labels per category). Method Multiple Heads Linear Pseudo Head Nonlinear Pseudo Head Worst Case Estimation Supervised Unsupervised Pre-training Pre-training Fix Match 53.1 51.4 Mutual Learning 53.4 52.5 DST w/o worst 58.2 59.0 DST w/o worst 60.6 60.9 DST 70.4 68.2 5.4 Analysis To further investigate how DST improves pseudo labeling and self-training performance, we conduct some analysis on CIFAR-100. For simplicity, we only give the results with supervised pre-trained models. More comparisons can be found in Appendix B.4. DST improves both the quantity and quality of pseudo labels. As shown in Figures 8(a) and 8(b), Fix Match exploits unlabeled data aggressively, on average producing more than 70% pseudo labels during training. But the cost is that the accuracy of pseudo labels continues to drop, eventually falling below 60%, which is consistent with our motivation in Section 3 that inappropriate utilization of pseudo labels will in turn enlarges the training bias. On the contrary, the accuracy of pseudo labels in DST suffers from a smaller drop. Rather, it keeps rising afterward and exceeds 70% throughout the training. Besides, DST generates more pseudo labels in the later stages of training. 10K 20K 30K 40K Iterations Fix Match DST w/o worst DST (a) Quantity 10K 20K 30K 40K Iterations Accuracy (%) Fix Match DST w/o worst DST (b) Quality 10K 20K 30K 40K Iterations Imbalance Ratio Fix Match DST w/o worst DST (c) Quantity of bad classes 10 20 Number of Classes Accuracy (%) Fix Match DST w/o worst DST (d) Quality of bad classes Figure 8: The quantity and quality of pseudo labels on CIFAR-100 (Res Net50, supervised pre-trained). DST generates better pseudo labels for poorly-behaved classes. To measure the quantity of pseudo labels on poorly-behaved classes, we calculate the class imbalance ratio I on a class-balanced validation set, I = maxc N(c)/minc N(c ), where N(c) denotes the number of predictions that fall into category c. As shown in Figure 8(c), the class imbalance ratio of Fix Match rises rapidly and reaches infinity after 5000 iterations, indicating that the model completely ignores those poorlylearned classes. To measure the quality of pseudo labels on poorly-behaved classes, we calculate the average accuracy of 10 or 20 worst-behaved classes in Figure 8(d). The average accuracy on the worst 20 classes of Fix Match is only 1.0%. By reducing training bias with the pseudo head and data bias with the worst-case estimation, the average accuracy balloons to 28.5% and 34.5%, respectively. 5.5 Convergence and computation cost of the min-max optimization We optimize ψ and h with stochastic gradient descent alternatively. The optimization can be viewed as an alternative form of GAN [18]. Figure 9 shows that the worst-case error rate of h and worstcase loss in Equation 7 first increase (h dominates), and then gradually decrease and converge (ψ dominates). When training 1000k iterations on CIFAR-100 using 4 2080 Ti GPUs, Fix Match takes 104 hours while DST takes 111 hours, only a 7% increase in time. Note that DST introduces no additional computation cost during inference. Figure 9: Empirical error rate and loss (CIFAR-100). Error rate of h Error rate of h' Worst-case Loss Worst-case Loss Figure 10: DST as a general add-on on CIFAR-100. Pre-training Supervised Unsupervised Label Amount 400 1000 400 1000 Mean Teacher Base 56.0 67.0 51.3 63.5 DST 62.7 70.7 60.7 69.3 Noisy Student Base 52.8 64.3 55.6 65.8 DST 68.9 74.8 66.6 75.2 Divide Mix Base 55.8 67.5 53.6 64.9 DST 69.1 75.1 65.0 74.2 Fix Match Base 53.1 67.8 51.4 64.2 DST 70.4 75.6 68.2 76.8 Flex Match Base 63.4 71.2 60.2 71.1 DST 71.2 77.3 68.9 77.5 5.6 DST as a general add-on We incorporate DST into several representative self-training methods, including Fix Match [47], Mean Teacher [51], Noisy Student [58], Divide Mix [31] and Flex Match [62]. Implementation details of DST versions of these methods can be found in Appendix A.3. Table 10 compares the original and DST versions of these methods on CIFAR-100 with both supervised pre-trained and unsupervised pre-trained models. Results show that the proposed DST yields large improvement on all these selftraining methods, indicating that self-training bias widely exists in existing vanilla or sophisticated self-training methods and DST can serve as a universal add-on to reduce the bias. 6 Conclusion To mitigate the requirement for labeled data, pseudo labels are widely used on the unlabeled data, yet they suffer from severe confirmation bias. In this paper, we systematically delved into the bias issues and present Debiased Self-Training (DST), a novel approach to decrease bias in self-training. Experimentally, DST achieves state-of-the-art performance on 13 semi-supervised learning tasks and can serve as a universal and beneficial add-on for existing self-training methods. Acknowledgements This work was supported by the National Key Research and Development Plan (2021YFB1715200), National Natural Science Foundation of China (62022050 and 62021002), Beijing Nova Program (Z201100006820041), BNRist Innovation Fund (BNR2021RC01002), and Kuaishou Research Fund. [1] Eric Arazo, Diego Ortego, Paul Albert, Noel E O Connor, and Kevin Mc Guinness. Pseudolabeling and confirmation bias in deep semi-supervised learning. In IJCNN, 2020. [2] Mahmoud Assran, Mathilde Caron, Ishan Misra, Piotr Bojanowski, Armand Joulin, Nicolas Ballas, and Michael Rabbat. Semi-supervised learning of visual features by non-parametrically predicting view assignments with support samples. In ICCV, 2021. [3] David Berthelot, Nicholas Carlini, Ekin D Cubuk, Alex Kurakin, Kihyuk Sohn, Han Zhang, and Colin Raffel. Remixmatch: Semi-supervised learning with distribution alignment and augmentation anchoring. In ICLR, 2020. [4] David Berthelot, Nicholas Carlini, Ian Goodfellow, Nicolas Papernot, Avital Oliver, and Colin Raffel. Mixmatch: A holistic approach to semi-supervised learning. In Neur IPS, 2019. [5] Avrim Blum and Tom Mitchell. Combining labeled and unlabeled data with co-training. In Proceedings of the eleventh annual conference on Computational learning theory, 1998. [6] Lukas Bossard, Matthieu Guillaumin, and Luc Van Gool. Food-101 mining discriminative components with random forests. In ECCV, 2014. [7] Ting Chen, Simon Kornblith, Kevin Swersky, Mohammad Norouzi, and Geoffrey Hinton. Big self-supervised models are strong semi-supervised learners. In Neur IPS, 2020. [8] Xinlei Chen, Haoqi Fan, Ross Girshick, and Kaiming He. Improved baselines with momentum contrastive learning. ar Xiv preprint ar Xiv:2003.04297, 2020. [9] Mircea Cimpoi, Subhransu Maji, Iasonas Kokkinos, Sammy Mohamed, and Andrea Vedaldi. Describing textures in the wild. In CVPR, 2014. [10] Adam Coates, Andrew Ng, and Honglak Lee. An analysis of single-layer networks in unsupervised feature learning. In AISTATS, 2011. [11] Ekin D Cubuk, Barret Zoph, Jonathon Shlens, and Quoc V Le. Randaugment: Practical automated data augmentation with a reduced search space. In CVPR, 2020. [12] Zihang Dai, Zhilin Yang, Fan Yang, William W Cohen, and Russ R Salakhutdinov. Good semi-supervised learning that requires a bad gan. In Neur IPS, 2017. [13] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In CVPR, 2009. [14] Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. Bert: Pre-training of deep bidirectional transformers for language understanding. In NAACL, 2019. [15] Vincent Dumoulin, Ishmael Belghazi, Ben Poole, Olivier Mastropietro, Alex Lamb, Martin Arjovsky, and Aaron Courville. Adversarially learned inference. In ICLR, 2017. [16] Li Fei-Fei, R. Fergus, and P. Perona. Learning generative visual models from few training examples: An incremental bayesian approach tested on 101 object categories. In CVPR, 2004. [17] Yixiao Ge, Dapeng Chen, and Hongsheng Li. Mutual mean-teaching: Pseudo label refinery for unsupervised domain adaptation on person re-identification. In ICLR, 2020. [18] Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial nets. In Neur IPS, 2014. [19] Ian J Goodfellow, Jonathon Shlens, and Christian Szegedy. Explaining and harnessing adversarial examples. In ICLR, 2015. [20] Yves Grandvalet and Yoshua Bengio. Semi-supervised learning by entropy minimization. In Neur IPS, 2005. [21] Kaiming He, Haoqi Fan, Yuxin Wu, Saining Xie, and Ross Girshick. Momentum contrast for unsupervised visual representation learning. In CVPR, 2020. [22] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In CVPR, 2016. [23] Ahmet Iscen, Giorgos Tolias, Yannis Avrithis, and Ondrej Chum. Label propagation for deep semi-supervised learning. In CVPR, 2019. [24] Junguang Jiang, Yang Shu, Jianmin Wang, and Mingsheng Long. Transferability in deep learning: A survey, 2022. [25] James Kirkpatrick, Razvan Pascanu, Neil Rabinowitz, Joel Veness, Guillaume Desjardins, Andrei A. Rusu, Kieran Milan, John Quan, Tiago Ramalho, Agnieszka Grabska-Barwinska, Demis Hassabis, Claudia Clopath, Dharshan Kumaran, and Raia Hadsell. Overcoming catastrophic forgetting in neural networks. Proceedings of the National Academy of Sciences, 2017. [26] Simon Kornblith, Jonathon Shlens, and Quoc V Le. Do better imagenet models transfer better? In CVPR, 2019. [27] Jonathan Krause, Jia Deng, Michael Stark, and Li Fei-Fei. Collecting a large-scale dataset of fine-grained cars. In FGVC, 2013. [28] Alex Krizhevsky, Geoffrey Hinton, et al. Learning multiple layers of features from tiny images. Technical report, University of Toronto, 2009. [29] Samuli Laine and Timo Aila. Temporal ensembling for semi-supervised learning. In ICLR, 2017. [30] Dong-Hyun Lee. Pseudo-label: The simple and efficient semi-supervised learning method for deep neural networks. In ICML, 2013. [31] Junnan Li, Richard Socher, and Steven CH Hoi. Dividemix: Learning with noisy labels as semi-supervised learning. In ICLR, 2020. [32] Junnan Li, Caiming Xiong, and Steven CH Hoi. Comatch: Semi-supervised learning with contrastive graph regularization. In ICCV, 2021. [33] Subhransu Maji, Esa Rahtu, Juho Kannala, Matthew Blaschko, and Andrea Vedaldi. Finegrained visual classification of aircraft. ar Xiv preprint ar Xiv:1306.5151, 2013. [34] Takeru Miyato, Shin-ichi Maeda, Masanori Koyama, and Shin Ishii. Virtual adversarial training: a regularization method for supervised and semi-supervised learning. In TPAMI, 2018. [35] Yuval Netzer, Tao Wang, Adam Coates, Alessandro Bissacco, Bo Wu, and Andrew Y Ng. Reading digits in natural images with unsupervised feature learning. In Neur IPS, 2011. [36] Maria-Elena Nilsback and Andrew Zisserman. Automated flower classification over a large number of classes. In ICVGIP, 2008. [37] Augustus Odena. Semi-supervised learning with generative adversarial networks. ar Xiv preprint ar Xiv:1606.01583, 2016. [38] Youngtaek Oh, Dong-Jin Kim, and In So Kweon. Daso: Distribution-aware semantics-oriented pseudo-label for imbalanced semi-supervised learning. In CVPR, 2022. [39] Sungrae Park, Jun Keon Park, Su-Jin Shin, and Il-Chul Moon. Adversarial dropout for supervised and semi-supervised learning. In AAAI, 2018. [40] Omkar M Parkhi, Andrea Vedaldi, Andrew Zisserman, and CV Jawahar. Cats and dogs. In CVPR, 2012. [41] Hieu Pham, Zihang Dai, Qizhe Xie, and Quoc V Le. Meta pseudo labels. In CVPR, 2021. [42] Mamshad Nayeem Rizve, Kevin Duarte, Yogesh S Rawat, and Mubarak Shah. In defense of pseudo-labeling: An uncertainty-aware pseudo-label selection framework for semi-supervised learning. In ICLR, 2021. [43] Chuck Rosenberg, Martial Hebert, and Henry Schneiderman. Semi-supervised self-training of object detection models. In WACV, 2005. [44] Sebastian Ruder and Barbara Plank. Strong baselines for neural semi-supervised learning under domain shift. In ACL, 2018. [45] Tim Salimans, Ian Goodfellow, Wojciech Zaremba, Vicki Cheung, Alec Radford, and Xi Chen. Improved techniques for training gans. In Neur IPS, 2016. [46] Weiwei Shi, Yihong Gong, Chris Ding, Zhiheng Ma Xiaoyu Tao, and Nanning Zheng. Transductive semi-supervised deep learning using min-max features. In ECCV, 2018. [47] Kihyuk Sohn, David Berthelot, Chun-Liang Li, Zizhao Zhang, Nicholas Carlini, Ekin D Cubuk, Alex Kurakin, Han Zhang, and Colin Raffel. Fixmatch: Simplifying semi-supervised learning with consistency and confidence. In Neur IPS, 2020. [48] Nitish Srivastava, Geoffrey Hinton, Alex Krizhevsky, Ilya Sutskever, and Ruslan Salakhutdinov. Dropout: a simple way to prevent neural networks from overfitting. In ICML, 2014. [49] Jong-Chyi Su, Zezhou Cheng, and Subhransu Maji. A realistic evaluation of semi-supervised learning for fine-grained classification. In CVPR, 2021. [50] Teppei Suzuki and Ikuro Sato. Adversarial transformations for semi-supervised learning. In AAAI, 2020. [51] Antti Tarvainen and Harri Valpola. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. In Neur IPS, 2017. [52] Catherine Wah, Steve Branson, Peter Welinder, Pietro Perona, and Serge Belongie. The caltechucsd birds-200-2011 dataset. Technical Report CNS-TR-2011-001, California Institute of Technology, 2011. [53] Ximei Wang, Jinghan Gao, Mingsheng Long, and Jianmin Wang. Self-tuning for data-efficient deep learning. In ICML, 2021. [54] Xudong Wang, Zhirong Wu, Long Lian, and Stella X Yu. Debiased learning from naturally imbalanced pseudo-labels for zero-shot and semi-supervised learning. In CVPR, 2022. [55] Colin Wei, Kendrick Shen, Yining Chen, and Tengyu Ma. Theoretical analysis of self-training with deep networks on unlabeled data. In ICLR, 2021. [56] Jianxiong Xiao, James Hays, Krista A Ehinger, Aude Oliva, and Antonio Torralba. Sun database: Large-scale scene recognition from abbey to zoo. In CVPR, 2010. [57] Qizhe Xie, Zihang Dai, Eduard Hovy, Minh-Thang Luong, and Quoc V Le. Unsupervised data augmentation for consistency training. In Neur IPS, 2020. [58] Qizhe Xie, Minh-Thang Luong, Eduard Hovy, and Quoc V Le. Self-training with noisy student improves imagenet classification. In CVPR, 2020. [59] Yi Xu, Lei Shang, Jinxing Ye, Qi Qian, Yu-Feng Li, Baigui Sun, Hao Li, and Rong Jin. Dash: Semi-supervised learning with dynamic thresholding. In ICML, 2021. [60] David Yarowsky. Unsupervised word sense disambiguation rivaling supervised methods. In ACL, 1995. [61] Sergey Zagoruyko and Nikos Komodakis. Wide residual networks. In BMVC, 2016. [62] Bowen Zhang, Yidong Wang, Wenxin Hou, Hao Wu, Jindong Wang, Manabu Okumura, and Takahiro Shinozaki. Flexmatch: Boosting semi-supervised learning with curriculum pseudo labeling. In Neur IPS, 2021. [63] Ying Zhang, Tao Xiang, Timothy M Hospedales, and Huchuan Lu. Deep mutual learning. In CVPR, 2018. [64] Nanxuan Zhao, Zhirong Wu, Rynson W. H. Lau, and Stephen Lin. What makes instance discrimination good for transfer learning? In ICLR, 2021. [65] Bolei Zhou, Agata Lapedriza, Aditya Khosla, Aude Oliva, and Antonio Torralba. Places: A 10 million image database for scene recognition. In TPAMI, 2018. 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] (b) Did you describe the limitations of your work? [No] (c) Did you discuss any potential negative societal impacts of your work? [N/A] (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [N/A] (b) Did you include complete proofs of all theoretical results? [N/A] 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] We include the code in the supplemental material and the data is publicly available. (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] See Section 5 and Appendix A. (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [No] To be consistent with previous paper. we report the average performance with 3 seeds. (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] We include the type of resources in Appendix A 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [Yes] We cite the creators in Section 5. (b) Did you mention the license of the assets? [Yes] (c) Did you include any new assets either in the supplemental material or as a URL? [No] We do not use new assets. (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [Yes] All datasets are public. (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [Yes] All datasets are legal. 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]