# recursivemix_mixed_learning_with_history__31c594eb.pdf Recursive Mix: Mixed Learning with History Lingfeng Yang1#, Xiang Li2,1#, Borui Zhao3, Renjie Song3, Jian Yang1 1Nanjing University of Science and Technology, 2Nankai University, 3Megvii Technology {yanglfnjust, csjyang}@njust.edu.cn, xiang.li.implus@nankai.edu.cn zhaoborui.gm@gmail.com, songrenjie@megvii.com Mix-based augmentation has been proven fundamental to the generalization of deep vision models. However, current augmentations only mix samples from the current data batch during training, which ignores the possible knowledge accumulated in the learning history. In this paper, we propose a recursive mixed-sample learning paradigm, termed Recursive Mix (RM), by exploring a novel training strategy that leverages the historical input-prediction-label triplets. More specifically, we iteratively resize the input image batch from the previous iteration and paste it into the current batch while their labels are fused proportionally to the area of the operated patches. Furthermore, a consistency loss is introduced to align the identical image semantics across the iterations, which helps the learning of scaleinvariant feature representations. Based on Res Net-50, RM largely improves classification accuracy by 3.2% on CIFAR-100 and 2.8% on Image Net with negligible extra computation/storage costs. In the downstream object detection task, the RM-pretrained model outperforms the baseline by 2.1 AP points and surpasses Cut Mix by 1.4 AP points under the ATSS detector on COCO. In semantic segmentation, RM also surpasses the baseline and Cut Mix by 1.9 and 1.1 m Io U points under Uper Net on ADE20K, respectively. Codes and pretrained models are available at https://github.com/implus/Recursive Mix. 1 Introduction Deep convolutional neural networks (CNN) have made great progress in many computer vision tasks such as image classification [26, 29, 62], object detection [25, 71, 42], semantic segmentation [72, 67], etc. Despite achieving better performances, the models inevitably become more complex and incur over-fitting risks. To overcome this problem, methods like regularization [53, 30] and data augmentation [15, 70, 68] are developed consequently. Recently, mixed-based data augmentations [70, 68, 59, 60, 34, 13] have been proven to enhance the model generalization capability. Following the popular Mixup [70] and Cut Mix [68], which optimize networks based on mixed pixels and fused labels of two images, improvement has been made on aspects of the mix level [59, 18, 39] and mix region [34, 13, 60]. However, existing methods only mix samples from the current data batch. As a result, they fail to capitalize on the potential knowledge in their learning history. To leverage the historical knowledge, we propose an efficient and effective augmentation approach termed Recursive Mix (RM) that makes use of the historical input-prediction-label triplets. Different Corresponding author. # Equal contributions. Lingfeng Yang, Xiang Li and Jian Yang are with Jiangsu Key Lab of Image and Video Understanding for Social Security, and Key Lab of Intelligent Perception and Systems for High-Dimensional Information of Ministry of Education, Nanjing University of Science and Technology, Nanjing, 210094, P.R. China. 36th Conference on Neural Information Processing Systems (Neur IPS 2022). Iteration 𝑡-1 Iteration 𝑡+1 Model Model Model Historical Label Historical Input Iteration 𝑡 Resize Fill Iteration 𝑡+2 One-hot Label 𝑦𝑡 1 One-hot Label 𝑦𝑡 One-hot Label 𝑦𝑡+1 One-hot Label 𝑦𝑡+2 Historical Prediction KL divergence KL divergence KL divergence GAP GAP Ro IAlign Ro IAlign GAP Ro IAlign GAP ℋ ℋ ℋ ℋ ℋ ℋ ℋ Figure 1: Illustration of the Recursive Mix, which leverages the historical input-prediction-label triplets in a recursive paradigm. The historical input images are resized and then mixed into current ones where labels are fused proportionally (λ and 1 λ) to the area of operating patches. from the conventional mix-based approaches which only fuse the training images in the current batch, RM further leverages the last mixed inputs and labels to form new ones in a recursive paradigm. It aims at continuously reusing the augmented data and supervisions in previous iterations (Fig. 1). In addition, a consistency loss is introduced to align two spatial representations derived from the identical instance. RM shows significant advantages over its competitive counterparts [68, 70] on classification tasks and its space-aware features can potentially benefit the downstream tasks such as object detection and segmentation as they are more sensitive to the spatial understanding of the image. The advanced performance of RM will be illustrated later in our methods and experiments. Mixup Cut Mix RM Consistency Figure 2: Three benefits of the proposed Recursive Mix, which iteratively leverages the historical input-predictionlabel triplets: 1) an enlarged diversity of data space and richer supervisions, 2) adequate training signals for an instance with multi-scale/-space views, and 3) explicit learning on the spatial semantic consistency which can further benefit the downstream tasks. Another notable property of the proposed RM is that it hardly increases the training/inference budget (Table 11) but only consumes an extremely negligible amount of storage complexity (only a mini-batch of the historical input, prediction, and label) and a lightweight additional module (only an Ro I [25] operator and an optional fully connected layer). Experimental evidence in CIFAR and Image Net classification benchmarks demonstrate its effectiveness. Specifically, RM shows a gain of absolutely 3.2% points over the baseline model and outperforms Cut Mix [68] by 1.1% points in the CIFAR-100 dataset under Dense Net161 [29]. In the Image Net dataset, RM improves by 2.9% and 2.7% points based on the Res Net-50 [26] and PVTv2B1 [61] baselines while outperforming Cut Mix by 0.6% and 0.7% points, respectively. In terms of the downstream tasks, RM-pretrained models improve the baseline by 1 3 box AP points (object detection) and 1 2 mask AP points (instance segmentation) on the COCO dataset and 1 3 m Io U (semantic segmentation) on ADE20K while consistently outperforming models pretrained by Cut Mix. Notably, under the one-stage detector ATSS [71], the RMpretrained model under Res Net-50 outperforms the baseline and Cut Mix by 2.1 AP and 1.4 AP points, respectively. 2 Related Work Regularization Methods. There are many other types of regularization techniques that focus on the feature [53, 20, 30], data [15, 8], label [49], data-label pair [70, 68, 34] and feature-label pair [59]. Dropout [53], Drop Block [20] and Stochastic Depth [30] introduce the stochastic feature drop during training in an element-wise, block-wise, and path-wise way, respectively. Cutout [15] and Grid Mask [8] randomly/systematically erase regional pixels on images while Label Smoothing [49] tries to prevent the classifiers from being too confident on a certain category by slightly modifying the training labels into a soft distribution [27]. Mixup [70] and Cut Mix [68] both combine two samples, where the corresponding training label is given by the linear interpolation of two one-hot labels. They differ in the detailed combining strategy of the input images: Mixup employs pairwise linear combination but Cut Mix adopts a regional crop-and-paste operation. Saliency Mix [58], Puzzle Mix [34], Co-Mixup [33], and Super Mix [13] further exploit the salient regions and local statistics for optimal Mixup. Attentive Cut Mix [60] and Snap Mix [31] take the activation map of the correct label as a confidence indicator for selecting the semantically meaningful mix regions. Trans Mix [5] requires the attention map of the self-attention module to re-weight the targets. Manifold Mixup [59], Patch Up [18] and Mo Ex [39] perform a feature-level interpolation over two samples to prevent overfitting the intermediate representations. Recently, Style Mix [28] employs style transfer to enrich the generated mixed images. The proposed Recursive Mix (RM) shares similarities with Cut Mix by combining data samples in a regional replacement manner, but differs essentially from all the existing regularizers in that RM is the first to exploit the historical information. In the experiment part, we will show the comparisons between the proposed RM and other regularization methods. Contrastive Learning. Contrastive learning [22] aims to minimize the distances between the positive pairs by consistency losses [22] which measure their similarities in a representation space or a probability distribution space. In recent years, contrastive learning has made a lot of progress in the fields of computer vision [24, 51, 2, 57, 1, 9, 66, 50, 4, 10, 19, 11, 21, 38] and natural language processing [64, 48, 32]. Positive pairs can be constructed in many ways. In the fields of selfsupervised learning and semi-supervised learning, many works [2, 24, 9, 4, 21] have used two augmented inputs of one instance to form positive pairs. In terms of generating positive pairs with different model structures or weights, [64, 52] employ Dropout [53] twice on one model, and [24, 57] make a difference in weights through Exponential Moving Average (EMA) and separate projection heads. The above methods essentially require Siamese networks [3] or the need to pass one instance through a model twice to obtain positive pairs, which involves additional computational costs and large architectural designs. Another way is to record historical representations or predictions and compare them with their current outputs [66, 35, 38, 65]. However, using a memory bank to store key statistics for all instances consumes a huge memory cost and is not applicable to even bigger datasets. Our method employs the idea of contrastive learning from the learning history while consuming extremely negligible storage and computational complexity. Learning with History. In the training history, the accumulated elements are a large amount of wealth for the improvement of learning deep models. Existing methods have explored the usage across model parameters [4, 10, 24, 19, 11, 21], gradients [54, 17, 69, 36, 47, 44], predictions [35, 38], and all-level feature maps [63, 65, 66, 24] applied during the training process. More related works refer to the Supplementary Material. Different from existing practices, we exploit a novel aspect of the historical input-prediction-label triplets to improve the generalization of vision models. In this section, we introduce the proposed Recursive Mix (RM) approach and discuss its properties. 3.1 Recursive Mix Incon sistency Cat Dog Bird Figure 3: Cut operation may lead to the inconsistency between input and label under the proposed historical operation, but Resize can correctly preserve the consistency. We attempt to exploit the historical input-predictionlabel triplets of a learning instance to improve the generalization of vision models. A practical idea is to reuse the past input-prediction-label triplets to construct diversely mixed training samples with the current batch (see Fig. 1). To be specific, the historical input images are resized and then filled into the current ones where the labels are fused proportionally to the area of the mixed patches. Interestingly, such an operation formulates an exact recursive paradigm where each instance can have a gradually shrunken size of view during training, thus we term the method Recursive Mix (RM). Specifically at iteration t, the preliminary of RM is to generate a new training sample (ext, eyt) by combining the current sample (xt, yt) and the past historical one (xh, yh). The new training sample (ext, eyt) participates in training with the original loss objectives LCE and then updates the historical pair (xh, yh) iteratively: ext = M Resize Fill(xh, M) + (1 M) xt, eyt = λtyh + (1 λt)yt, where M {0, 1}W H is the binary mask indicating the place for the historical input to fill, 1 is the binary mask filled with ones and is the element-wise product. The combination ratio λ (λt indicates λ at moment t) is sampled from the uniform distribution U[0, α] where α defaults to 0.5. This setting is different from the symmetric beta distribution in Mixup [70] and Cut Mix [68] because λ is not suggested to be quite large for it denotes the proportion of the historical images in our case. Otherwise, the information of the new data points has a risk of being discarded. The function Resize Fill(xh, M) means resizing the image xh and filling it back exactly into its rectangle sub-region where M = 1. The resized height and width are also determined by the rectangle area. This design is to make the recursive mechanism reasonable since the original cut operation in Cut Mix [68] could possibly lose former information, which would cause the inconsistency problem with the label (see Fig. 3). The binary mask M is generated by filling in with 1 inside the sampled bounding box coordinates B = (rx, ry, rw, rh), otherwise defaulting to 0. Similar to Cut Mix [68], the size and position of the rectangular region B are uniformly sampled according to: rx U(0, W), rw = W ry U(0, H), rh = H As depicted in Fig. 2, the resize-filled part within the current input is semantically consistent with the historical input regardless of proportion and scale. Inspired by contrastive learning [22, 24, 4], we optimize the KL divergence between the cross-iteration predictions of the two corresponding regions. To be specific, we record the historical prediction ph which is updated by the original outputs after network function F( ), global average pooling GAP( ), and a linear layer H( ) in the last iteration: ept 1 = H(GAP(F(ext 1))), ph = ept 1. (3) In the current iteration, we obtain the corresponding prediction of local features by 1 1 Ro IAlign [25] aligning with the computed coordinates B of Recursive Mix: ept roi = H (Ro IAlign(F(ext), B)), (4) where F denotes the backbone function and H, H denote the final linear classification layer. By default, the parameters are not shared between layers H and H , which is purely suggested by the experiments in Table 4a. The total loss function to train our model is: L = LCE(ext, eyt) + ωλt LKL(ept roi, ph), (5) where LCE denotes the cross-entropy loss and LKL denotes the consistency loss. Note that ph is the historical prediction from the last iteration, and the consistency loss weight ω is set to 0.1 unless otherwise stated. In addition, we also use the mixed ratio λ representing the proportion of the resize-filled historical image to weight LKL. It makes sense that the confidence of LKL relates to λ as a smaller λ (smaller image size to be filled) usually causes the Ro I feature to lose more spatial semantic information. 3.2 Discussion Similar to Cut Mix/Mixup, RM is simple and hardly introduces additional computational costs (see Table 11). During training, only a mini-batch of the historical input, label, and prediction (i.e., (xh, yh, ph)) needs to be stored and updated iteratively. The Ro I [25] operator and an individual fully connected layer are considerably lightweight and free of deployment. The additional storage cost is extremely negligible compared to the entire model, data, and feature maps, thus making RM very efficient to train any network architecture. As illustrated in Fig. 2, introducing the historical design of input, label, and prediction into augmenting the mixed training pairs recursively and bridging the spatial semantics consistency has three obvious Pyramid Net-200, α=240 (300 epochs) Top-1 Err (%) Baseline 3.85 + Label Smoothing [49] 3.74 + Drop Block [20] 3.27 + Stochastic Depth [30] 3.11 + Cutout [15] 3.10 + Mixup (α=1.0) [70] 3.09 + Manifold Mixup (α=1.0) [59] 3.15 + Cut Mix [68] 2.88 + Mo Ex [39] 3.44 + Style Cut Mix (auto-γ) [28] 2.55 + RM (ours) 2.35 Table 1: Comparison of state-of-the-art regularization methods in CIFAR-10 under 300 epochs settings, the same as that of the Cut Mix [68] paper. Results are reported as average over 3 runs. Model RS HIS CL Top-1 Err (%) Pyramid Net 16.67 + Cut Mix [68] 15.59 + RM (ours) 15.36 14.81 14.65 Table 2: Ablation studies on the individual gain by each component of RM in CIFAR-100 with Pyramid Net-164, α=270 under the 200epoch setting. The results are averaged over 3 runs. RS : Resize strategy. HIS : Historical mix. CL : Consistency loss. Model (200 epochs) Type Top-1 Err (%) Res Net-18 [26] Baseline 21.70 + Mixup [70] 20.99 + Cut Mix [68] 19.61 + RM (ours) 18.64 Res Net-34 [26] Baseline 20.62 + Mixup [70] 19.19 + Cut Mix [68] 17.89 + RM (ours) 17.15 Dense Net-121 [29] Baseline 19.51 + Mixup [70] 17.71 + Cut Mix [68] 17.21 + RM (ours) 16.22 Dense Net-161 [29] Baseline 18.78 + Mixup [70] 16.84 + Cut Mix [68] 16.64 + RM (ours) 15.54 Pyramid Net-164, α=270 [23] Baseline 16.67 + Mixup [70] 16.02 + Cut Mix [68] 15.59 + RM (ours) 14.65 Table 3: Performance on various architectures in CIFAR-100 under 200 epochs. The proposed RM (α=0.5, ω=0.1) shows consistent improvements over other competitive approaches. We report an average of 3 runs. 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 Hyper-parameter Top-1 Err (%) 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 Hyper-parameter Top-1 Err (%) Figure 4: The effect of hyperparameter α on CIFAR-10 and CIFAR100 datasets. Standard deviation is also plotted and another hyperparameter ω is fixed to 0.1. Hyper-parameter Top-1 Err (%) Hyper-parameter Top-1 Err (%) Figure 5: The effect of hyperparameter ω on CIFAR-10 and CIFAR100 datasets. Standard deviation is also plotted and another hyperparameter α is fixed to 0.5. advantages: 1) RM encourages more diversity in data space as it exhaustively explores exponential combinations of multiple instances along with their corresponding richer training signals within a single iteration. 2) RM can explicitly provide continuous training supervisions for an instance with multi-scale and spatial-variant views. 3) RM makes use of the identical part of inputs from two adjacent iterations through contrastive learning via the consistency loss, which helps the model learn spatial-correlative semantic representation. These properties thus lead to the better generalization of computer vision models and can potentially benefit downstream tasks, e.g., object detection and semantic segmentation with enhanced spatial representation ability. Interestingly these properties can be simultaneously achieved by a simple operation of iteratively using the historical input-predictionlabel triplets. 4 Experiment In this section, we evaluate RM on image recognition tasks to show the effectiveness of the usage of historical input-prediction-label triplets. More details can be found in the Supplementary Material. 4.1 CIFAR Classification Dataset. The two CIFAR datasets [37] consist of colored natural scene images, each with 32 32 pixels in total. The train and test sets have 50K images and 10K images respectively. CIFAR-10 has 10 classes and CIFAR-100 has 100. Setup. We conduct two major training settings: 200-epoch and 300-epoch training, respectively. For 200-epoch training, we employ SGD with a momentum of 0.9, a weight decay of 5 10 4, and Top-1 Err (%) CIFAR-10 CIFAR-100 Image Net Unshared weights 2.35 18.64 20.80 Shared weights 2.61 18.66 20.92 (a) Comparisons of linear layer weights. Top-1 Err (%) CIFAR-10 CIFAR-100 Image Net nearest 2.35 18.64 20.80 bilinear 2.58 18.78 20.86 (b) Comparisons of different interpolation modes. Table 4: Ablation studies on 1) CIFAR-10 with Pyramid Net-200, α=240 under 300-epoch training setting, 2) CIFAR-100 with Res Net-18 under 200-epoch training setting and 3) Image Net with Res Net-50 under 300-epoch training setting. Model (300 epochs) Top-1 Err (%) Top-5 Err (%) Res Net-50 [70] 23.68 7.05 + Mixup [70] 22.58 6.40 + Cut Mix [68] 21.40 5.92 + RM (ours) 20.80 5.42 PVTv2-B1 [61] 24.92 8.06 + Mixup [70] 23.23 6.65 + Cut Mix [68] 22.92 6.42 + RM (ours) 22.22 6.17 Table 5: Performance on various architectures in Image Net under 300 epochs. The proposed RM (α=0.5, ω=0.5) shows consistent improvements over other competitive counterparts. Res Net-50 (300 epochs) Top-1 Err (%) Top-5 Err (%) Baseline 23.68 7.05 + Cutout [15] 22.93 6.66 + Stochastic Depth [30] 22.46 6.27 + Mixup [70] 22.58 6.40 + Manifold Mixup [59] 22.50 6.21 + Drop Block [20] 21.87 5.98 + Feature Cut Mix [68] 21.80 6.06 + Cut Mix [68] 21.40 5.92 + Puzzle Mix [34] 21.24 5.71 + Mo Ex [39] 21.90 6.10 + Cut Mix + Mo Ex [39] 20.90 5.70 + RM (ours) 20.80 5.42 Table 6: Comparison of state-of-the-art regularization methods in Image Net under 300 epochs. 2 GPUs with a mini-batch size of 64 on each to optimize the models. The learning rate is set to 0.1 with a linear warmup [26] for five epochs and a cosine decay schedule [46]. For 300-epoch training, we align all the hyperparameters with the official Cut Mix [68] for fair comparisons. Following [68], the averaged best performances are reported via three trials of experiments. Comparisons with State-of-the-art Regularizers. Based on Pyramid Net-200, α=240 [23], Table 1 shows the comparisons against the state-of-the-art data augmentation and regularization approaches under 300 epochs. The proposed RM achieves a 2.35% Top-1 classification error in CIFAR-10, 1.5% better than the baseline (3.85%). It outperforms the two popular regularizers Mixup and Cut Mix by 0.74% and 0.53% points, respectively. Performance under Various Network Backbones. The effectiveness of RM is further validated across a variety of network architectures in CIFAR-100 under the 200-epoch training setting, including Res Net [26], Dense Net [29], and Pyramid Net [23]. From Table 3, we observe that RM has a consistent improvement in accuracy against the baselines (+1.5 3.5%) and other competitive counterparts (+0.3 1.1%). Notably, for Dense Net-161, RM shows an absolute gain of 3.2% points over the baseline model and outperforms the competitive Cut Mix by 1.1%. Ablation Study on Hyperparameter α. To reveal the impact of α, we conduct ablation study by varying α {0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9} in CIFAR-10 with Pyramid Net-200, α=240 backbone under 300-epoch setting and CIFAR-100 with Res Net-18 [26] backbone under 200-epoch setting, respectively. As shown in Fig. 4, RM improves upon the baseline (3.85% and 21.70%) for all considered α values. The best performance is achieved when α = 0.5. Ablation Study on Hyperparameter ω. Next, we examine what is the best value for the loss weight ω. We adopt the same experimental setting on CIFAR-10 and CIFAR-100 as that on α. By fixing α to 0.5, we change ω {0, 0.01, 0.05, 0.08, 0.1, 0.2, 0.3, 0.4, 0.5, 1} and report their performances in average value and standard deviation over 3 runs on each. Notably, when ω=0, the RM model is optimized without the consistency loss, but only leverages the historical input-label pairs. Fig. 5 shows that the best performance is achieved when ω = 0.1, while the performance is not sensitive (i.e., around 0.3%) in CIFAR-10 to a wide range of ω [0.1, 1]. Ablation Study on RM Components. To illustrate the performance improvement brought by each RM component individually, we first modify Cut Mix with the same resize strategy of RM, instead of the original cut operation. Then we add the historical mechanism and consistency loss, respectively. It is observed in Table 2 that the resize strategy can have a slightly positive effect (i.e., +0.23% accuracy) on Cut Mix for it is necessary to successfully formulate the correct historical (recursive) paradigm (Fig. 3). Further, the historical mechanism and consistency loss individually improve by +0.55% points and +0.16% points, respectively. Above all, RM achieves a total of 2.02% gain over the baseline and 0.94% over Cut Mix. Detector Pretrain Backbone AP AP50 AP75 Res Net-50 [26] 39.4 57.6 42.8 + Cut Mix [68] 40.1 58.4 43.4 + RM (ours) 41.5 59.9 45.1 PVTv2-B1 [61] 39.3 57.2 42.5 + Cut Mix [68] 41.8 60.3 45.5 + RM (ours) 42.3 61.0 45.6 Res Net-50 [26] 40.2 58.4 43.3 + Cut Mix [68] 41.3 59.5 44.6 + RM (ours) 41.9 60.2 45.6 PVTv2-B1 [61] 40.2 58.1 43.2 + Cut Mix [68] 42.1 60.7 45.5 + RM (ours) 43.0 61.6 46.5 Table 7: Object detection fine-tuned on COCO with the 1 schedule by ATSS [71] and GFL [42]. Detector Pretrain Backbone APbox APbox 50 APbox 75 APmask APmask 50 APmask 75 Mask R-CNN [25] Res Net-50 [26] 38.2 58.8 41.4 34.7 55.7 37.2 + Cut Mix [68] 38.5 58.9 42.2 34.8 56.0 37.4 + RM (ours) 39.6 60.4 43.1 35.8 57.3 38.2 PVTv2-B1 [61] 38.5 60.7 41.5 36.1 57.7 38.4 + Cut Mix [68] 40.6 63.1 44.1 37.6 59.9 40.1 + RM (ours) 41.2 63.5 44.4 38.1 60.5 40.9 Res Net-50 [26] 41.9 60.5 45.5 37.1 57.8 40.1 + Cut Mix [68] 42.2 60.7 46.0 37.4 58.2 40.4 + RM (ours) 42.8 61.4 46.5 37.7 58.5 40.8 PVTv2-B1 [61] 43.0 62.9 46.4 39.2 60.5 42.1 + Cut Mix [68] 45.2 65.0 49.2 40.7 62.4 44.2 + RM (ours) 45.8 65.6 49.8 41.0 63.0 44.6 Table 8: Object detection and instance segmentation fine-tuned on COCO with the 1 schedule by Mask R-CNN [25] and HTC [6]. Ablation Study on Linear Classifiers. As illustrated in Sec. 3, the historical predictions and the aligned current predictions are derived through separate linear classifiers H and H . We conducted comparisons on whether to share the parameters of the two linear layers. The training strategy on CIFAR datasets follows the above. Table 4a shows that unshared weights bring slightly better performance for it enhances model diversity which benefits the contrastive learning. Ablation Study on Interpolation Mode. As depicted in Table 4b, for the resizing operation on historical inputs, we compare two common down-sampling algorithms, i.e., interpolation methods, nearest and bilinear . As a result, nearest performs slightly better, so we fix the interpolation method to it by default. 4.2 Image Net Classification Dataset. The Image Net 2012 dataset [14] contains 1.28 million training images and 50K validation images from 1K classes. Networks are trained on the training set and the Top-1/-5 errors are reported on the validation set. Setup. For Image Net, we also conduct experiments with sufficient training settings (300 epochs) under two major backbone designs, i.e., CNNs [29, 26] and Transformers [61]. For training CNN backbone networks, we employ the augmentations described in Cut Mix [68] to the input images for a fair comparison. All networks are trained using SGD with a momentum of 0.9, a weight decay of 1 10 4, and 8 GPUs with a mini-batch size of 64 on each to optimize models. The initial learning rate is 0.2 with a linear warmup [26] for five epochs and is then decayed following a cosine schedule [46]. To optimize Transformer backbone networks, we use Adam W [47] with a learning rate of 5 10 4, a momentum of 0.9, a weight decay of 5 10 2, and 8 GPUs with a mini-batch size of 64 on each. We follow PVT [62] and apply random resizing/cropping of 224 224 pixels, random horizontal flipping [55], label-smoothing regularization [56], and random erasing [73] as the standard data augmentations. The hyperparameters for RM on Image Net are set to α=0.5, ω=0.5. Comparison with State-of-the-art Regularizers. Based on Res Net-50 [26], we compare RM against a series of popular regularizers in Table 6 under the 300-epoch training setting, where RM significantly improves the baseline by absolute 2.88% points in Top-1 accuracy and outperforms the strong Cut Mix and Puzzle Mix by 0.6% and 0.44%, respectively. Performance under Various Network Backbones. Since Transformer models [16, 45, 62, 61] have made significant progress in image classification due to the self-attention mechanism, we also demonstrate the effectiveness of RM on the popular Transformer network architecture PVTv2-B1 [61] in Image Net under the 300-epoch training setting in Table 5, where it remains the superiority. 4.3 Transfer Learning Benefiting from the explicit multi-scale/-space property and spatial semantic learning of the proposed Recursive Mix (RM), we suspect that the pretrained model with RM can transfer well to the downstream tasks, e.g., object detection, instance segmentation, and semantic segmentation, where multi-scale/-space semantic information plays an important role for identification. Object Detection and Instance Segmentation. We conduct experiments using the one-stage object detector ATSS [71], GFL [42, 41, 40], and two-stage detector Mask R-CNN [25], HTC [6] Segmentor Pretrain Backbone m Io U m Acc a Acc PSPNet [72] Res Net-50 [26] 40.90 51.11 79.52 + Cut Mix [68] 40.96 51.16 79.93 + RM (ours) 41.73 52.47 80.01 PVTv2-B1 [61] 36.48 46.26 76.79 + Cut Mix [68] 37.99 48.70 77.50 + RM (ours) 38.67 49.40 77.93 Uper Net [67] Res Net-50 [26] 40.40 51.00 79.54 + Cut Mix [68] 41.24 51.79 79.69 + RM (ours) 42.30 52.61 80.14 PVTv2-B1 [61] 39.94 50.75 79.02 + Cut Mix [68] 41.73 52.99 80.02 + RM (ours) 43.26 54.21 80.36 Table 9: Semantic segmentation fine-tuned on ADE20K [75] for 80k iterations by PSPNet [72] and Uper Net [67]. Detector CL AP AP50 AP75 ATSS [71] 41.1 59.4 44.5 41.5 59.9 45.1 GFL [42] 41.4 59.4 44.9 41.9 60.2 45.6 (a) Comparisons on object detection. Segmentor CL m Io U m Acc a Acc PSPNet [72] 41.09 51.72 79.99 41.73 52.47 80.01 Uper Net [67] 41.88 52.79 79.94 42.30 52.61 80.14 (b) Comparisons on semantic segmentation. Table 10: Comparisons on transferring downstream tasks w/ and w/o consistency loss (CL). Res Net-50 (300 epochs) Memory Flops #P (train) #P (deploy) Hours Top-1 Err (%) Baseline 5832.0 MB 4.12 G 25.56 M 25.56 M 73.0 23.68 + Mixup [70] 5870.0 MB 4.12 G 25.56 M 25.56 M 73.5 22.58 + Cut Mix [68] 5832.0 MB 4.12 G 25.56 M 25.56 M 73.8 21.40 + RM (ours) 5887.0 MB 4.12 G 27.61 M 25.56 M 73.8 20.80 Table 11: Comparisons of the training efficiency by hours, evaluated on 8 TITAN Xp GPUs. #P denotes the number of parameters. in COCO [43] dataset with comparisons to the vanilla model and Cut Mix [68]-pretrained model based on Res Net-50 [26] and PVTv2-B1 [61]. The training protocol follows the standard 1x (12 epochs) setting as described in MMDetection [7]. In Table 7, it is observed that under ATSS, the RM-pretrained models significantly improve baseline by +2.1 AP points with Res Net-50 and +3.0 AP points with PVTv2-B1, while outperforming Cut Mix [68] pretrained models by +1.4 AP and 0.5 AP. Table 8 indicates the superiority of RM in instance segmentation under Mask R-CNN and HTC, where RM boosts the bbox AP and mask AP by 2.8 and 2.0 points, respectively. Semantic Segmentation. Next, we experiment on ADE20K [75] using two popular algorithms, i.e., PSPNet [72] and Uper Net [67] following the semantic segmentation code of MMSegmentation [12]. Under Res Net-50 [26] and PVTv2-B1 [61] backbone networks, we fine-tune 80k iterations on ADE20K with a batch size of 16 and Adam W [47] optimization algorithm. Table 9 shows that among various pretrained models RM outperforms the baseline and Cut Mix [68]. Notably, we improve the Res Net-50 by 1.9 and 1.1 m Io U compared to baseline and Cut Mix under Uper Net. Also, there is an improvement of 3.3 m Io U over the baseline under PVTv2-B1 based on Uper Net. 4.4 Analyses Class Activation Mapping Visualization. To demonstrate the benefits of RM, we visualize the class activation map through CAM [74]. We select images that originally contain multiple categories and create CAM heatmaps for each (Fig. 6) to demonstrate the superiority of RM in spatial semantic representation. The visualization shows that RM locates the corresponding objects more accurately than Baseline and Cut Mix. Thanks to the recursive paradigm of RM, the algorithm can generate inputs that have multiple objects (>2), which increases the diversity of inputs in data space and enriches each instance with multi-scale and spatial-variant views. Thus, the model is more capable of recognizing complex images containing multiple categories or objects that differ greatly in size. Furthermore, the model trained under RM explicitly learns dense semantic representations and can locate the meaningful regions more accurately via the designed consistency loss. More visualization can be found in the Supplementary Material. RM Enables Efficient Self-distillation. Generally, the consistency loss resembles a variant of local self-distillation by explicitly restricting two spatial representations of one instance. However, it is worth mentioning that the proposed recursive augmentation paradigm and the consistency loss are not isolated from each other. Only in the form of the recursive mix , can we achieve self-distillation in a convenient, efficient, and memory-friendly way. Because in the recursive paradigm, the historical part in the current iteration can be supervised directly by the predictions of the former iteration as they share identical but differentially-scaled contents. It delicately fits the logic of local self-distillation, at a minimal cost of storing only the outputs of the last iteration. On the contrary, to implement local self-distillation without the recursive operation, one needs to forward one input through the networks repeatedly [64, 52, 4, 10, 24, 19, 11, 21, 57] or record the information of the whole data Mixup Cut Mix RM (ours) Natural Samples with Multiple Labels Ox Barn Picket Fence Orange Apple Banana Couch Pillow Lamp Stretcher Military Uniform Military Plane Figure 6: CAM [74] visualization comparing RM with Mixup [70] and Cut Mix [68] on natural samples with multiple labels. in an entire epoch [66, 35, 38, 65], which either requires a tremendous computation cost or large memory consumption. Experimentally, RM with consistency loss enhances accuracy by 0.16% over that without in Table 2. We also verify the effectiveness of consistency loss on downstream tasks in Table 10. With the models pretrained with consistency loss, GFL and ATSS improve 0.5m AP on COCO while PSPNet and Uper Net improve 0.7 m Io U on ADE20K. The results show that the introduced consistency loss enhances the performance on the downstream tasks via learning the spatial-correlative semantic representation. 10 5 10 4 10 3 10 2 10 1 Effective Classes Baseline Mixup Cut Mix RM (ours) Figure 7: Comparisons of the effective classes during training under different thresholds. Effective Class Number. To illustrate the diversity in the supervision signals of RM, we denote the number of classes in the regularized target whose supervision values are higher than a given threshold as effective classes. We then record the average effective number over 3 runs under 100 iterations on RM and other methods in the Image Net training set. Note that the label smoothing threshold is roughly set to 10 4 . We suspect that if a signal value above the label smoothing threshold can be used as effective supervision, then RM can achieve a learning signal containing an average of about 4 5 objects in a single image, which confirms the diversity of its supervision (Fig. 7). Efficiency. We improve performance with negligible additional memory cost, as shown in Table 11. Notably, the training time and computation complexity are competitive with existing methods, which are evaluated on 8 TITAN Xp GPUs. Limitations. Although our method shares the same parameters and FPS as the baseline model during employment, due to the introduction of an additional classification head and the operation of Ro IAlign, we inevitably increase a small number of parameters only during the training process. In addition, it is unclear how to effectively adopt the recursive paradigm to Mixup [70] since the images would be cluttered and unidentifiable via iterative mixup operations. 5 Conclusion In this paper, we propose a novel regularization method termed Recursive Mix (RM), which leverages the rarely exploited aspect of the historical information: the input-prediction-label triplets, to enhance the generalization of deep vision models. RM shares several good properties according to its historical mechanism, and it consistently improves the recognition accuracy on competitive vision benchmarks with considerably negligible additional budgets. We hope RM can serve as a simple yet effective baseline for the community. Acknowledgments and Disclosure of Funding This paper is supported by the National Key R&D Plan of the Ministry of Science and Technology (Project No.2020AAA0104400), the Young Scientists Fund of the National Natural Science Foundation of China (Grant No.62206134), and the National Science Fund of China (Grant No.U1713208). [1] Philip Bachman, R Devon Hjelm, and William Buchwalter. Learning representations by maximizing mutual information across views. ar Xiv preprint ar Xiv:1906.00910, 2019. 3 [2] David Berthelot, Nicholas Carlini, Ian Goodfellow, Nicolas Papernot, Avital Oliver, and Colin Raffel. Mixmatch: A holistic approach to semi-supervised learning. ar Xiv preprint ar Xiv:1905.02249, 2019. 3 [3] Jane Bromley, Isabelle Guyon, Yann Le Cun, Eduard Säckinger, and Roopak Shah. Signature verification using a" siamese" time delay neural network. Neur IPS, 1993. 3 [4] Mathilde Caron, Hugo Touvron, Ishan Misra, Hervé Jégou, Julien Mairal, Piotr Bojanowski, and Armand Joulin. Emerging properties in self-supervised vision transformers. In ICCV, 2021. 3, 4, 8 [5] Jie-Neng Chen, Shuyang Sun, Ju He, Philip Torr, Alan Yuille, and Song Bai. Transmix: Attend to mix for vision transformers. ar Xiv preprint ar Xiv:2111.09833, 2021. 3 [6] Kai Chen, Jiangmiao Pang, Jiaqi Wang, Yu Xiong, Xiaoxiao Li, Shuyang Sun, Wansen Feng, Ziwei Liu, Jianping Shi, Wanli Ouyang, et al. Hybrid task cascade for instance segmentation. In CVPR, 2019. 7 [7] Kai Chen, Jiaqi Wang, Jiangmiao Pang, Yuhang Cao, Yu Xiong, Xiaoxiao Li, Shuyang Sun, Wansen Feng, Ziwei Liu, Jiarui Xu, et al. Mmdetection: Open mmlab detection toolbox and benchmark. ar Xiv preprint ar Xiv:1906.07155, 2019. 8 [8] Pengguang Chen, Shu Liu, Hengshuang Zhao, and Jiaya Jia. Gridmask data augmentation. ar Xiv preprint ar Xiv:2001.04086, 2020. 2 [9] Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. A simple framework for contrastive learning of visual representations. In ICML, 2020. 3 [10] Xinlei Chen, Haoqi Fan, Ross Girshick, and Kaiming He. Improved baselines with momentum contrastive learning. ar Xiv preprint ar Xiv:2003.04297, 2020. 3, 8 [11] Xinlei Chen, Saining Xie, and Kaiming He. An empirical study of training self-supervised visual transformers. In ICCV, 2021. 3, 8 [12] MMSegmentation Contributors. MMSegmentation: Openmmlab semantic segmentation toolbox and benchmark. https://github.com/open-mmlab/mmsegmentation, 2020. 8 [13] Ali Dabouei, Sobhan Soleymani, Fariborz Taherkhani, and Nasser M Nasrabadi. Supermix: Supervising the mixing data augmentation. In CVPR, 2021. 1, 3 [14] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In CVPR, 2009. 7 [15] Terrance De Vries and Graham W Taylor. Improved regularization of convolutional neural networks with cutout. ar Xiv preprint ar Xiv:1708.04552, 2017. 1, 2, 5, 6 [16] Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. ar Xiv preprint ar Xiv:2010.11929, 2020. 7 [17] John Duchi, Elad Hazan, and Yoram Singer. Adaptive subgradient methods for online learning and stochastic optimization. JMLR, 2011. 3 [18] Mojtaba Faramarzi, Mohammad Amini, Akilesh Badrinaaraayanan, Vikas Verma, and Sarath Chandar. Patchup: A regularization technique for convolutional neural networks. ar Xiv preprint ar Xiv:2006.07794, 2020. 1, 3 [19] Weixin Feng, Yuanjiang Wang, Lihua Ma, Ye Yuan, and Chi Zhang. Temporal knowledge consistency for unsupervised visual representation learning. In ICCV, 2021. 3, 8 [20] Golnaz Ghiasi, Tsung-Yi Lin, and Quoc V Le. Dropblock: A regularization method for convolutional networks. ar Xiv preprint ar Xiv:1810.12890, 2018. 2, 5, 6 [21] Jean-Bastien Grill, Florian Strub, Florent Altché, Corentin Tallec, Pierre H Richemond, Elena Buchatskaya, Carl Doersch, Bernardo Avila Pires, Zhaohan Daniel Guo, Mohammad Gheshlaghi Azar, et al. Bootstrap your own latent: A new approach to self-supervised learning. In Neur IPS, 2020. 3, 8 [22] Raia Hadsell, Sumit Chopra, and Yann Le Cun. Dimensionality reduction by learning an invariant mapping. In CVPR, 2006. 3, 4 [23] Dongyoon Han, Jiwhan Kim, and Junmo Kim. Deep pyramidal residual networks. In CVPR, 2017. 5, 6 [24] Kaiming He, Haoqi Fan, Yuxin Wu, Saining Xie, and Ross Girshick. Momentum contrast for unsupervised visual representation learning. In CVPR, 2020. 3, 4, 8 [25] Kaiming He, Georgia Gkioxari, Piotr Dollár, and Ross Girshick. Mask r-cnn. In ICCV, 2017. 1, 2, 4, 7 [26] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In CVPR, 2016. 1, 2, 5, 6, 7, 8 [27] Tong He, Zhi Zhang, Hang Zhang, Zhongyue Zhang, Junyuan Xie, and Mu Li. Bag of tricks for image classification with convolutional neural networks. In CVPR, 2019. 3 [28] Minui Hong, Jinwoo Choi, and Gunhee Kim. Stylemix: Separating content and style for enhanced data augmentation. In CVPR, 2021. 3, 5 [29] Gao Huang, Zhuang Liu, Laurens Van Der Maaten, and Kilian Q Weinberger. Densely connected convolutional networks. In CVPR, 2017. 1, 2, 5, 6, 7 [30] Gao Huang, Yu Sun, Zhuang Liu, Daniel Sedra, and Kilian Q Weinberger. Deep networks with stochastic depth. In ECCV, 2016. 1, 2, 5, 6 [31] Shaoli Huang, Xinchao Wang, and Dacheng Tao. Snapmix: Semantically proportional mixing for augmenting fine-grained data. In AAAI, 2021. 3 [32] Dan Iter, Kelvin Guu, Larry Lansing, and Dan Jurafsky. Pretraining with contrastive sentence objectives improves discourse performance of language models. ar Xiv preprint ar Xiv:2005.10389, 2020. 3 [33] Jang-Hyun Kim, Wonho Choo, Hosan Jeong, and Hyun Oh Song. Co-mixup: Saliency guided joint mixup with supermodular diversity. ar Xiv preprint ar Xiv:2102.03065, 2021. 3 [34] Jang-Hyun Kim, Wonho Choo, and Hyun Oh Song. Puzzle mix: Exploiting saliency and local statistics for optimal mixup. In ICML, 2020. 1, 2, 3, 6 [35] Kyungyul Kim, Byeong Moon Ji, Doyoung Yoon, and Sangheum Hwang. Self-knowledge distillation with progressive refinement of targets. In ICCV, 2021. 3, 9 [36] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. In ICLR, 2015. 3 [37] Alex Krizhevsky, Geoffrey Hinton, et al. Learning multiple layers of features from tiny images. Citeseer, 2009. 5 [38] Samuli Laine and Timo Aila. Temporal ensembling for semi-supervised learning. In ICLR, 2017. 3, 9 [39] Boyi Li, Felix Wu, Ser-Nam Lim, Serge Belongie, and Kilian Q Weinberger. On feature normalization and data augmentation. In CVPR, 2021. 1, 3, 5, 6 [40] Xiang Li, Chengqi Lv, Wenhai Wang, Gang Li, Lingfeng Yang, and Jian Yang. Generalized focal loss: Towards efficient representation learning for dense object detection. TPAMI, 2022. 7 [41] Xiang Li, Wenhai Wang, Xiaolin Hu, Jun Li, Jinhui Tang, and Jian Yang. Generalized focal loss v2: Learning reliable localization quality estimation for dense object detection. In CVPR, 2021. 7 [42] Xiang Li, Wenhai Wang, Lijun Wu, Shuo Chen, Xiaolin Hu, Jun Li, Jinhui Tang, and Jian Yang. Generalized focal loss: Learning qualified and distributed bounding boxes for dense object detection. In Neur IPS, 2020. 1, 7, 8 [43] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C Lawrence Zitnick. Microsoft coco: Common objects in context. In ECCV, 2014. 8 [44] Liyuan Liu, Haoming Jiang, Pengcheng He, Weizhu Chen, Xiaodong Liu, Jianfeng Gao, and Jiawei Han. On the variance of the adaptive learning rate and beyond. In ICLR, 2020. 3 [45] Ze Liu, Yutong Lin, Yue Cao, Han Hu, Yixuan Wei, Zheng Zhang, Stephen Lin, and Baining Guo. Swin transformer: Hierarchical vision transformer using shifted windows. ar Xiv preprint ar Xiv:2103.14030, 2021. 7 [46] Ilya Loshchilov and Frank Hutter. Sgdr: Stochastic gradient descent with warm restarts. ar Xiv preprint ar Xiv:1608.03983, 2016. 6, 7 [47] Ilya Loshchilov and Frank Hutter. Decoupled weight decay regularization. In ICLR, 2019. 3, 7, 8 [48] Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Corrado, and Jeff Dean. Distributed representations of words and phrases and their compositionality. In Neur IPS, 2013. 3 [49] Rafael Müller, Simon Kornblith, and Geoffrey Hinton. When does label smoothing help? ar Xiv preprint ar Xiv:1906.02629, 2019. 2, 5 [50] Aaron van den Oord, Yazhe Li, and Oriol Vinyals. Representation learning with contrastive predictive coding. ar Xiv preprint ar Xiv:1807.03748, 2018. 3 [51] Byungseok Roh, Wuhyun Shin, Ildoo Kim, and Sungwoong Kim. Spatially consistent representation learning. In CVPR, 2021. 3 [52] Kuniaki Saito, Yoshitaka Ushiku, Tatsuya Harada, and Kate Saenko. Adversarial dropout regularization. ar Xiv preprint ar Xiv:1711.01575, 2017. 3, 8 [53] Nitish Srivastava, Geoffrey Hinton, Alex Krizhevsky, Ilya Sutskever, and Ruslan Salakhutdinov. Dropout: a simple way to prevent neural networks from overfitting. JMLR, 2014. 1, 2, 3 [54] Ilya Sutskever, James Martens, George Dahl, and Geoffrey Hinton. On the importance of initialization and momentum in deep learning. In ICML, 2013. 3 [55] Christian Szegedy, Wei Liu, Yangqing Jia, Pierre Sermanet, Scott Reed, Dragomir Anguelov, Dumitru Erhan, Vincent Vanhoucke, and Andrew Rabinovich. Going deeper with convolutions. In CVPR, 2015. 7 [56] Christian Szegedy, Vincent Vanhoucke, Sergey Ioffe, Jon Shlens, and Zbigniew Wojna. Rethinking the inception architecture for computer vision. In CVPR, 2016. 7 [57] Antti Tarvainen and Harri Valpola. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. In Neur IPS, 2017. 3, 8 [58] AFM Uddin, Mst Monira, Wheemyung Shin, Tae Choong Chung, Sung-Ho Bae, et al. Saliencymix: A saliency guided data augmentation strategy for better regularization. ar Xiv preprint ar Xiv:2006.01791, 2020. 3 [59] Vikas Verma, Alex Lamb, Christopher Beckham, Amir Najafi, Ioannis Mitliagkas, David Lopez-Paz, and Yoshua Bengio. Manifold mixup: Better representations by interpolating hidden states. In ICML, 2019. 1, [60] Devesh Walawalkar, Zhiqiang Shen, Zechun Liu, and Marios Savvides. Attentive cutmix: An enhanced data augmentation approach for deep learning based image classification. ar Xiv preprint ar Xiv:2003.13048, 2020. 1, 3 [61] Wenhai Wang, Enze Xie, Xiang Li, Deng-Ping Fan, Kaitao Song, Ding Liang, Tong Lu, Ping Luo, and Ling Shao. Pvtv2: Improved baselines with pyramid vision transformer. ar Xiv preprint ar Xiv:2106.13797, 2021. 2, 6, 7, 8 [62] Wenhai Wang, Enze Xie, Xiang Li, Deng-Ping Fan, Kaitao Song, Ding Liang, Tong Lu, Ping Luo, and Ling Shao. Pyramid vision transformer: A versatile backbone for dense prediction without convolutions. ar Xiv preprint ar Xiv:2102.12122, 2021. 1, 7 [63] Xun Wang, Haozhi Zhang, Weilin Huang, and Matthew R Scott. Cross-batch memory for embedding learning. In CVPR, 2020. 3 [64] Lijun Wu, Juntao Li, Yue Wang, Qi Meng, Tao Qin, Wei Chen, Min Zhang, Tie-Yan Liu, et al. R-drop: regularized dropout for neural networks. Neur IPS, 2021. 3, 8 [65] Zhirong Wu, Alexei A Efros, and Stella X Yu. Improving generalization via scalable neighborhood component analysis. In ECCV, 2018. 3, 9 [66] Zhirong Wu, Yuanjun Xiong, Stella X Yu, and Dahua Lin. Unsupervised feature learning via non-parametric instance discrimination. In CVPR, 2018. 3, 9 [67] Tete Xiao, Yingcheng Liu, Bolei Zhou, Yuning Jiang, and Jian Sun. Unified perceptual parsing for scene understanding. In ECCV, 2018. 1, 8 [68] Sangdoo Yun, Dongyoon Han, Seong Joon Oh, Sanghyuk Chun, Junsuk Choe, and Youngjoon Yoo. Cutmix: Regularization strategy to train strong classifiers with localizable features. In ICCV, 2019. 1, 2, 3, 4, 5, 6, 7, 8, 9 [69] Matthew D Zeiler. Adadelta: an adaptive learning rate method. ar Xiv preprint ar Xiv:1212.5701, 2012. 3 [70] Hongyi Zhang, Moustapha Cisse, Yann N Dauphin, and David Lopez-Paz. mixup: Beyond empirical risk minimization. In ICLR, 2017. 1, 2, 3, 4, 5, 6, 8, 9 [71] Shifeng Zhang, Cheng Chi, Yongqiang Yao, Zhen Lei, and Stan Z Li. Bridging the gap between anchorbased and anchor-free detection via adaptive training sample selection. In CVPR, 2020. 1, 2, 7, 8 [72] Hengshuang Zhao, Jianping Shi, Xiaojuan Qi, Xiaogang Wang, and Jiaya Jia. Pyramid scene parsing network. In CVPR, 2017. 1, 8 [73] Zhun Zhong, Liang Zheng, Guoliang Kang, Shaozi Li, and Yi Yang. Random erasing data augmentation. In AAAI, 2020. 7 [74] Bolei Zhou, Aditya Khosla, Agata Lapedriza, Aude Oliva, and Antonio Torralba. Learning deep features for discriminative localization. In CVPR, 2016. 8, 9 [75] Bolei Zhou, Hang Zhao, Xavier Puig, Tete Xiao, Sanja Fidler, Adela Barriuso, and Antonio Torralba. Semantic understanding of scenes through the ade20k dataset. IJCV, 2019. 8 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] (b) Did you describe the limitations of your work? [Yes] See Section 4.4. (c) Did you discuss any potential negative societal impacts of your work? [Yes] See the broader impacts in the Supplementary Material. (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [N/A] (b) Did you include complete proofs of all theoretical results? [N/A] 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [Yes] Results are reported over 3 runs on the CIFAR dataset (see Fig. 4 and Fig. 5). (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [Yes] (b) Did you mention the license of the assets? [N/A] They are publicly available for research purpose. (c) Did you include any new assets either in the supplemental material or as a URL? [No] There are no new assets. (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [N/A] They are publicly available for research purpose. (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [N/A] They are publicly available for research purpose. 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]