# crosstask_knowledge_distillation_in_multitask_recommendation__2311b262.pdf Cross-Task Knowledge Distillation in Multi-Task Recommendation Chenxiao Yang1, Junwei Pan2, Xiaofeng Gao1, Tingyu Jiang2, Dapeng Liu2, Guihai Chen1 1 Department of Computer Science and Engineering, Shanghai Jiao Tong University 2 Tencent Inc. chr26195@sjtu.edu.com, jonaspan@tencent.com, gao-xf@cs.sjtu.edu.cn, travisjiang@tencent.com, rocliu@tencent.com, gchen@cs.sjtu.edu.cn Multi-task learning (MTL) has been widely used in recommender systems, wherein predicting each type of user feedback on items (e.g, click, purchase) are treated as individual tasks and jointly trained with a unified model. Our key observation is that the prediction results of each task may contain task-specific knowledge about user s fine-grained preference towards items. While such knowledge could be transferred to benefit other tasks, it is being overlooked under the current MTL paradigm. This paper, instead, proposes a Cross-Task Knowledge Distillation framework that attempts to leverage prediction results of one task as supervised signals to teach another task. However, integrating MTL and KD in a proper manner is non-trivial due to several challenges including task conflicts, inconsistent magnitude and requirement of synchronous optimization. As countermeasures, we 1) introduce auxiliary tasks with quadruplet loss functions to capture cross-task fine-grained ranking information and avoid task conflicts, 2) design a calibrated distillation approach to align and distill knowledge from auxiliary tasks, and 3) propose a novel error correction mechanism to enable and facilitate synchronous training of teacher and student models. Comprehensive experiments are conducted to verify the effectiveness of our framework in real-world datasets. Introduction Online recommender systems often involve predicting various types of user feedback such as clicking and purchasing. Multi-Task Learning (MTL) (Caruana 1997) has emerged in this context as a powerful tool to explore the connection of tasks for improving user interest modeling (Ma et al. 2018b; Lu, Dong, and Smyth 2018; Wang et al. 2018). Common MTL models consist of low-level shared network and several high-level individual networks, as shown in Fig. 1(a), in the hope that the shared network could transfer the knowledge about how to encode the input features by sharing or enforcing similarity on parameters of different tasks (Ruder 2017). Most prior works (Ma et al. 2018a; Tang et al. 2020a; Ma et al. 2019) put efforts on designing different shared network architectures with ad-hoc parametersharing mechanisms such as branching and gating. In these models, each task is trained under the supervision of its own Copyright 2022, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. binary ground-truth label (1 or 0), attempting to rank positive items above negative ones. However, using such binary labels as training signals, the task may fail to accurately capture user s preference for items with the same label, despite that learning the auxiliary knowledge about these items relations may benefit the overall ranking performance. To address this limitation, we observe that the predictions of other tasks may contain useful information about how to rank same-labeled items. For example, given two tasks predicting Buy and Like , and two items labeled as Buy:0, Like:1 and Buy:0, Like:0 , the task Buy may not accurately distinguish their relative ranking since both of their labels are 0. In contrast, another task Like will identity the former item as positive with larger probability (e.g. 0.7), the latter with smaller probability (e.g. 0.1). Based on the fact that a user is more likely to purchase the item she likes 1, we could somehow take advantage of these predictions from other tasks as a means to transfer ranking knowledge. Knowledge Distillation (KD) (Hinton, Vinyals, and Dean 2015) is a teacher-student learning framework where the student is trained using the predictions of the teacher. As revealed by theoretical analysis in previous studies (Tang et al. 2020b; Phuong and Lampert 2019), the predictions of the teacher, also known as soft labels, are usually seen as more informative training signals than binary hard labels, since they could reflect whether the sample is true positive (negative) . On the perspective of backward gradient, KD can adaptively re-scale student model s training dynamics based on the values of soft labels. Specially, in the above example, we could incorporate predictions 0.7 and 0.1 in the training signals for task Buy . Consequently, the gradients w.r.t the sample labeled Buy:0 & Like:0 in the example will be larger, indicating it is a more confident negative sample. Through this process, the task Buy could hopefully give accurate rankings of same-labeled items. Motivated by these above observations, we proceed to design a new knowledge transfer paradigm on the optimization level of MTL models by leveraging KD. It is non-trivial due to three critical and fundamental challenges: How to address the task conflict problem during distillation? Not all knowledge from other tasks is use- 1The same applies to other types of user feedback, e.g., click, collect, forward. The Thirty-Sixth AAAI Conference on Artificial Intelligence (AAAI-22) User Feature Item Feature Context Feature Shared Layers Buy ? Like ? (b) Fine-grained ranking (FGR) (a) Existing MTL framework User Feature Item Feature Context Feature Buy ? Like ? FGR (Like) KD KD Shared Layers (c) Our framework FC Layer FC Layer FC Layer FC Layer FC Layer FC Layer Auxiliary Ranking Auxiliary Ranking Knowledge Transfer *NO* Knowledge Transfer Knowledge Transfer Figure 1: Illustration of the motivation of Cross Distil. ful (Yu et al. 2020). Specially, in online recommendation, the target task may believe that a user prefers item A since she bought item A instead of item B, while another task may reversely presume she prefers item B since she puts it in the collection rather than item A. Such conflicting ranking knowledge may be harmful for the target task and could empirically cause significant performance drop. How to align the magnitude of predictions for different tasks? Distinct from vanilla KD where teacher and student models have the same prediction target, different tasks may have different magnitude of positive ratio. Directly using another task s predictions as training signals without alignment could mislead the target task to yield biased predictions (Zhou et al. 2021). How to enhance training when teacher and student are synchronously optimized? The vanilla KD adopts asynchronous training where the teacher model is welltrained beforehand. However, MTL inherently requires synchronous training, where each task is jointly learned from scratch. This indicates the teacher may be poorlytrained and provide inaccurate or even erroneous training signals, causing slow convergence and local optima (Wen, Lai, and Qian 2019; Xu et al. 2020). In this paper, we propose a novel framework named as Cross-Task Knowledge Distillation (Cross Distil). Different from prior MTL models where knowledge transfer is achieved by sharing representations in bottom layers, Cross Distil also facilitates transferring ranking knowledge on the top layers, as shown in Fig. 1(c). To solve the aforementioned challenges: First, we introduce augmented tasks to learn the knowledge of the ranking orders of four types of samples as shown in Fig. 1(b). New tasks are trained based on a quadruplet loss function, and could fundamentally avoid conflicts by only preserving the useful knowledge and discarding the harmful one. Second, we consider a calibration process that is seamlessly integrated in the KD procedure to align predictions of different tasks, which is accompanied with a bi-level training algorithm to optimize parameters for prediction and calibration respectively. Third, teachers and students are trained in an end-to-end manner with a novel error correction mechanism to speed up model training and further enhance knowledge quality. We conduct comprehensive experiments on a large-scale public dataset and a real-world production dataset that is collected from our platform. The results demonstrate that Cross Distil achieves state-of-the-art performance. The ablation studies also thoroughly dissect the effectiveness of its modules. Preliminaries and Related Works Knowledge Distillation (Hinton, Vinyals, and Dean 2015) is a teacher-student learning framework where the student model is trained by mimicking outputs of the teacher model. For binary classification, the distillation loss function is formulated as LKD = CE(σ(r T /τ), σ(r S/τ)), (1) where CE(y, ˆy) = y log(ˆy) + (1 y) log(1 ˆy) is binary cross-entropy, r T and r S denote logits of the teacher and student model, and τ is the temperature hyper-parameter. Recent advances (Tang et al. 2020b; Yuan et al. 2020) show that KD performs instance-specific label smoothing regularization that re-scales backward gradient in logits space, and thus could hint to the student model about the confidence of the ground-truth, which explain the efficacy of KD for wider applications apart from traditional model compression (Kim et al. 2021; Yuan et al. 2020). Existing works in recommender systems adopt KD for its original purpose, i.e., distilling knowledge from a cumbersome teacher model into a lightweight student model targeting the same task (Tang and Wang 2018; Xu et al. 2020; Zhu et al. 2020). Distinct from theirs or other KD works in other fields, this paper leverages KD to transfer knowledge across different tasks, which is non-trivial due to the aforementioned three major challenges. Multi-Task Learning (Zhang and Yang 2021) is a machine learning framework that learns task-invariant representations by a shared bottom network, and yields predictions for each individual task by task-specific networks. It has received increasing interests in recommender systems (Ma et al. 2018b; Lu, Dong, and Smyth 2018; Wang et al. 2018; Pan et al. 2019) for modeling user interests by predicting different types of user feedback. A series of works seek for improvements by designing different shared network architectures, such as adding constraints on taskspecific parameters (Duong et al. 2015; Misra et al. 2016; Yang and Hospedales 2016) and separating shared and taskspecific parameters (Ma et al. 2018a; Tang et al. 2020a; Ma et al. 2019). Different from theirs, we resort to KD to transfer ranking knowledge across tasks on top of task-specific networks. Notably, our model is a general framework and could be leveraged as an extension of off-the-shelf MTL models. Proposed Model Task Augmentation for Ranking This paper focuses on multi-task learning for predicting different user feedback (e.g. click, like, purchase, lookthrough), and considers two tasks denoted as task A and Shared Layers Sigmoid Function: M Correction: Temperature: Optimization Optimization No backward gradient 1. Dataset Sampling 2. Multi-Task Model 3. Loss Functions Computation C Calibration: Figure 2: Illustration of computational graph for Cross Distil. task B to simplify illustration. As shown in the left panel of Fig. 2, we first split the set of training samples into multiple subsets according to combinations of tasks labels: D+ = {(xi, y A i , y B i ) D | y A i = 1, y B i = 0}, D + = {(xi, y A i , y B i ) D | y A i = 0, y B i = 1}, D = D D +, D+ = D+ D++, D = D D+ , D + = D + D++, where x is an input feature vector, y A and y B denote hard labels for task A and task B respectively. The goal is to rank positive samples before negative ones, which can be expressed a bipartite order x+ x for task A and x + x for task B, where x+ D+ and so forth. Note that these bipartite orders may be contradictory among different tasks, e.g., x+ x + for task A while x+ x + for task B. Due to the existence of such conflicts, directly conducting KD by treating one task as teacher and another task as student may cause inconsistent training signals and is empirically harmful for the overall ranking performance. To enable knowledge transfer across tasks via KD, we introduce auxiliary ranking-based tasks that could essentially avoid task conflicts while preserving useful ranking knowledge. In specific, we consider a quadruplet (x++, x+ , x +, x ) and the corresponding multipartite order x++ x+ x + x for task A. In contrast with the original bipartite order, the multipartite order reveals additional information about the ranking of samples, i.e., x++ x+ and x + x without introducing contradictions. Therefore, we refer such order as fine-grained ranking. Based on this, we introduce a new ranking-based task called augmented task A+ for enhancing knowledge transfer by additionally maximizing ln p(Θ| ) = ln p(x++ x+ |Θ) p(x + x |Θ) p(Θ) ln σ(ˆr++ + ) + ln σ(ˆr + ) Reg(Θ), where r is the logit value before activation in the last layer, ˆr++ + = ˆr++ ˆr+ , and sigmoid function σ(x) = 1/(1+ exp( x)). The loss function for augmented task A+ is βA 1 ln σ(ˆr++ + ) βA 2 ln σ(ˆr + ) (x+ ,x ) ln σ(ˆr+ ), which consists of three terms that respectively correspond to three pair-wise ranking relations of samples, where coefficients β1, β2 balance their importance. The loss function for augmented task B+ could be defined in a similar spirit. These augmented ranking-based tasks are jointly trained with original regression-based tasks in MTL framework as shown in the second panel of Fig. 2. The original regression-based loss function is formulated as: LA =CE(y A, ˆy A), LB = CE(y B, ˆy B), CE(y, ˆy) = X xi D yi ln ˆyi (1 yi) ln(1 ˆyi), (5) where ˆy = σ(r) is the predicted probability. The introduced auxiliary ranking-based tasks could avoid task conflicts and act as prerequisites for knowledge transfer through KD. Besides, the task augmentation approach itself is beneficial for the generalizability of main tasks (Hsieh and Tseng 2021) by introducing more related tasks that may provide hints about what shall be learned and transferred in shared layers. Calibrated Knowledge Distillation We next design a cross-task knowledge distillation approach that can transfer fine-grained ranking knowledge for MTL. Since the prediction results of another task may contain the information about unseen rankings between samples of the same label, a straightforward approach is to use soft labels of another task to teach the current task by the vanilla hint loss (i.e. distillation loss) as in Eqn. (1). Unfortunately, such naive approach may be problematic and even imposes negative effects in practice. This is because the labels of different tasks may have contradictory ranking information that would harm the learning of other tasks as mentioned previously. To avoid such conflicts, we instead treat augmented ranking-based tasks as teachers, original regression-based tasks as students, and adopt the following distillation loss functions: LA KD = CE(σ(ˆr A+/τ), σ(ˆr A/τ)), LB KD = CE(σ(ˆr B+/τ), σ(ˆr B/τ)). (6) Note that soft labels ˆy A+ = σ(ˆr A+/τ) and ˆy B+ = σ(ˆr B+/τ) are invariant when training the student model, and hence the student will not mislead the teacher. The loss functions for students are formulated as LA Stu = (1 αA)LA + αALA KD, LB Stu = (1 αB)LB + αBLB KD, (7) where αA [0, 1] is the hyper-parameter to balance two losses. The soft labels output by augmented ranking-based tasks are more informative training signals than hard labels. As an example, for samples x++, x+ , x +, x , the teacher model for augmented task A+ may give predictions 0.9, 0.8, 0.2, 0.1 which intrinsically contains auxiliary ranking orders x++ x+ and x + x that are not revealed in hard labels. Such knowledge is then explicitly transferred through the distillation loss and can meanwhile regularize task-specific layers from over-fitting the hard labels. However, an issue of the aforementioned approach is that augmented tasks are optimized with pair-wise loss functions and thus are not predicting a probability, i.e., the prediction σ(ˆr A+) does not agree with the actual probability that the input sample is a positive one. Directly using the soft labels of teachers may mislead students and cause performance deterioration. To solve this problem, we propose to calibrate the predictions so as to provide numerically sound and unbiased soft labels. Platt Scaling (Niculescu-Mizil and Caruana 2005; Platt et al. 1999) is a classic probability calibration method. We adopt it for calibration in this work. Still, one can replace it with any other more complex methods in practice. Formally, to get calibrated probabilities, we transform the logit values of teacher models through the following equation: r A+ = P A ˆr A+ + QA, y A+ = 1 1 + exp r A+ (8) where r and y are the logit value and probability after calibration, respectively. The same process is also used for task B+. P, Q are learnable parameters specific to each task. They are trained by optimizing the calibration loss LCal = LA Cal+LB Cal = CE(y A, y A+)+CE(y B, y B+). (9) We fix MTL model parameters when optimizing LCal as shown in the third panel of Fig. 2. Note that, since the calibrated outputs of the teacher model are linear projections of the original outputs, the ranking result is unaffected so that the latent fine-grained ranking knowledge in soft labels is preserved during the calibration process. Distillation losses in Eqn. (6) are then revised by replacing ˆr A+, ˆr B+ with r A+, r B+. Algorithm 1: Training Algorithm for Cross Distil Input: Training dataset D, learning rate γ1 and γ2, initial parameters Θ and Ω. 1 Construct set D++, D+ , D +, D , D+ , D , D +, D ; 2 while Not converged do 3 Sample x uniformly at random from D; 4 Sample x++, x+ , x +, x uniformly at random from D++, D+ , D +, D respectively; 5 Sample x+ , x , x +, x uniformly at random from D+ , D , D +, D respectively; 6 Model parameter Θ optimization: 7 Calculate LA+(x+ , x , x++, x+ , x +, x ; Θ); 8 Calculate LB+(x +, x , x++, x+ , x +, x ; Θ); 9 Calculate LA Stu(x; Θ), LB Stu(x; Θ); 10 LModel wighted Sum(LA+, LB+, LA Stu, LB Stu); 11 Θ Θ γ1 ΘLModel; 12 Calibration parameter Ωoptimization: 13 Calculate LCal(x; Ω); 14 Ω Ω γ2 ΩLCal; Model Training Conventional KD adopts a two-stage training process where the teacher model is trained in advance and its parameters are fixed when training the student model (Hinton, Vinyals, and Dean 2015). However, such asynchronous training procedure is not favorable for industrial applications such as online advertising. Instead, due to simplicity and easy maintenance, synchronous training procedure where teacher and student models are trained in an end-to-end manner is more desirable as done in (Xu et al. 2020; Anil et al. 2018; Zhou et al. 2018). In our framework, there are two sets of parameters for optimization, namely, parameters in MTL backbone for prediction (denoted as Θ) and parameters for calibration including P A, P B, QA and QB (denoted as Ω). To jointly optimize prediction parameters and calibration parameters, we propose a bi-level training procedure where Θ and Ωare optimized in turn for each iteration as shown in the training algorithm. For sampling, it is impractical to enumerate every combination of samples as in Eqn. (4). Instead, We adopt bootstrap sampling strategy as used in (Rendle et al. 2012; Shan, Lin, and Sun 2018) as unbiased approximation. Error Correction Mechanism In KD-based methods, the student model is trained according to predictions of the teacher model, without considering if they are accurate or not. However, inaccurate predictions of the teacher model that is contradictory with the hard label could harm the student model s performance in two aspects. First, at early stage of training when the teacher model is not well-trained, frequent errors in soft labels may distract the training process of the student model, causing slow convergence (Xu et al. 2020). Second, even at later stage of training when the teacher model is relatively well-trained, it is still likely that the teacher model would occasionally provide mistaken predictions that may cause performance deterioration (Wen, Lai, and Qian 2019). A previous work (Xu et al. 2020) adopts a warm-up scheme by removing distillation loss in the earliest k steps of training. However, it is not clear how to choose an appropriate hyper-parameter k, and it cannot prevent errors after k steps. In this work, we propose to adjust predictions of the teacher model y to align with the hard label y. Specifically, we clamp logit values for the teacher model (if the prediction is inconsistent with the ground truth) as follows: r T eacher(x) 1[y] Max 1[y] r T eacher(x), m (10) where r T eacher could be r A+ or r B+, 1[y] is an indicator function that returns 1 is y = 1 else returns 1, and m is the error correction margin, a hyper-parameter. This procedure could accelerate convergence by eliminating inaccurate predictions at the early stage of training, and further enhance knowledge quality at the later stage to improve student model s performance. The proposed error correction mechanism has the following properties: 1) It does not affect the predictions of the teacher model if they are sufficiently correct (that predicts the true label with at least probability σ(m)); 2) It does not affect training of the teacher model since the computation of distillation loss has no backward gradient for teachers as shown in Fig. 2. Experiments We conduct experiments on real-world datasets to answer the following research questions: RQ1: How do Cross Distil performs compared with the state-of-the-art multi-task learning frameworks; RQ2: Are the proposed modules in Cross Distil effective for improving the performance; RQ3: Does error correction mechanism help to accelerate convergence and enhance knowledge quality; RQ4: Does the student model really benefit from auxiliary ranking knowledge; RQ5: How do the hyper-parameters influence the performance? Datasets We conduct experiments on a publicly accessible dataset Tik Tok 2 and our We Chat dataset. Tiktok dataset is collected from a short-video app with two types of user feedback, i.e., Finish watching and Like . We Chat dataset is collected on We Chat Moments platform through sampling user logs during 5 consecutive days with two types of user feedback, i.e., Not interested and Click . For Tiktok, we randomly choose 80% samples as training set, 10% as validation set and the rest as test set. For We Chat, we split the data according to days and use the data of the first four days for training and the last day for validation and test. The statistics of datasets are given in Table ??. Evaluation Metrics We use two widely adopted metrics, i.e., AUC and Multi-AUC, for evaluation. AUC indicates the bipartite ranking (i.e., x+ x ) performance of the model. AUC = 1 N +N X xj D (I(p(xi) > p(xj))) (11) 2https://www.biendata.xyz/competition/icmechallenge2019/data/ Datasets #Samples #Fields #Features Density(A) Density(B) We Chat 9,381,820 10 447,002 1.510% 9.975% Tik Tok 19,622,340 9 4,691,483 37.994% 1.101% Table 1: Statistics of two datasets. where p(x) is the predicted probability of x being a positive sample and I( ) is the indicator function. The vanilla AUC measures the performance of bipartite ranking where a data point is labeled either as a positive sample or a negative one. However, we are also interested in multipartite ranking performance since samples have multiple classes with an order x++ x+ x + x (for task A). Therefore, following (Shan, Lin, and Sun 2018; Shan et al. 2017), we adopt multi-class area under ROC curve (Multi-AUC) to evaluate multipartite ranking performance on test set. Note that we use the weighted version which considers the class imbalance problem (Hand and Till 2001) and is defined as: Multi-AUC = 2 c(c 1) k>j p(j k) AUC(k, j), (12) where c is the number of classes, p() is the prevalenceweighting function as described in (Ferri, Hern andez-Orallo, and Modroiu 2009), AUC(k, j) is the AUC score with class k as the positive class and j as the negative class. Baseline Methods We choose the following MTL models with different shared network architectures for comparison: Shared-Bottom (Caruana 1997), Cross-Stitch (Misra et al. 2016), MMo E (Ma et al. 2018a), PLE (Tang et al. 2020a). We use two variants of our method: TAUG incorporates augmented tasks on top of MTL models, and Cross Distil extends TAUG by conducting calibrated knowledge distillation. Despite that Both TAUG and Cross Distil could be implemented on most state-of-the-art MTL models, we choose the best competitor (i.e. PLE) as the backbone. RQ1: Performance Comparison Table 2 and 3 show the experiment results of our methods versus other competitors on We Chat and Tik Tok datasets respectively. The bold value marks the best one in one column, while the underlined value corresponds to the best one among all the baselines. To show improvements over the single-task counterpart, we report results of Single-Model which uses a separate network for learning each task. As is shown in the tables, the proposed Cross Distil achieves the best performance improvements over Single-Model in terms of AUC and Multi-AUC 3. These results manifest that Cross Distil could indeed better leverage the knowledge from other tasks to improve both bipartite and multipartite ranking abilities on all tasks. Also, TAUG model alone, without calibrated KD, achieves better performance compared with the backbone model PLE, which validates the effectiveness of task augmentation. 3For large-scale datasets in online advertisement, the improvements of AUC in the table is considerable because of its hardness. Methods Task A-Student Task B-Student Task A-Teacher Task B-Teacher AUC Multi-AUC AUC Multi-AUC AUC Multi-AUC AUC Multi-AUC Single-Model .7528 .6270 .7597 .6024 .7535 .6708 .7604 .6705 Shared-Bottom .7540(+.0012) .6378(+.0108) .7587( .0010) .6145(+.0121) - - - - Cross-Stitch .7582(+.0054) .6360(+.0090) .7600(+.0003) .6195(+.0171) - - - - MMo E .7619(+.0091) .6431(+.0161) .7605(+.0008) .6226(+.0202) - - - - PLE .7625(+.0097) .6394(+.0124) .7607(+.0010) .6240(+.0216) - - - - TAUG .7632(+.0104) .6432(+.0162) .7612(+.0015) .6394(+.0370) .7625(+.0090) .6853(+.0145) .7608(+.0004) .6768(+.0063) Cross Distil .7644(+.0116) .6879(+.0609) .7618(+.0021) .6861(+.0837) .7618(+.0083) .6910(+.0202) .7609(+.0005) .6850(+.0145) Table 2: Experiment results of Cross Distil and competitors on We Chat dataset. Methods Task A-Student Task B-Student Task A-Teacher Task B-Teacher AUC Multi-AUC AUC Multi-AUC AUC Multi-AUC AUC Multi-AUC Single-Model .7456 .6335 .9491 .7966 .7453 .7140 .9481 .8297 Shared-Bottom .7375( .0081) .6344(+.0009) .9489( .0002) .8101(+.0135) - - - - Cross-Stitch .7468(+.0012) .6445(+.0110) .9488( .0003) .7985(+.0019) - - - - MMo E .7479(+.0023) .6474(+.0139) .9490( .0001) .7980(+.0014) - - - - PLE .7485(+.0029) .6464(+.0129) .9495(+.0004) .7983(+.0017) - - - - TAUG .7491(+.0035) .6743(+.0408) .9498(+.0007) .8081(+.0115) .7485(+.0032) .7408(+.0268) .9501(+.0020) .8335(+.0038) Cross Distil .7494(+.0038) .7411(+.1076) .9513(+.0022) .8341(+.0375) .7487(+.0034) .7403(+.0263) .9502(+.0021) .8324(+.0027) Table 3: Experiment results of Cross Distil and competitors on Tik Tok dataset. Variants AUC Multi-AUC w/o Auxiliary Rank .7488 ( .0006) .6510 ( .0901) w/o Calibration .7478 ( .0016) .7396 ( .0015) w/o Correction .7486 ( .0008) .7399 ( .0012) KD (same task) .7489 ( .0005) .6901 ( .0510) KD (cross task) .7269 ( .0225) .6120 ( .1291) Baseline .7494 .7411 Table 4: Ablation analysis for Task A on Tik Tok dataset. Variants AUC Multi-AUC w/o Auxiliary Rank .9501 ( .0012) .8005 ( .0336) w/o Calibration .9504 ( .0009) .8312 ( .0029) w/o Correction .9508 ( .0005) .8310 ( .0031) KD (same task) .9505 ( .0008) .8014 ( .0327) KD (cross task) .9184 ( .0329) .7520 ( .0821) Baseline .9513 .8341 Table 5: Ablation analysis for Task B on Tik Tok dataset. Besides, there are several observations in comparison tables. First, Single-Model on augmented ranking-based tasks (teacher) achieves better results in Multi-AUC compared with Single-Model on original regression-based task (student). It verifies that the proposed augmented tasks are capable of capturing task-specific fine-grained ranking information. Second, the student model exceeds the teacher model both in AUC and Multi-AUC performance in most cases, which is not strange since the student benefits from additional training signals that could act as label smoothing regu- larization and the teacher does not have such advantage. The same phenomenon is observed in many other works (Yuan et al. 2020; Tang et al. 2020b; Zhang and Sabuncu 2020) RQ2: Ablation Study We design a series of ablation studies to investigate the effectiveness of some key components. Four variants are considered to simplify Cross Distil by: i) removing BPR losses for learning auxiliary ranking relations, ii) directly employing the teacher model outputs for knowledge distillation without any calibration, iii) not applying the error correction mechanism, vi) using regressionbased teacher models that learn the same task as students and using the vanilla knowledge distillation that is similar with (Zhou et al. 2018), v) directly using the predictions of another task for distillation. Table 4 and 5 show the results for these variants on Tik Tok dataset and performance drops compared with the baseline (i.e. Cross Distil). For the first variant, teacher loss function degrades to traditional BPR loss with no auxiliary ranking information. Such auxiliary ranking information that contains cross-task knowledge is a key factor for good performance in AUC and Multi-AUC. The second variant without calibration may produce unreliable soft labels and result in performance deterioration. Also, it is worth mentioning that the calibration process could significantly improve the performance of Log Loss, which is a widely used regression-based metric. Concretely, Log Loss reduces from 0.5832 to 0.5703 for task A, and 0.0623 to 0.0337 for task B by using calibration. The results of the third variant indicate that the error correction mechanism can also bring up improvements for AUC and Multi-AUC. Another benefit of error correction is to accelerate model training, which will be further discussed. For the fourth variant, we can see that the proposed Cross Distil 0.734 0.736 0.738 0.740 0.832 0.834 0.836 0.838 Multi-AUC for Task A Multi-AUC for Task B -4 -2 -1 0 1 2 4 (a) Error correction margin m Multi-AUC for Task B Multi-AUC for Task A 0.740 0.742 0.736 0.738 0.832 0.836 0.840 0 0.1 0.2 0.3 0.4 0.5 (b) Coefficient β1 Multi-AUC for Task B Multi-AUC for Task A 0.738 0.740 0.742 0.830 0.832 0.834 0.836 0 0.3 0.1 0.2 0.4 0.5 (c) Coefficient β2 Multi-AUC for Task A Multi-AUC for Task B 0.738 0.740 0.828 0.832 0.836 0 1.0 0.8 0.6 0.4 0.2 (d) Distillation loss weight α Figure 3: Multi-AUC performance on Tik Tok dataset for Task A and Task B w.r.t. different hyper-parameters. w/o Correction with Correction w/o Correction with Correction 2000 0 4000 6000 8000 10000 (a) Task A (Finish Watching). 2000 2500 3000 3500 4000 w/o Correction with Correction 1000 2000 3000 4000 5000 0 0.0 (b) Task B (Like). Figure 4: Learning curves of Cross Distil with and without error correction mechanism on Tik Tok dataset. 0.743 0.745 0.747 0.749 0.67 0.69 0.71 0.73 0.75 0 0.1 0.3 0.5 0.7 0.9 (a) Task A (Finish Watching). 0.948 0.949 0.950 0.951 0.952 0.953 0.74 0.76 0.78 0.80 0.82 0.84 0 0.1 0.3 0.5 0.7 0.9 (b) Task B (Like). Figure 5: Impact of corrupted auxiliary ranking information on the student model performance for Tik Tok dataset. is better than the vanilla KD since it transfers fine-grained ranking knowledge across tasks. For the last variant, directly conducting KD could cause performance drop because of the ranking conflicts of tasks. RQ3: Does Error Correction Mechanism Help to Accelerate Convergence and Enhance Knowledge Quality? To answer this question, we plot the learning curves of test loss with (blue line) and without (red line) error correction in Fig. 4. As we can see, for both tasks, the test loss of Cross Distil with error correction significantly goes down faster at the beginning of the training process when the teacher is not well-trained. Plus, at later stage of training when the teacher becomes well-trained, the test loss of Cross Distil with error correction slowly keeps going down and achieves better optimal results compared with the variant, indicating that the proposed error correction mechanism could indeed help to improve knowledge quality. RQ4: Does the Student Model Really Benefit from Auxiliary Ranking Knowledge from Other Tasks? To answer this question, we conduct the following experiment: For a target task A, we randomly choose a certain ratio of pos- itive samples of task B, and then exchange their task B s label with the same number of randomly selected negative samples, to create a corrupted training set. Note that such data corruption process only has negative effects on the reliability of the auxiliary ranking information, so that we can investigate its impact on the student model s performance. Figure 5 shows the results of performance change when increasing the ratio from 10% to 90%. The results indicate that flawed auxiliary information has considerable negative effects on the overall performance, which again verifies Cross Distil could effectively transfer knowledge across tasks. RQ5: Hyper-parameter Study This subsection studies the performance variation of Cross Distil w.r.t. some key hyper-parameters (i.e. error correction margin m, auxiliary ranking loss coefficient β1 and β2, distillation loss weight α). Figure 3(a) shows the Multi-AUC performance with error correction margin ranges from 4 to 4. As we can see, the model performance first increases and then decreases. This is because extremely small m is equivalent to not conducting error correction, while extremely large m makes the soft labels degrade to hard labels. The results in Fig. 3(b) and Fig. 3(c) indicate a proper setting for β can help to capture the correct underlying fine-grained ranking information. The results in Fig. 3(d) reveal that a proper α from 0 to 1 can bring the best performance, which is reasonable since the distillation loss plays the role of label smoothing regularization and could not replace hard labels. In this paper, we propose a cross-task knowledge distillation framework for multi-task recommendation. First, augmented ranking-based tasks are designed to capture finegrained ranking knowledge, which could avoid conflicted information to alleviate negative transfer problem and prepare for subsequent knowledge distillation. Second, calibrated knowledge distillation is adopted to transfer knowledge from augmented tasks (teacher) to original tasks (student). Third, an additional error correction method is proposed to speed up the convergence and improve knowledge quality in the synchronous training process. Cross Distil could be incorporated in most off-the-shelf multi-task learning models, and is easy to be extended or modified for industrial applications such as online advertising. The core idea of Cross Distil could inspire a new paradigm for solving domain-specific task conflict problem and enhancing knowledge transfer in broader areas of data mining and machine learning. Acknowledgments This work was supported by the National Key R&D Program of China [2020YFB1707903]; the National Natural Science Foundation of China [61872238, 61972254], Shanghai Municipal Science and Technology Major Project [2021SHZDZX0102], the Tencent Marketing Solution Rhino-Bird Focused Research Program [FR202001], the CCF-Tencent Open Fund [RAGR20200105], and the Huawei Cloud [TC20201127009]. Xiaofeng Gao is the corresponding author. References Anil, R.; Pereyra, G.; Passos, A.; Ormandi, R.; Dahl, G. E.; and Hinton, G. E. 2018. Large scale distributed neural network training through online distillation. ar Xiv preprint ar Xiv:1804.03235. Caruana, R. 1997. Multitask learning. Machine learning, 28(1): 41 75. Duong, L.; Cohn, T.; Bird, S.; and Cook, P. 2015. Low resource dependency parsing: Cross-lingual parameter sharing in a neural network parser. In ACL, 845 850. Ferri, C.; Hern andez-Orallo, J.; and Modroiu, R. 2009. An experimental comparison of performance measures for classification. PRL, 30(1): 27 38. Hand, D. J.; and Till, R. J. 2001. A simple generalisation of the area under the ROC curve for multiple class classification problems. Machine Learning, 45(2): 171 186. Hinton, G.; Vinyals, O.; and Dean, J. 2015. Distilling the knowledge in a neural network. ar Xiv preprint ar Xiv:1503.02531. Hsieh, M.-E.; and Tseng, V. 2021. Boosting Multi-task Learning Through Combination of Task Labels-with Applications in ECG Phenotyping. In AAAI, volume 35, 7771 7779. Kim, K.; Ji, B.; Yoon, D.; and Hwang, S. 2021. Selfknowledge distillation with progressive refinement of targets. In ICCV, 6567 6576. Lu, Y.; Dong, R.; and Smyth, B. 2018. Why I like it: multitask learning for recommendation and explanation. In Rec Sys, 4 12. Ma, J.; Zhao, Z.; Chen, J.; Li, A.; Hong, L.; and Chi, E. H. 2019. Snr: Sub-network routing for flexible parameter sharing in multi-task learning. In AAAI, volume 33, 216 223. Ma, J.; Zhao, Z.; Yi, X.; Chen, J.; Hong, L.; and Chi, E. H. 2018a. Modeling task relationships in multi-task learning with multi-gate mixture-of-experts. In SIGKDD, 1930 1939. Ma, X.; Zhao, L.; Huang, G.; Wang, Z.; Hu, Z.; Zhu, X.; and Gai, K. 2018b. Entire space multi-task model: An effective approach for estimating post-click conversion rate. In SIGIR, 1137 1140. Misra, I.; Shrivastava, A.; Gupta, A.; and Hebert, M. 2016. Cross-stitch networks for multi-task learning. In CVPR, 3994 4003. Niculescu-Mizil, A.; and Caruana, R. 2005. Predicting good probabilities with supervised learning. In ICML, 625 632. Pan, J.; Mao, Y.; Ruiz, A. L.; Sun, Y.; and Flores, A. 2019. Predicting different types of conversions with multi-task learning in online advertising. In SIGKDD, 2689 2697. Phuong, M.; and Lampert, C. 2019. Towards understanding knowledge distillation. In ICML, 5142 5151. Platt, J.; et al. 1999. Probabilistic outputs for support vector machines and comparisons to regularized likelihood methods. Advances in Large Margin Classifiers, 10(3): 61 74. Rendle, S.; Freudenthaler, C.; Gantner, Z.; and Schmidt Thieme, L. 2012. BPR: Bayesian personalized ranking from implicit feedback. ar Xiv preprint ar Xiv:1205.2618. Ruder, S. 2017. An overview of multi-task learning in deep neural networks. ar Xiv preprint ar Xiv:1706.05098. Shan, L.; Lin, L.; and Sun, C. 2018. Combined regression and tripletwise learning for conversion rate prediction in real-time bidding advertising. In SIGIR, 115 123. Shan, L.; Lin, L.; Sun, C.; Wang, X.; and Liu, B. 2017. Optimizing ranking for response prediction via triplet-wise learning from historical feedback. International Journal of Machine Learning and Cybernetics, 8(6): 1777 1793. Tang, H.; Liu, J.; Zhao, M.; and Gong, X. 2020a. Progressive layered extraction (PLE): A novel multi-task learning model for personalized recommendations. In Rec Sys, 269 278. Tang, J.; Shivanna, R.; Zhao, Z.; Lin, D.; Singh, A.; Chi, E. H.; and Jain, S. 2020b. Understanding and improving knowledge distillation. ar Xiv preprint ar Xiv:2002.03532. Tang, J.; and Wang, K. 2018. Ranking distillation: Learning compact ranking models with high performance for recommender system. In SIGKDD, 2289 2298. Wang, N.; Wang, H.; Jia, Y.; and Yin, Y. 2018. Explainable recommendation via multi-task learning in opinionated text data. In SIGIR, 165 174. Wen, T.; Lai, S.; and Qian, X. 2019. Preparing lessons: Improve knowledge distillation with better supervision. ar Xiv preprint ar Xiv:1911.07471. Xu, C.; Li, Q.; Ge, J.; Gao, J.; Yang, X.; Pei, C.; Sun, F.; Wu, J.; Sun, H.; and Ou, W. 2020. Privileged features distillation at Taobao recommendations. In SIGKDD, 2590 2598. Yang, Y.; and Hospedales, T. 2016. Deep multi-task representation learning: A tensor factorisation approach. ar Xiv preprint ar Xiv:1605.06391. Yu, T.; Kumar, S.; Gupta, A.; Levine, S.; Hausman, K.; and Finn, C. 2020. Gradient surgery for multi-task learning. Neur IPS. Yuan, L.; Tay, F. E.; Li, G.; Wang, T.; and Feng, J. 2020. Revisiting knowledge distillation via label smoothing regularization. In CVPR, 3903 3911. Zhang, Y.; and Yang, Q. 2021. A survey on multi-task learning. TKDE. Zhang, Z.; and Sabuncu, M. R. 2020. Self-distillation as instance-specific label smoothing. ar Xiv preprint ar Xiv:2006.05065. Zhou, G.; Fan, Y.; Cui, R.; Bian, W.; Zhu, X.; and Gai, K. 2018. Rocket launching: A universal and efficient framework for training well-performing light net. In AAAI, volume 32. Zhou, H.; Song, L.; Chen, J.; Zhou, Y.; Wang, G.; Yuan, J.; and Zhang, Q. 2021. Rethinking soft labels for knowledge distillation: A bias-variance tradeoff perspective. ICLR. Zhu, J.; Liu, J.; Li, W.; Lai, J.; He, X.; Chen, L.; and Zheng, Z. 2020. Ensembled CTR prediction via knowledge distillation. In CIKM, 2941 2958.