# teaching_semisupervised_classifier_via_generalized_distillation__4430e2bd.pdf Teaching Semi-Supervised Classifier via Generalized Distillation Chen Gong1,2, Xiaojun Chang3, Meng Fang4, Jian Yang1,2 1 School of Computer Science and Engineering, Nanjing University of Science and Technology 2 Jiangsu Key Laboratory of Image and Video Understanding for Social Security 3 Language Technologies Institute, Carnegie Mellon University 4 Tencent AI Lab chen.gong@njust.edu.cn Semi-Supervised Learning (SSL) is able to build reliable classifier with very scarce labeled examples by properly utilizing the abundant unlabeled examples. However, existing SSL algorithms often yield unsatisfactory performance due to the lack of supervision information. To address this issue, this paper formulates SSL as a Generalized Distillation (GD) problem, which treats existing SSL algorithm as a learner and introduces a teacher to guide the learner s training process. Specifically, the intelligent teacher holds the privileged knowledge that explains the training data but remains unknown to the learner, and the teacher should convey its rich knowledge to the imperfect learner through a specific teaching function. After that, the learner gains knowledge by imitating the output of the teaching function under an optimization framework. Therefore, the learner in our algorithm learns from both the teacher and the training data, so its output can be substantially distilled and enhanced. By deriving the Rademacher complexity and error bounds of the proposed algorithm, the usefulness of the introduced teacher is theoretically demonstrated. The superiority of our algorithm to the stateof-the-art methods has also been demonstrated by the experiments on different datasets with various sources of privileged knowledge. 1 Introduction Semi-Supervised Learning (SSL) [Zhu and Goldberg, 2009] is proposed for solving the real-world problems where the labeled examples are inadequate, yet a large amount of unlabeled examples are relatively easy to obtain. For example, manually annotating web-scale images is intractable due to the unacceptable human labor cost. Acquiring sufficient labeled examples for protein structure categorization is also infeasible as it often takes months of laboratory work for experts to identify a single protein s 3D structure. Fortunately, the massive unlabeled data in these applications are usually available which carry rich distribution information and are helpful for training a reliable classifier. Regarding how to utilize such abundant unlabeled examples, current SSL algorithms are usually built under manifold assumption or cluster assumption. Manifold assumption [Belkin et al., 2006; Gong et al., 2015a] assumes that the massive unlabeled examples as well as the scarce labeled examples reveal the underlying manifold of the entire dataset, and the labels of all examples should vary smoothly along this manifold. Cluster assumption [Joachims, 1999; Li and Zhou, 2015] holds that the examples belonging to the same class usually form a cluster in the feature space, and the true decision boundary should fall into the low density region between the well-separated clusters. From above analyses, we see that existing methods exploit unlabeled examples via different ways to relieve the adverse impact caused by the insufficiency of labeled examples. However, their performances are not stable and are still far from perfect in many situations. Therefore, this paper aims to use machine teaching strategy so that some additional information can be incorporated to aid the training of semisupervised classifier. Recently, Lopez-Paz et al. [Lopez-Paz et al., 2016] proposed a novel Generalized Distillation (GD) framework which integrates the privileged information proposed in [Vapnik and Izmailov, 2015] and the distillation developed by [Hinton et al., 2015]. In GD, an intelligent teacher (i.e. a teaching function) is introduced to transfer its rich knowledge to the imperfect learner , during which the teacher s knowledge constitutes the privileged information that explains the training data to the learner. After that, the learner receives the teacher s knowledge by imitating the output of the teacher so that its learning performance can be enhanced, and this process is called distillation . By this way, a teacher-learner-collaborative framework is established which is also very similar to the educational and cognitive process of humans. The problem studied in this paper is mathematically defined as follows. Suppose we have a set of labeled examples L = {(xi, x i , Yi)}l i=1 and a set of unlabeled examples U = {(xi, x i , Yi)}l+u i=l+1 (typically l u), where xi (1 i n, n = l+u) are d-dimensional regular data features for the learner, and {x i }n i=1 are privileged features that are only Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence (IJCAI-18) accessible to the teacher. For any xi L, Yi R1 C (C is the total number of classes) is its label vector with (Yi)j = Yij being 1 if xi belongs to the j-th class, and 0 otherwise. For xi U, all the elements in the associated Yi are initially set to 1/C as its groundtruth label is unknown. Given the training set denoted by Z = L U, the target of learner is to find a suitable semi-supervised classifier fs based on Z {x i }n i=1 so that the label of an unseen test example x0 can be perfectly predicted as y0 = fs(x0). Note that {x i }n i=1 are deducted from Z as {x i }n i=1 are invisible to the learner. The function of teacher ft is to generate the soft label Si [0, 1]1 C for any xi U based on the privileged features {x i }n i=1 associated with the training examples. These additional soft labels provide the class prior that is not contained in Z for the learner, thereby they bias the learned fs to the output of ft on U and also effectively reduce the capacity of fs s hypothesis space. The algorithm proposed in this paper is termed Generalized Distillation for Semi-Supervised Learning (GDSSL). In GDSSL, we use a popular graph transduction model called Gaussian Field and Harmonic Functions (GFHF) [Zhu et al., 2003] as a teacher, and employ the induction method named Laplacian Regularized Least Squares (Lap RLS) [Belkin et al., 2006] as our learner. A transduction method only focuses on labeling the unlabeled examples that are observable in the training set, and does not care about the classification of unseen test data. Therefore, it will probably produce better results on U than the induction algorithm that pays more attention to the test accuracy than the training accuracy [Gammerman et al., 1998]. As a result, GFHF is adopted as the intelligent teacher ft to guide the training process of Lap RLS (i.e. fs). Our proposed GDSSL has several merits: Algorithmically, GDSSL enjoys the good property of GD that the privileged information can still be utilized for model training although they are originally withheld from the learner [Lopez-Paz et al., 2016]; Theoretically, we derive the transductive error bound and inductive error bound of the proposed GDSSL, which clearly indicate the usefulness of machine teaching to SSL; and Empirically, by comparing GDSSL with several related state-of-the-art methods, the effectiveness of the proposed method has also been confirmed. 2 Related Work As this paper aims to harness GD to teach an SSL algorithm, we review the representative prior works on these two topics. 2.1 Semi-Supervised Learning SSL is specifically designed for the applications where sufficient labeled examples are difficult to obtain while massive unlabeled examples are available. Existing SSL algorithms are usually developed under manifold assumption or cluster assumption. Cluster assumption assumes that the classes formed by both labeled and unlabeled examples are well-separated, such that the ideal decision boundary falls exactly into the low density area in the feature space. The representative algorithms based on cluster assumption include [Joachims, 1999; Grandvalet and Bengio, 2004; Li et al., 2009; 2016b; Yan et al., 2016]. Manifold assumption based methods explore the geometry of data distribution by postulating that it is supported by a Riemannian manifold. To describe this manifold, a graph is usually built where the vertices correspond to the labeled and unlabeled examples and the edges between the vertices encode the pairwise similarity between them. The label information can then be transmitted from the limited labeled examples to the remaining unlabeled examples through the edges, during which the labels of all examples are required to vary smoothly on the graph. For example, Zhu et al. [Zhu et al., 2003], Zhou et al. [Zhou et al., 2004] and Gong et al. [Gong et al., 2015a] respectively deploy the standard graph Laplacian, normalized graph Laplacian, and deformed graph Laplacian to smooth the labels of examples on the graph. Other works grounded on manifold assumption include [Wang et al., 2009; Belkin et al., 2006; Fang et al., 2014; Gong et al., 2015b; Liu et al., 2017; Gong et al., 2017a; 2017b]. The proposed GDSSL follows manifold assumption, as it is general and includes the cluster assumption as one of its special case [Chapelle et al., 2006]. Apart from harnessing the unlabeled examples for improving the performance like above SSL methods, our GDSSL also incorporates privileged information carried by a teacher via generalized distillation. 2.2 Generalized Distillation The technique of distillation was initially presented by [Hinton et al., 2015] to reduce the computational cost in test stage for an ensemble of large deep neural networks. Besides, [Vapnik and Izmailov, 2015] proposed a new paradigm named Learning Using Privileged Information (LUPI), which incorporates a knowledgable teacher with privileged information and a traditional learner. Here the privileged information, which serve as the additional descriptions to the training data, can only be used by the teacher in training the learner. LUPI has been widely used in many machine learning scenarios such as metric learning [Yang et al., 2016], transfer hashing [Zhou et al., 2016], multi-label learning [You et al., 2017], and deep learning [Hu et al., 2016]. By combining distillation and LUPI into a unified framework, [Lopez-Paz et al., 2016] proposed a novel framework called Generalized Distillation (GD). In GD, the teacher transfers its privileged knowledge to the learner, and then the learner absorbs these knowledge by mimicking the outputs of the teacher on the training data. Algorithmically, GD has been employed for domain adaptation when the data in source domain are not accessible [Ao et al., 2017]. Practically, Celik et al. [Celik et al., 2016] use GD for disease diagnosis when the patients are not willing to reveal their privacy-sensitive data. To the best of our knowledge, this paper is the first devoted work to adapt GD to solving general SSL problem. Thanks to the efforts of teacher, our method is demonstrated to be more effective than the traditional SSL algorithms from both theoretical and empirical aspects. 3 The Proposed Method Our GDSSL contains two roles, namely a teacher and a learner . The overall framework of GDSSL is presented in Fig. 1. In training stage, the teacher ft uses the privileged Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence (IJCAI-18) regular input features 1 0 x test example 0 0 ( ) " " s y f dog x Label: 1 2 , , T T T T n Y Y Y Y privileged features Training Test Figure 1: The framework of our algorithm. In training stage, the privileged information (e.g. the textual explanations in Wikipedia for the interested objects) is employed by teacher to generate a soft label matrix S, which is further combined with the given labels Y to form the auxiliary label matrix P. Next, the learner utilizes the regular input features (e.g. the color feature of images) as well as P to train a decision function fs. In test stage, fs classifies a test image into the correct class. information of training data to generate a soft label matrix S over Z, where each row of S corresponds to an example. After that, S and the initial labels in Y are fused to the auxiliary label matrix P in a weighted sum way. Then the learner uses P and the regular input features to establish the decision function fs. In test stage, a test example x0 is fed into fs to get its label fs(x0). Next we detail the training process of GDSSL accomplished by the teacher and the learner. 3.1 Teaching Model In GDSSL, we use the off-the-shelf transductive method GFHF [Zhu et al., 2003] as the teacher ft because this method is effective, efficient, and easy to implement. To this end, we build a K-NN graph Gt = Vt, Et where Vt is the vertex set consisted of all n examples in Z, and Et is the edge set encoding the similarity between these examples in terms of the privileged features {x i }n i=1. Here the similarity between two examples x i and x j is computed by the Gaussian kernel [Wt]ij = exp x i x j 2/(2ξ2) with ξ being the kernel width if x i and x j are linked by an edge in Gt, and [Wt]ij =0 otherwise. To mathematically quantify Gt, an n n adjacency matrix Wt is computed where its (i, j)-th element is [Wt]ij. Therefore, the diagonal degree matrix is defined by [Dt]ii = Pn j=1 [Wt]ij and the graph Laplacian matrix is Lt = Dt Wt. By recalling the definition of label vectors {Yi}l i=1, the objective function of GFHF is [Zhu et al., 2003] min F tr(F Lt F) s.t. FL = YL, (1) where YL = Y 1 , Y 2 , , Y l , and F = FL FU Rn C is the variable to be optimized with each row corresponding to an example in L or U. By partitioning Lt as Lt = LL,L LL,U LU,L LU,U , the solution of (1) is FL = YL, FU = L 1 U,ULU,LYL. (2) To squash the elements in F to [0, 1], we use a softmax operation to process every Fi (i=1, 2, , n), and obtain Sij = exp([Fi]j/T) PC i=1 exp([Fi]j/T) , (3) where [Fi]j denotes the j-th element in the row vector Fi, and T > 0 is the temperature parameter which decides how much we want to soften the S output by the teacher ft. Therefore, the element Sij can be interpreted as the probability of the i-th example belonging to the j-th class. 3.2 Learning Model This section details how to train an inductive learner fs : Rd+1 RC with fs(x)=x Θ based on S and Z {x i }n i=1. Here Θ is the coefficient matrix compatible with the augmented x as x := x 1 . According to [Lopez-Paz et al., 2016], the loss function for training the learner fs on any training example xi Z is V (Yi, Si, fs(xi)) = λℓ(Yi, fs(xi))+(1 λ)ℓ(Si, fs(xi)) , (4) where the first term is the traditional fidelity loss enforcing the output of fs(xi) to be close to the given label Yi, and the second term is called imitation loss that pushes the learner to generate similar result with the soft label Si produced by the teacher. The relative weights of such two terms are governed by a trade-off parameter λ [0, 1]. Let ℓ(Yi, fs(xi)) = Yi fs(xi) 2 be the squared loss in this paper, it can be easily verified that (4) is equivalent to V (Yi, Si, fs(xi)) = ℓ(λYi + (1 λ)Si, fs(xi)) + Const ℓ(Pi, fs(xi)), (5) where Const is irrelevant to fs, and Pi =λYi+(1 λ)Si is the auxiliary label in which every element is within [0, 1]. The cross-entropy loss mentioned in [Lopez-Paz et al., 2016] is not employed here as it makes the learning model do not have closed-form solution. Thereby, inspired by Lap RLS [Belkin et al., 2006], the model for pursuing fs is expressed as min Θ J (Θ)= 1 i=1 ℓ(Pi, fs(xi)) + αtr(Θ XLs X Θ)+β Θ 2, (6) where X = (x1, x2, , xn) Rd n is the training data matrix, and Ls is the graph Laplacian matrix of learner that can be computed via a similar way with Lt. In (6), the first term is exactly (5), the second term enables the labels of all examples to vary smoothly along the potential manifold, and the third term controls the model complexity to prevent overfitting. Above three terms are weighted by the nonnegative tuning parameters α and β. To solve (6), we set the derivative of J (Θ) to Θ to zero, and then obtain the optimal Θ as Θ = XX + αXLs X + βI 1XP, (7) where I denotes the identity matrix. Based on the optimized Θ, we may compute the label vector of a test example x0 as F0 = x 0 Θ, and then x0 is classified into the j-th class with j =arg maxj {1, ,C} [F0]j . 4 Theoretical Analyses For a binary classification problem, the learner s objective function for generating the final decision function becomes min θ J (θ)= 1 θ xi pi 2 +αθ XLs X θ+β θ 2 # where pi [ 1, 1] is the target auxiliary label of xi, and θ is a d-dimensional coefficient vector to be optimized. Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence (IJCAI-18) 4.1 Semi-Supervised Rademacher Complexity In computational learning theory, Rademacher complexity is employed to measure the complexity of a class of functions with respect to a certain data distribution. [El-Yaniv and Pechyony, 2009] provide the formulation of Rademacher complexity for SSL, which is Definition 1. Let Z = {x1, , xl, xl+1, , xl+u} be n = l+u examples drawn i.i.d. from a distribution PZ, and F be a hypothesis space of semi-supervised classifier f : Z R, then the empirical Rademacher complexity of F is defined by Eσ h supf F Xn i=1 σif(xi) i , (9) where σ1, , σl are independent Rademacher variables randomly chosen from {1, 1, 0} with the probabilities p, p, and 1 2 p ( p [0, 1/2]), respectively. Based on above definition, we have the upper bound of Rademacher complexity of our GDSSL, which is Theorem 2. Suppose θ is the optimal solution of (8), and Fs is the hypothesis space of learner that is guided by a teacher in GDSSL. Then the empirical Rademacher complexity of GDSSL satisfies RZ(Fs) n2 βlumax i xi . (10) Proof. Since θ is the minimizer of (8), we have J (θ ) J (θ = 0) = 1 2 Pn i=1 p2 i n 2 . By noticing that every term in (8) is nonnegative, we know β 2 θ 2 J (θ ) n leads to θ q n β . Besides, from Definition 1 we have n RZ(Fs) = Eσ h supf Fs Xn i=1 σif(xi) i = Eσ h supf: θ n i=1 σi θ, xi i = Eσ h supf: θ n i=1 σixi Ei i=1 σixi i , (11) where Eσ( ) computes the expectation over σ = (σ1, , σn), and the last inequality holds due to the Cauchy-Schwarz inequality. By using Jensen s inequality, we obtain Eσ[ Pn i=1 σixi ] = Eσ h Pn i=1 σixi 2 1/2i Eσ h Pn i=1 σixi 2i 1/2 . Since σ1, , σn are independent, it is straightforward that i=1 σixi 2 = Eσ h X i,j σiσj xi, xj i i =j xi, xj Eσ [σiσj] + Xn i=1 xi, xi Eσ σ2 i = 0 + 2 p Xn i=1 xi 2 nmaxi xi 2, which further proves (10) by recalling (11). 4.2 Error Bounds For an SSL decision function f, its empirical error on the labeled set L with size l is ℜl(f) = 1 l Pl i=1 ℓ(f(xi), yi) where yi is the given label of xi. The transductive error on the unlabeled set U with size u is defined by ℜu(f) = 1 u Pl+u i=l+1 ℓ(f(xi), yi). Similarly, the training error of f on L U is ℜn(f) = 1 n Pn i=1 l(f(xi), yi). Besides, the inductive error (i.e. generalization error) of f is ℜ(f) = Ex P [ℓ(f(x), y)] where x is generated from a certain distribution P. Based on above definitions, we introduce two useful lemmas for bounding the transductive error and inductive error of the proposed GDSSL: Lemma 3. [El-Yaniv and Pechyony, 2009] Let F be the hypothesis space of f on the training set Z = L U with |L| = l and |U| = u. Let c = p 32ln(4e)/3 < 5.05, Q = 1/l + 1/u and B = l+u (l+u 1/2 )(1 1/(2 max(l,u))). For any fixed γ, with the probability at leat 1 δ, we have ℜu(f) ℜγ l (f)+ 1 γ RZ(F)+c Q p Lemma 4. [Bartlett and Mendelson, 2002] Suppose Z = {x1, , xn} are generated i.i.d. from a distribution P, and F is the hypothesis space, then for any f F with the probability at least 1 δ, we have ℜ(f) ℜn(f) + 2RZ(F) + 3 By directly putting the result of (10) into (13) and (14), we can easily obtain the transductive error bound and inductive error bound of our GDSSL, which are Theorem 5. (transductive error bound) Given the notations c , γ, Q, and B defined in Lemma 3 with γ fixed to 1, then with the probability at leat 1 δ, the transductive error of the student fs in GDSSL satisfies ℜu(fs) ℜl(fs)+ n2 βlumax i xi +c Q Theorem 6. (inductive error bound) Let Fs be the hypothesis space of learner fs that is guided by a teacher, then for any fs Fs, with the probability at least 1 δ, we have ℜ(fs) ℜn(fs) + 2n2 βlumax i xi + 3 From above Theorems 5 and 6, we see that both the transductive error on unlabeled training examples and the inductive error on test examples are upper bounded, so our GDSSL is guaranteed to produce very encouraging performance. 4.3 Usefulness of Teaching From the general bounds displayed in Eqs. (13) and (14), we see that for a certain training set Z, the Rademacher complexity of hypothesis space F regarding Z critically determines the upper bounds for transduction and induction. Next we will theoretically show that the teacher introduced by our GDSSL is able to decrease the Rademacher complexity of the space of the generated decision function, so the transductive error bound and inductive error bound of GDSSL can be effectively reduced when compared with the vanilla Lap RLS learner that is not equipped with a teacher. Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence (IJCAI-18) Suppose the decision function produced by Lap RLS is f Lap, which is in the hypothesis space FLap, then it is straightforward that |FLap| |Fs|. This is because the prediction of the learner fs in GDSSL should be not only consistent with the known labels in L, but also mimic the teacher s label output on U. In contrast, the output of f Lap is only required to be similar to the labels of L. Based on this observation, we have the following theorem: Theorem 7. Suppose Fs and FLap are respectively the hypothesis spaces of GDSSL and Lap RLS, then given a training set Z, their empirical Rademacher complexities satisfy RZ(Fs) RZ(FLap). (17) Proof. Suppose that FLap = Fs +F where F is the complementary space of Fs in FLap, and Q = 1 u, then according to (9), we have RZ(FLap) = QEσ h supf FLap i=1 σif(xi) i = QEσ h supf Fs+F Xn i=1 σif(xi) i i=1 σif(xi) +QEσ i=1 σif(xi) # = RZ(Fs) + QEσ h supf F Xn i=1 σif(xi) i . (18) By using the Jensen s inequality, it is obvious that Eσ h supf F, (Pn i=1 σif(xi)) i supf F Eσ [Pn i=1 σif(xi)] = supf F Pn i=1 E (σi) f(xi) = 0. The last equation holds because E(σi) = 0 based on the Definition 1. Consequently, the Theorem 7 is proved. Therefore, the teacher in our GDSSL plays an important role in reducing the learner s empirical Rademacher complexity, which further leads to the decreased transductive error bound and inductive error bound. As a result, GDSSL consistently performs better than the plain Lap RLS algorithm that is the learner of GDSSL. 5 Experimental Results In this section, we investigate the classification ability of GDSSL on various datasets by using different sources of privileged information. The compared algorithms include: State-of-the-art SSL algorithms: Reliable Label Inference via Smoothness Hypothesis (Re LISH) [Gong et al., 2014], Label Prediction via Deformed Graph Laplacian (LPDGL) [Gong et al., 2015a], and Laplacian Regularized Least Squares (Lap RLS) [Belkin et al., 2006], among which Lap RLS is the basic learner in our GDSSL that is not taught by a teacher. State-of-the-art algorithms that utilize privileged information: Original SVM+ (SVM+) [Vapnik and Izmailov, 2015], and the improved ℓ2-SVM+ [Li et al., 2016a]. 5.1 Face Recognition: High-Resolution Images as Privileged Information In this experiment, we investigate the ability of compared methods on face recognition. Specifically, the ORL dataset Table 1: Performance of compared methods on ORL dataset. The best result is marked in bold. indicates that GDSSL is significantly better than the corresponding method via paired t-test. training accuracy test accuracy Lap RLS 0.713 0.032 0.640 0.066 Re LISH 0.738 0.021 0.700 0.057 LPDGL 0.754 0.019 0.710 0.065 SVM+ - 0.735 0.066 ℓ2-SVM+ - 0.738 0.058 GDSSL (ours) 0.776 0.016 0.755 0.049 [Cai et al., 2006] is employed, which contains totally 400 face images belonging to 40 subjects. Every image has been respectively resized to the resolutions of 32 32 and 64 64, so the 32 32 face images serve as regular input for the learner while the 64 64 images are treated as privileged information that are used for teaching. We directly re-arrange every two-dimensional image example to a long feature vector in which the elements are grayscale values of the corresponding pixels. In this paper, every compared method is evaluated by the 5-fold cross validation on each dataset, and the average accuracy over the outputs of the five independent runs are reported to assess the performance of a certain algorithm. Therefore, the training set in ORL contains 320 examples, in which we randomly select 80 examples into labeled set L and the remaining 240 examples constitute the unlabeled set U. The labeled set and unlabeled set in each fold are identical for all the compared methods throughout the experiment. For fair comparison, we build the same 10-NN graph for the graph-based methods including Lap RLS, LPDGL, Re LISH, and our GDSSL. In GDSSL, the parameter λ is set to 0.4 by searching the grid {0.2, 0.4, 0.6, 0.8}. The temperature parameter T is tuned to 0.01. Besides, the trade-off parameters in (6) are α = β = 0.1. The regularization parameters C and γ in SVM+ and ℓ2-SVM+ are set to 1 and 0.01 to achieve the best performance. The results obtained by all the compared methods are presented in Table 1. We see that our GDSSL achieves the highest accuracy among all the compared methods. The training accuracies of SVM+ and ℓ2SVM+ are not reported as they are fully supervised methods that are only trained on L, so they cannot be directly compared with other semi-supervised methods that are trained on Z = L U. The superiority of GDSSL has also been statistically verified by the paired t-test with the significance level 0.05. Furthermore, by comparing the performances of Lap RLS and GDSSL, we see that GDSSL leads the plain learner Lap RLS with 6.3% on training accuracy and 11.5% on test accuracy, which firmly demonstrate the usefulness of privileged information and the strength of GD-based teacher. 5.2 Document Categorization: Illustrations as Privileged Information We apply the proposed GDSSL to text categorization to see its strength in dealing with non-image tasks. To this end, we use a recent Wikipedia dataset [Pereira et al., 2014] that contains 2866 documents across 10 categories such as history , music , and biology . Apart from the texts, the articles in this dataset are also accompanied by the illustrative images, Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence (IJCAI-18) Table 2: Performance of compared methods on Wikipedia dataset. The best result is marked in bold. indicates that GDSSL is significantly better than the corresponding method via paired t-test. training accuracy test accuracy Lap RLS 0.643 0.010 0.613 0.013 Re LISH 0.607 0.032 0.604 0.030 LPDGL 0.634 0.018 0.630 0.020 SVM+ - 0.569 0.031 ℓ2-SVM+ - 0.597 0.035 GDSSL (ours) 0.663 0.006 0.658 0.018 Table 3: Performance of compared methods on CIFAR100 dataset. The best result is marked in bold. indicates that GDSSL is significantly better than the corresponding method via paired t-test. training accuracy test accuracy Lap RLS 0.565 0.002 0.438 0.003 Re LISH 0.497 0.003 0.266 0.003 LPDGL 0.558 0.001 0.359 0.005 SVM+ - 0.452 0.001 ℓ2-SVM+ - 0.414 0.002 GDSSL (ours) 0.593 0.002 0.460 0.002 which can be naturally regarded as privileged information to aid the classification of text examples. The image and textual features in each document example are directly provided by [Pereira et al., 2014], namely the text is represented by bag-of-words (BOW) encoding and the associated image is characterized by the SIFT descriptor. Similar to the above experiment, we also conduct 5-fold cross validation on this Wikipedia dataset and see the average outputs of the compared methodologies. The split of training set and test set for each fold has been kept identical for all the algorithms. In the training set of size 2293, we only select 20 examples from each class as labeled and thus the supervision information is very limited for GDSSL to achieve satisfactory performance. At this time, the guidance of teacher would be of great value to make up the shortage of labeled examples. The accuracies obtained by various algorithms are illustrated in Table 2, which clearly indicate that our GDSSL achieves the best results among all approaches. In this dataset, the SSL methods (e.g. Lap RLS, Re LISH, LPDGL and GDSSL) perform consistently better than the supervised models such as SVM+ and ℓ2-SVM+. This is because supervised models are trained only based on the insufficient labeled examples, while SSL methods exploit both labeled and unlabeled examples to build a more powerful classifier. 5.3 Natural Image Classification: Different Feature as Privileged Information In this section, we evaluate our GDSSL on classifying general images from everyday life. Specifically, a very challenging dataset CIFAR100 [Krizhevsky and Hinton, 2009] is employed here, which contains 60000 32 32 color images across 100 classes with 600 images per class. Practically, an image can be characterized by more than one feature modalities, and the features from additional modality can be used as privileged information to enhance the classification results. Here we use the features extracted by Alex Net [Krizhevsky et al., 2012] as regular input, and employ the features extracted Figure 2: The parametric sensitivity of T and λ. The first column presents the results by changing T, and the second column plots the results w.r.t. the variation of λ. by VGGNet-16 [Simonyan and Zisserman, 2014] as the privileged information that are only known by the teacher. The 10-NN graph with kernel width ξ = 10 is built for Lap RLS, Re LISH, LPDGL, and GDSSL. From the experimental results provided in Table 3, we see that the test accuracies of all compared methods are below 50% because this dataset is very challenging and the labeled examples are too scarce to train a good classifier. However, our GDSSL is still significantly better than other methodologies in terms of both training accuracy and test accuracy. 5.4 Parametric Sensitivity This section studies how the variations of T in (3) and λ in (4) influence the accuracy of the proposed GDSSL. The training accuracy and test accuracy of GDSSL under different choices of T and λ are observed on the ORL, Wikipedia and CIFAR100 datasets. In every dataset, we investigate the effects of T and λ by fixing one of them and then examining the output w.r.t. the change of the other one. From Fig. 2, we see that the training accuracy of GDSSL will not be largely influenced by the choice of these two parameters, however the test accuracy might drop when T gradually increases, so we simply fix T to 0.01 throughout this paper. The determination of λ depends on the quality of teacher. If the teacher is able to provide accurate label prior on U, one may choose a small λ to put more emphasize on the output of teacher in (4), otherwise this parameter should be tuned up. 6 Conclusion To handle the shortage of labeled data in SSL, this paper proposes to incorporate a teacher that brings the additional privileged information to the training process of learner. By designing the imitation loss under the framework of Generalized Distillation [Lopez-Paz et al., 2016], the learner can study the knowledge imparted by the teacher as well as from the regular features of training data, and thus it will generate a better classifier than the pure learner without the guidance of a teacher. As the proposed GDSSL is a general framework for teaching a semi-supervised classifier, we plan to apply GDSSL to more SSL algorithms to enhance their classification ability. Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence (IJCAI-18) Acknowledgments This research is supported by NSF of China (Nos: 61602246, U1713208 and 61472187), NSF of Jiangsu Province (No: BK20171430), the Fundamental Research Funds for the Central Universities (No: 30918011319), the Summit of the Six Top Talents Program (No: DZXX-027), the 973 Program (No: 2014CB349303), and Program for Changjiang Scholars. [Ao et al., 2017] S. Ao, X. Li, and C. Ling. Fast generalized distillation for semi-supervised domain adaptation. In AAAI, pages 1719 1725, 2017. [Bartlett and Mendelson, 2002] P. Bartlett and S. Mendelson. Rademacher and Gaussian complexities: risk bounds and structural results. JMLR, 3:463 482, 2002. [Belkin et al., 2006] M. Belkin, P. Niyogi, and V. Sindhwani. Manifold regularization: A geometric framework for learning from labeled and unlabeled examples. JMLR, 7(11):2399 2434, 2006. [Cai et al., 2006] D. Cai, X. He, J. Han, and H. Zhang. Orthogonal laplacianfaces for face recognition. TIP, 15(11):3608 3614, 2006. [Celik et al., 2016] Z. Celik, D. Lopez-Paz, and P. Mc Daniel. Patient-driven privacy control through generalized distillation. ar Xiv:1611.08648, 2016. [Chapelle et al., 2006] O. Chapelle, B. Sch olkopf, and A. Zien. Semi-Supervised Learning. MIT Press, Cambridge, MA, 2006. [El-Yaniv and Pechyony, 2009] R. El-Yaniv and D. Pechyony. Transductive rademacher complexity and its applications. JAIR, 35(1):193, 2009. [Fang et al., 2014] Y. Fang, K. Chang, and H. Lauw. Graph-based semi-supervised learning: realizing pointwise smoothness probabilistically. In ICML, pages 406 414, 2014. [Gammerman et al., 1998] A. Gammerman, V. Vovk, and V. Vapnik. Learning by transduction. In UAI, pages 148 155, 1998. [Gong et al., 2014] C. Gong, D. Tao, K. Fu, and J. Yang. Re LISH: Reliable label inference via smoothness hypothesis. In AAAI, pages 1840 1846, 2014. [Gong et al., 2015a] C. Gong, T. Liu, D. Tao, K. Fu, E. Tu, and J. Yang. Deformed graph Laplacian for semisupervised learning. TNNLS, 26(10):2261 2274, Oct. 2015. [Gong et al., 2015b] C. Gong, D. Tao, K. Fu, and J. Yang. Fick s law assisted propagation for semisupervised learning. TNNLS, 26(9):2148 2162, 2015. [Gong et al., 2017a] C. Gong, D. Tao, W. Liu, L. Liu, and J. Yang. Label propagation via teaching-to-learn and learning-to-teach. TNNLS, 28(6):1452 1465, 2017. [Gong et al., 2017b] C. Gong, H. Zhang, J. Yang, and D. Tao. Learning with inadequate and incorrect supervision. In ICDM, pages 889 894. IEEE, 2017. [Grandvalet and Bengio, 2004] Y. Grandvalet and Y. Bengio. Semisupervised learning by entropy minimization. In NIPS, volume 17, pages 529 536, 2004. [Hinton et al., 2015] G. Hinton, O. Vinyals, and J. Dean. Distilling the knowledge in a neural network. ar Xiv:1503.02531, 2015. [Hu et al., 2016] Z. Hu, X. Ma, Z. Liu, E. Hovy, and E. Xing. Harnessing deep neural networks with logic rules. In ACL, pages 2410 2420, 2016. [Joachims, 1999] T. Joachims. Transductive inference for text classification using support vector machines. In ICML, volume 99, pages 200 209, 1999. [Krizhevsky and Hinton, 2009] A. Krizhevsky and G. Hinton. Learning multiple layers of features from tiny images. 2009. [Krizhevsky et al., 2012] A. Krizhevsky, I. Sutskever, and G. Hinton. Image Net classification with deep convolutional neural networks. In NIPS, pages 1097 1105, 2012. [Li and Zhou, 2015] Y. Li and Z. Zhou. Towards making unlabeled data never hurt. TPAMI, 37(1):175 188, 2015. [Li et al., 2009] Y. Li, J. Kwok, and Z. Zhou. Semi-supervised learning using label mean. In ICML, pages 633 640. ACM, 2009. [Li et al., 2016a] W. Li, D. Dai, M. Tan, D. Xu, and L. Gool. Fast algorithms for linear and kernel svm+. In CVPR, pages 2258 2266, 2016. [Li et al., 2016b] Y. Li, J. Kwok, and Z. Zhou. Towards safe semisupervised learning for multivariate performance measures. In AAAI, pages 1816 1822, 2016. [Liu et al., 2017] Y. Liu, Y. Guo, H. Wang, F. Nie, and H. Huang. Semi-supervised classifications via elastic and robust embedding. In AAAI, pages 2294 2300, 2017. [Lopez-Paz et al., 2016] D. Lopez-Paz, B. Sch olkopf, L. Bottou, and V. Vapnik. Unifying distillation and privileged information. In ICLR, 2016. [Pereira et al., 2014] J. Pereira, E. Coviello, G. Doyle, N. Rasiwasia, G. Lanckriet, R. Levy, and N. Vasconcelos. On the role of correlation and abstraction in cross-modal multimedia retrieval. TPAMI, 36(3):521 535, 2014. [Simonyan and Zisserman, 2014] K. Simonyan and A. Zisserman. Very deep convolutional networks for large-scale image recognition. ar Xiv: 1409.1556v6, 2014. [Vapnik and Izmailov, 2015] V. Vapnik and R. Izmailov. Learning using privileged information: similarity control and knowledge transfer. JMLR, 16:2023 2049, 2015. [Wang et al., 2009] J. Wang, F. Wang, and C. Zhang. Linear neighborhood propagation and its applications. TPAMI, 31(9):1600 1615, 2009. [Yan et al., 2016] Y. Yan, Z. Xu, I. Tsang, G. Long, and Y. Yang. Robust semi-supervised learning through label aggregation. In AAAI, pages 2244 2250, 2016. [Yang et al., 2016] X. Yang, M. Wang, L. Zhang, and D. Tao. Empirical risk minimization for metric learning using privileged information. In IJCAI, pages 2266 2272, 2016. [You et al., 2017] S. You, C. Xu, Y. Wang, C. Xu, and D. Tao. Privileged multi-label learning. ar Xiv:1701.07194, 2017. [Zhou et al., 2004] D. Zhou, O. Bousquet, T. Lal, J. Weston, and B. Sch olkopf. Learning with local and global consistency. In NIPS, pages 321 328, 2004. [Zhou et al., 2016] J. Zhou, X. Xu, S. Pan, I. Tsang, Z. Qin, and R. Goh. Transfer hashing with privileged information. ar Xiv:1605.04034, 2016. [Zhu and Goldberg, 2009] X. Zhu and B. Goldberg. Introduction to Semi-Supervised Learning. Morgan & Claypool Publishers, 2009. [Zhu et al., 2003] X. Zhu, X. Ghahramani, and J. Lafferty. Semisupervised learning using Gaussian fields and harmonic functions. In ICML, pages 912 919, 2003. Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence (IJCAI-18)