# learning_from_complementary_labels__07d55193.pdf Learning from Complementary Labels Takashi Ishida1,2,3 Gang Niu2,3 Weihua Hu2,3 Masashi Sugiyama3,2 1 Sumitomo Mitsui Asset Management, Tokyo, Japan 2 The University of Tokyo, Tokyo, Japan 3 RIKEN, Tokyo, Japan {ishida@ms., gang@ms., hu@ms., sugi@}k.u-tokyo.ac.jp Collecting labeled data is costly and thus a critical bottleneck in real-world classification tasks. To mitigate this problem, we propose a novel setting, namely learning from complementary labels for multi-class classification. A complementary label specifies a class that a pattern does not belong to. Collecting complementary labels would be less laborious than collecting ordinary labels, since users do not have to carefully choose the correct class from a long list of candidate classes. However, complementary labels are less informative than ordinary labels and thus a suitable approach is needed to better learn from them. In this paper, we show that an unbiased estimator to the classification risk can be obtained only from complementarily labeled data, if a loss function satisfies a particular symmetric condition. We derive estimation error bounds for the proposed method and prove that the optimal parametric convergence rate is achieved. We further show that learning from complementary labels can be easily combined with learning from ordinary labels (i.e., ordinary supervised learning), providing a highly practical implementation of the proposed method. Finally, we experimentally demonstrate the usefulness of the proposed methods. 1 Introduction In ordinary supervised classification problems, each training pattern is equipped with a label which specifies the class the pattern belongs to. Although supervised classifier training is effective, labeling training patterns is often expensive and takes a lot of time. For this reason, learning from less expensive data has been extensively studied in the last decades, including but not limited to, semisupervised learning [4, 38, 37, 13, 1, 21, 27, 20, 35, 16, 18], learning from pairwise/triple-wise constraints [34, 12, 6, 33, 25], and positive-unlabeled learning [7, 11, 32, 2, 8, 9, 26, 17]. In this paper, we consider another weakly supervised classification scenario with less expensive data: instead of any ordinary class label, only a complementary label which specifies a class that the pattern does not belong to is available. If the number of classes is large, choosing the correct class label from many candidate classes is laborious, while choosing one of the incorrect class labels would be much easier and thus less costly. In the binary classification setup, learning with complementary labels is equivalent to learning with ordinary labels, because complementary label 1 (i.e., not class 1) immediately means ordinary label 2. On the other hand, in K-class classification for K > 2, complementary labels are less informative than ordinary labels because complementary label 1 only means either of the ordinary labels 2, 3, . . . , K. The complementary classification problem may be solved by the method of learning from partial labels [5], where multiple candidate class labels are provided to each training pattern complementary label y can be regarded as an extreme case of partial labels given to all K 1 classes other than class y. Another possibility to solve the complementary classification problem is to consider a multi-label 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA. setup [3], where each pattern can belong to multiple classes complementary label y is translated into a negative label for class y and positive labels for the other K 1 classes. Our contribution in this paper is to give a direct risk minimization framework for the complementary classification problem. More specifically, we consider a complementary loss that incurs a large loss if a predicted complementary label is not correct. We then show that the classification risk can be empirically estimated in an unbiased fashion if the complementary loss satisfies a certain symmetric condition the sigmoid loss and the ramp loss (see Figure 1) are shown to satisfy this symmetric condition. Theoretically, we establish estimation error bounds for the proposed method, showing that learning from complementary labels is also consistent; the order of these bounds achieves the optimal parametric rate Op(1/pn), where Op denotes the order in probability and n is the number of complementarily labeled data. We further show that our proposed complementary classification can be easily combined with ordinary classification, providing a highly data-efficient classification method. This combination method is particularly useful, e.g., when labels are collected through crowdsourcing [14]: Usually, crowdworkers are asked to give a label to a pattern by selecting the correct class from the list of all candidate classes. This process is highly time-consuming when the number of classes is large. We may instead choose one of the classes randomly and ask crowdworkers whether a pattern belongs to the chosen class or not. Such a yes/no question can be much easier and quicker to be answered than selecting the correct class out of a long list of candidates. Then the pattern is treated as ordinarily labeled if the answer is yes; otherwise, the pattern is regarded as complementarily labeled. Finally, we demonstrate the practical usefulness of the proposed methods through experiments. 2 Review of ordinary multi-class classification Suppose that d-dimensional pattern x 2 Rd and its class label y 2 {1, . . . , K} are sampled independently from an unknown probability distribution with density p(x, y). The goal of ordinary multi-class classification is to learn a classifier f(x) : Rd ! {1, . . . , K} that minimizes the classification risk with multi-class loss L R(f) = Ep(x,y) where E denotes the expectation. Typically, a classifier f(x) is assumed to take the following form: f(x) = arg max y2{1,...,K} where gy(x) : Rd ! R is a binary classifier for class y versus the rest. Then, together with a binary loss (z) : R ! R that incurs a large loss for small z, the one-versus-all (OVA) loss1 or the pairwise-comparison (PC) loss defined as follows are used as the multi-class loss [36]: LOVA(f(x), y) = gy(x) gy0(x) Finally, the expectation over unknown p(x, y) in Eq.(1) is empirically approximated using training samples to give a practical classification formulation. 3 Classification from complementary labels In this section, we formulate the problem of complementary classification and propose a risk minimization framework. We consider the situation where, instead of ordinary class label y, we are given only complementary label y which specifies a class that pattern x does not belong to. Our goal is to still learn a classifier 1We normalize the rest loss by K 1 to be consistent with the discussion in the following sections. that minimizes the classification risk (1), but only from complementarily labeled training samples {(xi, yi)}n i=1. We assume that {(xi, yi)}n i=1 are drawn independently from an unknown probability distribution with density:2 p(x, y) = 1 K 1 p(x, y). (5) Let us consider a complementary loss L(f(x), y) for a complementarily labeled sample (x, y). Then we have the following theorem, which allows unbiased estimation of the classification risk from complementarily labeled samples: Theorem 1. The classification risk (1) can be expressed as R(f) = (K 1)Ep(x,y) M1 + M2, (6) if there exist constants M1, M2 0 such that for all x and y, the complementary loss satisfies Proof. According to (5), (K 1)Ep(x,y)[L(f(x), y)] = (K 1) L(f(x), y)p(x, y)dx L(f(x), y)p(x, y)dx 5 = Ep(x,y)[M1 L(f(x), y)] = M1 Ep(x,y)[L(f(x), y)], where the fifth equality follows from the first constraint in (7). Subsequently, (K 1)Ep(x,y)[L(f(x), y)] Ep(x,y)[L(f(x), y)] = M1 Ep(x,y)[L(f(x), y) + L(f(x), y)] = M1 Ep(x,y)[M2] = M1 M2, where the second equality follows from the second constraint in (7). The first constraint in (7) can be regarded as a multi-class loss version of a symmetric constraint that we later use in Theorem 2. The second constraint in (7) means that the smaller L is, the larger L should be, i.e., if pattern x belongs to class y is correct, pattern x does not belong to class y should be incorrect. With the expression (6), the classification risk (1) can be naively approximated in an unbiased fashion by the sample average as b R(f) = K 1 M1 + M2. (8) Let us define the complementary losses corresponding to the OVA loss LOVA(f(x), y) and the PC loss LPC LOVA(f(x), y) = 1 K 1 gy(x) gy(x) Then we have the following theorem (its proof is given in Appendix A): 2The coefficient 1/(K 1) is for the normalization purpose: it would be natural to assume p(x, y) = (1/Z) P y6=y p(x, y) since all p(x, y) for y 6= y equally contribute to p(x, y); in order to ensure that p(x, y) is a valid joint density such that Ep(x,y)[1] = 1, we must take Z = K 1. Figure 1: Examples of binary losses that satisfy the symmetric condition (11). Theorem 2. If binary loss (z) satisfies (z) + ( z) = 1, (11) then LOVA satisfies conditions (7) with M1 = K and M2 = 2, and LPC satisfies conditions (7) with M1 = K(K 1)/2 and M2 = K 1. For example, the following binary losses satisfy the symmetric condition (11) (see Figure 1): Zero-one loss: 0-1(z 0 if z > 0, 1 if z 0, (12) Sigmoid loss: S(z = 1 1 + ez , (13) Ramp loss: R Note that these losses are non-convex [8]. In practice, the sigmoid loss or ramp loss may be used for training a classifier, while the zero-one loss may be used for tuning hyper-parameters (see Section 6 for the details). 4 Estimation Error Bounds In this section, we establish the estimation error bounds for the proposed method. Let G = {g(x)} be a function class for empirical risk minimization, σ1, . . . , σn be n Rademacher variables, then the Rademacher complexity of G for X of size n drawn from p(x) is defined as follows [23]: Rn(G) = EX Eσ1,...,σn define the Rademacher complexity of G for X of size n drawn from p(x) as Rn(G) = EX Eσ1,...,σn Note that p(x) = p(x) and thus Rn(G) = Rn(G), which enables us to express the obtained theoretical results using the standard Rademacher complexity Rn(G). To begin with, let e (z) = (z) (0) be the shifted loss such that e (0) = 0 (in order to apply the Talagrand s contraction lemma [19] later), and e LOVA and e LPC be losses defined following (9) and (10) but with e instead of ; let L be any (not necessarily the best) Lipschitz constant of . Define the corresponding function classes as follows: HOVA = {(x, y) 7! e LOVA(f(x), y) | g1, . . . , g K 2 G}, HPC = {(x, y) 7! e LPC(f(x), y) | g1, . . . , g K 2 G}. Then we can obtain the following lemmas (their proofs are given in Appendices B and C): Lemma 3. Let Rn(HOVA) be the Rademacher complexity of HOVA for S of size n drawn from p(x, y) defined as Rn(HOVA) = ESEσ1,...,σn 4 sup h2HOVA σih(xi, yi) Rn(HOVA) KL Rn(G). Lemma 4. Let Rn(HPC) be the Rademacher complexity of HPC defined similarly to Rn(HOVA). Then, Rn(HPC) 2K(K 1)L Rn(G). Based on Lemmas 3 and 4, we can derive the uniform deviation bounds of b R(f) as follows (its proof is given in Appendix D): Lemma 5. For any δ > 0, with probability at least 1 δ, sup g1,...,g K2G 666 b R(f) R(f) 666 2K(K 1)L Rn(G) + (K 1) where b R(f) is w.r.t. LOVA, and sup g1,...,g K2G 666 b R(f) R(f) 666 4K(K 1)2L Rn(G) + (K 1)2 where b R(f) is w.r.t. LPC. 1, . . . , g K) be the true risk minimizer and (bg1, . . . , bg K) be the empirical risk minimizer, i.e., 1, . . . , g K) = arg min g1,...,g K2G R(f) and (bg1, . . . , bg K) = arg min g1,...,g K2G f (x) = arg max y2{1,...,K} y(x) and bf(x) = arg max y2{1,...,K} Finally, based on Lemma 5, we can establish the estimation error bounds as follows: Theorem 6. For any δ > 0, with probability at least 1 δ, R( bf) R(f ) 4K(K 1)L Rn(G) + (K 1) if (bg1, . . . , bg K) is trained by minimizing b R(f) is w.r.t. LOVA, and R( bf) R(f ) 8K(K 1)2L Rn(G) + (K 1)2 if (bg1, . . . , bg K) is trained by minimizing b R(f) is w.r.t. LPC. Proof. Based on Lemma 5, the estimation error bounds can be proven through R( bf) R(g ) = b R( bf) b R(f ) R( bf) b R( bf) b R(f ) R(f ) 0 + 2 sup g1,...,g K2G 666 b R(f) R(f) where we used that b R( bf) b R(f ) by the definition of bf. Theorem 6 also guarantees that learning from complementary labels is consistent: as n ! 1, R( bf) ! R(f ). Consider a linear-in-parameter model defined by G = {g(x) = hw, φ(x)i H | kwk H Cw, kφ(x)k H Cφ}, where H is a Hilbert space with an inner product h , i H, w 2 H is a normal, φ : Rd ! H is a feature map, and Cw > 0 and Cφ > 0 are constants [29]. It is known that Rn(G) Cw Cφ/pn [23] and thus R( bf) ! R(f ) in Op(1/pn) if this G is used, where Op denotes the order in probability. This order is already the optimal parametric rate and cannot be improved without additional strong assumptions on p(x, y), and G jointly. 5 Incorporation of ordinary labels In many practical situations, we may also have ordinarily labeled data in addition to complementarily labeled data. In such cases, we want to leverage both kinds of labeled data to obtain more accurate classifiers. To this end, motivated by [28], let us consider a convex combination of the classification risks derived from ordinarily labeled data and complementarily labeled data: R(f) = Ep(x,y)[L(f(x), y)] + (1 ) (K 1)Ep(x,y)[L(f(x), y)] M1 + M2 where 2 [0, 1] is a hyper-parameter that interpolates between the two risks. The combined risk (15) can be naively approximated by the sample averages as L(f(xj), yj) + (1 )(K 1) L(f(xi), yi), (16) where {(xj, yj)}m j=1 are ordinarily labeled data and {(xi, yi)}n i=1 are complementarily labeled data. As explained in the introduction, we can naturally obtain both ordinarily and complementarily labeled data through crowdsourcing [14]. Our risk estimator (16) can utilize both kinds of labeled data to obtain better classifiers3. We will experimentally demonstrate the usefulness of this combination method in Section 6. 6 Experiments In this section, we experimentally evaluate the performance of the proposed methods. 6.1 Comparison of different losses Here we first compare the performance among four variations of the proposed method with different loss functions: OVA (9) and PC (10), each with the sigmoid loss (13) and ramp loss (14). We used the MNIST hand-written digit dataset, downloaded from the website of the late Sam Roweis4 (with all patterns standardized to have zero mean and unit variance), with different number of classes: 3 classes (digits 1 to 3 ) to 10 classes (digits 1 to 9 and 0 ). From each class, we randomly sampled 500 data for training and 500 data for testing, and generated complementary labels by randomly selecting one of the complementary classes. From the training dataset, we left out 25% of the data for validating hyperparameter based on (8) with the zero-one loss plugged in (9) or (10). 3 Note that when pattern x has already been equipped with ordinary label y, giving complementary label y does not bring us any additional information (unless the ordinary label is noisy). 4See http://cs.nyu.edu/~roweis/data.html. Table 1: Means and standard deviations of classification accuracy over five trials in percentage, when the number of classes ( cls ) is changed for the MNIST dataset. PC is (10), OVA is (9), Sigmoid is (13), and Ramp is (14). Best and equivalent methods (with 5% t-test) are highlighted in boldface. Method 3 cls 4 cls 5 cls 6 cls 7 cls 8 cls 9 cls 10 cls OVA Sigmoid For all the methods, we used a linear-in-input model gk(x) = w> k x + bk as the binary classifier, where > denotes the transpose, wk 2 Rd is the weight parameter, and bk 2 R is the bias parameter for class k 2 {1, . . . , K}. We added an 2-regularization term, with the regularization parameter chosen from {10 4, 10 3, . . . , 104}. Adam [15] was used for optimization with 5,000 iterations, with mini-batch size 100. We reported the test accuracy of the model with the best validation score out of all iterations. All experiments were carried out with Chainer [30]. We reported means and standard deviations of the classification accuracy over five trials in Table 1. From the results, we can see that the performance of all four methods deteriorates as the number of classes increases. This is intuitive because supervised information that complementary labels contain becomes weaker with more classes. The table also shows that there is no significant difference in classification accuracy among the four losses. Since the PC formulation is regarded as a more direct approach for classification [31] (it takes the sign of the difference of the classifiers, instead of the sign of each classifier as in OVA) and the sigmoid loss is smooth, we use PC with the sigmoid loss as a representative of our proposed method in the following experiments. 6.2 Benchmark experiments Next, we compare our proposed method, PC with the sigmoid loss (PC/S), with two baseline methods. The first baseline is one of the state-of-the-art partial label (PL) methods [5] with the squared hinge loss5: = (max(0, 1 z))2. The second baseline is a multi-label (ML) method [3], where every complementary label y is translated into a negative label for class y and positive labels for the other K 1 classes. This yields the following loss: LML(f(x), y) = where we used the same sigmoid loss as the proposed method for . We used a one-hidden-layer neural network (d-3-1) with rectified linear units (Re LU) [24] as activation functions, and weight decay candidates were chosen from {10 7, 10 4, 10 1}. Standardization, validation and optimization details follow the previous experiments. We evaluated the classification performance on the following benchmark datasets: WAVEFORM1, WAVEFORM2, SATIMAGE, PENDIGITS, DRIVE, LETTER, and USPS. USPS can be downloaded from the website of the late Sam Roweis6, and all other datasets can be downloaded from the UCI machine learning repository7. We tested several different settings of class labels, with equal number of data in each class. 5We decided to use the squared hinge loss (which is convex) here since it was reported to work well in the original paper [5]. 6See http://cs.nyu.edu/~roweis/data.html. 7See http://archive.ics.uci.edu/ml/. Table 2: Means and standard deviations of classification accuracy over 20 trials in percentage. PC/S is the proposed method for the pairwise comparison formulation with the sigmoid loss, PL is the partial label method with the squared hinge loss, and ML is the multi-label method with the sigmoid loss. Best and equivalent methods (with 5% t-test) are highlighted in boldface. Class denotes the class labels used for the experiment and Dim denotes the dimensionality d of patterns to be classified. # train denotes the total number of training and validation samples in each class. # test denotes the number of test samples in each class. Dataset Class Dim # train # test PC/S PL ML WAVEFORM1 1 3 21 1226 398 85.8(0.5) 85.7(0.9) 79.3(4.8) WAVEFORM2 1 3 40 1227 408 84.7(1.3) 84.6(0.8) 74.9(5.2) SATIMAGE 1 7 36 415 211 68.7(5.4) 60.7(3.7) 33.6(6.2) 719 336 87.0(2.9) 76.2(3.3) 44.7(9.6) 6 10 719 335 78.4(4.6) 71.1(3.3) 38.4(9.6) even # 719 336 90.8(2.4) 76.8(1.6) 43.8(5.1) odd # 719 335 76.0(5.4) 67.4(2.6) 40.2(8.0) 1 10 719 335 38.0(4.3) 33.2(3.8) 16.1(4.6) 3955 1326 89.1(4.0) 77.7(1.5) 31.1(3.5) 6 10 3923 1313 88.8(1.8) 78.5(2.6) 30.4(7.2) even # 3925 1283 81.8(3.4) 63.9(1.8) 29.7(6.3) odd # 3939 1278 85.4(4.2) 74.9(3.2) 27.6(5.8) 1 10 3925 1269 40.8(4.3) 32.0(4.1) 12.7(3.1) 565 171 79.7(5.3) 75.1(4.4) 28.3(10.4) 6 10 550 178 76.2(6.2) 66.8(2.5) 34.0(6.9) 11 15 556 177 78.3(4.1) 67.4(3.3) 28.6(5.0) 16 20 550 184 77.2(3.2) 68.4(2.1) 32.7(6.4) 21 25 585 167 80.4(4.2) 75.1(1.9) 32.0(5.7) 1 25 550 167 5.1(2.1) 5.0(1.0) 5.2(1.1) 652 166 79.1(3.1) 70.3(3.2) 44.4(8.9) 6 10 542 147 69.5(6.5) 66.1(2.4) 37.3(8.8) even # 556 147 67.4(5.4) 66.2(2.3) 35.7(6.6) odd # 542 147 77.5(4.5) 69.3(3.1) 36.6(7.5) 1 10 542 127 30.7(4.4) 26.0(3.5) 13.3(5.4) In Table 2, we summarized the specification of the datasets and reported the means and standard deviations of the classification accuracy over 10 trials. From the results, we can see that the proposed method is either comparable to or better than the baseline methods on many of the datasets. 6.3 Combination of ordinary and complementary labels Finally, we demonstrate the usefulness of combining ordinarily and complementarily labeled data. We used (16), with hyperparameter fixed at 1/2 for simplicity. We divided our training dataset by 1 : (K 1) ratio, where one subset was labeled ordinarily while the other was labeled complementarily8. From the training dataset, we left out 25% of the data for validating hyperparameters based on the zero-one loss version of (16). Other details such as standardization, the model and optimization, and weight-decay candidates follow the previous experiments. We compared three methods: the ordinary label (OL) method corresponding to = 1, the complementary label (CL) method corresponding to = 0, and the combination (OL & CL) method with = 1/2. The PC and sigmoid losses were commonly used for all methods. We reported the means and standard deviations of the classification accuracy over 10 trials in Table 3. From the results, we can see that OL & CL tends to outperform OL and CL, demonstrating the usefulnesses of combining ordinarily and complementarily labeled data. 8We used K 1 times more complementarily labeled data than ordinarily labeled data since a single ordinary label corresponds to (K 1) complementary labels. Table 3: Means and standard deviations of classification accuracy over 10 trials in percentage. OL is the ordinary label method, CL is the complementary label method, and OL & CL is a combination method that uses both ordinarily and complementarily labeled data. Best and equivalent methods are highlighted in boldface. Class denotes the class labels used for the experiment and Dim denotes the dimensionality d of patterns to be classified. # train denotes the number of ordinarily/complementarily labeled data for training and validation in each class. # test denotes the number of test data in each class. Dataset Class Dim # train # test OL CL OL & CL ( = 1) ( = 0) ( = 1 WAVEFORM1 1 3 21 413/826 408 85.3(0.8) 86.0(0.4) 86.9(0.5) WAVEFORM2 1 3 40 411/821 411 82.7(1.3) 82.0(1.7) 84.7(0.6) SATIMAGE 1 7 36 69/346 211 74.9(4.9) 70.1(5.6) 81.2(1.1) 144/575 336 91.3(2.1) 84.7(3.2) 93.1(2.0) 6 10 144/575 335 86.3(3.5) 78.3(6.2) 87.8(2.8) even # 144/575 336 94.3(1.7) 91.0(4.3) 95.8(0.6) odd # 144/575 335 85.6(2.0) 75.9(3.1) 86.9(1.1) 1 10 72/647 335 61.7(4.3) 41.1(5.7) 66.9(2.0) 780/3121 1305 92.1(2.6) 89.0(2.1) 94.2(1.0) 6 10 795/3180 1290 87.0(3.0) 86.5(3.1) 89.5(2.1) even # 657/3284 1314 91.4(2.9) 81.8(4.6) 91.8(3.3) odd # 790/3161 1255 91.1(1.5) 86.7(2.9) 93.4(0.5) 1 10 397/3570 1292 75.2(2.8) 40.5(7.2) 77.6(2.2) 113/452 171 85.2(1.3) 77.2(6.1) 89.5(1.6) 6 10 110/440 178 81.0(1.7) 77.6(3.7) 84.6(1.0) 11 15 111/445 177 81.1(2.7) 76.0(3.2) 87.3(1.6) 16 20 110/440 184 81.3(1.8) 77.9(3.1) 84.7(2.0) 21 25 117/468 167 86.8(2.7) 81.2(3.4) 91.1(1.0) 1 25 22/528 167 11.9(1.7) 6.5(1.7) 31.0(1.7) 130/522 166 83.8(1.7) 76.5(5.3) 89.5(1.3) 6 10 108/434 147 79.2(2.1) 67.6(4.3) 85.5(2.4) even # 108/434 166 79.6(2.7) 67.4(4.4) 84.8(1.4) odd # 111/445 147 82.7(1.9) 72.9(6.2) 87.3(2.2) 1 10 54/488 147 43.7(2.6) 28.5(3.6) 59.3(2.2) 7 Conclusions We proposed a novel problem setting called learning from complementary labels, and showed that an unbiased estimator to the classification risk can be obtained only from complementarily labeled data, if the loss function satisfies a certain symmetric condition. Our risk estimator can easily be minimized by any stochastic optimization algorithms such as Adam [15], allowing large-scale training. We theoretically established estimation error bounds for the proposed method, and proved that the proposed method achieves the optimal parametric rate. We further showed that our proposed complementary classification can be easily combined with ordinary classification. Finally, we experimentally demonstrated the usefulness of the proposed methods. The formulation of learning from complementary labels may also be useful in the context of privacyaware machine learning [10]: a subject needs to answer private questions such as psychological counseling which can make him/her hesitate to answer directly. In such a situation, providing a complementary label, i.e., one of the incorrect answers to the question, would be mentally less demanding. We will investigate this issue in the future. It is noteworthy that the symmetric condition (11), which the loss should satisfy in our complementary classification framework, also appears in other weakly supervised learning formulations, e.g., in positive-unlabeled learning [8]. It would be interesting to more closely investigate the role of this symmetric condition to gain further insight into these different weakly supervised learning problems. Acknowledgements GN and MS were supported by JST CREST JPMJCR1403. We thank Ikko Yamane for the helpful discussions. [1] M. Belkin, P. Niyogi, and V. Sindhwani. Manifold regularization: a geometric framework for learning from labeled and unlabeled examples. Journal of Machine Learning Research, 7:2399 2434, 2006. [2] G. Blanchard, G. Lee, and C. Scott. Semi-supervised novelty detection. Journal of Machine Learning Research, 11:2973 3009, 2010. [3] M. R. Boutell, J. Luo, X. Shen, and C. M. Brown. Learning multi-label scene classification. Pattern Recognition, 37(9):1757 1771, 2004. [4] O. Chapelle, B. Schölkopf, and A. Zien, editors. Semi-Supervised Learning. MIT Press, 2006. [5] T. Cour, B. Sapp, and B. Taskar. Learning from partial labels. Journal of Machine Learning Research, 12:1501 1536, 2011. [6] J. Davis, B. Kulis, P. Jain, S. Sra, and I. Dhillon. Information-theoretic metric learning. In ICML, 2007. [7] F. Denis. PAC learning from positive statistical queries. In ALT, 1998. [8] M. C. du Plessis, G. Niu, and M. Sugiyama. Analysis of learning from positive and unlabeled data. In NIPS, 2014. [9] M. C. du Plessis, G. Niu, and M. Sugiyama. Convex formulation for learning from positive and unlabeled data. In ICML, 2015. [10] C. Dwork. Differential privacy: A survey of results. In TAMC, 2008. [11] C. Elkan and K. Noto. Learning classifiers from only positive and unlabeled data. In KDD, 2008. [12] J. Goldberger, S. Roweis, G. Hinton, and R. Salakhutdinov. Neighbourhood components analysis. In NIPS, 2004. [13] Y. Grandvalet and Y. Bengio. Semi-supervised learning by entropy minimization. In NIPS, 2004. [14] J. Howe. Crowdsourcing: Why the power of the crowd is driving the future of business. Crown Publishing Group, 2009. [15] D. P. Kingma and J. L. Ba. Adam: A method for stochastic optimization. In ICLR, 2015. [16] T. N. Kipf and M. Welling. Semi-supervised classification with graph convolutional networks. In ICLR, [17] R. Kiryo, G. Niu, M. C. du Plessis, and M. Sugiyama. Positive-unlabeled learning with non-negative risk estimator. In NIPS, 2017. [18] S. Laine and T. Aila. Temporal ensembling for semi-supervised learning. In ICLR, 2017. [19] M. Ledoux and M. Talagrand. Probability in Banach Spaces: Isoperimetry and Processes. Springer, [20] Y.-F. Li and Z.-H. Zhou. Towards making unlabeled data never hurt. IEEE Transactions on Pattern Analysis and Machine Intelligence, 37(1):175 188, 2015. [21] G. Mann and A. Mc Callum. Simple, robust, scalable semi-supervised learning via expectation regular- ization. In ICML, 2007. [22] C. Mc Diarmid. On the method of bounded differences. In J. Siemons, editor, Surveys in Combinatorics, pages 148 188. Cambridge University Press, 1989. [23] M. Mohri, A. Rostamizadeh, and A. Talwalkar. Foundations of Machine Learning. MIT Press, 2012. [24] V. Nair and G. Hinton. Rectified linear units improve restricted boltzmann machines. In ICML, 2010. [25] G. Niu, B. Dai, M. Yamada, and M. Sugiyama. Information-theoretic semi-supervised metric learning via entropy regularization. Neural Computation, 26(8):1717 1762, 2014. [26] G. Niu, M. C. du Plessis, T. Sakai, Y. Ma, and M. Sugiyama. Theoretical comparisons of positiveunlabeled learning against positive-negative learning. In NIPS, 2016. [27] G. Niu, W. Jitkrittum, B. Dai, H. Hachiya, and M. Sugiyama. Squared-loss mutual information regular- ization: A novel information-theoretic approach to semi-supervised learning. In ICML, 2013. [28] T. Sakai, M. C. du Plessis, G. Niu, and M. Sugiyama. Semi-supervised classification based on classifica- tion from positive and unlabeled data. In ICML, 2017. [29] B. Schölkopf and A. Smola. Learning with Kernels. MIT Press, 2001. [30] S. Tokui, K. Oono, S. Hido, and J. Clayton. Chainer: a next-generation open source framework for deep learning. In Proceedings of Workshop on Machine Learning Systems in NIPS, 2015. [31] V. N. Vapnik. Statistical learning theory. John Wiley and Sons, 1998. [32] G. Ward, T. Hastie, S. Barry, J. Elith, and J. Leathwick. Presence-only data and the EM algorithm. Biometrics, 65(2):554 563, 2009. [33] K. Weinberger, J. Blitzer, and L. Saul. Distance metric learning for large margin nearest neighbor classi- fication. Journal of Machine Learning Research, 10:207 244, 2009. [34] E. P. Xing, A. Y. Ng, M. I. Jordan, and S. Russell. Distance metric learning with application to clustering with side-information. In NIPS, 2002. [35] Z. Yang, W. W. Cohen, and R. Salakhutdinov. Revisiting semi-supervised learning with graph embed- dings. In ICML, 2016. [36] T. Zhang. Statistical analysis of some multi-category large margin classification methods. Journal of Machine Learning Research, 5:1225 1251, 2004. [37] D. Zhou, O. Bousquet, T. Navin Lal, J. Weston, and B. Schölkopf. Learning with local and global consistency. In NIPS, 2003. [38] X. Zhu, Z. Ghahramani, and J. Lafferty. Semi-supervised learning using Gaussian fields and harmonic functions. In ICML, 2003.