# vetoconsensus_multiple_kernel_learning__bbe1c50b.pdf Veto-Consensus Multiple Kernel Learning Yuxun Zhou Department of EECS UC Berkeley yxzhou@berkeley.edu Ninghang Hu Amsterdam Machine Learning Lab University of Amsterdam huninghang@gmail.com Costas J. Spanos Department of EECS UC Berkeley spanos@berkeley.edu We propose Veto-Consensus Multiple Kernel Learning (VCMKL), a novel way of combining multiple kernels such that one class of samples is described by the logical intersection (consensus) of base kernelized decision rules, whereas the other classes by the union (veto) of their complements. The proposed configuration is a natural fit for domain description and learning with hidden subgroups. We first provide generalization risk bound in terms of the Rademacher complexity of the classifier, then a large margin multi-ν learning objective with tunable training error bound is formulated. Seeing that the corresponding optimization is non-convex and existing methods severely suffer from local minima, we establish a new algorithm, namely Parametric Dual Descent Procedure(PDDP), that can approach global optimum with guarantees. The bases of PDDP are two theorems that reveal the global convexity and local explicitness of the parameterized dual optimum, for which a series of new techniques for parametric program have been developed. The proposed method is evaluated on extensive set of experiments, and the results show significant improvement over the state-of-the-art approaches. 1 Introduction In recent years, multiple kernel learning (MKL) has shown promising results in a variety of applications and has attracted much attention in machine learning community. Given a set of base kernels, MKL finds an optimal combination of them with which an appropriate hypothesis is determined on the training data. A large body of literature has been addressing the arising issues of MKL, mainly from three perspectives and their intersections, i.e. theoretical learning bound, related optimization algorithm, and alternative MKL settings. To list a few, the generalization bounds for learning linear combination of multiple kernels have been extensively studied in (Ying and Campbell 2009; Cortes, Mohri, and Rostamizadeh 2010; Hussain and Shawe-Taylor 2011) by analyzing various complexity metrics. Following the pioneer work (Lanckriet et al. 2004) that formulates linear MKL as a semi-definite program (SDP), a series of work is devoted to improve the efficiency with various optimization techniques, such as reduced gradient (Rakotomamonjy et al. 2008), Newtown s Copyright c 2016, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. Figure 1: 2D VCMK with non-linear/all linear base kernels method (Kloft et al. 2011), and mirror descent (Jagarlapudi et al. 2009). Also data related issues such as sample adaptability and missing channels (Liu et al. 2014; 2015) have been addressed. Despite the substantial theoretical advancement and algorithmic progress made in linear MKL, few of the results could be directly applied to MKL that incorporates nonlinear combinations. Indeed non-linear MKL is usually studied on a case-by-case basis, such as polynomial combination (Cortes, Mohri, and Rostamizadeh 2009), hierarchical kernel (Bach 2009), hyperkernels (Ong, Williamson, and Smola 2005), etc. In this work we propose a novel non-linear combination of multiple kernels. To motivate the configuration, Figure 1 illustrates a practical problem where part of the classes contains hidden structures. In this example, the positive class is labeled. In contrast, the negative class contains several subgroups but only a single label is provided. To compensate for this implicit information, we propose to describe the positive class by the intersection of the acceptance region of multiple base kernel decision rules, and the negative class by the union of their complements. Hence a sample is classified as negative as long as one or more rules votes negative (Veto), and a positive assignment is made for a sample if and only if all of the rules agree (Consensus). With this intuition, VCMKL is a natural solution for applications involving hidden structures. Moreover because the construction inherently emphasizes the sensitivity to negative class and the specificity to positive class, it is also a promising tool for domain description problems. We discuss the proposed VCMKL from both theoretical and algorithmic perspectives. Firstly, we formalize the Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence (AAAI-16) the construction of the classifier and provide Rademacher complexity analysis. Then a large margin multi-ν learning formulation is proposed with training error controlled by hyperparameters. Another main contribution of this paper is a learning algorithm that can approach global optimum for the non-convex learning objective. We first show that the objective could be reformulated with hidden variables, and then we show that the dual solution is locally explicit and globally convex piece-wise quadratic in hidden variables. Based on this, a subgradient algorithm is proposed with guaranteed convergence to global optimum. We call the overall framework Parametric Dual Descent Procedure (PDDP). Although we develop it for the purpose of VCMKL, the method can be readily adapted to many other situations such as the training of latent structured SVM, ramp loss robust Regression/SVM and approximate L0 regularized problems (Yu and Joachims 2009; Xu, Crammer, and Schuurmans 2006). Please refer to the supplementary material for details. To the best of our knowledge, PDDP is the first method that is able to find global optimum for this class of non-convex learning problems without resorting to combinatorial search. 2 Related Work With only linear kernels, VCMKL is reduced to constructing convex polyhedrons in the original feature space for classification, as is shown in the right subplot of Figure 1. Recently several attempts were made for this purpose: Based on the AND-OR model, (Dundar et al. 2008) proposed an alternating algorithm, assuming some subclass labels are available. In (Manwani and Sastry 2010; 2011) the author proposed two methods, one with logistic link function and the other based on perception algorithm. In (Kantchelian et al. 2014) a Stochastic Gradient Descent (SGD) method is applied with a notion of entropy to alleviate collapsed solutions. Improved SGD with better initialization method is proposed in (Zhou, Li, and Spanos 2015; Zhou, Jin, and Spanos 2015). However, all of these methods only deal with the primal, which cannot be directly adopted in a general MKL setting. Moreover, even with searching heuristics these learning methods still suffer severely from local minimums. As will be shown in section 3, with a slight reformulation, the VCMKL objective resembles a multi-kernel extension of SVMs with hidden/latent variables. Although the idea of including hidden variables in SVM variations has been suggested in literature (Yu and Joachims 2009; Felzenszwalb et al. 2010), the proposed VCMKL is novel in that multiple base kernels are combined and each hidden subgroup is dealt with by an optimal one. Regarding the learning algorithm, the aforementioned work uses either group alternating optimization or Concave-Convex Procedure (CCCP). Again these methods only converge to local minimas and the solution could be very deteriorated with large number of latent variables (Stephen Boyd 2004; Floudas 1995). 3 Problem Formulation 3.1 The Classifier and Generalization Bound Let S = {xi, yi}l i=1 be a training data set, where x Rd and d is the dimension of features. Without loss of generality let y {+1, 1} indicate class labels, with negative class contains hidden subgroups. Consider a feature mapping φ : Rd H, in the new Hilbert space a hyperplane can be written as f(x) = w, φ(x) H + b. A decision rule for data x is given by the sign of f(x ). Formalizing the idea of using intersections of M base kernel mappings for positive class and the union of their complement for negative class, the composed classifier g( ) is such that {g(x) > 0} = {f1(x) > 0} {f M(x) > 0} = min 1, ,M{f1(x), , f M(x)} > 0 On the other hand, the acceptance region for negative class is {min1, ,M{f1(x), , f M(x)} 0}. For short notation, let us denote w, φ(x) H = w φ(x) as the inner product and || || as the corresponding norm in H. Then the combined classifier is simply g(x) = min{w1 φ(x) + b1, , w M φ(x) + b M} Before proceeding to any method to learn this classifier, we conduct complexity analysis in MKL framework for generalization bound and model selection purpose. Let the function class of g be denoted as G, and that of fj be denoted as Fj. As a classic measure of richness, the Empirical Rademacher Complexity for a function class F is defined as R(F(xl 1)) Eσ supf F 2 l l i=1 σif(xi) where σ1, , σl are i.i.d. Rademacher variables such that P(σi = +1) = P(σi = 1) = 1/2. The definition measures the complexity/richness of function class F in terms of its ability to match Rademancher variables. With Talagrand s Lemma and an induction argument, we show that Theorem 1. The function class G of VCMKL has R(G xl 1) 2 j=1 R Fj(xl 1) Further assume Fj forms a bounded function class with kernel κj( , ) and kernel matrix Kj such that Fj = x l i=1 αiκj(xi, x) | αT Kjα Bj then R G(xl 1) 4 Note that in general it is hard to tighten the additive nature of the complexity1. With the above results at hand, the generalization guarantee of the MKVCL can be obtained immediately from the classic results in the PAC learning framework. Let L(g) = ES[1sgn(g(x)) =y] be the generalization error of the MKVC classifier g, and let Lρ(g) 1 l l i=1 Ψρ (yig(xi)) be the empirical ρ margin loss with 1In fact in the case of all linear kernels, it is shown in (Zhou, Li, and Spanos 2015) that the VC dimension has additive lower bound. Ψρ(t) = [min{1, 1 t/ρ}]+. Then we have with probability at least 1 δ L(g) Lρ(g) + 8 3.2 Large Margin Learning Objective To learn the multi-kernels classifier from data, we adopt a learning objective that maximize the total margin defined in (Kantchelian et al. 2014) while minimize the hinge loss of misclassifications. Inspired by the advantages of νSVM (Scholkopf B 2000), the following large margin multi-ν learning formulation is proposed: min wm,bm,ρm 1 2 m=1 ||wm||2 i I+ max m {[ρm yi(wm xi + bm)]+} i I min m {[ρm yi(wm xi + bm)]+} (OPT1) where I+ and I are index sets for positive and negative class, respectively. The hyperparameters ν1, , νM [0, 1] weight the margins, and two types of losses are treated differently by introducing γ [0, 1]. The multi-ν formulation still reflects the Veto-Consensus intuition: the loss for positive class is the maximum over all decision boundaries, while for negative class only the one with minimum loss is counted. The first three terms in the above minimization problem are convex, however the last term is non-convex as the minimum of M truncated hyperplanes. We handle this term by introducing new variables. Consider a weighted version of the M losses over a simplex: Lavg(w, xi, λi) = m=1 λim[ρm yi(wm xi +bm)]+ (1) with λi λi : M m=1 λim = 1, λim 0 SM, a row vector in the |I | M matrix λ containing the loss weighting parameters of negative samples. Denote Lmin(w, xi) = minm {[ρm yi(wm xi + bm)]+} as the original loss for xi, it is straightforward that Lmin(w, xi) = min λi SM Lavg(w, xi, λi) (2) With this trick we reformulate the learning objective as Proposition 1. (OPT1) is equivalent to min λi SM min wm,bm, ρm 0 m=1 ||wm||2 i I+ max m {[ρm yi(wm xi + bm)]+} m=1 λim[ρm yi(wm xi + bm)]+ The newly introduced variables λ can be viewed as hidden subgroup indicators, hence VCMKL can indeed be thought of as a multi-kernel extension of learning with latent variables. Considering the form of OPT1 and Primal, it is tempting to apply CCCP and alternating heuristics. Yet in this work a rigorous optimization algorithm will be developed to approach global optimum. But before that let s look into the relation between training error and the hyperparameters νm, γ in the learning formulation. Replacing the inner optimization of the Primal with its dual, we obtain that the Primal is equivalent to min λi SMJd(λ) where 1 2 M m=1 l i,j=1 αimyiκm(xi, xj)yjαjm subject to αim 0 i, m αim 1 γ l λim i I , m M m=1 αim γ l i I+ l i=1 αim νm m l i=1 αimyi = 0 m (Dual) To see the effect of hyperparameters, it is useful to define partition of samples similar to the classical SVM: Definition 1. Partition of Samples : Based on the value of αim at optimal, the ith sample is called positive support vector if i I+ and m αim > 0. positive bounded support vector if i I+ and l . negative support vector of class m if i I and αim > 0. negative bounded support vector of class m if i I and αim = 1 γ l . All the other samples are called non-support vectors. The following proposition relates the choice of hyperparameters to the training error tolerance. Proposition 2. Define ν+ = l m νm 2γ|I+| and ν m = lνm 2(1 γ)|I |, and denote N sv+, N sv m , N bsv+, N bsv m as the number of all postive/negative support vectors, positive/negative bounded support vectors, respectively, then |I+| ν+ N sv+ |I+| N bsv m |I | ν m N sv m |I | Form the right hand side, ν+ and ν m give a lower bound on the fraction of positive support vectors and negative support vectors of class m, respectively. The left hand side upper bound is more interesting: by definition, N bsv+/|I+| and N bsv m /|I | are respectively the training false negative error and false positive error of class m. Hence the bound implies that one can impose smaller training error of different types by decreasing corresponding ν. The role of γ is also significant: it can incorporate an uneven consideration of training errors committed in two classes, which can be harnessed to handle imbalanced availability of positive/negative samples. In short, the advantage of the multi-ν formulation is that the training result can be controlled simply by tuning bounds as a function of hyperparameters. 4 PDDP Learning Method We introduce the new global optimization method PDDP in this section with VCMKL as a special case. The first step involves writing the optimal solution α and the optimal value J d as a locally explicit function of λ. In the terminology of mathematical programming, this non-trivial task is referred to as parametric program, or sensitivity analysis. For the ease of notation, we write Jd(λ) in a more compact matrix form min α1, ,αMJ (α) = 1 m=1 αT m Qmαm Group I: Cααm Cλλm + Cm m Group II: [1, , 1]αT i γ OHP: y T αm = 0 m where Qm = Km yy T with Km the Gram matrix for mth base kernel. The inequalities addressing each subgroup m are contained in Group I, and constant matrices Cα, Cλ, Cm encapsulate the coefficients. Group II includes the third inequality in the dual. The equality constraint (4c) is called orthogonal hyperplane property due to its geometric interpretation. With the solution α (λ), the ith row of the mth constraint (4a) is active if Cα i α m(λ) = Cλ i λm + Cm i , and inactive if Cα i α m(λ) < Cλ i λm + Cm i . We have used column vector convention for αm, λm, and row/column selection of a matrix with subscripts. In order to determine the relation between the optimum of OPT2 and its parameters λ, we develop a series of new techniques for parametric program. Interested readers are referred to supplementary material for detailed proofs of the main theorems in this section. In particular, we find that the existence of local explicit expression is guaranteed if the following sample partition property is satisfied. Definition 2. (Non-degeneracy) A solution of the dual is called Non-degenerate if there exist at least one positive unbounded support vector and one negative unbounded support vector for each class m. Since the unbounded support vectors are essentially the sample points that lie on the margin boundaries, except for deliberately designed cases, in practice with large sample size the non-degeneracy is expected to be satisfied. With that we provide the local explicitness theorem. Theorem 2. Assume that the solution of the VCMKL inner optimization is non-degenerate, then 1. For all m, the optimal solution α m is a continuous piecewise affine (PWA) function of λ = [λ1, , λM] 2. The optimal objective J (α ) is a continuous piece-wise quadratic (PWQ) function of λ. 3. At optimal, let the index set of group I active constraints be denoted as Am, m = {1, , M}, and the index set of group II active constraints be denoted as B. Then in the critical region defined by inequalities (5) (8), the optimal solution α admits a closed form α m =F m(Cα Am)T G 1 m Cλ Amλ m + Cm Am m=1 HBm Cλ Amλ m + Cm Am γ where the involved matrices are computed by F m Q 1 m yy T Q 1 m y T Q 1 m y Q 1 m ; Gm Cα Am F m(Cα Am)T ; Hm F m(Cα Am)T G 1 m ; Rm F m(Cα Am)T G 1 m Cα Am F m F m; m=1 [Rm]B B ; Rm [Rm0] B ; and Rm0 is obtained from matrix Rm by setting rows that are not in B to 0 In essence the theorem indicates that each time the inner optimization OPT2 of the dual is solved, full information (closed form solution) in a well-defined neighborhood (critical region) can be retrieved as a function of outer optimization variable λ. Besides this local explicitness result, we characterize the overall geometric structure of the optimality, and show in the next theorem that globally the optimal objective is convex in λ, which serves as the underpinning for PDDP to be a global optimization algorithm. Theorem 3. Assuming non-degeneracy, then 1. The dual optimization has finite number of polyhedron critical regions CR1, , CRNr which constitute a partition of the feasible set of λ, i.e. each feasible λ belongs to one and only one critical region. 2. The optimal objective J (α (λ)) is a convex PWQ function of λ, and is almost everywhere differentiable. Now the learning problem is reduced to minimizing a non-smooth but convex function J (α (λ)). Projected subgradient based method is a natural choice. At each step, OPT2 is solved with current λ, and with Theorem 2 one can directly compute the critical region boundaries (5) (8), as well as the gradients with respect to λm: λm J (α ) = DT m Qmα m ET m m =m R T m Qm α m Em = U 1HBm Cλ Am; Dm = F m(Cα Am)T G 1 m Cλ Am Rm Em The function is not differentiable only on the boundary of critical regions. In this case the subgradient set is the convex hull of the gradients of adjacent regions, as the dual can be viewed as point-wisely optimizing α with each λ (Stephen Boyd 2004). Thus for both cases one can simply use the gradient induced by current optimal objective and proceed to update λ. Since each λi is a M dimensional simplex, the updated value is projected row by row onto this space. By theorem 2, if λt is in the critical regions that have been explored before, all information could be retrieved in G 1 m Cλ Amλ m + Cm Am G 1 m Cα Am F m B U 1 M m=1 HBm Cλ Amλ m + Cm Am γ Cα Acm F m(Cα Am)T G 1 m Cλ Amλ m + Cm Am Cα Acm Rm U 1 M m =1 HBm Cλ Am λ m + Cm A m l 1B Cλ Acmλ m + Cm Acm 0 (6) U 1 M m=1 HBm Cλ Amλ m + Cm Am γ M m=1 F Bm(Cα Am)T G 1 m Cλ Amλ m + Cm Am M m=1 RBm U 1 M m =1 HBm Cλ Am λ m + Cm A m Algorithm 1 PDDP for VCMKL Input S, ν, γ, K = {κ1, , κ} λ1 init Lambda(S); {Q, C} cal Matrix(S, ν, γ, K) CRexplored , n itermax, τ stepsize, t 1 while Improved & t n do if λt R then {α , AM 1 , B} sol OPT2(Q, C) CRnew get Region(Q, C, AM 1 , B) % (5)-(8) CRexplored CRexplored CRnew else α alpha Inside R(λt, AM 1 , B) % theorem 2 end if g get Grad(α , AM 1 , B) % (9) λt+1 Proj(λt τg); t t + 1 end while return α ,λ an explicit form and there is no need to solve OPT2 again. However, when the variable goes to a new critical region, a QP solver for OPT2 has to be invoked again for optimal solution and sample partition. The overall PDDP for VCMKL is summarized in Algorithm 1. The following convergence result to the global optimal is a consequence of theorem 3: Theorem 4. Convergence Guarantee Let supλ ||λ1 λ|| = B, and the Lipschitz constant of J (α (λ)) be G, then algorithm I with iteration n and optimal step size τi = B/G n i converges to global optimum within O (1/ n). To be specific, let O be the global optimum of the learning objective (Dual form), then J (α (λn best)) O BG n , where J (α (λn best)) min J (α (λ1)), , J (α (λn)) Hence in order to get J (α (λn best)) O ϵ, the algorithm needs O(1/ϵ2) iterations. B is bounded because the feasible set of λ is simplexes. Also as J (α (λ)) is globally convex (hence continuous) and locally quadratic (hence has bounded gradient), G must be bounded as well. The constant step size is optimal in the sense that it minimizes the gap upper bound. Other choices, such as a diminishing step size that is square summable but not summable, could also be used if faster convergence is a concern. The most computational expensive step is the inner QP solver. However note that since OPT2 shares a similar structure as the classic SVM dual, variety of existing methods could be reused for acceleration (Chang and Lin 2011). In the functions get Region(), alpha Insider Region(), and get Grad(), the computational overhead is mostly matrix inver- Table 1: Comparison of algorithms Method PDDP AM SGD CCCP Objective (10 8) 6.247 10.21 13.55 8.784 Testing Accuracy (%) 78.91 70.17 67.54 71.38 # of Iterations 79 17 237 31 Elapsed Time (s) 417 166 34 226 sions of Q, G, U. Fortunately, from the proof of the theorem 2 they are either symmetric positive definite or symmetric negative definite matrices and various of decomposition methods are thusly available for efficient inversion. 5 Experiment 5.1 PDDP Optimization Performance Firstly we compare the performance of the proposed PDDP with existing optimization methods, including alternating minimization (Dundar et al. 2008), stochastic gradient descent (Kantchelian et al. 2014), and concave-convex procedure (Yu and Joachims 2009). The public UCI (Lichman 2013) pima data set is used in this experiment, and the learning problem OPT1 is solved with different algorithms. Since the other methods can only be applied to the primal, for comparison purpose we restrict to all linear kernels. Hyperparameters are set with M = 5, νm = 0.02 m and γ = |I |/l 2. For PDDP initialization a simple K-mean is applied to the negative class and λ1 is assigned according to cluster labels. The final value of the objective function, the corresponding testing accuracy, number of iterations and the time consumed are shown in Table 1, for which |Jt+1 Jt|/Jt 10 4 is chosen as the stopping criteria. We observer that PDDP achieves a much better objective value, about 28.9% lower than the runner-up CCCP. The improved training also leads to 7.53% increase in testing accuracy. As a global optimization algorithm, it is expected that PDDP consumes more time than algorithms that only converge to local minimums, as is shown in the last row of the table. One possible acceleration is the design of approximate larger region that can reduce the number of invocations of the quadratic solver. Figure 2 shows the corresponding objective value, gradient norm, λ value, and testing accuracy of PDDP in each iteration. Note that for clear presentation only a subset of λ with same initial value is shown. We see that the iterative 2Here the choice of these hyperparameters is not a concern, as the primary goal is to compare optimization algorithms. Table 2: Testing Cost = c1 FP P(y = 1) + c2 FN P(y = +1) and their rankings Data Set c2:c1 l-VCML rbf-VCML csrbf-SVM 2LMKL Lasso-LR NN Ada Boost 1:1 0.197 (5) 0.144 (1) 0.176 (3) 0.169 (2) 0.210 (7) 0.189 (4) 0.206 (6) 1:4 0.289 (2) 0.245 (1) 0.315 (4) 0.314 (3) 0.337 (5) 0.366 (7) 0.352 (6) 1:1 0.277 (4) 0.229 (1) 0.250 (2) 0.264 (3) 0.317 (7) 0.297 (5) 0.315 (6) 1:4 0.443 (4) 0.330 (1) 0.410 (3) 0.398 (2) 0.514 (5) 0.519 (6) 0.556 (7) 1:1 0.131 (2) 0.121 (1) 0.140 (4) 0.133 (3) 0.168 (7) 0.152 (5) 0.157 (6) 1:4 0.214 (4) 0.156 (1) 0.218 (5) 0.211 (3) 0.239 (6) 0.192 (2) 0.332 (7) 1:1 0.132 (5) 0.025 (1) 0.113 (3) 0.122 (4) 0.159 (7) 0.111 (2) 0.142 (6) 1:4 0.184 (3) 0.051 (1) 0.181 (2) 0.199 (4) 0.276 (6) 0.228 (5) 0.293 (7) 0 20 40 60 80 0.5 Objective Value 0 20 40 60 80 0 Gradient Norm 0 20 40 60 80 Iteration Variables λ 0 20 40 60 80 Iteration Testing Accuracy Figure 2: PDDP Convergence Results result is similar to the general subgradient descent for nonsmooth convex functions (Alber, Iusem, and Solodov 1998): The learning objective is in general decreasing, with some fluctuations due to non-smoothness between two critical regions. The pattern of gradient norm and λ evolution also reflects this character of insider region exploitation and beyond region exploration . Note that although the algorithm takes 79 steps to converge, only 38 calls of the quadratic solver is involved, as in the critical region that has been explored before, explicit computation can be done with theorem 2. The testing accuracy in each iteration is shown at the bottom right of Figure 2, which increases correspondingly with the decrease of learning objective. 5.2 VCMKL Classification Performance We test VCMKL with the PDDP solver on various data set. To begin with, two data sets (i.e. the UCI robot execution and vowel) are used to demonstrate the effect of number of kernels M on classification performance. Note that labels of the raw data are transformed into a binary case, e.g. for the robot data we set y = +1 if the label is NM, and all the other samples with label LP1-LP4 are pooled together as y = 1. Similar for the vowel data, y = +1 if the label is h Od or hod , and y = 1 for the rest of 8 classes. Figure 3 shows the testing accuracy versus the number of kernels with 3 commonly used kernel families (linear, polynomial with dif- 0 2 4 6 8 Number of Kernels Testing Accuracy Linear Kernels Polynomial Kernels RBF Kernels 2 4 6 8 10 12 Number of Kernels Testing Accuracy Linear Kernels Polynomial Kernels RBF Kernels Figure 3: Testing Accuracy vs. M number of kernels. Left:Robot, Right:Vowel ferent orders and RBF with various σ). Interestingly, in all cases it is seen that the testing accuracy is improved in the first few steps as M increases, however, further combining more kernels does not help, or even lead to a degraded performance due to overfitting (for linear and nonlinear kernels, respectively). In particular, the saturated point for linear kernels is just the number of hidden subgroups, i.e. M = 4 for robot data and M = 8 for vowel data, while for nonlinear kernels the optimal M is smaller as in the transformed space the subgroups may merge. This observation not only justifies the veto-consensus intuition to describe hidden subgroups, but also is consistent with the implications of theorem 1, which provides additive upper bound and does not encourage the use of many kernels. Finally, VCMKL with linear and RBF kernels is compared to other methods, including cost sensitive SVM with RBF kernel, Two-layer MKL (Zhuang, Tsang, and Hoi 2011), Lasso Logistic Regression, multi-layer Neural Network, and Ada Boost. We compare a cost sensitive loss with c1 and c2 the cost rate for false positive (FP) and false negative (FN), respectively. Accordingly we set γ = c1|I |/(c1|I |+c2|I+|). As is suggested by preceding discussion, an incremental CV is applied for selecting M. The other hyperparameters for all methods are chosen with 10 folds CV. Due to page limits, only the results for 4 data sets with hidden subgroups are shown in Table 2. Looking at the testing error (c1 = c2 = 1), it is seen that rbf-VCMKL outperforms all the other methods. Especially on the vowel data set the improvement is quite significant (testing accuracy from 88.9% to 97.5%). The performance of linear VCMKL is somewhere among those existing methods (average rank 4). When the cost of false positive is higher (c1 = 4, c2 = 1), the performance improvement in using VCMKL is even more significant, as can be seen in the second row for each experiment. This justifies our intuition that the veto-consensus construction is sensitive to false positive while is robust to false negative, making it a promising tool for cost sensitive domain description problems. 6 Conclusion and Discussion We study the veto-consensus combination of multiple kernels for classification. Our contributions are three folds (1) An upper bound on the Rademacher complexity of the veto-consensus multi-kernel classifier is provided. Previous bound obtained for polyhedron classifiers (Zhou, Jin, and Spanos 2015) can be viewed as a special case with only linear kernels. (2) A large margin multi-ν learning objective is formulated, which has provable training error bound in terms of hyperparameters. (3) A new global optimization algorithm, namly PDDP, is established. Sufficient conditions and new techniques for parametric program have been derived along the way. Although PDDP is developed in the context of VCMKL, it is also adaptable for a class of learning problems with similar structures. Experimental results demonstrate the effectiveness of the proposed approach. Future work consists of (1) Apply PDDP to other learning problems such as hidden structured SVM, robust SVM, and piecewise linear function learning. (2) Obtain data dependent convergence bound for PDDP, in particular detailed bound for the Lipschtiz constant. (3) Design accelerated algorithm that uses enlarged approximate critical regions to reduce invocations of quadratic solver. The key issue is the trade off between region approximation and convexity of the dual optimality. 7 Appendix The proof of Theorem I is given here. The proofs for PDDP could be found in supplementary material. Proof. For the first part, we need the following lemma Lemma 1. Talagrand s Lemma Let Φ : R R be η Lipschitz, and Υ : R R be convex and nondecreasing. Then for any function class F of real-valued functions, the following inequality holds: ˆR(Υ Φ F(xl 1)) η ˆR(Υ F(xl 1)) Now for the main theorem, consider the case M = 2. The following inequality is straightforward: i=1 σig(xi) i=1 σig(xi) i=1 σig(xi) + Noticing that σ1, , σl has the same distribution as σ1, , σl, we get ˆR G(xl 1) = Eσ i=1 σig(xi) i=1 σig(xi) Writing g = min{f1, f2} = 1 2(f1 + f2) 1 2|f1 f2|, the last term yields i=1 σig(xi) i=1 σi 1 2(f1(xi) + f2(xi)) i=1 σi 1 2|f1(xi) f2(xi)| i=1 σif1(xi) i=1 σif2(xi) i=1 σi 1 2|f1(xi) f2(xi)| where (a) and (b) are due to the upper additive property of sup and [ ] = max{0, } function. Taking expectations for this upper bound and applying Talagrand s Lemma with Υ = [ ] and Φ = | | yields i=1 σi 1 2|f1(xi) f2(xi)| i=1 σif1(xi) i=1 σif2(xi) + Putting two inequalities together we have i=1 σig(xi) i=1 σif1(xi) i=1 σif2(xi) i=1 σif1(xi) i=1 σif2(xi) i=1 σif1(xi) i=1 σif2(xi) = ˆR F1(xl 1) + ˆR F2(xl 1) hence finally, ˆR G(xl 1) 2 ˆR F1(xl 1) + ˆR F2(xl 1) Also it is straightforward to generalize the above argument to M > 2 with simple induction. Finally we get ˆR(G xl 1) 2 j=1 ˆR Fj(xl 1) The second part of the theorem can be obtained with a standard approach in bounding empirical Rademacher complexity of Kernels. 8 Acknowledgments This research is funded by the Republic of Singapore s National Research Foundation through a grant to the Berkeley Education Alliance for Research in Singapore (BEARS) for the Singapore-Berkeley Building Efficiency and Sustainability in the Tropics (Sin Ber BEST) Program. BEARS has been established by the University of California, Berkeley as a center for intellectual excellence in research and education in Singapore. References Alber, Y. I.; Iusem, A. N.; and Solodov, M. V. 1998. On the projected subgradient method for nonsmooth convex optimization in a hilbert space. Mathematical Programming 81(1):23 35. Bach, F. 2009. High-dimensional non-linear variable selection through hierarchical kernel learning. ar Xiv preprint ar Xiv:0909.0844. Chang, C.-C., and Lin, C.-J. 2011. LIBSVM: A library for support vector machines. ACM Transactions on Intelligent Systems and Technology 2:27:1 27:27. Cortes, C.; Mohri, M.; and Rostamizadeh, A. 2009. Learning non-linear combinations of kernels. In Advances in neural information processing systems, 396 404. Cortes, C.; Mohri, M.; and Rostamizadeh, A. 2010. Generalization bounds for learning kernels. In Proceedings of the 27th International Conference on Machine Learning (ICML-10), 247 254. Dundar, M. M.; Wolf, M.; Lakare, S.; Salganicoff, M.; and Raykar, V. C. 2008. Polyhedral classifier for target detection: a case study: colorectal cancer. In Proceedings of the 25th international conference on Machine learning, 288 295. ACM. Felzenszwalb, P. F.; Girshick, R. B.; Mc Allester, D.; and Ramanan, D. 2010. Object detection with discriminatively trained part-based models. Pattern Analysis and Machine Intelligence, IEEE Transactions on 32(9):1627 1645. Floudas, C. A. 1995. Nonlinear and mixed-integer optimization: Fundamentals and applications. Oxford University Press. Hussain, Z., and Shawe-Taylor, J. 2011. Improved loss bounds for multiple kernel learning. In International Conference on Artificial Intelligence and Statistics, 370 377. Jagarlapudi, S. N.; Dinesh, G.; Raman, S.; Bhattacharyya, C.; Ben-Tal, A.; and Kr, R. 2009. On the algorithmics and applications of a mixed-norm based kernel learning formulation. In Advances in neural information processing systems, 844 852. Kantchelian, A.; Tschantz, M. C.; Huang, L.; Bartlett, P. L.; Joseph, A. D.; and Tygar, J. 2014. Large-margin convex polytope machine. In Advances in Neural Information Processing Systems, 3248 3256. Kloft, M.; Brefeld, U.; Sonnenburg, S.; and Zien, A. 2011. Lp-norm multiple kernel learning. The Journal of Machine Learning Research 12:953 997. Lanckriet, G. R.; Cristianini, N.; Bartlett, P.; Ghaoui, L. E.; and Jordan, M. I. 2004. Learning the kernel matrix with semidefinite programming. The Journal of Machine Learning Research 5:27 72. Lichman, M. 2013. UCI machine learning repository. Liu, X.; Wang, L.; Zhang, J.; and Yin, J. 2014. Sampleadaptive multiple kernel learning. Twenty-Eighth AAAI Conference on Artificial Intelligence (AAAI-14). Liu, X.; Wang, L.; Yin, J.; Dou, Y.; and Zhang, J. 2015. Absent multiple kernel learning. Twenty-Ninth AAAI Conference on Artificial Intelligence (AAAI-15). Manwani, N., and Sastry, P. 2010. Learning polyhedral classifiers using logistic function. In ACML, 17 30. Manwani, N., and Sastry, P. 2011. Polyceptron: A polyhedral learning algorithm. ar Xiv preprint ar Xiv:1107.1564. Ong, C. S.; Williamson, R. C.; and Smola, A. J. 2005. Learning the kernel with hyperkernels. In Journal of Machine Learning Research, 1043 1071. Rakotomamonjy, A.; Bach, F.; Canu, S.; and Grandvalet, Y. 2008. Simplemkl. Journal of Machine Learning Research 9:2491 2521. Scholkopf B, Smola AJ, W. R. B. P. 2000. New support vector algorithms. Advances In Neural Information Processing Systems (NIPS). Stephen Boyd, L. V. 2004. Convex optimization. ISBN:0521833787. Cambridge University Press New York. Xu, L.; Crammer, K.; and Schuurmans, D. 2006. Robust support vector machine training via convex outlier ablation. In AAAI, volume 6, 536 542. Ying, Y., and Campbell, C. 2009. Generalization bounds for learning the kernel. In 22nd Annual Conference on Learning Theory (COLT 2009). Yu, C.-N. J., and Joachims, T. 2009. Learning structural svms with latent variables. In Proceedings of the 26th Annual International Conference on Machine Learning, 1169 1176. ACM. Zhou, Y.; Jin, B.; and Spanos, C. 2015. Learning convex piecewise linear machine for data-driven optimal control. In International conference on Machine Learning and Applications (ICMLA). IEEE. Zhou, Y.; Li, D.; and Spanos, C. 2015. Learning optimization friendly comfort model for hvac model predictive control. In International conference on Data Mining (ICDMW 2015). IEEE. Zhuang, J.; Tsang, I. W.; and Hoi, S. 2011. Two-layer multiple kernel learning. In International Conference on Artificial Intelligence and Statistics, 909 917.