# learning_theory_for_distribution_regression__e0d1a202.pdf Journal of Machine Learning Research 17 (2016) 1-40 Submitted 12/14; Revised 1/16; Published 9/16 Learning Theory for Distribution Regression Zolt an Szab o zoltan.szabo@gatsby.ucl.ac.uk ORCID 0000-0001-6183-7603 Gatsby Unit, University College London Sainsbury Wellcome Centre, 25 Howland Street London - W1T 4JG, UK Bharath K. Sriperumbudur bks18@psu.edu Department of Statistics Pennsylvania State University University Park, PA 16802, USA Barnab as P oczos bapoczos@cs.cmu.edu Machine Learning Department School of Computer Science Carnegie Mellon University 5000 Forbes Avenue Pittsburgh PA 15213 USA Arthur Gretton arthur.gretton@gmail.com ORCID 0000-0003-3169-7624 Gatsby Unit, University College London Sainsbury Wellcome Centre, 25 Howland Street London - W1T 4JG, UK Editor: Ingo Steinwart We focus on the distribution regression problem: regressing to vector-valued outputs from probability measures. Many important machine learning and statistical tasks fit into this framework, including multi-instance learning and point estimation problems without analytical solution (such as hyperparameter or entropy estimation). Despite the large number of available heuristics in the literature, the inherent two-stage sampled nature of the problem makes the theoretical analysis quite challenging, since in practice only samples from sampled distributions are observable, and the estimates have to rely on similarities computed between sets of points. To the best of our knowledge, the only existing technique with consistency guarantees for distribution regression requires kernel density estimation as an intermediate step (which often performs poorly in practice), and the domain of the distributions to be compact Euclidean. In this paper, we study a simple, analytically computable, ridge regression-based alternative to distribution regression, where we embed the distributions to a reproducing kernel Hilbert space, and learn the regressor from the embeddings to the outputs. Our main contribution is to prove that this scheme is consistent in the two-stage sampled setup under mild conditions (on separable topological domains enriched with kernels): we present an exact computational-statistical efficiency trade-off analysis showing that our estimator is able to match the one-stage sampled minimax op- . Now at Applied Mathematics Department, Center for Applied Mathematics, Ecole Polytechnique, University of Paris-Saclay, Route de Saclay, 91128 Palaiseau Cedex, France. c 2016 Zolt an Szab o, Bharath K. Sriperumbudur, Barnab as P oczos, and Arthur Gretton. Szab o et al. timal rate (Caponnetto and De Vito, 2007; Steinwart et al., 2009). This result answers a 17-year-old open question, establishing the consistency of the classical set kernel (Haussler, 1999; G artner et al., 2002) in regression. We also cover consistency for more recent kernels on distributions, including those due to Christmann and Steinwart (2010). Keywords: Two-Stage Sampled Distribution Regression, Kernel Ridge Regression, Mean Embedding, Multi-Instance Learning, Minimax Optimality 1. Introduction We address the learning problem of distribution regression in the two-stage sampled setting, where we only have bags of samples from the probability distributions: we regress from probability measures to real-valued (P oczos et al., 2013) responses, or more generally to vector-valued outputs (belonging to an arbitrary separable Hilbert space). Many classical problems in machine learning and statistics can be analysed in this framework. On the machine learning side, multiple instance learning (Dietterich et al., 1997; Ray and Page, 2001; Dooly et al., 2002) can be thought of in this way, where each instance in a labeled bag is an i.i.d. (independent identically distributed) sample from a distribution. On the statistical side, tasks might include point estimation of statistics on a distribution without closed form analytical expressions (e.g., its entropy or a hyperparameter). Intuitive description of our goal: Let us start with a somewhat informal definition of the distribution regression problem and an intuitive phrasing of our goals. Suppose that our data consist of z = {(xi, yi)}l i=1, where xi is a probability distribution, yi is its label (in the simplest case yi R or yi Rd) and each (xi, yi) pair is i.i.d. sampled from a meta distribution M. However, we do not observe xi directly; rather, we observe a sample xi,1, . . . , xi,Ni i.i.d. xi. Thus the observed data are ˆz = {({xi,n}Ni n=1, yi)}l i=1. Since xi,j is sampled from xi, and xi is sampled from M, we call this process two-stage sampling. Our goal is to predict a new yl+1 from a new batch of samples xl+1,1, . . . , xl+1,Nl+1 drawn from a new distribution xl+1 M. For example, in a medical application, the ith patient might be identified with a probability distribution (xi), which can be periodically assessed by blood tests ({xi,n}Ni n=1). We are also given some health indicator of the patient (yi), which might be inferred from his/her blood measurements. Based on the observations (ˆz), we might wish to learn the mapping from the set of blood tests to the health indicator, and the hope is that by observing more patients (larger l) and performing a larger number of tests (larger Ni) the estimated mapping ( ˆf = ˆf(ˆz)) becomes more precise . Briefly, we consider the following questions: Can the distribution regression problem be solved consistently under mild conditions? What is the exact computationalstatistical efficiency trade-offimplied by the two-stage sampling? In our work the estimated mapping ( ˆf) is the analytical solution of a kernel ridge regression (KRR) problem.1 The performance of ˆf depends on the assumed function class (H), the 1. Beyond its simple analytical formula, kernel ridge regression also allows efficient distributed (Zhang et al., 2015; Richt arik and Tak a c, 2016), sketch (Alaoui and Mahoney, 2015; Yang et al., 2016) and Nystr om based approximations (Rudi et al., 2015). Learning Theory for Distribution Regression family of ˆf candidates used in the ridge formulation. We shall focus on the analysis of two settings: 1. Well-specified case (f H): In this case we assume that the regression function f belongs to H. We focus on bounding the goodness of ˆf compared to f . In other words, if R[f ] denotes the prediction error (expected risk) of f , then our goal is to derive a finite-sample bound for the excess risk, E( ˆf, f ) = R[ ˆf] R[f ] that holds with high probability. We make use of this bound to establish the consistency of the estimator (i.e., drive the excess risk to zero) and to derive the exact computationalstatistical efficiency trade-offof the estimator as a function of the sample number (l, N = Ni, i) and the problem difficulty (see Theorem 5 and its corresponding remarks for more details). 2. Misspecified case (f L2\H): Since in practise it might be hard to check whether f H, we also study the misspecified setting of f L2; the relevant case is when f L2\H. In the misspecified setting the richness of H has crucial importance, in other words the size of D2 H = inff H f f 2 2, the approximation error from H. Our main contributions consist of proving a finite-sample excess risk bound, using which we show that the proposed estimator can attain the ideal performance, i.e., E( ˆf, f ) D2 H can be driven to zero. Moreover, on smooth classes of f -s, we give a simple and explicit description for the computational-statistical efficiency trade-offof our estimator (see Theorem 9 and its corresponding remarks for more details). There exist a vast number of heuristics to tackle learning problems on distributions; we will review them in Section 5. However, to the best of our knowledge, the only prior work addressing the consistency of regression on distributions requires kernel density estimation (P oczos et al., 2013; Oliva et al., 2014; Sutherland et al., 2016), which assumes that the response variable is scalar-valued,2 and the covariates are nonparametric continuous distributions on Rd. As in our setting, the exact forms of these distributions are unknown; they are available only through finite sample sets. P oczos et al. estimated these distributions through a kernel density estimator (assuming these distributions have a density) and then constructed a kernel regressor that acts on these kernel density estimates.3 Using the classical bias-variance decomposition analysis for kernel regressors, they showed the consistency of the constructed kernel regressor, and provided a polynomial upper bound on the rates, assuming the true regressor to be H older continuous, and the meta distribution that generates the covariates xi to have finite doubling dimension (Kpotufe, 2011).4 One can define kernel learning algorithms on bags based on set kernels (G artner et al., 2002) by computing the similarity of the sets/bags of samples representing the input distributions; set kernels are also called called multi-instance kernels or ensemble kernels, and are examples of convolution kernels (Haussler, 1999). In this case, the similarity of two sets 2. Oliva et al. (2013, 2015) consider the case where the responses are also distributions or functions. 3. We would like to clarify that the kernels used in their work are classical smoothing kernels extensively studied in non-parametric statistics (Gy orfiet al., 2002) and not the reproducing kernels that appear throughout our paper. 4. Using a random kitchen sinks approach, with orthonormal basis projection estimators Oliva et al. (2014); Sutherland et al. (2016) propose distribution regression algorithms that can computationally handle large scale datasets; as with P oczos et al. (2013), these approaches are based on density estimation in Rd. Szab o et al. is measured by the average pairwise point similarities between the sets. From a theoretical perspective, nothing is known about the consistency of set kernel based learning method since their introduction in 1999 (Haussler, 1999; G artner et al., 2002): i.e. in what sense (and with what rates) is the learning algorithm consistent, when the number of items per bag, and the number of bags, are allowed to increase? It is possible, however, to view set kernels in a distribution setting, as they represent valid kernels between (mean) embeddings of empirical probability measures into a reproducing kernel Hilbert space (RKHS; Berlinet and Thomas-Agnan, 2004). The population limits are well-defined as being dot products between the embeddings of the generating distributions (Altun and Smola, 2006), and for characteristic kernels the distance between embeddings defines a metric on probability measures (Sriperumbudur et al., 2011; Gretton et al., 2012). When bounded kernels are used, mean embeddings exist for all probability measures (Fukumizu et al., 2004). When we consider the distribution regression setting, however, there is no reason to limit ourselves to set kernels. Embeddings of probability measures to RKHS are used by Christmann and Steinwart (2010) in defining a yet larger class of easily computable kernels on distributions, via operations performed on the embeddings and their distances. Note that the relation between set kernels and kernels on distributions was also applied by Muandet et al. (2012) for classification on distribution-valued inputs, however consistency was not studied in that work. We also note that motivated by the current paper, Lopez-Paz et al. (2015) have recently presented the first theoretical results about surrogate risk guarantees on a class (relying on uniformly bounded Lipschitz functionals) of soft distribution-classification problems. Our contribution in this paper is to establish the learning theory of a simple, mean embedding based ridge regression (MERR) method for the distribution regression problem. This result applies both to the basic set kernels of Haussler (1999); G artner et al. (2002), the distribution kernels of Christmann and Steinwart (2010), and additional related kernels. We provide finite-sample excess risk bounds, prove consistency, and show how the two-stage sampled nature of the problem (bag size) governs the computational-statistical efficiency of the MERR estimator. More specifically, in the 1. well-specified case: We (a) derive finite-sample bounds on the excess risk: We construct R[ ˆf] R[f ] r(l, N, λ) bounds holding with high probability, where λ is the regularization parameter in the ridge problem (λ 0, l , N = Ni ). (b) establish consistency and computational-statistical efficiency trade-offof the MERR estimator on a general prior family P(b, c) as defined by Caponnetto and De Vito (2007), where b captures the effective input dimension, and larger c means smoother f (1 < b, c (1, 2]). In particular, when the number of samples per bag is chosen as N = la log(l) and a b(c+1) bc+1 , then the learning rate saturates at l bc bc+1 , which is known to be one-stage sampled minimax optimal (Caponnetto and De Vito, 2007). In other words, by choosing a = b(c+1) bc+1 < 2, we suffer no loss in statistical performance compared with the best possible one-stage sampled estimator. Note: the advantage of considering the P(b, c) family is two-fold. It does not assume parametric distributions, yet certain complexity terms can be explicitly upper bounded in the family. This property will be exploited in our analysis. Moreover, (for special Learning Theory for Distribution Regression input distributions) the parameter b can be related to the spectral decay of Gaussian Gram matrices, and existing analysis techniques (Steinwart and Christmann, 2008) may be used in interpreting these decay conditions. 2. misspecified case: We establish consistency and convergence rates even if f / H. Particularly, by deriving finite-sample bounds on the excess risk we (a) prove that the MERR estimator can achieve the best possible approximation accuracy from H, i.e. the R[ ˆf] R[f ] D2 H quantity can be driven to zero (recall that DH = inff H f f 2). Specifically, this result implies that if H is dense in L2 (DH = 0), then the excess risk R[ ˆf] R[f ] converges to zero. (b) analyse the computational-statistical efficiency trade-off: We show that by choosing the bag size as N = l2a log(l) (a > 0) one can get rate l 2sa s+1 for a s+1 s+2, and the rate saturates for a s+1 s+2 at l 2s s+2 , where the difficulty of the regression problem is captured by s (0, 1] (a larger s means an easier problem). This means that easier tasks give rise to faster convergence (for s = 1, the rate is l 2 3 ), the bag size N can again be sub-quadratic in l (2a 2(s+1) 3 < 2), and the rate at saturation is close to r(l) = l 2s 2s+1 , which is the asymptotically optimal rate in the one-stage sampled setup, with real-valued output and stricter eigenvalue decay conditions (Steinwart et al., 2009). Due to the differences in the assumptions made and the loss function used, a direct comparison of our theoretical result and that of P oczos et al. (2013) remains an open question, however we make three observations. First, our approach is more general, since we may regress from any probability measure defined on separable, topological domains endowed with kernels. P oczos et al. s work is restricted to compact domains of finite dimensional Euclidean spaces, and requires the distributions to admit probability densities; distributions on strings, graphs, and other structured objects are disallowed. Second, in our analysis we will allow separable Hilbert space valued outputs, in contrast to the real-valued output considered by P oczos et al. (2013). Third, density estimates in high dimensional spaces suffer from slow convergence rates (Wasserman, 2006, Section 6.5). Our approach mitigates this problem, as it works directly on distribution embeddings, and does not make use of density estimation as an intermediate step. The principal challenge in proving theoretical guarantees arises from the two-stage sampled nature of the inputs. In our analysis of the well-specified case, we make use of Caponnetto and De Vito (2007) s results, which focus (only) on the one-stage sample setup. These results will make our analysis somewhat shorter (but still rather challenging) by giving upper bounds for some of the objective terms. Even the verification of these conditions requires care since the inputs in the ridge regression are themselves distribution embeddings (i.e., functions in a reproducing kernel Hilbert space). In the misspecified case, RKHS methods alone are not sufficient to obtain excess risk bounds: one has to take into account the richness of the modelling RKHS class (H) in the embedding L2 space. The fundamental challenge is whether it is possible to achieve the best possible performance dictated by H; or in the special case when further smoothness conditions hold on f , what convergence rates can yet be attained, and what computationalstatistical efficiency trade-offrealized. The second smoothness property could be modelled Szab o et al. for example by range spaces of (fractional) powers of integral operators associated to H. Indeed, there exist several results along these lines with KRR for the case of real-valued outputs: see for example (Sun and Wu, 2009a, Theorem 1.1), (Sun and Wu, 2009b, Corollary 3.2), (Mendelson and Neeman, 2010, Theorem 3.7 with Assumption 3.2). The question of optimal rates has also been addressed for the semi-supervised KRR setting (Caponnetto, 2006, Theorem 1), and for clipped KRR estimators (Steinwart et al., 2009) with integral operators of rapidly decaying spectrum. Our results apply more generally to the two-stage sampled setting and to vector valued outputs belonging to separable Hilbert spaces. Moreover, we obtain a general consistency result without range space assumptions, showing that the modelling power of H can be fully exploited, and convergence to the best approximation available from H can be realized.5 There are numerous areas in machine learning and statistics, where estimating vectorvalued functions has crucial importance. Often in statistics, one is not only confronted with the estimation of a scalar parameter, but with a vector of parameters. On the machine learning side, multi-task learning (Evgeniou et al., 2005), functional response regression (Kadri et al., 2016), or structured output prediction (Brouard et al., 2011; Kadri et al., 2013) fall under the same umbrella: they can be naturally phrased as learning vectorvalued functions (Micchelli and Pontil, 2005). The idea underlying all these tasks is simple and intuitive: if multiple prediction problems have to be solved simultaneously, it might be beneficial to exploit their dependencies. Imagine for example that the task is to predict the motion of a dancer: taking into account the interrelation of the actor s body parts is likely to lead to more accurate estimation, as opposed to predicting the individual parts one by one, independently. Successful real-world applications of a multi-task approach include for example preference modelling of users with similar demographics (Evgeniou et al., 2005), prediction of the daily precipitation profiles of weather stations (Kadri et al., 2010), acoustic-to-articulatory speech inversion (Kadri et al., 2016), identifying biomarkers capable of tracking the progress of Alzheimer s disease (Zhou et al., 2013), personalized human activity recognition based on i Pod/i Phone accelerometer data (Sun et al., 2013), finger trajectory prediction in brain-computer interfaces (Kadri et al., 2012) or ecological inference (Flaxman et al., 2015); for a recent review on multi-output prediction methods see ( Alvarez et al., 2011; Borchani et al., 2015). A mathematically sound way of encoding prior information about the relation of the outputs can be realized by operator-valued kernels and the associated vector-valued RKHS-s (Pedrick, 1957; Micchelli and Pontil, 2005; Carmeli et al., 2006, 2010); this is the tool we use to allow vector-valued learning tasks. Finally, we note that the current work extends our earlier conference paper (Szab o et al., 2015) in several important respects: we now show that the MERR method can attain the one-stage sampled minimax optimal rate; we generalize the analysis in the well-specified setting to allow outputs belonging to an arbitrary separable Hilbert spaces (in contrast to the original scalar-valued output domain); and we tackle the misspecified setting, obtaining finite sample guarantees, consistency, and computational-statistical efficiency trade-offs. The paper is structured as follows: The distribution regression problem and the MERR technique are introduced in Section 2. Our assumptions are detailed in Section 3. We present our theoretical guarantees (finite-sample bounds on the excess risk, consistency, 5. Specializing our result, we get explicit rates and an exact computational-statistical efficiency description for MERR as a function of sample numbers and problem difficulty, for smooth regression functions. Learning Theory for Distribution Regression computational-statistical efficiency trade-offs) in Section 4: the well-specified case is considered in Section 4.1, and the misspecified setting is the focus of Section 4.2. Section 5 is devoted to an overview of existing heuristics for learning on distributions. Conclusions are drawn in Section 6. Section 7 contains proof details. In Section 8 we discuss our assumptions with concrete examples. 2. The Distribution Regression Problem Below we first introduce our notation (Section 2.1), then formally define the distribution regression task (Section 2.2). 2.1 Notation We use the following notations throughout the paper: Sets, topology, measure theory: Let K be a Hilbert space; cl V is the closure of a set V K. i ISi is the direct product of sets Si. f g is the composition of function f and g. Let (X, τ) be a topological space and let B(X) := B(τ) be the Borel σ-algebra induced by the topology τ. If (X, d) is a metric space, then B = B(d) is the Borel σ-algebra generated by the open sets induced by metric d. M+ 1 (X) denotes the set of Borel probability measures on the (X, B(X)) measurable space. Given measurable spaces (U1, S1) and (U2, S2), the S1 S2 product σ-algebra (Steinwart and Christmann, 2008, page 480) on the product space U1 U2 is the σ-algebra generated by the cylinder sets U1 S2, S1 U2 (S1 S1, S2 S2). The weak topology (τw = τw(X, τ)) on M+ 1 (X) is defined as the weakest topology such that the Lh : (M+ 1 (X), τw) R, Lh(x) = R X h(u)dx(u) mapping is continuous for all h Cb(X) = {(X, τ) R bounded, continuous functions}. Functional analysis: Let (N1, N1) and (N2, N2) denote two normed spaces, then L(N1, N2) stands for the space of N1 N2 bounded linear operators; if N1 = N2, we will use the L(N1) = L(N1, N2) shorthand. For M L(N1, N2) the operator norm is defined as M L(N1,N2) = sup0 =h N1 Mh N2 / h N1, Im(M) = {Mn1}n1 N1 denotes the range of M, Ker(M) = {n1 N1 : Mn1 = 0} is the null space of M. Let K be a Hilbert space. The adjoint operator M L(K) of an operator M L(K) is the operator such that Ma, b K = a, M b K for all a and b in K. M L(K) is called positive if Ma, a K 0 ( a K), self-adjoint if M = M , and trace class if P j J |M|ej, ej K < for an (ej)j J ONB (orthonormal basis) of K (|M| := (M M) 1 2 ), in which case Tr(M) := P j J Mej, ej K < ; compact if cl [Ma : a K, a K 1] is a compact set. Let K1 and K2 be Hilbert spaces. M L(K1, K2) is called Hilbert-Schmidt if M 2 L2(K1,K2) = Tr(M M) = P j J Mej, Mej K2 < for some (ej)j J ONB of K1. The space of Hilbert-Schmidt operators is denoted by L2(K1, K2) = {M L(K1, K2) : M L2(K1,K2) < }. We use the shorthand notation L2(K) = L2(K, K) if K := K1 = K2; L2(K) is separable if and only if K is separable (Steinwart and Christmann, 2008, page 506). Trace class and Hilbert-Schmidt operators over a K Hilbert space are compact operators (Steinwart and Christmann, 2008, page 505-506); moreover, A L(K) A L2(K) , A L2(K), (1) AB L2(K) A L2(K) B L(K) , A, B L2(K). (2) Szab o et al. I is the identity operator; Il Rl l is the identity matrix. RKHS, mean embedding: Let H = H(k) be an RKHS (Steinwart and Christmann, 2008) with k : X X R as the reproducing kernel. Denote by X = µ M+ 1 (X) = {µx : x M+ 1 (X)} H, µx = Z X k( , u)dx(u) = Eu x[k( , u)] H the set of mean embeddings (Berlinet and Thomas-Agnan, 2004) of the distributions to the space H.6 Let Y be a separable Hilbert space, where the inner product is denoted by , Y ; the associated norm is Y . H = H(K) is the Y -valued RKHS (Pedrick, 1957; Micchelli and Pontil, 2005; Carmeli et al., 2006, 2010) of X Y functions with K : X X L(Y ) as the reproducing kernel (we will present some concrete examples of K in Section 3; see Table 1); Kµx L(Y, H) is defined as K(µx, µt)(y) = (Kµty)(µx), ( µx, µt X), or K( , µt)(y) = Kµty. (3) Further, f(µx) = K µxf ( µx X, f H). Regression function: Let ρ be the µ-induced probability measure on the Z = X Y product space, and let ρ(µx, y) = ρ(y|µx)ρX(µx) be the factorization of ρ into conditional and marginal distributions.7 The regression function of ρ with respect to the (µx, y) pair is denoted by Y y dρ(y|µa) (µa X) (4) and for f L2 ρX let f ρ = q f, f ρ := f L2ρX = h R X f(µa) 2 Y dρX(µa) i 1 assume that the operator-valued kernel K : X X L(Y ) is a Mercer kernel (that is H = H(K) C(X, Y ) = {X Y continuous functions}), is bounded ( BK < such that K(x, x) L(Y ) BK), and is a compact operator for all x X. These requirements will be guaranteed by our assumptions, see Section 7.2.6. In this case, the inclusion S K: H , L2 ρX is bounded, and its adjoint SK : L2 ρX H is given by (SKg)(µu) = Z X K(µu, µt)g(µt)dρX(µt). (5) We further define T as T = S KSK : L2 ρX L2 ρX; (6) in other words, the result of operation (5) belongs to H, which is embedded in L2 ρX. T is a compact, positive, self-adjoint operator (Carmeli et al., 2010, Proposition 3), thus by the spectral theorem T s exists, where s 0. 6. The x 7 µx mapping is defined for all x M+ 1 (X) if k is bounded, i.e., supu X k(u, u) < . 7. Our assumptions will guarantee the existence of ρ (see Section 3). Since Y is a Polish space (because it is separable Hilbert) the ρ(y|µa) conditional distribution (y Y , µa X) is also well-defined (Steinwart and Christmann, 2008, Lemma A.3.16, page 487). Learning Theory for Distribution Regression 2.2 Distribution Regression We now formally define the distribution regression task. Let us assume that M+ 1 (X) is endowed with S1 = B(τw), the weak-topology generated σ-algebra; thus (M+ 1 (X), S1) is a measurable space. In the distribution regression problem, we are given samples ˆz = {({xi,n}Ni n=1, yi)}l i=1 with xi,1, . . . , xi,Ni i.i.d. xi where z = {(xi, yi)}l i=1 with xi M+ 1 (X) and yi Y drawn i.i.d. from a joint meta distribution M defined on the measurable space (M+ 1 (X) Y, S1 B(Y )), the product space enriched with the product σ-algebra. Unlike in classical supervised learning problems, the problem at hand involves two levels of randomness, wherein first z is drawn from M, and then ˆz is generated by sampling points from xi for all i = 1, . . . , l. The goal is to learn the relation between the random distribution x and response y based on the observed ˆz. For notational simplicity, we will assume that N = Ni ( i). As in the classical regression problem (Rd R), distribution regression can be tackled via kernel ridge regression (using a squared loss as the discrepancy criterion). The kernel (say KG) is defined on M+ 1 (X), and the regressor is then modelled by an element in the RKHS G = G(KG) of functions mapping from M+ 1 (X) to Y . In this paper, we choose KG(x, x ) = K(µx, µx ) where x, x M+ 1 (X) and so that the function (in G) to describe the (x, y) random relation is constructed as a composition f µx, i.e. M+ 1 (X) µ X( H = H(k)) f H=H(K) Y. (7) In other words, the distribution x M+ 1 (X) is first mapped to X H by the mean embedding µ, and the result is composed with f, an element of the RKHS H. Let the expected risk for a f : X Y (measurable) function be defined as R f = E(x,y) M f(µx) y 2 Y , which is minimized by the fρ regression function. The classical regularization approach is to optimize fλ z = arg min f H i=1 f(µxi) yi 2 Y + λ f 2 H (8) instead of R, based on samples z. Since z is not available, we consider the objective function defined by the observable quantity ˆz, fλ ˆz = arg min f H i=1 f(µˆxi) yi 2 Y + λ f 2 H , (9) where ˆxi = 1 N PN n=1 δxi,n is the empirical distribution determined by {xi,n}N i=1. The ridge regression objective function has an analytical solution: given training samples ˆz, the prediction for a new t test distribution is (fλ ˆz µ)(t) = k(K + lλI) 1[y1; . . . ; yl], (10) where k = [K(µˆx1, µt), . . . , K(µˆxl, µt)] L(Y )1 l, K = [K(µˆxi, µˆxj)] L(Y )l l, [y1; . . . ; yl] Y l. Szab o et al. It is important to note that the algorithm has access to the sample points only via their mean embeddings {µˆxi}l i=1 in Eq. (9). There is a two-stage sampling difficulty to tackle: The transition from fρ to fλ z represents the fact that we have only l distribution samples (z); the transition from fλ z to fλ ˆz means that the xi distributions can be accessed only via samples (ˆz). While ridge regression can be performed using the kernel KG, the two-stage sampling makes it difficult to work with arbitrary KG. By contrast, our choice of KG(x, x ) = K(µx, µx ) enables us to handle the two-stage sampling by estimating µx with an empirical estimator, and using it in the algorithm as shown above. In case of scalar output (Y = R), L(Y ) = L(R) = R and (10) is a standard linear equation with K Rl l, k R1 l. More generally, if Y = Rd, then L(Y ) = L(Rd) = Rd d and (10) is still a finite-dimensional linear equation with K R(dl) (dl) and k Rd (dl). One could also formulate the problem (and get guarantees) for more abstract X H Y regression tasks [see Eq. (7)] on a convex set X with H and Y being general, separable Hilbert spaces. Since distribution regression is probably the most accessible example where two-stage sampling appears, and in order to keep the presentation simple, we do not consider such extended formulations in this work. Our main goals in this paper are as follows: first, to analyse the excess risk E fλ ˆz , fρ := R[fλ ˆz ] R[fρ], both when fρ H (the well-specified case) and fρ L2 ρX\H (the misspecified case); second, to establish consistency (E fλ ˆz , fρ 0, or in the misspecified case E fλ ˆz , fρ D2 H 0, where D2 H := infq H fρ S Kq 2 ρ is the approximation error of fρ by a function in H); and third, to derive an exact computational-statistical efficiency trade-offas a function of the (l, N, λ) triplet, and of the difficulty of the problem. 3. Assumptions In this section, we detail our assumptions on the (X, Y, k, K) quartet. Our analysis for the well-specified case uses existing ridge regression results (Caponnetto and De Vito, 2007) focusing on problem (8) where only a single-stage sampling is present, hence we have to verify the associated conditions. Though we make use of these results, the analysis still remains challenging; the available bounds can moderately shorten our proof. We must take particular care in verifying that Caponnetto and De Vito (2007) s conditions are met, since they need to hold for the space of mean embeddings of the distributions (X = µ M+ 1 (X) ), whose properties as a function of X and H must themselves be established. Our assumptions are as follows: 1. (X, τ) is a separable, topological space. Learning Theory for Distribution Regression 2. Y is a separable Hilbert space. 3. k is bounded, in other words Bk < such that supu X k(u, u) Bk, and continuous. 4. The {Kµa}µa X operator family is uniformly bounded in Hilbert-Schmidt norm and H older continuous in operator norm. Formally, BK < such that Kµa 2 L2(Y,H) = Tr K µa Kµa BK, ( µa X), (11) and L > 0, h (0, 1] such that the mapping K( ) : X L(Y, H) is H older continuous: Kµa Kµb L(Y,H) L µa µb h H , (µa, µb) X X. (12) 5. y is bounded: C < such that y Y C almost surely. These requirements hold under mild conditions: in Section 8, we provide insight into the consequences of our assumptions, with several concrete illustrations (e.g. regression with setand RBF-type kernels). 4. Error Bounds, Consistency & Computational-Statistical Efficiency Trade-off In this section, we present our analysis of the consistency of the mean embedding based ridge regression (MERR) method. Given the estimator (fλ ˆz ) in Eq. (9), we derive finite-sample high probability upper bounds (see Theorems 2 and 7) for the excess risk E fλ ˆz , fρ , and in the misspecified setting, for the excess risk compared to the best attainable value from H, i.e., E fλ ˆz , fρ D2 H. We illustrate the bounds for particular classes of prior distributions, and work through special cases to obtain consistency conditions and computational-statistical efficiency trade-offs (see Theorems 4, 9 and the 3rd bullet of Remark 8). The main challenge is how to turn the convergence rates of the mean embeddings into those for an error E of the predictor. Although the main ideas of the proofs can be summarized relatively briefly, the full details are more demanding. High-level ideas with the sketches of the proofs and the obtained results are presented in Section 4.1 (well-specified case) and Section 4.2 (misspecified case). The derivations of some technical details of Theorems 2 and 7 are available in Section 7. 4.1 Results for the Well-specified Case We first focus on the well-specified case (fρ H) and present our first main result. We derive a high probability upper bound for the excess risk E fλ ˆz , fρ of the MERR method (Theorem 2). The upper bound is instantiated for a general class of prior distributions (Theorem 4), which leads to a simple computational-statistical efficiency description (Theorem 5); this shows (among others) conditions when the MERR technique is able to achieve the one-stage sampled minimax optimal rate. We first give a high-level sketch of our convergence analysis and an intuitive interpretation of the results. An outline of the main proof ideas is given below, with technical details in Section 7. Let us define x = {xi}l i=1 and ˆx = {{xi,n}N n=1}l i=1 as the x-part of z and ˆz, respectively. One can express fλ z [Eq. (8)] (Caponnetto and De Vito, 2007), and similarly fλ ˆz [Eq. (9)], Szab o et al. fλ z = (Tx + λ) 1gz, Tx = 1 i=1 Tµxi, gz = 1 i=1 Kµxiyi, (13) fλ ˆz = (Tˆx + λ) 1gˆz, Tˆx = 1 i=1 Tµˆxi, gˆz = 1 i=1 Kµˆxiyi, (14) where Tµa = Kµa K µa L(H) (µa X), Tx, Tˆx : H H, gz, gˆz H. By these explicit expressions, one can decompose the excess risk into 5 terms (Szab o et al., 2015, Section A.1.8): E fλ ˆz , fρ = R fλ ˆz R [fρ] 5 [S 1 + S0 + A(λ) + S1 + S2] , S 1 = S 1(λ, z, ˆz) = T(Tˆx + λI) 1(gˆz gz) 2 H, (15) S0 = S0(λ, z, ˆz) = T(Tˆx + λI) 1(Tx Tˆx)fλ z 2 H, (16) T(fλ fρ) 2 H, S1 = S1(λ, z) = T(Tx + λI) 1(gz Txfρ) 2 H, S2 = S2(λ, z) = T(Tx + λI) 1(T Tx)(fλ fρ) 2 H, fλ = arg min f H (R[f] + λ f 2 H), T = Z X TµadρX(µa) = SKS K : H H. (17) Three of the terms (S1, S2, A(λ)) are identical to the terms in Caponnetto and De Vito (2007), hence the earlier bounds can be applied. The two new terms (S 1, S0) resulting from two-stage sampling will be upper bounded by making use of the convergence of the empirical mean embeddings. These bounds will lead to the following results: Theorem 2 (Finite-sample excess risk bounds; well-specified case) Let K( ) : X L(Y, H) be H older continuous with constants L, h. Let l Z+, N Z+, 0 < λ, 0 < η < 1, 0 < δ, Cη = 32 log2(6/η), y Y C (a.s.) and A(λ) be the residual as defined above. Define M = 2(C + fρ H BK), Σ = M 2 , T as in (17), B(λ) = fλ fρ 2 H as the reconstruction error, and N(λ) = Tr[(T + λI) 1T] as the effective dimension. Then with probability at least 1 η e δ, the excess risk can be upper bounded as E fλ ˆz , fρ 5 log(l) + δ 2h (2Bk)h l2λ + Σ2N(λ) l2 + BKA(λ) + B(λ) + fρ 2 H l2λ + BKA(λ) l2λ + Σ2N(λ) if l 2CηBKN(λ)/λ, λ T L(H) and N 1 + p log(l) + δ 22 h+6 h Bk(BK) 1 h L 2 h /λ 2 h . Learning Theory for Distribution Regression Below we specialize our excess risk bound for a general prior class, which captures the difficulty of the regression problem as defined in Caponnetto and De Vito (2007). This P(b, c) class is described by two parameters b and c: larger b means faster decay of the eigenvalues of the covariance operator T [in Eq. (17)], hence smaller effective input dimension; larger c corresponds to a smoother regression function. Formally: Definition of the P(b, c) class: Let us fix the positive constants R, α, β. Then given 1 < b, c (1, 2], the P(b, c) class is the set of probability distributions ρ on Z = X Y such that 1. a range space assumption is satisfied: g H s.t. fρ = T c 1 2 g with g 2 H R, 2. in the spectral decomposition of T = P n=1 λn , en H en, where (en) n=1 is a basis of Ker(T) , the eigenvalues of T satisfy α nbλn β ( n 1). Remark 3 We make few remarks about the P(b, c) class: Range space assumption on fρ: The smoothness of fρ is expressed as a range space assumption, which is slightly different from the standard smoothness conditions appearing in non-parametric function estimation. By the spectral decomposition of T given above [λ1 λ2 . . . > 0, limn λn = 0], T ru = P n=1(λn)r u, en H en (r = c 1 2 0, u H) and Im(T r) = n X n=1 cnen : X n=1 c2 nλ 2r n < o . (18) Specifically, in the limit as r 0, we obtain fρ Im(T 0) = Im(I) = H (no constraint); larger values of r give rise to faster decay of the (cn) n=1 Fourier coefficients. This is the concrete meaning of fρ Im(T r). Spectral decay condition: We can provide a simple illustration of when the spectral decay conditions hold, in the event that the distributions are normal with means mi and identical variance (xi = N(mi, σ2I)). When Gaussian kernels (k) are used with linear K, then K(µxi, µxj) = e c mi mj 2 (Muandet et al., 2012, Table 1, line 2) (Gaussian, with arguments equal to the difference in means). Thus, this Gram matrix will correspond to the Gram matrix using a Gaussian kernel between points mi. The spectral decay of the Gram matrix will correspond to that of the Gaussian kernel, with points drawn from the meta-distribution over the mi. Thus, the source conditions are analysed in the same manner as for Gaussian Gram matrices: see e.g. Steinwart and Christmann (2008) for a discussion of these spectral decay properties. In the P(b, c) family, the behaviour of A(λ), B(λ) and N(λ) is known: A(λ) Rλc, B(λ) Rλc 1, N(λ) β b b 1λ 1 b . Specializing Theorem 2 and retaining its assumptions, we get: Theorem 4 (Finite-sample excess risk bound for ρ P(b, c)) Suppose the conditions in Theorem 2 hold. Let ρ P(b, c), where 1 < b and c (1, 2]. Szab o et al. E fλ ˆz , fρ 5 log(l) + δ 2h (2Bk)h (b 1)lλ 1 b + Rλc 1 + fρ 2 H " B2 KRλc 2 l2 + BKRλc 1 (b 1)lλ 1 b Discarding the constants in Theorem 4, the study of convergence of the excess risk E(fλ ˆz , fρ) to 0 boils down to finding N and λ (as a function of l) where N , λ 0 and r(l, N, λ) = logh(l) λ2l2 + 1 + 1 lλ 1 b 0, s.t. lλ b+1 b 1, log(l) as l . Let us choose N = l a h log(l); in this case Eq. (19) reduces to r(l, λ) = 1 l2+aλ3 + 1 lλ 1 b 0, s.t. lλ b+1 b 1, laλ2 1. (20) One can assume that a > 0, otherwise r(l, λ) 0 fails to hold; in other words, N should grow faster than log(l). Matching the bias (λs) and variance (other) terms in r(l, λ) to choose λ, and guaranteeing that the matched terms dominate and the constraints in Eq. (20) hold, one gets the following simple description for the computational-statistical efficiency trade-off:8 Theorem 5 (Computational-statistical efficiency trade-off; well-specified case; ρ P(b, c)) Suppose the conditions in Theorem 2 hold. Let ρ P(b, c) and N = l a h log(l), where 0 < a, 1 < b, c (1, 2]. If bc+1 , then E fλ ˆz , fρ = Op l ac c+1 with λ = l a c+1 , bc+1 then E fλ ˆz , fρ = Op l bc bc+1 with λ = l b bc+1 . Remark 6 Theorem 5 formulates an exact computational-statistical efficiency trade-offfor the choice of the bag size (N) as a function of the number of distributions (l) and problem difficulty (b, c). a-dependence: A smaller bag size (smaller a; N = l a h log(l)) means computational savings, but reduced statistical efficiency. It is not worth increasing a above b(c+1) bc+1 since from that point the rate becomes r(l) = l bc bc+1 ; remarkably, this rate is minimax in the one-stage sampled setup (Caponnetto and De Vito, 2007). The sensible choice a = b(c+1) bc+1 < 2 means that the one-stage sampled minimax rate can be achieved in the two-stage sampled setting with bag size N sub-quadratic in l. 8. The derivations are available in the supplement of http://arxiv.org/pdf/1411.2066. Learning Theory for Distribution Regression h-dependence: In accord with our smoothness assumptions it is rewarding to use smoother K kernels (larger h (0, 1]) since this reduces the bag size [N = l a h log(l)]. c-dependence: The strictly decreasing property of c 7 b(c+1) bc+1 implies that for smoother problems (larger c) fewer samples (N) are sufficient. Below we elaborate on the sketched high-level idea and prove Theorem 2. Proof of Theorem 2 (detailed derivations of each step can be found in Section 7.1) 1. Decomposition of the excess risk: We have the following upper bound for the excess risk E fλ ˆz , fρ = R fλ ˆz R [fρ] 5 [S 1 + S0 + A(λ) + S1 + S2] . (21) 2. It is sufficient to upper bound S 1 and S0: Caponnetto and De Vito (2007) have shown that for η > 0 if l 2CηBKN(λ)/λ, λ T L(H), then P(Θ(λ, z) 1/2) 1 η/3, where Θ(λ, z) = (T Tx)(T + λI) 1 L(H) , (22) using which upper bounds on S1 and S2 that hold with probability 1 η are obtained. It is known that A(λ) Rλc. 3. Probabilistic bounds on gˆz gz 2 H, Tx Tˆx 2 L(H), T(Tˆx + λI) 1 2 L(H), fλ z 2 H: One can bound S 1 and S0 as T(Tˆx + λI) 1 2 L(H) gˆz gz 2 H T(Tˆx + λI) 1 2 L(H) Tx Tˆx 2 L(H) fλ z 2 H. For the terms on the r.h.s., we derive upper bounds [for the definition of α, see Eq. (24)] gˆz gz 2 H L2C2 (1 + α)2h (2Bk)h T(Tˆx + λI) 1 L(H) 2 Tx Tˆx 2 L(H) (1 + α)2h 2h+2(Bk)h BKL2 l2λ + Σ2N(λ) l2 + BKA(λ) + B(λ) + fρ 2 H The bounds hold under the following conditions: Szab o et al. gˆz gz 2 H (see Section 7.1.1): if the empirical mean embeddings are close to their population counterparts, i.e., µxi µˆxi H (1 + α) 2Bk N , ( i = 1, . . . , l). (24) This event has probability 1 le α over all i = 1, . . . , l samples; see (Altun and Smola, 2006) and (Szab o et al., 2015, Section A.1.10). Tx Tˆx 2 L(H) (see Section 7.1.2): (24) is assumed. T(Tˆx + λI) 1 2 L(H) (Szab o et al., 2015, Section A.1.11): (24), Θ(λ, z) 1 (1 + α)2 2 h+6 h Bk(BK) 1 h L 2 h λ 2 h N. (25) fλ z 2 H: The bound is guaranteed to hold under the conditions of the bounds of S1 and S2.8 4. Union bound: By applying an α = log(l) + δ reparameterization, and combining the received upper bounds with Caponnetto and De Vito (2007) s results for S1 and S2, Theorem 2 follows (Section 7.1.3) with a union bound. Finally, we note that existing results/ideas were used at two points to simplify our analysis: bounding S1, S2, Θ(λ, z), fλ z 2 H (Caponnetto and De Vito, 2007) and µxi µˆxi H (Altun and Smola, 2006).9 4.2 Results for the Misspecified Case In this section, we focus on the misspecified case (fρ L2 ρX\H) and present our second main result, which was inspired by the proof technique of Sriperumbudur et al. (2014, Theorem 12). We derive a high probability upper bound for E fλ ˆz , fρ , i.e., the excess risk of the MERR method (Theorem 7) which gives rise to consistency results (3rd bullet of Remark 8) and precise computational-statistical efficiency trade-off(Theorem 9). Theorem 7 consists of two finite-sample bounds: 1. The first, more general bound [Eq. (27)] will be used to show consistency in the misspecified case (see the 3rd bullet of Remark 8), in other words that E fλ ˆz , fρ can be driven to its smallest possible value determined by the richness of H: D2 H := inf q H fρ S Kq 2 ρ . (26) The value of DH equals the approximation error of fρ by a function from H. Specifically, if H [precisely S K(H) = {S Kq : q H} L2 ρX] is dense in L2 ρX, then DH = 0. 9. We also corrected some constants in the previous works (Altun and Smola, 2006; Caponnetto and De Vito, 2007). Learning Theory for Distribution Regression 2. The second, specialized result [Eq. (28)] under additional smoothness assumptions on fρ will give rise to a precise computational-statistical efficiency trade-offin terms of the problem difficulty (s) and sample numbers (l, N); this result can be seen as the misspecified analogue of Theorem 5. After stating our results, the main ideas of the proof follow; further technical details are available in Section 7.2. Our main theorem for bounding the excess risk is as follows: Theorem 7 (Finite-sample excess risk bounds; misspecified case) Let l Z+, N Z+, 0 < λ, 0 < η < 1, 0 < δ and Cη = log 6 η . Assume that 12BK log(l) + δ 22 h+6 h Bk(BK) 1 h L 2 h /λ 2 h N. 1. Then for arbitrary q H with probability at least 1 η e δ E fλ ˆz , fρ 2LC 1 + p log(l) + δ h (2Bk) h 2 BK l + C BK λ fρ ρ Da(λ, q) + Da(λ, q), where Da(λ, q) = fρ S Kq ρ + max(1, T L(H))λ 1 2 q H. 2. In addition, suppose fρ Im( T s) for some s > 0, where T is defined in Eq. (6). Then with probability at least 1 η e δ, we have E fλ ˆz , fρ 2LC 1 + p log(l) + δ h (2Bk) h 2 BK l + C BK λ T sfρ ρ Db(λ, s) + Db(λ, s), (28) where Db(λ, s) = max(1, T s 1 L(L2ρX ))λmin(1,s) T sfρ ρ. Remark 8 We give a short insight into the assumptions of Theorem 7, followed by consequences of the theorem. Range space assumption on fρ: The range space assumption for the compact, positive, self-adjoint operator, T = T(K) : L2 ρX L2 ρX in the 2nd part of Theorem 7 can be interpreted similarly to that on T; see Eq. (18). One can also prove alternative descriptions for Im( T s) in terms of interpolation spaces (Steinwart and Scovel, 2012, Theorem 4.6, page 387), or the decay of the 2-approximation error function, A2(λ) = inff H(K) λ f 2 H(K) + R[f] R[fρ] (Smale and Zhou, 2003; Steinwart et al., 2009). Szab o et al. E fλ ˆz , fρ : Notice that in the bounds [ (27), (28)], instead of the excess risk, its square root appears; this has technical reasons, as it is easier to have the Da(λ, q) quantity (without multiplicative constants) appear on the r.h.s. of Eq. (27) with this form. Consistency in the misspecified case: The consequence of Theorem 7(1) is as follows. Discarding the constants in Eq. (27), we obtain the upper bound (notice that the constant multiplier of fρ S Kq ρ in the last term was one): r(l, N, λ, q) = log h 2 (l) N h 2 λ + 1 q fρ S Kq ρ + l + fρ S Kq ρ + By choosing N = l1/h log l, p r(l, λ) is bounded by fρ S Kq ρ + q fρ S Kq ρ Our goal is to investigate the behavior of the bound as l , λ 0 and λ l . Define K(α, β, γ) := infq H n fρ S Kq ρ + α q fρ S Kq ρ + β p q H + γ q H o . K(α, β, γ) is the pointwise infimum of affine functions, therefore it is upper semicontinuous and concave on R3 (Aliprantis and Border, 2006, Lemmas 2.41 and 5.40 ); it is continuous on 3 i=1R>0 (Rockafellar and Wets, 2008, Theorem 2.35). Moreover, by applying (Rockafellar and Wets, 2008, Corollary 2.37) it extends continuously to 3 i=1R 0; specifically it is continuous at (α, β, γ) = 0. In other words, as l , λ 0 λ DH and we get consistency in the misspecified r(N, l, λ) DH. Discarding the constants in Eq. (28) we get10 r(l, N, λ) = log h 2 (l) N h 2 λ + 1 l + λmin(1,s), subject to 1 Our goal is to drive r(l, N, λ) to zero with a suitable choice of the (l, N, λ) triplet under the stronger range space assumption. Since in Eq. (29) min(1, s) appears, one can assume without loss of generality that s (0, 1]; consequently 1 s l 1 2 λ 1 2 1 Let us choose N = l2a/h log(l); in this case using the previous dominance note, Eq. (29) reduces to the study of p r(l, λ) = 1 laλ + 1 2 l 1 2 + λs 0, s.t. lλ2 1. (30) One can assume that a > 0, otherwise r(l, λ) 0 fails to hold: in other words, N should grow faster than log(l). Matching the bias (λs) and variance (other) terms in r(l, λ) to choose λ, guaranteeing that the matched terms dominate and the constraint in Eq. (30) hold, one can arrive at the following computational-statistical efficiency trade-off:8 10. We have discarded the log(l)/λ 2 h N constraint implied by the convergence of the first term in r. Learning Theory for Distribution Regression Theorem 9 (Computational-statistical efficiency trade-off; misspecified case, fρ Im( T s)) Suppose that fρ Im( T s) and N = l 2a h log(l), where s (0, 1], a > 0. If s+2, then E fλ ˆz , fρ = Op l 2sa s+1 with λ = l a s+1 , s+2, then E fλ ˆz , fρ = Op l 2s s+2 with λ = l 1 s+2 . Remark 10 Theorem 9 provides a complete computational-statistical efficiency trade-off description for the choice of the bag size (N) as a number of the distributions (l). a-dependence: A smaller value of a (smaller bags N = l2a/h log(l)) leads to a computational advantage, but one looses in statistical efficiency. As a reaches s+1 s+2, the rate becomes r(l) = l 2s s+2 and one does not gain from further increasing the value of a. The sensible choice of a = s+1 3 means that N can again be sub-quadratic (2a < 4 3 < 2) in l. h-dependence: By using smoother K kernels (larger h (0, 1]) one can reduce the size of the bags: h 7 2a/h is decreasing in h. This is compatible with our smoothness requirement on fρ. s-dependence: Easier tasks (larger s) give rise to faster convergence. Indeed, in the r(l) = l 2s s+2 rate the s 7 2s s+2 exponent is strictly increasing function of the problem difficulty (s). For example, for extremely non-smooth regression problems (s 0) the convergence can be arbitrary slow (lims 0 2s s+2 = 0). In the smooth case (s = 1) lims 1 2s s+2 = 2 3 and one can achieve the r(l) = l 2 We may compare our r(l) = l 2s s+2 result with the ro(l) = l 2s 2s+1 (one-stage sampled) rate (Steinwart et al., 2009, β/2 := s, q := 2, p := 1 in Corollary 6), which was shown to be asymptotically optimal on Y = R for continuous k on compact metric X. Steinwart et al. s result is more general in terms of q ( f q H based regularization) and p ( f C f p H f 1 p ρ , f H; in our case p = 1), although it imposes an additional eigenvalue constraint [(Steinwart et al., 2009, Eq. (6))] as well as fρ Im( T s). Moreover, one can observe that ro(l) r(l) with a small gap, and that for s 0 and s = 1, ro(l) = r(l); see Fig. 1. We further remind the reader that our MERR analysis also holds for separable Hilbert output spaces Y , separable topological domains X enriched with a bounded, continuous kernel k, and that we handle the twostage sampled setting. The main steps of the proof of Theorem 7 are as follows: Proof of Theorem 7 (the details of the derivation are available in Section 7.2) Steps 1-7 will be identical in both proofs,11 and we present them jointly. 1. Decomposition of the excess risk: By the triangle inequality, we have q E fλ ˆz , fρ = S Kfλ ˆz fρ ρ S K fλ ˆz fλ z ρ + S Kfλ z fρ ρ. (31) 11. Importantly, with a slight modification of the more general, first part of Theorem 7, one can get the specialized second setting of the theorem (see Step 8). Szab o et al. 0 0.2 0.4 0.6 0.8 1 0 Smoothness (s) logl[ro(l)]=2s/(2s+1) logl[r(l)]=2s/(s+2) Figure 1: Comparison of the ro(l) = l 2s 2s+1 and r(l) = l 2s s+2 rates as function of the problem difficulty/smoothness (s). 2. Bound on S K fλ ˆz fλ z ρ: Using12 the fact that Th 2 H ( h H), (32) and the definitions of S 1 and S0 [see Eqs. (15)-(16)], we obtain S K fλ ˆz fλ z ρ = T fλ ˆz fλ z H p through an application of triangle inequality. One can derive without a P(b, c) prior assumption (Section 7.2.1) the upper bound13 S0 2LC(1 + α)h(2Bk) h 2 for the r.h.s. of Eq. (33) under the conditions that Θ(λ, z) 1 2 (which holds with probability 1 η if [12BK log(2/η)/λ]2 l), and that Eqs. (24)-(25) hold. 3. Decomposition of S Kfλ z fρ ρ: By the triangle inequality and Eq. (32), we have S Kfλ z fρ ρ = S K fλ z fλ + S Kfλ fρ ρ S K fλ z fλ ρ + S Kfλ fρ ρ = T fλ z fλ H + S Kfλ fρ ρ. (34) 4. Decomposition of T fλ z fλ H: Making use of the analytical expressions for fλ z and fλ [see Eq. (13) and Eq. (17)], and the operator Woodbury formula (Ding and Zhou, 2008, Theorem 2.1, page 724) we arrive at the decomposition (see Section 7.2.2) T fλ z fλ H T(Tx + λI) 1 L(H) T Tx L(H) λ 1 SK fρ ( T + λI) 1S KSKfρ H where gρ = SKfρ. As it is known (Caponnetto and De Vito, 2007, page 348) T(Tx + λI) 1 L(H) 1/ λ provided that Θ(λ, z) 1 12. See for example de Vito et al. (2006) on page 88 with the (H, G, A, T) := (H, L2 ρX , S K, T) choice. 13. See the remark at the end of Section 7.2.1. Learning Theory for Distribution Regression 5. Bound on gz gρ H, T Tx L(H): By concentration arguments the bounds BK l + 2C BK , T Tx L(H) 4BK hold with probability at least 1 η, each (see Section 7.2.3, 7.2.4). 6. Decomposition of SK fρ ( T +λI) 1S KSKfρ 2 H: Exploiting the analytical formula for fλ, one can construct (Section 7.2.5) the upper bound SK fρ ( T + λI) 1S KSKfρ 2 H T fρ ( T + λI) 1S KSKfρ ρ S Kfλ fρ ρ. 7. Bound on T fρ ( T + λI) 1S KSKfρ ρ: Using our assumptions that fρ Im( T s) (s 0)14 and exploiting the separability of L2 ρX, by Lemma 7.3.2 (K = L2 ρX, f = fρ, M = T, a = 1) and T = S KSK we obtain the upper bound T fρ ( T + λI) 1S KSKfρ ρ = T fρ ( T + λI) 1 Tfρ ρ max 1, T s L(L2ρX ) λmin(1,s+1) T sfρ ρ = max 1, T s L(L2ρX) where we used at the last step that min(1, s + 1) = 1; this follows from s 0. 8. Bound on S Kfλ fρ ρ: (a) No range space assumption: One can construct (Section 7.2.6) the bound S Kfλ fρ ρ fρ S Kq ρ + max 1, T L(H) λ 1 2 q H , which holds for arbitrary q H. (b) Range space assumption in L2 ρX: Using the S Kfλ = ( T + λI) 1 Tfρ identity [see Eq. (43)], and Lemma 7.3.2 (M = T, K = L2 ρX, a = 0), we get S Kfλ fρ ρ = ( T + λI) 1 Tfρ fρ ρ max 1, T s 1 L(L2ρX ) λmin(1,s) T sfρ ρ. 9. Union bound: Applying an α = log(l) + δ reparameterization, changing η to η 3 and combining the derived results (in case of the first statement with s = 0) with a union bound, Theorem 7 follows. Remark 11 To contrast the derivation of the welland the misspecified cases, we note that previous results [Section 4.1, or Caponnetto and De Vito (2007) s bound] were used at two points: (a) In Step 2 by using Eq. (32) and transforming the L2 ρX error S K fλ ˆz fλ z ρ to H, we could rely on our previous bounds for S 1 and S0. However, we were required to use a different concentration argument to guarantee Θ(λ, z) 1 2 since we no longer assume the P(b, c) prior class. 14. Note that we choose s = 0 and s > 0 in the first and second theorem part, respectively. Szab o et al. (b) In Step 4 the first term could be bounded by Caponnetto and De Vito (2007). Its Θ(λ, z) 1 2 condition was guaranteed by Step 2; and see Section 7.2.1. We note that our misspecified proof method was inspired by Sriperumbudur et al. (2014, Theorem 12), where the authors focused on the consistency of an infinite-dimensional exponential family estimator. 5. Related Work In this section we discuss existing approaches and heuristic techniques to tackle learning problems on distributions. Methods based on parametric assumptions: A number of methods have been proposed to compute the similarity of distributions or bags of samples. As a first approach, one could fit a parametric model to the bags, and estimate the similarity of the bags based on the obtained parameters. It is then possible to define learning algorithms on the basis of these similarities, which often take analytical form. Typical examples with explicit formulas include Gaussians, finite mixtures of Gaussians, and distributions from the exponential family (with known log-normalizer function and zero carrier measure, see Kondor and Jebara, 2003; Jebara et al., 2004; Wang et al., 2009; Nielsen and Nock, 2012). A major limitation of these methods, however, is that they apply quite simple parametric assumptions, which may not be sufficient or verifiable in practise. Methods based on parametric assumption in a RKHS: A heuristic related to the parametric approach is to assume that the training distributions are Gaussians in a reproducing kernel Hilbert space (see for example Jebara et al., 2004; Zhou and Chellappa, 2006, and references therein). This assumption is algorithmically appealing, as many divergence measures for Gaussians can be computed in closed form using only inner products, making them straightforward to kernelize. A fundamental shortfall of kernelized Gaussian divergences is the lack of their consistency analysis in specific learning algorithms. Kernels based techniques: A more theoretically grounded approach to learning on distributions has been to define positive definite kernels on the basis of statistical divergence measures on distributions, or by metrics on non-negative numbers; these can then be used in kernel algorithms. This category includes work on semigroup kernels (Cuturi et al., 2005), non-extensive information theoretical kernel constructions (Martins et al., 2009), and kernels based on Hilbertian metrics (Hein and Bousquet, 2005). For example, the intuition of semigroup kernels (Cuturi et al., 2005) is as follows: if two measures or sets of points overlap, then their sum is expected to be more concentrated. The value of dispersion can be measured by entropy or inverse generalized variance. In the second type of approach (Hein and Bousquet, 2005), homogeneous Hilbert metrics on the non-negative real line are used to define the similarity of probability distributions. While these techniques guarantee to provide valid kernels on certain restricted domains of measures, the performance of learning algorithms based on finite-sample estimates of these kernels remains a challenging open question. One might also plug into learning algorithms (based on similarities of distributions) consistent R enyi and Tsallis divergence estimates (P oczos et al., 2011, 2012), but these similarity indices are not kernels, and their consistency in specific learning tasks remains an open question. Learning Theory for Distribution Regression Multi-instance learning: An alternative paradigm in learning when the inputs are bags of objects is to simply treat each input as a finite set: this leads to the multi-instance learning task (MIL, see Dietterich et al., 1997; Ray and Page, 2001; Dooly et al., 2002). In MIL one is given a set of labelled bags, and the task of the learner is to find the mapping from the bags to the labels. Many important examples fit into the MIL framework: for example, different configurations of a given molecule can be handled as a bag of shapes, images can be considered as a set of patches or regions of interest, a video can be seen as a collection of images, a document might be described as a bag of words or paragraphs, a web page can be identified by its links, a group of people on a social network can be captured by their friendship graphs, in a biological experiment a subject can be identified by his/her time series trials, or a customer might be characterized by his/her shopping records. The MIL approach has been applied in several domains; see the reviews from Babenko (2004); Zhou (2004); Foulds and Frank (2010); Amores (2013). Bag-of-objects methods (MIL, classification): Despite the large number of MIL applications and the spate of heuristic solution techniques, there exist few theoretical results in the area (Auer, 1998; Long and Tan, 1998; Blum and Kalai, 1998; Babenko et al., 2011; Zhang et al., 2013; Sabato and Tishby, 2012) and they focus on the multi-instance classification (MIC) task. In particular, let us first consider the standard MIC assumption (Dietterich et al., 1997), where a bag is declared to be positive (labelled with 1 ) if at least one of its instances is positive ( 1 ); otherwise, the bag is negative ( 0 ).15 In other words, if the instances (xi,n) in the ith bag {xi,1, . . . , xi,N} have hidden label L(xi,n) {0, 1}, then the observed label of the bag is yi = h(xi,1, . . . , xi,N) = max(L(xi,1), . . . , L(xi,N)) {0, 1}. In case of the original APR (axis-aligned rectangles; Dietterich et al., 1997) hypothesis class, function L is equal to the indicator of an unknown rectangle R (L = IR). In other words, a bag is declared to be positive if there exists at least one instance in the bag, which belongs to R.16 The goal is to learn R with high probability given the bags ({xi,1, . . . , xi,N}- s) and their labels (yi-s). Long and Tan (1998) proved the PAC learnability (probably approximately correct; Valiant, 1984) of the APR hypothesis class, if all instances in each bag are i.i.d. and follow the same product distribution over the instance coordinates. On the other hand, for arbitrary distributions over bags, when the instances within a bag might be statistically dependent, APR learning under MIC is NP-hard (Auer, 1998); the same authors also showed that the product property (Long and Tan, 1998) on the coordinates is not required to obtain PAC results. Blum and Kalai (1998) extended PAC learnability of APR-s to hypothesis classes learnable from one-sided classification noise. In contrast to the previous approaches (Long and Tan, 1998; Auer, 1998; Blum and Kalai, 1998), Babenko et al. (2011) modelled the bags as low-dimensional manifolds, and proved PAC bounds. By relaxing the standard MIC assumption, Sabato and Tishby (2012) showed PAC-learnability for general MIC hypothesis classes with extended max functions. Zhang et al. (2013) derived high-probability generalization bounds in the MIC setting, when local and global representations are combined. Our work falls outside this setting since the label and bag generation mechanisms we consider are different: we do not assume an exact form of the 15. The motivation of this assumption comes from drug discovery: if a molecule has at least one well-binding configuration, then it is considered to bind well. 16. In terms of drug binding prediction, this means that a molecule binds to a target iffat least one of its configurations falls within a fixed, but unknown rectangle. Szab o et al. labelling mechanism (function L and max in h). Rather, the labelling is presumed to be stochastically determined by the underlying true distribution, not deterministically by the instance realizations in the bags (these are presumed i.i.d., and may be bag-specific). Bag-of-objects methods (MIL, not classification): Beyond classification, there exist several heuristics without consistency guarantees for many other multi-instance problems in the literature, including regression (Ray and Page, 2001; Dooly et al., 2002; Zhou et al., 2009; Kwok and Cheung, 2007), clustering (Zhang and Zhou, 2009; Zhang et al., 2009, 2011; Chen and Wu, 2012), ranking (Bergeron et al., 2008; Hu et al., 2008; Bergeron et al., 2012), outlier detection (Wu et al., 2010), transfer learning (Raykar et al., 2008; Zhang and Si, 2009), and feature selection, -weighting and -extraction (also called dimensionality reduction, low-dimensional embedding, manifold learning, see Raykar et al., 2008; Ping et al., 2010; Sun et al., 2010; Carter et al., 2011; Zafra et al., 2013; Chai et al., 2014a,b, and references therein). Approaches using set metrics: Adapting the bag viewpoint of MIL, one can come up with set metric based learning algorithms.17 Probably one of the most well-known set metrics is the Hausdorffmetric (Edgar, 1995), which is defined for non-empty compact sets of metric spaces, specifically for sets containing finitely many points. There also exist other (semi)metric constructions on points sets (Eiter and Mannila, 1997; Ramon and Bruynooghe, 2001). Unfortunately, the classical Hausdorffmetric is highly sensitive to outliers, seriously limiting its practical applicability. In order to mitigate this deficiency, several variants of the Hausdorffmetric have been designed in the MIL literature, such as the maximal-, the minimaland the ranked Hausdorffmetrics, with successful applications in MIC (Wang and Zucker, 2000) and multi-instance outlier detection (Wu et al., 2010); and the average Hausdorffmetric (Zhang and Zhou, 2009) and contextual Hausdorff dissimilarity (Chen and Wu, 2012), which have been found useful in multi-instance clustering. Unfortunately, these methods lack any theoretical guarantee when applied in specific learning problems. Functional data analysis techniques: Finally, the distribution regression task might also be interpreted as a functional data analysis problem (Ramsay and Silverman, 2002, 2005; M uller, 2005), by considering the probability measures xi as functions. This is a highly non-standard setup, however, since these functions (xi) are defined on σ-algebras and are non-negative, σ-additive. 6. Conclusion We have established a learning theory of distribution regression, where the inputs are probability measures on separable, topological domains endowed with reproducing kernels, and the outputs are elements of a separable Hilbert space. We studied a ridge regression scheme defined on embeddings of the input distributions to a reproducing kernel Hilbert space, which has a simple analytical solution, as well as theoretically sound, efficient methods for approximation (Zhang et al., 2015; Richt arik and Tak a c, 2016; Alaoui and Mahoney, 2015; Yang et al., 2016; Rudi et al., 2015). We derived explicit bounds on the excess risk as a function of the number of samples and problem difficulty. We tackled both the well- 17. Often these metrics are only semi-metrics, as they do not satisfy the triangle inequality. Learning Theory for Distribution Regression specified case (when the regression function belongs to the assumed RKHS modelling class), and the more general misspecified setup. As a special case of our results, we proved the consistency of regression for set kernels (Haussler, 1999; G artner et al., 2002), which was a 17-year-old open problem, and for a recent kernel family (Christmann and Steinwart, 2010), which we have expanded upon (Table 1). We proved an exact computational-statistical efficiency trade-offfor the MERR estimator: in the well-specified setting, we showed how to choose the bag size in the two-stage sampled setup to match the one-stage sampled minimax optimal rate (Caponnetto and De Vito, 2007); and in the misspecified setting, our rates approximate closely an asymptotically optimal estimator imposing stricter eigenvalue decay conditions (Steinwart et al., 2009). Several exciting open questions remain, including whether improved/optimal rates can be derived in the misspecified case, whether we can obtain consistency guarantees for non-point estimates, and how to handle non-ridge extensions. Finally, we note that although the primary focus of the current paper was theoretical, we have applied the MERR method (Szab o et al., 2015, Section A.2) to supervised entropy learning and aerosol prediction based on multispectral satellite images.18 In future work, we will address applications with vector-valued outputs. We provide proofs for our results detailed in Section 4: Section 7.1 (resp. Section 7.2) focuses on the well-specified case (resp. misspecified setting). The used lemmas are enlisted in Section 7.3. 7.1 Proofs of the Well-specified Case We give proof details concerning the excess risk in the well-specified case (Theorem 2). 7.1.1 Proof of the bound on gˆz gz 2 H By (13), (14) we get gˆz gz = 1 l Pl i=1 Kµˆxi Kµxi yi; hence by applying the H older property of K( ), the boundedness of yi ( yi Y C) and (24), we obtain gˆz gz 2 H 1 Kµˆxi Kµxi yi 2 H 1 Kµˆxi Kµxi 2 L(Y,H) yi 2 Y i=1 yi 2 Y µˆxi µxi 2h H L2C2 (1 + α) 2Bk 2h = L2C2 (1 + α)2h (2Bk)h with probability at least 1 le α, based on a union bound. 18. For code, see https://bitbucket.org/szzoli/ite/. Szab o et al. 7.1.2 Proof of the bound on Tx Tˆx 2 L(H) Using the definition of Tx and Tˆx, and exploiting (with L(H)) that in a normed space19 (N, ), fi N, (i = 1, . . . , n) i=1 fi 2 n Xn i=1 fi 2 , (35) Tx Tˆx 2 L(H) 1 L(H) . (36) To upper bound Tµxi Tµˆxi 2 L(H), let us see how Tµu = Kµa K µa acts. The existence of an E 0 constant satisfying (Tµu Tµv)(f) H E f H implies Tµu Tµv L(H) E. We continue with the l.h.s. of this equation using Eq. (35): (Tµu Tµv)(f) 2 H = Kµu K µu(f) Kµv K µv(f) 2 H = Kµu K µu(f) K µv(f) + (Kµu Kµv) K µv(f) 2 H 2 h Kµu K µu(f) K µv(f) 2 H + (Kµu Kµv) K µv(f) 2 H By Eq. (45) and the H older continuity of K( ), one arrives at Kµu K µu(f) K µv(f) 2 H Kµu 2 L(Y,H) K µu(f) K µv(f) 2 Y Kµu 2 L(Y,H) K µu K µv 2 L(H,Y ) f 2 H = Kµu 2 L(Y,H) (Kµu Kµv) 2 L(H,Y ) f 2 H = Kµu 2 L(Y,H) Kµu Kµv 2 L(Y,H) f 2 H BKL2 µu µv 2h H f 2 H , (Kµu Kµv) K µv(f) 2 H Kµu Kµv 2 L(Y,H) K µv(f) 2 Y Kµu Kµv 2 L(Y,H) K µv 2 L(H,Y ) f 2 H BKL2 µu µv 2h H f 2 H . Hence (Tµu Tµv)(f) 2 H 4BKL2 µu µv 2h H f 2 H E2 = 4BKL2 µu µv 2h H . Exploiting this property in (36) with Eq. (24) we arrive to the bound Tx Tˆx 2 L(H) 4BKL2 i=1 µxi µˆxi 2h H 4BKL2 (1 + α)2h (2Bk)h = (1 + α)2h 2h+2(Bk)h BKL2 7.1.3 Proof: final union bound in Theorem 2 Until now, we obtained that if (i) the sample number N satisfies Eq. (25), (ii) (24) holds (which has probability at least 1 le α = 1 e [α log(l)] = 1 e δ applying a union bound 19. Eq. (35) holds since 2 is convex function, thus 1 n Pn i=1 fi 2 1 n Pn i=1 fi 2. Learning Theory for Distribution Regression argument; α = log(l) + δ), and (iii) Θ(λ, z) 1 2 is fulfilled [see Eq. (22)], then L2C2 (1 + α)2h (2Bk)h Nh + (1 + α)2h 2h+2(Bk)h BKL2 l2λ + Σ2N(λ) l2 + BKA(λ) + B(λ) + fρ 2 H = 4L2 (1 + α)2h (2Bk)h l2λ + Σ2N(λ) l2 + BKA(λ) + B(λ) + fρ 2 H By taking into account Caponnetto and De Vito (2007) s bounds for S1 and S2, S1 32 log2 6 η h BKM2 l2λ + Σ2N(λ) l i , S2 8 log2 6 η h 4B2 KB(λ) l2λ + BKA(λ) lλ i , plugging all the expressions to (21), we obtain Theorem 2 with a union bound. 7.2 Proofs of the Misspecified Case We present the proof details concerning the excess risk in the misspecified case (Theorem 7). 7.2.1 Proof of the bound on S 1 + S0 without P(b, c) The upper bounds on S 1 and S0 [which are defined in Eqs. (15), (16)] remain valid without modification provided that (i) Θ(λ, z) = (T Tx)(T + λ) 1 L(H) (T Tx)(T + λ) 1 L2(H) 1 2, where we used Eq. (1), (ii) Eq. (24) is satisfied (which has probability 1 le α) and (iii) Eq. (25) holds. Our goal below is to guarantee the Θ(λ, z) 1 2 condition with high probability without assuming that the prior belongs to P(b, c). Requirement Θ(λ, z) 1 2: Let us define ξi = Tµxi(T +λ) 1 L2(H), (i = 1, . . . , l). With this choice we get E[ξi] = T(T + λ) 1, (T Tx)(T + λ) 1 = E[ξi] 1 l Pl i=1 ξi and ξi L2(H) Tµxi L2(H) (T + λ) 1 L(H) BK/λ E ξi 2 L2(H) (BK)2/λ2, where we made use of (2), the Tµxi L2(H) BK identity following from the boundedness of K (Caponnetto and De Vito, 2007, page 341, Eq. (13)), and the spectral theorem. Consequently, by the Bernstein s inequality (Lemma 7.3.1 with K = L2(H), B = 2BK/λ, σ = BK/λ) we obtain that for η (0, 1) P (T Tx)(T + λ) 1 L2(H) 2 2BK Thus, for Θ(λ, z) 1 2 with probability 1 η it is sufficient to have Szab o et al. Under these conditions, we arrived at the upper bound 4L2C2 (1 + α)2h (2Bk)h = 2LC(1 + α)h(2Bk) h 2 where as opposed to Section 7.1.3 and Eq. (23) we used a slightly cruder fλ z 2 H C2 λ bound; it holds without the P(b, c) assumption by the definition of fλ z and the boundedness of y since λ fλ z 2 H 1 l Pl i=1 yi 2 Y C2. Remark: Notice that the price we pay for not assuming that the prior belongs to the P(b, c) class (b > 1) is a slightly tighter 1 λ2 l constraint [Eq. (38)] instead of 1 Eq. (19), and a somewhat looser fλ z 2 H bound. 7.2.2 Proof of the decomposition of T fλ z fλ H Using the analytical formula of fλ z [see Eq. (13)] and that of fλ [see Eq.(17)] fλ = (SKS K + λI) 1SKfρ = (T + λI) 1SKfρ (39) one gets (T + λI)fλ = SKfρ λfλ = SKfρ Tfλ and fλ z fλ = (Tx + λI) 1gz fλ = (Tx + λI) 1gz (Tx + λI) 1(Tx + λI)fλ = (Tx + λI) 1 gz (Tx + λI)fλ = (Tx + λI) 1 gz Txfλ λfλ = (Tx + λI) 1 gz Txfλ SKfρ + Tfλ = (Tx + λI) 1 (gz SKfρ) + (Tx + λI) 1(T Tx)fλ = (Tx + λI) 1 (gz SKfρ) + (Tx + λI) 1(T Tx)(T + λI) 1SKfρ. (40) Let us rewrite (T +λI) 1 by the (A+UV ) 1 = A 1 A 1U I + V A 1U 1 V A 1 operator Woodbury formula (Ding and Zhou, 2008, Theorem 2.1, page 724) (T + λI) 1 = (λI + SKS K) 1 = (λ 1I) (λ 1I)SK I + S K(λ 1I)SK 1 S K(λ 1I) = (λ 1I) λ 1SK(λI + T) 1S K. By the derived expression for (T +λI) 1, we get (T +λI) 1SKfρ = λ 1SKfρ λ 1SK(λI + T) 1S KSKfρ = λ 1SK fρ ( T + λI) 1S KSKfρ . Plugging this result to Eq. (40), introducing the gρ = SKfρ notation, using the triangle inequality we get T fλ z fλ H = T(Tx + λI) 1 n (gz SKfρ) + (T Tx)λ 1SK h fρ ( T + λI) 1S KSKfρ io H T(Tx + λI) 1 L(H) gz gρ H + T Tx L(H) λ 1 SK fρ ( T + λI) 1S KSKfρ H Learning Theory for Distribution Regression 7.2.3 Proof of the bound on gz gρ H As is known gz = 1 l Pl i=1 Kµxiyi [see Eq. (13)] and gρ = R X Kµxfρ(µx)dρX(µx) (Caponnetto and De Vito, 2007, Eq. (23), page 344). Let ξi = Kµxiyi H (i = 1, . . . , l). In this case E[ξi] = gρ, gρ gz = E[ξi] 1 l Pl i=1 ξi, and ξi 2 H = Kµxiyi 2 H Kµxi 2 L(Y,H) yi 2 Y BKC2 ξi H C BK E ξi 2 H C2BK using the boundedness of kernel K ( Kµxi 2 L(Y,H) BK) and the boundedness of output y ( |y Y C). Applying the Bern- stein inequality (see Lemma 7.3.1 with K = H, B = 2C BK, σ = C BK) one gets that for any η (0, 1) BK l + C BK 7.2.4 Proof of the bound on T Tx L(H) Let ξi = Tµxi L2(H) (i = 1, . . . , l), then E[ξi] = T, T Tx = T 1 l Pl i=1 Tµxi, ξi L2(H) = Tµxi L2(H) BK, E h ξi 2 L2(H) i B2 K. Applying the T Tx L(H) T Tx L2(H) relation [see Eq. (1)] and the Bernstein inequality (see Lemma 7.3.1 with K = L2(H), B = 2BK, σ = BK), we obtain that for any η (0, 1) P T Tx L(H) 2 2BK 7.2.5 Proof of the decomposition of SK fρ ( T + λI) 1S KSKfρ 2 H Since SKa 2 H = SKa, SKa H = S KSKa, a ρ = Ta, a ρ ( a L2 ρX) by the definition of the adjoint operator and T = S KSK [see Eq. (6)], we can rewrite the target term as SK h fρ ( T + λI) 1S KSKfρ i 2 = D T h fρ ( T + λI) 1S KSKfρ i , fρ ( T + λI) 1S KSKfρ E T h fρ ( T + λI) 1S KSKfρ i ρ fρ ( T + λI) 1S KSKfρ ρ , where the CBS (Cauchy-Bunyakovsky-Schwarz) inequality was applied. Since (S KSK + λI)S K = S K(SKS K + λI) S K(SKS K + λI) 1 = (S KSK + λI) 1S K S K(SKS K + λI) 1SK = (S KSK + λI) 1S KSK (41) S K(T + λI) 1SK = ( T + λI) 1 T (42) using Eq. (41) and the analytical expression for fλ [see Eq. (39)] we have ( T + λI) 1 Tfρ = ( T + λI) 1S KSKfρ = (S KSK + λI) 1 S KSKfρ = S K(SKS K + λI) 1SKfρ = S Kfλ (43) and SK fρ ( T + λI) 1S KSKfρ 2 H T fρ ( T + λI) 1S KSKfρ ρ S Kfλ fρ ρ. Szab o et al. 7.2.6 Proof of the bound on S Kfλ fρ ρ Let us apply (i) the Af f = Af f q + q = (A I)(f q ) + Aq q relation with A = ( T + λI) 1 T, f = fρ and q = S Kq, where q H is an arbitrary element from H, (ii) Eq. (43) and (iii) the triangle inequality to arrive at S Kfλ fρ ρ = ( T + λI) 1 Tfρ fρ ρ = ( T + λI) 1 T I (fρ S Kq) + ( T + λI) 1 TS Kq S Kq ρ ( T + λI) 1 T I (fρ S Kq) ρ + ( T + λI) 1 TS Kq S Kq ρ. Below we give upper bounds on these two terms. First, notice that µx X 7 K(µx, µx) L(Y ) BK. This boundedness with the strong continuity of K( ) imply (Carmeli et al., 2006, Proposition 12) that H C(X, Y ), i.e., K is a Mercer kernel. Since Kµx is a Hilbert-Schmidt operator for all µx X [see Eq. (11)], it is also a compact operator ( µx X). The compactness of Kµx-s with the bounded and Mercer property of K guarantees the boundedness of S K and that T is a compact, positive, self-adjoint operator (Carmeli et al., 2010, Proposition 3). Bound on [( T + λI) 1 T I](fρ S Kq) ρ: Since T is a compact positive self-adjoint operator, by the spectral theorem (Steinwart and Christmann, 2008, Theorem 4.27, page 127) there exist an (ui)i I countable ONB in cl Im( T) , and a1 a2 . . . > 0 such that Tf = P i I ai f, ui ρ ui ( f L2 ρX) and let (vj)j J (J is also countable by the separability20 of L2 ρX) an ONB in Ker( T ) = Ker( T); L2 ρX = cl Im( T) Ker( T). Thus, ( T + λI) 1 T I (fρ S Kq) 2 ρ = X ai ai + λ 1 2 fρ S Kq, ui 2 ρ + X j J fρ S Kq, vj 2 ρ i I fρ S Kq, ui 2 ρ + X j J fρ S Kq, vj 2 ρ = fρ S Kq 2 ρ exploiting the Parseval s identity and that λi λi+λ 1 2 1. Bound on ( T + λI) 1 TS Kq S Kq ρ: By using Eq. (42), Eq. (32), and Lemma 7.3.2 (M = T = SKS K, K = H, f = q, a = 1 2), the target quantity can be bounded as ( T + λI) 1 TS Kq S Kq ρ = S K(T + λI) 1SKS Kq S Kq ρ = T (T + λI) 1SKS Kq q H T (T + λI) 1Tq q H max 1, T L(H) λ 1 2 q H . Making use of the two derived bounds, we get S Kfλ fρ ρ fρ S Kq ρ + max 1, T L(H) λ 1 2 q H. 20. L2 ρX = L2(X, B(H)|X , ρX; Y ) is isomorphic to L2(X, B(H)|X , ρX; R) Y , where is the tensor product of Hilbert spaces. The separability follows from that of Y and L2(X, B(H)|X , ρX; R); the latter holds (Cohn, 2013, Proposition 3.4.5) since B(H)|X is countably generated since X H is separable. Learning Theory for Distribution Regression 7.3 Supplementary Lemmas In this section, we list two lemmas used in the proofs. 7.3.1 Bernstein s inequality (Caponnetto and De Vito, 2007, Prop. 2, p. 345) Let ξi (i = 1, . . . , l) be i.i.d. realizations of a random variable on a (Ω, A, P) probability space with values in a separable Hilbert space K. If there exist B > 0, σ > 0 constants such that ξ(ω) K B 2 a.s., E h ξ 2 K i σ2, then for all l 1 and η (0, 1) we have i=1 ξi E[ξ1] K 2 B 7.3.2 Lemma on bounded, self-adjoint compact operators; Sriperumbudur et al. (2014, Proposition A.2, page 39) Let M be a bounded, self-adjoint compact operator on a separable Hilbert space K. Let a 0, λ > 0, and s 0. Let f K such that f Im (Ms). If s + a > 0, then Ma (M + λI) 1Mf f K max 1, M s+a 1 L(K) λmin(1,s+a) M sf K . Note: specifically for s = 0 we have Im (Ms) = Im (I) = K, in other words, there is no additional range space constraint. 8. Discussion of Our Assumptions We give a short insight into the consequences of our assumptions (detailed in Section 3) and present some concrete examples. Well-definedness of ρ: The boundedness and continuity of k imply the measurability of µ : (M+ 1 (X), B(τw)) (H, B(H)). Let τ denote the open sets on H = H(k), τ|X = {A X : A τ} the subspace topology on X, and B(H)|X = {A X : A B(H)} the subspace σ-algebra on X. By noting (Schwartz, 1998, Corollary 5.2.13) that B (τ|X) = B(H)|X = {A B(H) : A X} B(H), the H-measurability of µ guarantees the measurability of µ : (M+ 1 (X), B(τw)) (X, B(H)|X), and hence the well-definedness of ρ, the measure induced by M on X Y ; for further details see (Szab o et al., 2015, Section A.1.1).21 Separability of X: separability of X and the continuity of k implies the separability of H = H(k) (Steinwart and Christmann, 2008, Lemma 4.33, page 130). Also, since X H, X is separable. Finiteness of Bk: If X is compact, then the continuity of k implies Bk < . Finiteness of BK, compact metricness of X: Let X be a compact metric space. In this case M+ 1 (X) is also compact metric (Parthasarathy, 1967, Theorem 6.4, page 55). 21. Note that the referred proof also holds for separable Hilbert Y , and by the simplified reasoning above the original X B(H) condition could be avoided. Szab o et al. Hence if µ : (M+ 1 (X), τw) H(k) is continuous22 (not just measurable), then X is compact metric and thus by the H older property of K( ), it is continuous implying that BK < . K properties: It is known (Caponnetto and De Vito, 2007, page 339-340) that K(µa, µb) = K µa Kµb, ( µa, µb X) (44) Kµa L(Y,H) = K µa L(H,Y ) p BK, ( µa X). (45) Remark: In terms of Eq. (44), the Eq. (11) assumption means that the {K(µa, µa)}µa X operators are trace class, specifically they are compact operators. Separability of H: The separability of X and the continuity of K imply the separability of H. Indeed, since µa 7 Kµa is H older continuous w.r.t. the Hilbert-Schmidt norm it is also continuous. As a result it is continuous w.r.t. the operator norm, and thus also w.r.t. the strong topology. Using this property with the finiteness of BK the separability of H follows (Carmeli et al., 2006, Proposition 5.1, Corollary 5.2). Our assumptions imply Caponnetto and De Vito (2007) s conditions (not considering the P(b, c) prior requirement). Indeed 1. Y is a separable Hilbert space by assumption; the same property also holds for H as we have seen. 2. The measurability of (µx, µt) 7 Kµxw, Kµtv H for w, v Y is guaranteed by the continuity of K( ) w.r.t. the strong topology. 3. We have R X Y y 2 Y dρX(µx, y) R X Y C2dρX(µx, y) = C2 < due to the boundedness of y, and hence Σ > 0, M > 0 such that for ρX-almost µx X Z Y y fρ(µx) m Y dρ(y|µx) m!Σ2Mm 2 2 ( m 2). (46) Indeed, by (Caponnetto and De Vito, 2007, Eq. (33)) the Bernstein condition (46) holds if y fρ(µx) Y M Y y fρ(µx) 2 Y dρ(y|µx) Σ2. In our case using the boundedness of y, the regression function is also bounded and the same holds for y fρ(µx) Y by the triangle inequality: y fρ(µx) Y C + fρ H BK; thus, M = 2(C + fρ H BK), Σ = M 2 is a suitable choice. 4. The Polishness of X Y was used by Caponnetto and De Vito (2007) to assure the existence of ρ(y|µa); we guaranteed this existence under somewhat milder conditions (see footnote 7). Real-valued outputs: We now consider the specific case of Y = R, when the following simplifications and results hold. By noting that in this case Tr(K µa Kµa) = K(µa, µa), Eq. (11) simplifies to the boundedness of kernel K in the traditional sense K(µa, µa) BK ( µa X). (47) 22. For example, if k is universal, then µ metrizes the weak topology τw (Sriperumbudur et al., 2010, Theorem 23, page 1552), hence µ is continuous. Learning Theory for Distribution Regression KG Ke KC Kt Ki e µa µb 2 H 2θ2 e µa µb H 2θ2 1 + µa µb 2 H /θ2 1 1 + µa µb θ H 1 µa µb 2 H + θ2 1 h = 1 h = 1 2 h = 1 h = θ 2 (θ 2) h = 1 Table 1: Nonlinear kernels on mean embedded distributions: K = K(µa, µb); θ > 0. For the H older continuity of ΨK, we assume that X is a compact metric space and µ is continuous. Eq. (12) reduces to the H older continuity of the canonical feature map ΨK(µc) := K( , µc) : X H, in other words L > 0, h (0, 1] such that ΨK(µa) ΨK(µb) H L µa µb h H , (µa, µb) X X. In case of a linear kernel, K(µa, µb) = µa, µb H, (µa, µb X), the H older continuity of ΨK holds with L = 1, h = 1, and BK = Bk is a suitable choice. Evaluating the kernel K at the empirical embeddings µˆxi = R X k( , u)dˆxi(u) = 1 N PN n=1 k( , xi,n) H yields the standard set kernel K(µˆxi, µˆxj) = µˆxi, µˆxj n=1 k( , xi,n), 1 m=1 k( , xj,m) n,m=1 k(xi,n, xj,m) by the bilinearity of , H and the reproducing property of k. Remark: One can define many nonlinear kernels (see Table 1) on mean embedded distributions. These kernels are the natural extensions to distributions of the Gaussian (Christmann and Steinwart, 2010), exponential, Cauchy, generalized t-student and inverse multiquadric kernels. If X is a compact metric space and µ is continuous, then the ΨK canonical feature maps, associated to K-s in Table 1, can be shown to satisfy our H older continuity requirement [Eq. (12)]; for details, see (Szab o et al., 2015, Section A.1.5-A.1.6). Acknowledgments We would like to thank the anonymous reviewers for their highly valuable, constructive suggestions to improve the manuscript. This work was supported by the Gatsby Charitable Foundation, NSF grant 1247658, and DOE grant DE-SC001114. A part of the work was carried out while Bharath K. Sriperumbudur was a research fellow in the Statistical Laboratory, Department of Pure Mathematics and Mathematical Statistics at the University of Cambridge, UK. Ahmed Alaoui and Michael W. Mahoney. Fast randomized kernel ridge regression with statistical guarantees. In Advances in Neural Information Processing Systems (NIPS), pages 775 783, 2015. Charalambos D. Aliprantis and Kim C. Border. Infinite Dimensional Analysis: A Hitchhiker s Guide. Springer, 2006. Szab o et al. Yasemin Altun and Alexander Smola. Unifying divergence minimization and statistical inference via convex duality. In Conference on Learning Theory (COLT), pages 139 153, 2006. Mauricio A. Alvarez, Lorenzo Rosasco, and Neil D. Lawrence. Kernels for vector-valued functions: A review. Foundations and Trends in Machine Learning, 4:195 266, 2011. Jaume Amores. Multiple instance classification: Review, taxonomy and comparative study. Artificial Intelligence, 201:81 105, 2013. Peter Auer. Approximating hyper-rectangles: Learning and pseudorandom sets. Journal of Computer and System Sciences, 57:376 388, 1998. Boris Babenko. Multiple instance learning: Algorithms and applications. Technical report, Department of Computer Science and Engineering, University of California, San Diego, 2004. (http://cms.brookes.ac.uk/research/visiongroup/talks/rg_dec_11_ 09/bbabenko_re.pdf). Boris Babenko, Nakul Verma, Piotr Doll ar, and Serge Belongie. Multiple instance learning with manifold bags. In International Conference on Machine Learning (ICML), pages 81 88, 2011. Charles Bergeron, Jed Zaretzki, Curt Breneman, and Kristin P. Bennett. Multiple instance ranking. In International Conference on Machine Learning (ICML), pages 48 55, 2008. Charles Bergeron, Gregory Moore, Jed Zaretzki, Curt M. Breneman, and Kristin P. Bennett. Fast bundle algorithm for multiple-instance learning. IEEE Transactions on Pattern Analysis and Machine Intelligence, 34:1068 1079, 2012. Alain Berlinet and Christine Thomas-Agnan. Reproducing Kernel Hilbert Spaces in Probability and Statistics. Kluwer, 2004. Avrim Blum and Adam Kalai. A note on learning from multiple-instance examples. Machine Learning, 30:23 29, 1998. Hanen Borchani, Gherardo Varando, Concha Bielza, and Pedro Larranaga. A survey on multi-output regression. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery, 5:216 233, 2015. C eline Brouard, Florence d Alch e Buc, and Marie Szafranski. Semi-supervised penalized output kernel regression for link prediction. In International Conference on Machine Learning (ICML), pages 593 600, 2011. Andrea Caponnetto. Optimal rates for regularization operators in learning theory. Technical report, Massachusetts Institute of Technology, 2006. (http://www6.cityu.edu.hk/ma/ doc/people/caponnettoa/regop_TR%28TR11%29.pdf). Andrea Caponnetto and Ernesto De Vito. Optimal rates for regularized least-squares algorithm. Foundations of Computational Mathematics, 7:331 368, 2007. Learning Theory for Distribution Regression Claudio Carmeli, Ernesto De Vito, and Alessandro Toigo. Vector valued reproducing kernel Hilbert spaces of integrable functions and Mercer theorem. Analysis and Applications, 4: 377 408, 2006. Claudio Carmeli, Ernesto De Vito, Alessandro Toigo, and Veronica Umanit a. Vector valued reproducing kernel Hilbert spaces and universality. Analysis and Applications, 8:19 61, 2010. Kevin M. Carter, Raviv Raich, William G. Finn, and Alfred O. Hero. Information-geometric dimensionality reduction. IEEE Signal Processing Magazine, 28:89 99, 2011. Jing Chai, Hongtao Chen, Lixia Huang, and Fanhua Shang. Maximum margin multipleinstance feature weighting. Pattern Recognition, 47:2091 2103, 2014a. Jing Chai, Xinghao Ding, Hongtao Chen, and Tingyu Li. Multiple-instance discriminant analysis. Pattern Recognition, 47:2517 2531, 2014b. Ying Chen and Ou Wu. Contextual Hausdorffdissimilarity for multi-instance clustering. In International Conference on Fuzzy Systems and Knowledge Discovery (FSKD), pages 870 873, 2012. Andreas Christmann and Ingo Steinwart. Universal kernels on non-standard input spaces. In Advances in Neural Information Processing Systems (NIPS), pages 406 414, 2010. Donald L. Cohn. Measure Theory: Second Edition. Birkh auser Basel, 2013. Marco Cuturi, Kenji Fukumizu, and Jean-Philippe Vert. Semigroup kernels on measures. Journal of Machine Learning Research, 6:11691198, 2005. Ernesto de Vito, Lorenzo Rosasco, and Andrea Caponnetto. Discretization error analysis for Tikhonov regularization. Analysis and Applications, 4:81 99, 2006. Thomas G. Dietterich, Richard H. Lathrop, and Tom as Lozano-P erez. Solving the multiple instance problem with axis-parallel rectangles. Artificial Intelligence, 89:31 71, 1997. Jiu Ding and Aihui Zhou. A spectrum theorem for perturbed bounded linear operators. Applied Mathematics and Computation, 201:723 728, 2008. Daniel R. Dooly, Qi Zhang, Sally A. Goldman, and Robert A. Amar. Multiple-instance learning of real-valued data. Journal of Machine Learning Research, 3:651 678, 2002. Gerald Edgar. Measure, Topology and Fractal Geometry. Springer-Verlag, 1995. Thomas Eiter and Heikki Mannila. Distance measures for point sets and their computation. Acta Informatica, 34:109 133, 1997. Theodoros Evgeniou, Charles A. Micchelli, and Massimiliano Pontil. Learning multiple tasks with kernel methods. Journal of Machine Learning Research, 6:615 637, 2005. Seth Flaxman, Yu-Xiang Wang, and Alex Smola. Who supported Obama in 2012? ecological inference through distribution regression. In ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD), pages 289 298, 2015. Szab o et al. James Foulds and Eibe Frank. A review of multi-instance learning assumptions. The Knowledge Engineering Review, 25:1 25, 2010. Kenji Fukumizu, Francis Bach, and Michael Jordan. Dimensionality reduction for supervised learning with reproducing kernel Hilbert spaces. Journal of Machine Learning Research, 5:73 99, 2004. Thomas G artner, Peter A. Flach, Adam Kowalczyk, and Alexander Smola. Multi-instance kernels. In International Conference on Machine Learning (ICML), pages 179 186, 2002. Arthur Gretton, Karsten M. Borgwardt, Malte J. Rasch, Bernhard Sch olkopf, and Alexander Smola. A kernel two-sample test. Journal of Machine Learning Research, 13:723 773, 2012. L aszl o Gy orfi, Michael Kohler, Adam Krzyzak, and Harro Walk. A Distribution-Free Theory of Nonparametric Regression. Springer, New-york, 2002. David Haussler. Convolution kernels on discrete structures. Technical report, Department of Computer Science, University of California at Santa Cruz, 1999. (http://cbse.soe. ucsc.edu/sites/default/files/convolutions.pdf). Matthias Hein and Olivier Bousquet. Hilbertian metrics and positive definite kernels on probability measures. In International Conference on Artificial Intelligence and Statistics (AISTATS), pages 136 143, 2005. Yang Hu, Mingjing Li, and Nenghai Yu. Multiple-instance ranking: Learning to rank images for image retrieval. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 1 8, 2008. Tony Jebara, Risi Kondor, and Andrew Howard. Probability product kernels. Journal of Machine Learning Research, 5:819 844, 2004. Hachem Kadri, Emmanuel Duflos, Philippe Preux, St ephane Canu, and Manuel Davy. Nonlinear functional regression: a functional RKHS approach. International Conference on Artificial Intelligence and Statistics (AISTATS; JMLR W&CP), 9:374 380, 2010. Hachem Kadri, Alain Rakotomamonjy, Francis Bach, and Philippe Preux. Multiple operator-valued kernel learning. In Advances in Neural Information Processing Systems (NIPS), pages 2429 2437, 2012. Hachem Kadri, Mohammad Ghavamzadeh, and Philippe Preux. A generalized kernel approach to structured output learning. International Conference on Machine Learning (ICML; JMLR W&CP), 28:471 479, 2013. Hachem Kadri, Emmanuel Duflos, Philippe Preux, St ephane Canu, Alain Rakotomamonjy, and Julien Audiffren. Operator-valued kernels for learning from functional response data. Journal of Machine Learning Research, 17:1 54, 2016. Risi Kondor and Tony Jebara. A kernel between sets of vectors. In International Conference on Machine Learning (ICML), pages 361 368, 2003. Learning Theory for Distribution Regression Samory Kpotufe. k-NN regression adapts to local intrinsic dimension. Technical report, Max Planck Institute for Intelligent Systems, 2011. (http://arxiv.org/abs/1110.4300). James T. Kwok and Pak-Ming Cheung. Marginalized multi-instance kernels. In International Joint Conferences on Artificial Intelligence (IJCAI), pages 901 906, 2007. Philip M. Long and Lei Tan. PAC learning of axis-aligned rectangles with respect to product distributions from multiple-instance examples. Machine Learning, 30:7 21, 1998. David Lopez-Paz, Krikamol Muandet, Bernhard Sch olkopf, and Iliya Tolstikhin. Towards a learning theory of cause-effect inference. International Conference on Machine Learning (ICML; JMLR W&CP), 37:1452 1461, 2015. Andr e F. T. Martins, Noah A. Smith, Eric P. Xing, Pedro M. Q. Aguiar, and M ario A. T. Figueiredo. Nonextensive information theoretical kernels on measures. Journal of Machine Learning Research, 10:935 975, 2009. Shahar Mendelson and Joseph Neeman. Regularization in kernel learning. The Annals of Statistics, 38:526 565, 2010. Charles A. Micchelli and Massimiliano Pontil. On learning vector-valued functions. Neural Computation, 17:177 204, 2005. Krikamol Muandet, Kenji Fukumizu, Francesco Dinuzzo, and Bernhard Sch olkopf. Learning from distributions via support measure machines. In Advances in Neural Information Processing Systems (NIPS), pages 10 18, 2012. Hans-Georg M uller. Functional modelling and classification of longitudinal data. Scandinavian Journal of Statistics, 32:223 240, 2005. Frank Nielsen and Richard Nock. A closed-form expression for the Sharma-Mittal entropy of exponential families. Journal of Physics A: Mathematical and Theoretical, 45:032003, 2012. Junier Oliva, Barnab as P oczos, and JeffSchneider. Distribution to distribution regression. International Conference on Machine Learning (ICML; JMLR W&CP), 28:1049 1057, 2013. Junier Oliva, William Neiswanger, Barnab as P oczos, Eric Xing, Hy Trac, Shirley Ho, and JeffSchneider. Fast function to function regression. International Conference on Artificial Intelligence and Statistics (AISTATS; JMLR W&CP), 38:717 725, 2015. Junier B. Oliva, Willie Neiswanger, Barnab as P oczos, JeffSchneider, and Eric Xing. Fast distribution to real regression. International Conference on Artificial Intelligence and Statistics (AISTATS; JMLR W&CP), 33:706 714, 2014. Kalyanapuram R. Parthasarathy. Probability Measures on Metric Spaces. Academic Press, 1967. George Pedrick. Theory of reproducing kernels for Hilbert spaces of vector valued functions. Technical report, 1957. Szab o et al. Wei Ping, Ye Xu, Kexin Ren, Chi-Hung Chi, and Shen Furao. Non-I.I.D. multi-instance dimensionality reduction by learning a maximum bag margin subspace. In AAAI Conference on Artificial Intelligence, pages 551 556, 2010. Barnab as P oczos, Liang Xiong, and JeffSchneider. Nonparametric divergence estimation with applications to machine learning on distributions. In Uncertainty in Artificial Intelligence (UAI), pages 599 608, 2011. Barnab as P oczos, Liang Xiong, Dougal Sutherland, and JeffSchneider. Support distribution machines. Technical report, Carnegie Mellon University, 2012. (http://arxiv.org/abs/ 1202.0302). Barnab as P oczos, Alessandro Rinaldo, Aarti Singh, and Larry Wasserman. Distribution-free distribution regression. International Conference on Artificial Intelligence and Statistics (AISTATS; JMLR W&CP), 31:507 515, 2013. Jan Ramon and Maurice Bruynooghe. A polynomial time computable metric between point sets. Acta Informatica, 37:765 780, 2001. James O. Ramsay and Bernard W. Silverman. Applied Functional Data Analysis. Springer Verlag, New York, 2002. James O. Ramsay and Bernard W. Silverman. Functional Data Analysis. Springer Verlag, New York, 2005. Soumya Ray and David Page. Multiple instance regression. In International Conference on Machine Learning (ICML), pages 425 432, 2001. Vikas C. Raykar, Balaji Krishnapuram, Jinbo Bi, Murat Dundar, and R. Bharat Rao. Bayesian multiple instance learning: Automatic feature selection and inductive transfer. In International Conference on Machine Learning (ICML), pages 808 815, 2008. Peter Richt arik and Martin Tak a c. Distributed coordinate descent method for learning with big data. Journal of Machine Learning Research, 17:1 25, 2016. R. Tyrrell Rockafellar and Roger J-B Wets. Variational Analysis. Springer, 2008. Alessandro Rudi, Raffaello Camoriano, and Lorenzo Rosasco. Less is more: Nystr om computational regularization. In Advances in Neural Information Processing Systems (NIPS), pages 1648 1656, 2015. Sivan Sabato and Naftali Tishby. Multi-instance learning with any hypothesis class. Journal of Machine Learning Research, 13:2999 3039, 2012. Laurent Schwartz. Analyse III, Calcul Integr el. Hermann, 1998. Steve Smale and Ding-Xuan Zhou. Estimating the approximation error in learning theory. Analysis and Applications, 1:17 41, 2003. Learning Theory for Distribution Regression Bharath Sriperumbudur, Arthur Gretton, Kenji Fukumizu, Gert Lanckriet, and Bernhard Sch olkopf. Hilbert space embeddings and metrics on probability measures. Journal of Machine Learning Research, 11:1517 1561, 2010. Bharath K. Sriperumbudur, Kenji Fukumizu, and Gert R. G. Lanckriet. Universality, characteristic kernels and RKHS embedding of measures. Journal of Machine Learning Research, 12:2389 2410, 2011. Bharath K. Sriperumbudur, Kenji Fukumizu, Revant Kumar, Arthur Gretton, and Aapo Hyv arinen. Density estimation in infinite dimensional exponential families. Technical report, 2014. (http://arxiv.org/pdf/1312.3516). Ingo Steinwart and Andres Christmann. Support Vector Machines. Springer, 2008. Ingo Steinwart and Clint Scovel. Mercer s theorem on general domains: On the interaction between measures, kernels, and RKHSs. Constructive Approximation, 35:363 417, 2012. Ingo Steinwart, Don R. Hush, and Clint Scovel. Optimal rates for regularized least squares regression. In Conference on Learning Theory (COLT), 2009. Hongwei Sun and Qiang Wu. Application of integral operator for regularized least-square regression. Mathematical and Computer Modelling, 49:276 285, 2009a. Hongwei Sun and Qiang Wu. A note on application of integral operator in learning theory. Applied and Computational Harmonic Analysis, 26:416 421, 2009b. Xu Sun, Hisashi Kashima, and Naonori Ueda. Large-scale personalized human activity recognition using online multitask learning. IEEE Transactions on Knowledge and Data Engine, 25:2551 2563, 2013. Yu-Yin Sun, Michael K. Ng, and Zhi-Hua Zhou. Multi-instance dimensionality reduction. In AAAI Conference on Artificial Intelligence, pages 587 592, 2010. Dougal J. Sutherland, Junier B. Oliva, Barnab as P oczos, and JeffSchneider. Linear-time learning on distributions with approximate kernel embeddings. In AAAI Conference on Artifical Intelligence (AAAI), pages 2073 2079, 2016. Zolt an Szab o, Arthur Gretton, Barnab as P oczos, and Bharath Sriperumbudur. Two-stage sampled learning theory on distributions. In International Conference on Artificial Intelligence and Statistics (AISTATS), pages 948 957, 2015. Leslie Valiant. A theory of the learnable. Communications of the ACM, 27:1134 1142, 1984. Fei Wang, Tanveer Syeda-Mahmood, Baba C. Vemuri, David Beymer, and Anand Rangarajan. Closed-form Jensen-R enyi divergence for mixture of Gaussians and applications to group-wise shape registration. Medical Image Computing and Computer-Assisted Intervention, 12:648 655, 2009. Jun Wang and Jean-Daniel Zucker. Solving the multiple-instance problem: A lazy learning approach. In International Conference on Machine Learning (ICML), pages 1119 1126, 2000. Szab o et al. Larry Wasserman. All of Nonparametric Statistics. Springer, 2006. Ou Wu, Jun Gao, Weiming Hu, Bing Li, and Mingliang Zhu. Identifying multi-instance outliers. In SIAM International Conference on Data Mining (SDM), pages 430 441, 2010. Yun Yang, Mert Pilanci, and Martin J. Wainwright. Randomized sketches for kernels: Fast and optimal non-parametric regression. Annals of Statistics, 2016. (to appear; ar Xiv: http://arxiv.org/abs/1501.06195). Amelia Zafra, Mykola Pechenizkiy, and Sebasti an Ventura. Hy DR-MI: A hybrid algorithm to reduce dimensionality in multiple instance learning. Information Sciences, 222:282 301, 2013. Dan Zhang and Luo Si. Multiple instance transfer learing. In IEEE International Conference on Data Mining Workshops (ICDMW), pages 406 411, 2009. Dan Zhang, Fei Wang, Luo Si, and Tao Li. M3IC: Maximum margin multiple instance clustering. In International Joint Conferences on Artificial Intelligence (IJCAI), pages 1339 1344, 2009. Dan Zhang, Fei Wang, Luo Si, and Tao Li. Maximum margin multiple instance clustering with applications to image and text clustering. IEEE Transactions on Neural Networks, 22:739 751, 2011. Dan Zhang, Jingrui He, Luo Si, and Richard D. Lawrence. MILEAGE: Multiple Instance LEArning with Global Embedding. International Conference on Machine Learning (ICML; JMLR W&CP), 28:82 90, 2013. Min-Ling Zhang and Zhi-Hua Zhou. Multi-instance clustering with applications to multiinstance prediction. Applied Intelligence, 31:47 68, 2009. Yuchen Zhang, John C. Duchi, and Martin J. Wainwright. Divide and conquer kernel ridge regression: A distributed algorithm with minimax optimal rates. 16:3299 3340, 2015. Jiayu Zhou, Jun Liu, Vaibhav A. Narayan, and Jieping Ye. Modeling disease progression via multi-task learning. Neuro Image, 78:233 248, 2013. Shaohua Kevin Zhou and Rama Chellappa. From sample similarity to ensemble similarity: Probabilistic distance measures in reproducing kernel Hilbert space. IEEE Transactions on Pattern Analysis and Machine Intelligence, 28:917 929, 2006. Zhi-Hua Zhou. Multi-instance learning: A survey. Technical report, AI Lab, Department of Computer Science & Technology, Nanjing University, Nanjing, China, 2004. (http: //cs.nju.edu.cn/zhouzh/zhouzh.files/publication/techrep04.pdf). Zhi-Hua Zhou, Yu-Yin Sun, and Yu-Feng Li. Multi-instance learning by treating instances as non-I.I.D. samples. In International Conference on Machine Learning (ICML), pages 1249 1256, 2009.