# greedy_feature_construction__799536a4.pdf Greedy Feature Construction Dino Oglic dino.oglic@uni-bonn.de Institut für Informatik III Universität Bonn, Germany Thomas Gärtner thomas.gaertner@nottingham.ac.uk School of Computer Science The University of Nottingham, UK We present an effective method for supervised feature construction. The main goal of the approach is to construct a feature representation for which a set of linear hypotheses is of sufficient capacity large enough to contain a satisfactory solution to the considered problem and small enough to allow good generalization from a small number of training examples. We achieve this goal with a greedy procedure that constructs features by empirically fitting squared error residuals. The proposed constructive procedure is consistent and can output a rich set of features. The effectiveness of the approach is evaluated empirically by fitting a linear ridge regression model in the constructed feature space and our empirical results indicate a superior performance of our approach over competing methods. 1 Introduction Every supervised learning algorithm with the ability to generalize from training examples to unseen data points has some type of inductive bias [5]. The bias can be defined as a set of assumptions that together with the training data explain the predictions at unseen points [25]. In order to simplify theoretical analysis of learning algorithms, the inductive bias is often represented by a choice of a hypothesis space (e.g., the inductive bias of linear regression models is the assumption that the relationship between inputs and outputs is linear). The fundamental limitation of learning procedures with an a priori specified hypothesis space (e.g., linear models or kernel methods with a preselected kernel) is that they can learn good concept descriptions only if the hypothesis space selected beforehand is large enough to contain a good solution to the considered problem and small enough to allow good generalization from a small number of training examples. As finding a good hypothesis space is equivalent to finding a good set of features [5], we propose an effective supervised feature construction method to tackle this problem. The main goal of the approach is to embed the data into a feature space for which a set of linear hypotheses is of sufficient capacity. The motivation for this choice of hypotheses is in the desire to exploit the scalability of existing algorithms for training linear models. It is for their scalability that these models are frequently a method of choice for learning on large scale data sets (e.g., the implementation of linear SVM [13] has won the large scale learning challenge at ICML 2008 and KDD CUP 2010). However, as the set of linear hypotheses defined on a small or moderate number of input features is usually of low capacity these methods often learn inaccurate descriptions of target concepts. The proposed approach surmounts this and exploits the scalability of existing algorithms for training linear models while overcoming their low capacity on input features. The latter is achieved by harnessing the information contained in the labeled training data and constructing features by empirically fitting squared error residuals. We draw motivation for our approach by considering the minimization of the expected squared error using functional gradient descent (Section 2.1). In each step of the descent, the current estimator is updated by moving in the direction of the residual function. We want to mimic this behavior by constructing a feature representation incrementally so that for each step of the descent we add a feature which approximates well the residual function. In this constructive process, we select our features from a predetermined set of basis functions which can be chosen so that a high capacity set 30th Conference on Neural Information Processing Systems (NIPS 2016), Barcelona, Spain. of linear hypotheses corresponds to the constructed feature space (Section 2.2). In our theoretical analysis of the approach, we provide a convergence rate for this constructive procedure (Section 2.3) and give a generalization bound for the empirical fitting of residuals (Section 2.4). The latter is needed because the feature construction is performed based on an independent and identically distributed sample of labeled examples. The approach, presented in Section 2.5, is highly flexible and allows for an extension of a feature representation without complete re-training of the model. As it performs similar to gradient descent, a stopping criteria based on an accuracy threshold can be devised and the algorithm can then be simulated without specifying the number of features a priori. In this way, the algorithm can terminate sooner than alternative approaches for simple hypotheses. The method is easy to implement and can be scaled to millions of instances with a parallel implementation. To evaluate the effectiveness of our approach empirically, we compare it to other related approaches by training linear ridge regression models in the feature spaces constructed by these methods. Our empirical results indicate a superior performance of the proposed approach over competing methods. The results are presented in Section 3 and the approaches are discussed in Section 4. 2 Greedy feature construction In this section, we present our feature construction approach. We start with an overview where we introduce the problem setting and motivate our approach by considering the minimization of the expected squared error using functional gradient descent. Following this, we define a set of features and demonstrate that the approach can construct a rich set of hypotheses. We then show that our greedy constructive procedure converges and give a generalization bound for the empirical fitting of residuals. The section concludes with a pseudo-code description of our approach. 2.1 Overview We consider a learning problem with the squared error loss function where the goal is to find a mapping from a Euclidean space to the set of reals. In these problems, it is typically assumed that a sample z = ((x1, y1) , . . . , (xm, ym)) of m examples is drawn independently from a Borel probability measure ρ defined on Z = X Y , where X is a compact subset of a finite dimensional Euclidean space with the dot product , and Y R. For every x X let ρ (y | x) be the conditional probability measure on Y and ρX be the marginal probability measure on X. For the sake of brevity, when it is clear from the context, we will write ρ instead of ρX. Let fρ(x) = R y dρ (y | x) be the bounded target/regression function of the measure ρ. Our goal is to construct a feature representation such that there exists a linear hypothesis on this feature space that approximates well the target function. For an estimator f of the function fρ we measure the goodness of fit with the expected squared error in ρ, Eρ (f) = R (f (x) y)2 dρ. The empirical counterpart of the error, defined over a sample z Zm, is denoted with Ez (f) = 1 m Pm i=1 (f (xi) yi)2. Having defined the problem setting, we proceed to motivate our approach by considering the minimization of the expected squared error using functional gradient descent. For that, we first review the definition of functional gradient. For a functional F defined on a normed linear space and an element p from this space, the functional gradient F (p) is the principal linear part of a change in F after it is perturbed in the direction of q, F (p + q) = F (p) + ψ (q) + ϵ q , where ψ (q) is the linear functional with F (p) as its principal linear part, and ϵ 0 as q 0 [e.g., see Section 3.2 in 16]. In our case, the normed space is the Hilbert space of square integrable functions L2 ρ (X) and for the expected squared error functional on this space we have that it holds Eρ (f + ϵq) Eρ (f) = 2 (f fρ), ϵq L2ρ(X) + O ϵ2 . Hence, an algorithm for the minimization of the expected squared error using functional gradient descent on this space could be specified as ft+1 = νft + 2 (1 ν) (fρ ft) , where 0 ν 1 denotes the learning rate and ft is the estimate at step t. The functional gradient direction 2 (fρ ft) is the residual function at step t and the main idea behind our approach is to iteratively refine our feature representation by extending it with a new feature that matches the current residual function. In this way, for a suitable choice of learning rate ν, the functional descent would be performed through a convex hull of features and in each step we would have an estimate of the target function fρ expressed as a convex combination of the constructed features. 2.2 Greedy features We introduce now a set of features parameterized with a ridge basis function and hyperparameters controlling the smoothness of these features. As each subset of features corresponds to a set of hypotheses, in this way we specify a family of possible hypothesis spaces. For a particular choice of ridge basis function we argue below that the approach outlined in the previous section can construct a highly expressive feature representation (i.e., a hypothesis space of high capacity). Let C (X) be the Banach space of continuous functions on X with the uniform norm. For a Lipschitz continuous function φ : R R, φ 1, and constants r, s, t > 0 let FΘ C (X), Θ = (φ, r, s, t), be a set of ridge-wave functions defined on the set X, FΘ = a φ ( w, x + b) | w Rd, a, b R, |a| r, w 2 s, |b| t . From this definition, it follows that for all g FΘ it holds g r. As a ridge-wave function g FΘ is bounded and Lipschitz continuous, it is also square integrable in the measure ρ and g L2 ρ (X). Therefore, FΘ is a subset of the Hilbert space of square integrable functions defined on X with respect to the probability measure ρ, i.e., FΘ L2 ρ (X). Taking φ ( ) = cos ( ) in the definition of FΘ we obtain a set of cosine-wave features Fcos = a cos ( w, x + b) | w Rd, a, b R, |a| r, w 2 s, |b| t . For this set of features the approach outlined in Section 2.1 can construct a rich set of hypotheses. To demonstrate this we make a connection to shift-invariant reproducing kernel Hilbert spaces and show that the approach can approximate any bounded function from any shift-invariant reproducing kernel Hilbert space. This means that a set of linear hypotheses defined by cosine features can be of high capacity and our approach can overcome the problems with the low capacity of linear hypotheses defined on few input features. A proof of the following theorem is provided in Appendix B. Theorem 1. Let Hk be a reproducing kernel Hilbert space corresponding to a continuous shiftinvariant and positive definite kernel k defined on a compact set X. Let µ be the positive and bounded spectral measure whose Fourier transform is the kernel k. For any probability measure ρ defined on X, it is possible to approximate any bounded function f Hk using a convex combination of n ridge-wave functions from Fcos such that the approximation error in ρ decays with rate O (1/ n). 2.3 Convergence For the purpose of this paper, it suffices to show the convergence of ϵ-greedy sequences of functions (see Definition 1) in Hilbert spaces. We, however, choose to provide a stronger result which holds for ϵ-greedy sequences in uniformly smooth Banach spaces. In the remainder of the paper, co (S) and S will be used to denote the convex hull of elements from a set S and the closure of S, respectively. Definition 1. Let B be a Banach space with norm and let S B. An incremental sequence is any sequence {fn}n 1 of elements of B such that f1 S and for each n 1 there is some g S so that fn+1 co ({fn, g}). An incremental sequence is greedy with respect to an element f co (S) if for all n N it holds fn+1 f = inf { h f | h co ({fn, g}) , g S} . Given a positive sequence of allowed slack terms {ϵn}n 1, an incremental sequence {fn}n 1 is called ϵ-greedy with respect to f if for all n N it holds fn+1 f < inf { h f | h co ({fn, g}) , g S} + ϵn. Having introduced the notion of an ϵ-greedy incremental sequence of functions, let us now relate it to our feature construction approach. In the outlined constructive procedure (Section 2.1), we proposed to select new features corresponding to the functional gradient at the current estimate of the target function. Now, if at each step of the functional gradient descent there exists a ridge-wave function from our set of features which approximates well the residual function (w.r.t. fρ) then this sequence of functions defines a descent through co (FΘ) which is an ϵ-greedy incremental sequence of functions with respect to fρ co (FΘ). In Section 2.1, we have also demonstrated that FΘ is a subset of the Hilbert space L2 ρ (X) and this is by definition a Banach space. In accordance with Definition 1, we now consider under what conditions an ϵ-greedy sequence of functions from this space converges to any target function fρ co (FΘ). Note that this relates to Theorem 1 which confirms the strength of the result by showing that the capacity of co (FΘ) is large. Before we show the convergence of our constructive procedure, we need to prove that an ϵ-greedy incremental sequence of functions/features can be constructed in our setting. For that, we characterize the Banach spaces in which it is always possible to construct such sequences of functions/features. Definition 2. Let B be a Banach space, B the dual space of B, and f B, f = 0. A peak functional for f is a bounded linear operator F B such that F B = 1 and F (f) = f B. The Banach space B is said to be smooth if for each f B, f = 0, there is a unique peak functional. The existence of at least one peak functional for all f B, f = 0, is guaranteed by the Hahn-Banach theorem [27]. For a Hilbert space H, for each element f H, f = 0, there exists a unique peak functional F = f, H/ f H. Thus, every Hilbert space is a smooth Banach space. Donahue et al. [12, Theorem 3.1] have shown that in smooth Banach spaces and in particular in the Hilbert space L2 ρ (X) an ϵ-greedy incremental sequence of functions can always be constructed. However, not every such sequence of functions converges to the function with respect to which it was constructed. For the convergence to hold, a stronger notion of smoothness is needed. Definition 3. The modulus of smoothness of a Banach space B is a function τ : R+ 0 R+ 0 defined as τ (r) = 1 2 sup f = g =1 ( f + rg + f rg ) 1, where f, g B. The Banach space B is said to be uniformly smooth if τ (r) o (r) as r 0. We need to observe now that every Hilbert space is a uniformly smooth Banach space [12]. For the sake of completeness, we provide a proof of this proposition in Appendix B. Proposition 2. For any Hilbert space the modulus of smoothness is equal to τ (r) = 1 + r2 1. Having shown that Hilbert spaces are uniformly smooth Banach spaces, we proceed with two results giving a convergence rate of an ϵ-greedy incremental sequence of functions. What is interesting about these results is the fact that a feature does not need to match exactly the residual function in a greedy descent step (Section 2.1); it is only required that condition (ii) from the next theorem is satisfied. Theorem 3. [Donahue et al., 12] Let B be a uniformly smooth Banach space having modulus of smoothness τ (u) γut, with γ being a constant and t > 1. Let S be a bounded subset of B and let f co (S). Let K > 0 be chosen such that f g K for all g S, and let ϵ > 0 be a fixed slack value. If the sequences {fn}n 1 co (S) and {gn}n 1 S are chosen recursively so that: (i) f1 S, (ii) Fn (gn f) 2γ((K+ϵ)t Kt)/nt 1 fn f t 1, and (iii) fn+1 = n/n+1 fn + 1/n+1 gn, where Fn is the peak functional of fn f, then it holds fn f (2γt) 1/t(K+ϵ) n1 1/t h 1 + (t 1) log2 n The following corollary gives a convergence rate for an ϵ-greedy incremental sequence of functions constructed according to Theorem 3 with respect to fρ co (FΘ). As this result (a proof is given in Appendix B) holds for all such sequences of functions, it also holds for our constructive procedure. Corollary 4. Let {fn}n 1 co (FΘ) be an ϵ-greedy incremental sequence of functions constructed according to the procedure described in Theorem 3 with respect to a function f co (FΘ) . Then, it holds fn f ρ (K+ϵ) 2+log2 n/2n n . 2.4 Generalization bound In step t + 1 of the empirical residual fitting, based on a sample {(xi, yi ft (xi))}m i=1, the approach selects a ridge-wave function from FΘ that approximates well the residual function (fρ ft). In the last section, we have specified in which cases such ridge-wave functions can be constructed and provided a convergence rate for this constructive procedure. As the convergence result is not limited to target functions from FΘ and co (FΘ), we give a bound on the generalization error for hypotheses from F = co (FΘ), where the closure is taken with respect to C (X). Before we give a generalization bound, we show that our hypothesis space F is a convex and compact set of functions. The choice of a compact hypothesis space is important because it guarantees that a minimizer of the expected squared error Eρ and its empirical counterpart Ez exists. In particular, a continuous function attains its minimum and maximum value on a compact set and this guarantees the existence of minimizers of Eρ and Ez. Moreover, for a hypothesis space that is both convex and compact, the minimizer of the expected squared error is unique as an element of L2 ρ (X). A simple proof of the uniqueness of such a minimizer in L2 ρ (X) and the continuity of the functionals Eρ and Ez can be found in [9]. For the sake of completeness, we provide a proof in Appendix A as Proposition A.2. The following proposition (a proof is given in Appendix B) shows that our hypothesis space is a convex and compact subset of the metric space C (X). Algorithm 1 GREEDYDESCENT Input: sample z = {(xi, yi)}s i=1, initial estimates at sample points {f0,i}s i=1, ridge basis function φ, maximum number of descent steps p, regularization parameter λ, and precision ϵ 1: W 2: for k = 1, 2, . . . , p do 3: wk, ck arg minw,c=(c ,c ) Ps i=1 c fk 1,i + c φ w xi yi 2 + λΩ(c, w) 4: W W {wk} and fk,i c kfk 1,i + c kφ w k xi , i = 1, . . . , s 5: if |Ez(fk) Ez(fk 1)|/max{Ez(fk),Ez(fk 1)} < ϵ then EXIT FOR LOOP end if 6: end for 7: return W Proposition 5. The hypothesis space F is a convex and compact subset of the metric space C(X). Moreover, the elements of this hypothesis space are Lipschitz continuous functions. Having established that the hypothesis space is a compact set, we can now give a generalization bound for learning with this hypothesis space. The fact that the hypothesis space is compact implies that it is also a totally bounded set [27], i.e., for all ϵ > 0 there exists a finite ϵ-net of F. This, on the other hand, allows us to derive a sample complexity bound by using the ϵ-covering number of a space as a measure of its capacity [21]. The following theorem and its corollary (proofs are provided in Appendix B) give a generalization bound for learning with the hypothesis space F. Theorem 6. Let M > 0 such that, for all f F, |f (x) y| M almost surely. Then, for all ϵ > 0 P [Eρ (fz) Eρ (f ) ϵ] 1 N (F, ϵ/24M, ) exp ( mϵ/288M 2) , where fz and f are the minimizers of Ez and Eρ on the set F, z Zm, and N (F, ϵ, ) denotes the ϵ-covering number of F w.r.t. C (X). Corollary 7. For all ϵ > 0 and all δ > 0, with probability 1 δ, a minimizer of the empirical squared error on the hypothesis space F is (ϵ, δ)-consistent when the number of samples m Ω r (Rs + t) Lφ 1 δ . Here, R is the radius of a ball containing the set of instances X in its interior, Lφ is the Lipschitz constant of a function φ, and r, s, and t are hyperparameters of FΘ. 2.5 Algorithm Algorithm 1 is a pseudo-code description of the outlined approach. To construct a feature space with a good set of linear hypotheses the algorithm takes as input a set of labeled examples and an initial empirical estimate of the target function. A dictionary of features is specified with a ridge basis function and the smoothness of individual features is controlled with a regularization parameter. Other parameters of the algorithm are the maximum allowed number of descent steps and a precision term that defines the convergence of the descent. As outlined in Sections 2.1 and 2.3, the algorithm works by selecting a feature that matches the residual function at the current estimate of the target function. For each selected feature the algorithm also chooses a suitable learning rate and performs a functional descent step (note that we are inferring the learning rate instead of setting it to 1/n+1 as in Theorem 3). To avoid solving these two problems separately, we have coupled both tasks into a single optimization problem (line 3) we fit a linear model to a feature representation consisting of the current empirical estimate of the target function and a ridge function parameterized with a d-dimensional vector w. The regularization term Ωis chosen to control the smoothness of the new feature and avoid over-fitting. The optimization problem over the coefficients of the linear model and the spectrum of the ridge basis function is solved by casting it as a hyperparameter optimization problem [20]. For the sake of completeness, we have provided a detailed derivation in Appendix C. While the hyperparameter optimization problem is in general non-convex, Theorem 3 indicates that a globally optimal solution is not (necessarily) required and instead specifies a weaker condition. To account for the non-convex nature of the problem and compensate for the sequential generation of features, we propose to parallelize the feature construction process by running several instances of the greedy descent simultaneously. A pseudo-code description of this parallelized approach is given in Algorithm 2. The algorithm takes as input parameters required for running the greedy descent and some parameters specific to the parallelization scheme number of data passes and available machines/cores, regularization parameter for the fitting of linear models in the constructed feature space, and cut-off parameter for the elimination of redundant features. The whole process is started by adding a bias feature and setting the initial empirical estimates at sample points to the mean value of the outputs (line 1). Following this, the algorithm mimics stochastic gradient descent and makes Algorithm 2 GREEDY FEATURE CONSTRUCTION (GFC) Input: sample z = {(xi, yi)}m i=1, ridge basis function φ, number of data passes T, maximum number of greedy descent steps p, number of machines/cores M, regularization parameters λ and ν, precision ϵ, and feature cut-off threshold η 1: W {0} and f0,k 1 m Pm i=1 yi, k = 1, . . . , m 2: for i = 1, . . . , T do 3: for j = 1, 2, . . . , M parallel do 4: Sj U{1,2,...,m} and W W GREEDYDESCENT {(xk, yk)}k Sj , fi 1,k k Sj , φ, p, λ, ϵ 6: a arg mina Pm k=1 P|W | l=1 alφ w l xk yk 2 + ν a 2 2 7: W W \ wl W | |a l | < η, 1 l |W| and fi,k P|W | l=1 a l φ w l xk , k = 1, . . . , m 8: end for 9: return (W, a ) a specified number of passes through the data (line 2). In the first step of each pass, the algorithm performs greedy functional descent in parallel using a pre-specified number of machines/cores M (lines 3-5). This step is similar to the splitting step in parallelized stochastic gradient descent [32]. Greedy descent is performed on each of the machines for a maximum number of iterations p and the estimated parameter vectors are added to the set of constructed features W (line 4). After the features have been learned the algorithm fits a linear model to obtain the amplitudes (line 6). To fit a linear model, we use least square regression penalized with the l2-norm because it can be solved in a closed form and cross-validation of the capacity parameter involves optimizing a 1-dimensional objective function [20]. Fitting of the linear model can be understood as averaging of the greedy approximations constructed on different chunks of the data. At the end of each pass, the empirical estimates at sample points are updated and redundant features are removed (line 7). One important detail in the implementation of Algorithm 1 is the data splitting between the training and validation samples for the hyperparameter optimization. In particular, during the descent we are more interested in obtaining a good spectrum than the amplitude because a linear model will be fit in Algorithm 2 over the constructed features and the amplitude values will be updated. For this reason, during the hyperparameter optimization over a k-fold splitting in Algorithm 1, we choose a single fold as the training sample and a batch of folds as the validation sample. 3 Experiments In this section, we assess the performance of our approach (see Algorithm 2) by comparing it to other feature construction approaches on synthetic and real-world data sets. We evaluate the effectiveness of the approach with the set of cosine-wave features introduced in Section 2.2. For this set of features, our approach is directly comparable to random Fourier features [26] and á la carte [31]. The implementation details of the three approaches are provided in Appendix C. We address here the choice of the regularization term in Algorithm 1: To control the smoothness of newly constructed features, we penalize the objective in line 3 so that the solutions with the small L2 ρ (X) norm are preferred. For this choice of regularization term and cosine-wave features, we empirically observe that the optimization objective is almost exclusively penalized by the l2 norm of the coefficient vector c. Following this observation, we have simulated the greedy descent with Ω(c, w) = c 2 2. We now briefly describe the data sets and the experimental setting. The experiments were conducted on three groups of data sets. The first group contains four UCI data sets on which we performed parameter tuning of all three algorithms (Table 1, data sets 1-4). The second group contains the data sets with more than 5000 instances available from Luís Torgo [28]. The idea is to use this group of data sets to test the generalization properties of the considered algorithms (Table 1, data sets 5-10). The third group contains two artificial and very noisy data sets that are frequently used in regression tree benchmark tests. For each considered data set, we split the data into 10 folds; we refer to these splits as the outer cross-validation folds. In each step of the outer cross-validation, we use nine folds as the training sample and one fold as the test sample. For the purpose of the hyperparameter tuning we split the training sample into five folds; we refer to these splits as the inner cross-validation folds. We run all algorithms on identical outer cross-validation folds and construct feature representations with 100 and 500 features. The performance of the algorithms is assessed by comparing the root mean squared error of linear ridge regression models trained in the constructed feature spaces and the average time needed for the outer cross-validation of one fold. Table 1: To facilitate the comparison between data sets we have normalized the outputs so that their range is one. The accuracy of the algorithms is measured using the root mean squared error, multiplied by 100 to mimic percentage error (w.r.t. the range of the outputs). The mean and standard deviation of the error are computed after performing 10-fold cross-validation. The reported walltime is the average time it takes a method to cross-validate one fold. To assess whether a method performs statistically significantly better than the other on a particular data set we perform the paired Welch t-test [29] with p = 0.05. The significantly better results for the considered settings are marked in bold. DATASET m d n = 100 n = 500 GFC ALC GFC ALC ERROR WALLTIME ERROR WALLTIME ERROR WALLTIME ERROR WALLTIME parkinsons tm (total) 5875 21 2.73 ( 0.19) 00 : 03 : 49 0.78 ( 0.13) 00 : 05 : 19 2.20 ( 0.27) 00 : 04 : 15 0.31 ( 0.17) 00 : 27 : 15 ujindoorloc (latitude) 21048 527 3.17 ( 0.15) 00 : 21 : 39 6.19 ( 0.76) 01 : 21 : 58 3.04 ( 0.19) 00 : 36 : 49 6.99 ( 0.97) 02 : 23 : 15 ct-slice 53500 380 2.93 ( 0.10) 00 : 52 : 05 3.82 ( 0.64) 03 : 31 : 25 2.59 ( 0.10) 01 : 24 : 41 2.73 ( 0.29) 06 : 11 : 12 Year Prediction MSD 515345 90 10.06 ( 0.09) 01 : 20 : 12 9.94 ( 0.08) 05 : 29 : 14 10.01 ( 0.08) 01 : 30 : 28 9.92 ( 0.07) 11 : 58 : 41 delta-ailerons 7129 5 3.82 ( 0.24) 00 : 01 : 23 3.73 ( 0.20) 00 : 05 : 13 3.79 ( 0.25) 00 : 01 : 57 3.73 ( 0.24) 00 : 25 : 14 kinematics 8192 8 5.18 ( 0.09) 00 : 04 : 02 5.03 ( 0.23) 00 : 11 : 28 4.65 ( 0.11) 00 : 04 : 44 5.01 ( 0.76) 00 : 38 : 53 cpu-activity 8192 21 2.65 ( 0.12) 00 : 04 : 23 2.68 ( 0.27) 00 : 09 : 24 2.60 ( 0.16) 00 : 04 : 24 2.62 ( 0.15) 00 : 25 : 13 bank 8192 32 9.83 ( 0.27) 00 : 01 : 39 9.84 ( 0.30) 00 : 12 : 48 9.83 ( 0.30) 00 : 02 : 01 9.87 ( 0.42) 00 : 49 : 48 pumadyn 8192 32 3.44 ( 0.10) 00 : 02 : 24 3.24 ( 0.07) 00 : 13 : 17 3.30 ( 0.06) 00 : 02 : 27 3.42 ( 0.15) 00 : 57 : 33 delta-elevators 9517 6 5.26 ( 0.17) 00 : 00 : 57 5.28 ( 0.18) 00 : 07 : 07 5.24 ( 0.17) 00 : 01 : 04 5.23 ( 0.18) 00 : 32 : 30 ailerons 13750 40 4.67 ( 0.18) 00 : 02 : 56 4.89 ( 0.43) 00 : 16 : 34 4.51 ( 0.12) 00 : 02 : 11 4.77 ( 0.40) 01 : 05 : 07 pole-telecom 15000 26 7.34 ( 0.29) 00 : 10 : 45 7.16 ( 0.55) 00 : 20 : 34 5.55 ( 0.15) 00 : 11 : 37 5.20 ( 0.51) 01 : 39 : 22 elevators 16599 18 3.34 ( 0.08) 00 : 03 : 16 3.37 ( 0.55) 00 : 21 : 20 3.12 ( 0.20) 00 : 04 : 06 3.13 ( 0.24) 01 : 20 : 58 cal-housing 20640 8 11.55 ( 0.24) 00 : 05 : 49 12.69 ( 0.47) 00 : 11 : 14 11.17 ( 0.25) 00 : 06 : 16 12.70 ( 1.01) 01 : 01 : 37 breiman 40768 10 4.01 ( 0.03) 00 : 02 : 46 4.06 ( 0.04) 00 : 13 : 52 4.01 ( 0.03) 00 : 03 : 04 4.03 ( 0.03) 01 : 04 : 16 friedman 40768 10 3.29 ( 0.09) 00 : 06 : 07 3.37 ( 0.46) 00 : 18 : 43 3.16 ( 0.03) 00 : 07 : 04 3.25 ( 0.09) 01 : 39 : 37 An extensive summary containing the results of experiments with the random Fourier features approach (corresponding to Gaussian, Laplace, and Cauchy kernels) and different configurations of á la carte is provided in Appendix D. As the best performing configuration of á la carte on the development data sets is the one with Q = 5 components, we report in Table 1 the error and walltime for this configuration. From the walltime numbers we see that our approach is in both considered settings with 100 and 500 features always faster than á la carte. Moreover, the proposed approach is able to generate a feature representation with 500 features in less time than required by á la carte for a representation of 100 features. In order to compare the performance of the two methods with respect to accuracy, we use the Wilcoxon signed rank test [30, 11]. As our approach with 500 features is on all data sets faster than á la carte with 100 features, we first compare the errors obtained in these experiments. For 95% confidence, the threshold value of the Wilcoxon signed rank test with 16 data sets is T = 30 and from our results we get the T-value of 28. As the T-value is below the threshold, our algorithm can with 95% confidence generate in less time a statistically significantly better feature representation than á la carte. For the errors obtained in the settings where both methods have the same number of features, we obtain the T-values of 60 and 42. While in the first case for the setting with 100 features the test is inconclusive, in the second case our approach is with 80% confidence statistically significantly more accurate than á la carte. To evaluate the performance of the approaches on individual data sets, we perform the paired Welch t-test [29] with p = 0.05. Again, the results indicate a good/competitive performance of our algorithm compared to á la carte. 4 Discussion In this section, we discuss the advantages of the proposed method over the state-of-the-art baselines in learning fast shift-invariant kernels and other related approaches. Flexibility. The presented approach is a highly flexible supervised feature construction method. In contrast to an approach based on random Fourier features [26], the proposed method does not require a spectral measure to be specified a priori. In the experiments (details can be found in Appendix D), we have demonstrated that the choice of spectral measure is important as, for the considered measures (corresponding to Gaussian, Laplace, and Cauchy kernels), the random Fourier features approach is outperformed on all data sets. The second competing method, á la carte, is more flexible when it comes to the choice of spectral measure and works by approximating it with a mixture of Gaussians. However, the number of components and features per component needs to be specified beforehand or cross-validated. In contrast, our approach mimics functional gradient descent and can be simulated without specifying the size of the feature representation beforehand. Instead, a stopping criteria (see, e.g., Algorithm 1) based on the successive decay of the error can be devised. As a result, the proposed approach terminates sooner than the alternative approaches for simple concepts/hypothesis. The proposed method is also easy to implement (for the sake of completeness, the hyperparameter gradients are provided in Appendix C.1) and allows us to extend the existing feature representation without complete re-training of the model. We note that the approaches based on random Fourier features are also simple to implement and can be re-trained efficiently with the increase in the number of features [10]. Á la carte, on the other hand, is less flexible in this regard due to the number of hyperparameters and the complexity of gradients it is not straightforward to implement this method. Scalability. The fact that our greedy descent can construct a feature in time linear in the number of instances m and dimension of the problem d makes the proposed approach highly scalable. In particular, the complexity of the proposed parallelization scheme is dominated by the cost of fitting a linear model and the whole algorithm runs in time O T n3 + n2m + nmd , where T denotes the number of data passes (i.e., linear model fits) and n number of constructed features. To scale this scheme to problems with millions of instances, it is possible to fit linear models using the parallelized stochastic gradient descent [32]. As for the choice of T, the standard setting in simulations of stochastic gradient descent is 5-10 data passes. Thus, the presented approach is quite robust and can be applied to large scale data sets. In contrast to this, the cost of performing a gradient step in the hyperparameter optimization of á la carte is O n3 + n2m + nmd . In our empirical evaluation using an implementation with 10 random restarts, the approach needed at least 20 steps per restart to learn an accurate model. The required number of gradient steps and the cost of computing them hinder the application of á la carte to large scale data sets. In learning with random Fourier features which also run in time O n3 + n2m + nmd , the main cost is the fitting of linear models one for each pair of considered spectral and regularization parameters. Other approaches. Beside fast kernel learning approaches, the presented method is also related to neural networks parameterized with a single hidden layer. These approaches can be seen as feature construction methods jointly optimizing over the whole feature representation. A detailed study of the approximation properties of a hypothesis space of a single layer network with the sigmoid ridge function has been provided by Barron [4]. In contrast to these approaches, we construct features incrementally by fitting residuals and we do this with a set of non-monotone ridge functions as a dictionary of features. Regarding our generalization bound, we note that the past work on single layer neural networks contains similar results but in the context of monotone ridge functions [1]. As the goal of our approach is to construct a feature space for which linear hypotheses will be of sufficient capacity, the presented method is also related to linear models working with low-rank kernel representations. For instance, Fine and Scheinberg [14] investigate a training algorithm for SVMs using low-rank kernel representations. The difference between our approach and this method is in the fact that the low-rank decomposition is performed without considering the labels. Side knowledge and labels are considered by Kulis et al. [22] and Bach and Jordan [3] in their approaches to construct a low-rank kernel matrix. However, these approaches are not selecting features from a set of ridge functions, but find a subspace of a preselected kernel feature space with a good set of hypothesis. From the perspective of the optimization problem considered in the greedy descent (Algorithm 1) our approach can be related to single index models (SIM) where the goal is to learn a regression function that can be represented as a single monotone ridge function [19, 18]. In contrast to these models, our approach learns target/regression functions from the closure of the convex hull of ridge functions. Typically, these target functions cannot be written as single ridge functions. Moreover, our ridge functions do not need to be monotone and are more general than the ones considered in SIM models. In addition to these approaches and considered baseline methods, the presented feature construction approach is also related to methods optimizing expected loss functions using functional gradient descent [23]. However, while Mason et al. [23] focus on classification problems and hypothesis spaces with finite VC dimension, we focus on the estimation of regression functions in spaces with infinite VC dimension (e.g., see Section 2.2). In contrast to that work, we provide a convergence rate for our approach. Similarly, Friedman [15] has proposed a gradient boosting machine for greedy function estimation. In their approach, the empirical functional gradient is approximated by a weak learner which is then combined with previously constructed learners following a stagewise strategy. This is different from the stepwise strategy that is followed in our approach where previously constructed estimators are readjusted when new features are added. The approach in [15] is investigated mainly in the context of regression trees, but it can be adopted to feature construction. To the best of our knowledge, theoretical and empirical properties of this approach in the context of feature construction and shift-invariant reproducing kernel Hilbert spaces have not been considered so far. Acknowledgment: We are grateful for access to the University of Nottingham High Performance Computing Facility. A part of this work was also supported by the German Science Foundation (grant number GA 1615/1-1). [1] Martin Anthony and Peter L. Bartlett. Neural Network Learning: Theoretical Foundations. Cambridge University Press, 2009. [2] Nachman Aronszajn. Theory of reproducing kernels. Transactions of the American Math. Society, 1950. [3] Francis R. Bach and Michael I. Jordan. Predictive low-rank decomposition for kernel methods. In Proceedings of the 22nd International Conference on Machine Learning. [4] Andrew R. Barron. Universal approximation bounds for superpositions of a sigmoidal function. IEEE Transactions on Information Theory, 39(3), 1993. [5] Jonathan Baxter. A model of inductive bias learning. Journal of Artificial Intelligence Research, 12, 2000. [6] Alain Bertinet and Thomas C. Agnan. Reproducing Kernel Hilbert Spaces in Probability and Statistics. Kluwer Academic Publishers, 2004. [7] Salomon Bochner. Vorlesungen über Fouriersche Integrale. In Akademische Verlagsgesellschaft, 1932. [8] Bernd Carl and Irmtraud Stephani. Entropy, Compactness, and the Approximation of Operators. Cambridge University Press, 1990. [9] Felipe Cucker and Steve Smale. On the mathematical foundations of learning. Bulletin of the American Mathematical Society, 39, 2002. [10] Bo Dai, Bo Xie, Niao He, Yingyu Liang, Anant Raj, Maria-Florina Balcan, and Le Song. Scalable kernel methods via doubly stochastic gradients. In Advances in Neural Information Processing Systems 27. [11] Janez Demšar. Statistical comparisons of classifiers over multiple data sets. Journal of Machine Learning Research, 7, 2006. [12] Michael J. Donahue, Christian Darken, Leonid Gurvits, and Eduardo Sontag. Rates of convex approximation in non-Hilbert spaces. Constructive Approximation, 13(2), 1997. [13] Rong-En Fan, Kai-Wei Chang, Cho-Jui Hsieh, Xiang-Rui Wang, and Chih-Jen Lin. LIBLINEAR: A library for large linear classification. Journal of Machine Learning Research, 9, 2008. [14] Shai Fine and Katya Scheinberg. Efficient SVM training using low-rank kernel representations. Journal of Machine Learning Research, 2, 2002. [15] Jerome H. Friedman. Greedy function approximation: A gradient boosting machine. The Annals of Statistics, 29, 2000. [16] Israel M. Gelfand and Sergei V. Fomin. Calculus of variations. Prentice-Hall Inc., 1963. [17] Marc G. Genton. Classes of kernels for machine learning: A statistics perspective. Journal of Machine Learning Research, 2, 2002. [18] Sham M. Kakade, Varun Kanade, Ohad Shamir, and Adam T. Kalai. Efficient learning of generalized linear and single index models with isotonic regression. In Advances in Neural Information Processing Systems 24, 2011. [19] Adam T. Kalai and Ravi Sastry. The isotron algorithm: High-dimensional isotonic regression. In Proceedings of the Conference on Learning Theory, 2009. [20] Sathiya Keerthi, Vikas Sindhwani, and Olivier Chapelle. An efficient method for gradient-based adaptation of hyperparameters in SVM models. In Advances in Neural Information Processing Systems 19, 2006. [21] Andrey N. Kolmogorov and Vladimir M. Tikhomirov. ϵ-entropy and ϵ-capacity of sets in function spaces. Uspehi Matematicheskikh Nauk, 14(2), 1959. [22] Brian Kulis, Mátyás Sustik, and Inderjit Dhillon. Learning low-rank kernel matrices. In Proceedings of the 23rd International Conference on Machine Learning, 2006. [23] Llew Mason, Jonathan Baxter, Peter L. Bartlett, and Marcus Frean. Functional gradient techniques for combining hypotheses. In Advances in large margin classifiers. MIT Press, 1999. [24] Sebastian Mayer, Tino Ullrich, and Jan Vybiral. Entropy and sampling numbers of classes of ridge functions. Constructive Approximation, 42(2), 2015. [25] Tom M. Mitchell. Machine Learning. Mc Graw-Hill, 1997. [26] Ali Rahimi and Benjamin Recht. Random features for large-scale kernel machines. In Advances in Neural Information Processing Systems 20. [27] Walter Rudin. Functional Analysis. Int. Series in Pure and Applied Mathematics. Mc Graw-Hill Inc., 1991. [28] Luís Torgo. Repository with regression data sets. http://www.dcc.fc.up.pt/~ltorgo/ Regression/Data Sets.html, accessed September 22, 2016. [29] Bernard L. Welch. The generalization of student s problem when several different population variances are involved. Biometrika, 34(1/2), 1947. [30] Frank Wilcoxon. Individual comparisons by ranking methods. Biometrics Bulletin, 1(6), 1945. [31] Zichao Yang, Alexander J. Smola, Le Song, and Andrew G. Wilson. Á la carte Learning fast kernels. In Proceedings of the 18th International Conference on Artificial Intelligence and Statistics, 2015. [32] Martin A. Zinkevich, Alex J. Smola, Markus Weimer, and Lihong Li. Parallelized stochastic gradient descent. In Advances in Neural Information Processing Systems 23, 2010.