# kernel_interpolation_with_sparse_grids__80b5202d.pdf Kernel Interpolation with Sparse Grids Mohit Yadav University of Massachusetts Amherst ymohit@cs.umass.edu Daniel Sheldon University of Massachusetts Amherst sheldon@cs.umass.edu Cameron Musco University of Massachusetts Amherst cmusco@cs.umass.edu Structured kernel interpolation (SKI) accelerates Gaussian process (GP) inference by interpolating the kernel covariance function using a dense grid of inducing points, whose corresponding kernel matrix is highly structured and thus amenable to fast linear algebra. Unfortunately, SKI scales poorly in the dimension of the input points, since the dense grid size grows exponentially with the dimension. To mitigate this issue, we propose the use of sparse grids within the SKI framework. These grids enable accurate interpolation, but with a number of points growing more slowly with dimension. We contribute a novel nearly linear time matrixvector multiplication algorithm for the sparse grid kernel matrix. We also describe how sparse grids can be combined with an efficient interpolation scheme based on simplices. With these modifications, we demonstrate that SKI can be scaled to higher dimensions while maintaining accuracy. 1 Introduction Gaussian processes (GPs) are popular prior distributions over continuous functions for use in Bayesian inference [14]. Due to their simple mathematical structure, closed form expressions can be given for posterior inference [20]. Unfortunately, a well-established limitation of GPs is that they are difficult to scale to large datasets. In particular, for both exact posterior inference and the exact log-likelihood computation for hyperparameter learning, one must invert a dense kernel covariance matrix K Rn n, where n is the number of training points. Naively, this operation requires O(n3) time and O(n2) memory. Structured Kernel Interpolation. Many techniques have been proposed to mitigate this scalability issue [23, 19, 24, 9]. Recently, structured kernel interpolation (SKI) has emerged as a promising approach [9]. In SKI, the kernel matrix is approximated via interpolation onto a dense rectilinear grid of m inducing points. In particular, K is approximated as WKGW T , where KG Rm m is the kernel matrix on the inducing points and W Rn m is an interpolation weight matrix mapping training points to nearby grid points. Typically, W is sparse, and KG is highly structured e.g., for shift invariant kernels, KG is multi-level Toeplitz. Thus, W, KG, and in turn the approximate kernel matrix WKGW T admit fast matrix-vector multiplication. This allows fast approximate inference and log-likelihood computation via the use of iterative methods, e.g., the conjugate gradient algorithm. SKI s Curse of Dimensionality. Unfortunately, SKI does not scale well to high-dimensional input data: the number of points in the dense grid, and hence the size of KG, grows exponentially in the dimension d. Moreover, SKI typically employs local cubic interpolation, which leads to an interpolation weight matrix W with row sparsity that also scales exponentially in d. This curse of 36th Conference on Neural Information Processing Systems (Neur IPS 2022). dimensionality is a well-known issue with the use of dense grid interpolation. It has been studied extensively, e.g., in the context of high-dimensional interpolation and numerical integration [17, 5]. In the computational mathematics community, an important technique for interpolating functions in high dimensions is sparse grids [2]. Roughly a sparse grid is a union of rectilinear grids with different resolutions in each dimension. In particular, it is a union of all 2ℓ1 2ℓ2 . . . 2ℓd sized grids, where Pd i=1 ℓi ℓ, for some maximum total resolution ℓ. This upper bound on the total resolution limits the number of points in each grid while a grid can be dense in a few dimensions, no grid can be dense in all dimensions. See Figure 1 for an illustration. Sparse grids have interpolation accuracy comparable to dense grids under certain smoothness assumptions on the interpolated function [22], while using significantly fewer points. Concretely, for any function f : Rd R with bounded mixed partial derivatives, a sparse grid containing O(2ℓℓd 1) points can interpolate as accurately as a dense grid with O(2ℓd) points, where ℓis the maximum grid resolution [21]. Combining Sparse Grids with SKI. Our main contribution is to demonstrate that sparse grids can be used within the SKI framework to significantly improve scaling with dimension. Doing so requires several algorithmic developments. When the inducing point grid is sparse, the kernel matrix on the grid, KG, no longer has simple structure. E.g., it is not Toeplitz when the kernel is shift invariant. Thus, naive matrix-vector multiplications with KG require time that scales quadratically, rather than near-linearly in the grid size. This would significantly limit the scope of performance improvement from using a sparse grid. To handle this issue, we develop a near-linear1 time matrixvector multiplication (MVM) algorithm for any sparse grid kernel matrix. Our algorithm is recursive, and critically leverages the fact that sparse grids can be constructed from smaller dense grids and that they have tensor product structure across dimensions. For an illustration of our algorithm s complexity versus that of naive quadratic time MVMs, see Figure 1 (c). 2 4 6 8 10 12 14 16 d order of growth Dense-MVM Sparse-MVM Ours-MVM Figure 1: Illustration of sparse grid construction for d = 2 and maximum resolution ℓ= 2. (a) A dense grid with resolution (l1, l2) has 2l1 and 2l2 points in each dimension; the grids with red square points have total resolution l1 + l2 2. (b) The sparse grid G2,2 is the union of rectilinear grids with total resolution at most 2. The 17 points with red squares belong to the sparse grid. The dense grid includes the additional points with blue dots, for a total of 49 points. (c) The asymptotic order of growth in performing a single kernel MVM operation for both grids with 28 unique points in each dimension, ignoring constants. For sparse grids, Ours-MVM (i.e., the proposed MVM algorithm) improves significantly over the naive implementation (Sparse-MVM). A second key challenge is that, while sparse grids allow for a grid size that grows as a much more mild exponential function of the dimension d, the bottleneck for applying SKI on large datasets can come in computing MVMs with the interpolation weight matrix W Rn m. For classic high-dimensional interpolation schemes, like cubic interpolation, each row of W has O(2d) non-zero entries, i.e., the kernel covariance for each training point is approximated by a weighted sum of the covariance at O(2d) grid points. When the number of training points n is large, storing W in 1 Near-linear here means running in time O(m log m) for a sparse grid with m points. memory, and multiplying by it, can become prohibitively expensive. To handle this issue, we take an approach similar to that of Kapoor et al. [13] and employ simplicial basis functions for interpolation, whose support grows linearly with d. Combined with our fast MVM algorithm for sparse grid kernel matrices, this interpolation scheme lets us scale SKI to higher dimensions. In summary, we propose the use of sparse grids to improve the scalability of kernel interpolation for GP inference relative to the number of dimensions. To this end, we develop an efficient nearly linear time matrix-vector multiplication algorithm for the sparse grid kernel matrix. Furthermore, we also propose the use of simplicial interpolation to improve scalability of SKI for both dense and sparse grids. We show empirically that these ideas allow SKI to scale to at least 10 dimensions and perform competitively with state-of-the art GP regression methods. We provide an efficient GPU implementation of the proposed algorithm compatible with GPy Torch [9], which is available at https://github.com/ymohit/skisg and licensed under the MIT license. 2 Background Notation. We let N and N0 denote the natural numbers and N {0} respectively. Matrices are represented by capital letters, and vectors by bold letters. I denotes the identity matrix, with dimensions apparent from context. For a matrix M, mvm(M) denotes the number of operations required to multiply M by any admissible vector. In GP regression, the training data are modeled as noisy measurements of a random function f drawn from a GP prior, denoted f N(0, k( , )), where k : Rd Rd R is a covariance kernel. For an input xi, the observed value is modeled as yi = f(xi) + ϵi, with ϵi N(0, σ2). Observed training pairs (xi, yi) are collected as X = [x1, . . . , xn] Rn d and y = [y1, . . . , yn] Rn. The kernel matrix (on training data) is KX = [k(xi, xj)]n i,j=1 Rn n. The GP inference tasks are to compute the posterior distribution of f given (X, y), which itself is a Gaussian process, and to compute the marginal log-likelihood log p(y). Naive approaches rely on the Cholesky decomposition of the matrix KX = KX + σ2I, which takes Θ(n3) time; see Rasmussen [20] for more details. To avoid the Θ(n3) running time of naive inference, many modern methods use iterative algorithms such as the conjugate gradient (CG) algorithm to perform GP inference in a way that accesses the kernel matrix only through matrix vector multiplication (MVM), i.e., the mapping v 7 KXv [9]. These methods support highly accurate approximate solutions to GP posterior inference task as well as hyper-parameter optimization. The complexity of posterior inference and one step of hyper-parameter optimization is Θ(pn2) for p CG iterations, as mvm(KX) = n2. In practice, p n suffices [27, 9]. 2.1 SKI: Structured Kernel Interpolation SKI further accelerates iterative GP inference by approximating the kernel matrix in a way that makes matrix-vector multiplications faster [27]. Given a set of inducing points U Rd, SKI approximates the kernel function as k(x, x ) w T x KUwx , where KU R|U| |U| is the kernel matrix for the set of inducing points U, and the vector wx R|U| contains interpolation weights to interpolate from U to any x. The SKI approximate kernel matrix is KX = WKUW T , where W Rn |U| is the matrix with ith row equal to wxi. To accelerate matrix-vector multiplications with the approximate kernel matrix, SKI places inducing points on a regular grid and uses grid-based interpolation. This leads to a sparse interpolation weight matrix W for example, with cubic interpolation there are O(4d) entries per row, so that mvm(W) = O(n4d) and to a kernel matrix KU that is multi-level Toeplitz (if k is stationary) [27], so that mvm(KU) = O(|U| log |U|). Overall, mvm( KX) = O(n4d + |U| log |U|), which is much faster than n2 for small d. However, SKI quickly becomes infeasible in higher dimensions due to the 4d entries per row of W and curse of dimensionality for the number of points in a high-dimensional grid: specifically, |U| = md for a grid with m points in each dimension. 2.2 Sparse Grids Rectilinear grids. We first give a formal construction of rectilinear grids, which will later be the foundation for sparse grids [7]. For a resolution index l N0, define the 1-d grid Ωl as the centers of 2l equal partitions of the interval [0, 1], which gives Ωl := {i/2l+1 | 1 i 2l+1 and i is odd}. The fact that the position index i must be odd implies that grids for any two different resolutions are disjoint.2 Moreover, resolution-position index pairs (l, i) uniquely specify grid points in the union S l N0 Ωl of 1-d rectilinear grids. To extend rectilinear grids to d dimensions, let l Nd 0 denote a resolution vector. The corresponding rectilinear grid is given by Ωl := d j=1Ωlj, where denotes the Cartesian product. A grid point in Ωl is indexed by the pair (l, i) of a resolution vector l and position vector i, where (lj, ij) gives the position in the 1-d grid Ωj for dimension j. This construction of rectilinear grids yields three essential properties that will facilitate formalizing sparse grids: (1) a grid Ωl is uniquely determined by its resolution vector l, (2) grids Ωl and Ωl with different resolution vectors are disjoint, (3) the size |Ωl| = 2 l 1 of a grid is determined by the L1 norm of its resolution vector. Construction of Sparse Grids. Sparse grids use rectilinear grids as their fundamental building block [22] and exploit the fact that resolution vectors uniquely identify different rectilinear grids. Larger grids are formed as the union of rectilinear grids with different resolution vectors. Sparse grids use all rectilinear grids with resolution vector having L1 norm below a specified threshold. Formally, for a resolution index ℓ N0, the sparse grid Gℓ,d in d dimensions is Gℓ,d := S l: l 1 ℓΩl. Figure 1 illustrates the construction of the sparse grid G2,2 from smaller 2-d rectilinear grids with maximum resolution 2, i.e., {Ωl | l 1 2}. The figure also illustrates another important fact: the sparse grid Gℓ,d has a total of Θ(2ℓ) distinct and equally spaced coordinates in each dimension, but many fewer total points than a dense d-fold Cartesian product of such 1-d grids, which would have Θ(2ℓd) points. Sparse grids have a number of formal properties that are useful in algorithms and applications [22, 7]. Proposition 1 below summarizes the most relevant ones for our work. For completeness, a proof appears in appendix A. For more details, see Valentin [25]. Proposition 1 (Properties of Sparse Grid). Let Gℓ,d [0, 1]d be a sparse grid with any resolution ℓ N0 and dimension d N. Then the following properties hold: (P1) |Gd ℓ| = O(2ℓℓd 1), (P2) ℓ N, 0 ℓ ℓ = Gℓ ,d Gℓ,d, (P3) Gℓ,d = [ ℓ i=0 Ωi Gℓ i,d 1 and Gℓ,1 = [ ℓ i=0Ωi. Property P1 shows that the size of sparse grid with Θ(2ℓ) points in each dimension grows more slowly than a dense grid with the same number of points in each dimension, since O(2ℓℓd 1) O(2ℓd). Properties P2 and P3 are consequences of the structure of the set { l 1 ℓ} and the sparse grid construction. Property P2 says that a sparse grid with smaller resolution is contained in one with higher resolution. Property P3 is a crucial property, and says that a d-dimensional sparse grid can be constructed via Cartesian products of 1-dimensional dense grids with sparse grids in d 1 dimensions. 3 Structured Kernel Interpolation on Sparse Grids To scale kernel interpolation to higher dimensions, we propose to select inducing points U = Gℓ,d on a sparse grid and approximate the kernel matrix as WKGℓ,d W T for a suitable interpolation matrix W adapted to sparse grids. This will require fast matrix-vector multiplications with the sparse grid kernel matrix KGℓ,d and the interpolation matrix W. We show how to accomplish these two tasks in Sections 3.1 and 3.2 for the important case of stationary product kernels [10]. 3.1 Fast Multiplication with the Sparse Grid Kernel Matrix Algorithm 1 is an algorithm to compute KGℓ,dv for any vector v. The algorithm uses the following definitions. For any finite set U, let KU = k(x, x ) x,x U. The rows and columns of KU are U-indexed , meaning the entries correspond to elements of U under some arbitrary fixed ordering. 2Suppose i/2l+1 = j/2k+1 are both grid points and k > l. Then i = j2k l is even, a contradiction. For U V , we introduce a selection matrix SU,V to map between U-indexed and V -indexed vectors. It has entries (SU,V )ij equal to one if the ith element of U is equal to the jth element of V , and zero otherwise. Also, SV,U := ST U,V . For a V -indexed vector z V , the multiplication SU,V z V produces a U-indexed vector by selecting entries corresponding to elements in U, and for a U-indexed vector z U, the multiplication SV,Uz U produces a V -indexed vector by inserting zeros for elements not in U. Algorithm 1 Sparse Grid Kernel-MVM Algorithm Input: v R|Gℓ,d| and KGℓ,d R|Gℓ,d| |Gℓ,d| Output: u = mvm KGℓ,d, v , where, mvm (K, v) denotes Kv obtained using this algorithm. 1: Let Vi be the result of reshaping SΩi Gℓ i,d 1,Gℓ,d v into a |Ωi| |Gℓ i,d 1| matrix; this contains entries of v corresponding to the ith grid in the decomposition of P3. 2: if d = 1 then 3: return u = KGℓ,1v Base case 4: end if 5: Pre-computation 6: for i = 0 to ℓdo 7: Ai = KGi,1SGi,1,Ωi Vi 8: B T i = mvm KGℓ i,d 1, V T i Recursively multiply KGℓ i,d 1 with columns of V T i . 9: end for 10: Main loop 11: for i = 0 to ℓdo 12: AT i = mvm j>i SΩi,Gj,1Aj SGℓ j,d 1,Gℓ i,d 1 Recurse like line 8. 13: Bi = SΩi,Gi,1KGi,1 j i SGi,1,Ωj Bj SGℓ j,d 1,Gℓ i,d 1 14: ui = vec [Ai] + vec [Bi] 15: end for Theorem 1. Let KGℓ,d be the kernel matrix for a d-dimensional sparse grid with resolution ℓ for a stationary product kernel. For any v R|Gℓ,d|, Algorithm 1 computes KGℓ,dv in O(ℓd2ℓ) time. The formal analysis and proof of Theorem 1 appears in Appendix B. The running time of O(ℓd2ℓ) is nearly linear in |Gℓ,d| and much faster asymptotically than the naive MVM algorithm that materializes the full matrix and has running time quadratic in |Gℓ,d|. Algorithm 1 is built on two high-level observations. First, the decomposition of Property P3 from Proposition 1 and the fact that the kernel follows product structure across dimensions are used to decompose the MVM into blocks, each of which is between sub-grids which are the product of a 1-dimensional rectilinear grid and a sparse grid in d 1 dimensions. Therefore, the overall MVM computation can be recursively decomposed into MVMs with sparse grid kernel matrices in d 1 dimensions. This observation is in part inspired by Zeiser [28], which also decomposes computation with the matrix on sparse grids by the resolution of first dimension. The base case occurs when d = 1. We assume that Toeplitz structure, which arises due to the kernel being stationary, is leveraged to perform this base case MVM in O(ℓ2ℓ) time. Algorithm 1 can also be extended to non-stationary product kernels by using a standard MVM routine for the base case, which changes the overall running-time analysis but is still more efficient than the naive algorithm. Secondly, by Property P2, the kernel matrix multiplication for any grid of resolution ℓalso includes the result of the multiplication for grids of lower resolution and the same number of dimensions. Thus, the results of the multiplications for many individual blocks can be obtained by using the appropriate selection operators with the result of the multiplication with the kernel matrix KGℓ i,d 1 in Line 12 of the algorithm. Further intuition and explanation are provided in the Appendix B.1. Improving batching efficiency. The recursions in Lines 8 and 12 can be batched for efficiency, since both are multiplications with the same symmetric kernel matrix KGℓ,d. Similarly, the recursion spawns many recursive multiplications with kernel matrices of the form Gℓ ,d for 0 ℓ < ℓand 1 d < d, and the calculation can be reorganized to batch all multiplications with each KGℓ ,d . This is a significant savings, because there are only d(ℓ+1) distinct kernel matrices, but the recursion has a branching factor of (ℓ+ 1), so spawns many recursive calls with the same kernel matrices. 3.2 Sparse Interpolation For Sparse Grids We now seek to construct the matrix W, which interpolates function values from the sparse grid Gℓ,d to training points xi Rd, while ensuring that each row of W is sparse enough to preserve efficiency of matrix-vector multiplications with W. This requires a sparse interpolation rule for sparse grids. To set up the problem, we consider interpolating a function f observed at points in a generic set U. Let f U = (f(x))u U. A linear interpolation rule for U is a mapping x 7 wx R|U| used to approximate f(x) at an arbitrary point as f(x) w T x f U. The density of an interpolation rule is the maximum number of non-zeros in wx for any x. The combination technique for sparse grids constructs an interpolation rule by combining interpolation rules for the constituent rectilinear grids. Proposition 2. For each l, let wl x be an interpolation rule for the rectilinear grid Ωl with maximum density c. The combination technique gives an interpolation rule wx for the sparse grid Gℓ,d with density at most c ℓ+d 1 d 1 . The factor ℓ+d 1 d 1 is the number of rectilinear grids included in Gℓ,d. We use the combination technique to construct the sparse-grid interpolation coefficients wxi for each training point xi and stack them in the rows of W, which can be done in time proportional to the number of nonzeros. Details are given in Appendix A. The combination technique can use any base interpolation rule for the rectilinear grids, such as multilinear, cubic, or simplicial interpolation. 3.3 Simplicial Interpolation The density of the interpolation rule is a critical consideration for kernel interpolation techniques with or without sparse grids. Linear and cubic interpolation in d dimensions have density Θ(2d) and Θ(4d), respectively. This second curse of dimensionality makes computations with W intractable in higher dimensions independently of operations with the grid kernel matrix. For sparse grids, the density of W increases by an additional factor of ℓ+d 1 d 1 . We propose to use simplicial interpolation [11] for the underlying interpolation rule to avoid exponential growth of the density. Simplicial interpolation refers to a scheme where Rd is partitioned into simplices and a point x is interpolated using only the d+1 extreme points of the enclosing simplex, so the density of the interpolation rule is exactly d + 1. Simplicial interpolation was previously proposed for sparse grid classifiers in [8]. In work closely related to ours, Kapoor et al. [13] used simplicial interpolation for GP kernel interpolation, with the key difference that they use the permutohedral lattice as the underlying grid, which has a number of nice properties but does not come equipped with fast specialized routines for kernel matrix multiplication. In contrast to Kapoor et al. [13], we maintain rectilinear and/or sparse underlying grids, which preserve structure that enables fast kernel matrix multiplication. For rectilinear grids, this requires partitioning each hyper-rectangle into simplices, so the entire space is partitioned by simplices whose extreme points belong to the rectilinear grid. Then, within each simplex, the values at the extreme points are interpolated linearly. In general, there are different ways to partition hyper-rectangles into simplices we use the specific scheme detailed in [11]. For sparse grids, we then use the combination technique, leading to overall density of (d + 1) ℓ+d 1 d 1 . Table 1 provides the MVM complexities of W and kernel matrices for different interpolation schemes and both grids. More details on how to perform simplicial interpolation with rectilinear and sparse grids are given in the Appendix B.2. Table 1: The MVM complexities (i.e., mvm( )) of the interpolation matrix W and the kernel matrix KG for different interpolation bases and d-dimensional grids with 2ℓunique points in each dimension and n data points. A star indicates approaches proposed in this work. Dense denotes a rectilinear grid. Grid Basis mvm(W) mvm(KG) Cubic O(n 4d) O(ℓ d 2ℓ d) Linear O(n 2d) O(ℓ d 2ℓ d) Simplicial O(n d2) O(ℓ d 2ℓ d) Cubic O(n 4d ℓ+d 1 d 1 ) O(ℓd 2ℓ) Linear O(n 2d ℓ+d 1 d 1 ) O(ℓd 2ℓ) Simplicial O(n d2 ℓ+d 1 d 1 ) O(ℓd 2ℓ) 4 Experiments In this section, we empirically evaluate the time and memory taken by Algorithm 1 for matrix-vector multiplication with the sparse grid kernel matrix, the accuracy of sparse grid interpolation and GP regression as the data dimension d increases, and the accuracy of sparse grid kernel interpolation for GP regression on real higher-dimensional datasets from UCI. Hyper-parameters, data processing steps, and optimization details are given in Appendix C.1. 102 104 106 |G| MVM time (in secs) Sparse-MVM Ours-MVM-I Ours-MVM-R 102 104 106 |G| Build time (in secs) 102 104 106 |G| Memory (in GB) 102 104 106 |G| CG time (in secs) Figure 2: Matrix-vector multiplication (MVM) resource usage relative to sparse grid size for d = 6 and increasing resolution ℓ. From top-left to bottom-right: time for one MVM operation, build (pre-processing) time for the kernel matrix, peak memory usage, and typical time to solve a linear system using CG (time for build plus 50 MVMs). Each plot shows the naive quadratic MVM algorithm (Sparse-MVM), the recursive implementation of Algorithm 1 (Ours-MVM-R), and the efficiently batched iterative implementation of Algorithm 1 (Ours MVM-I). All measurements are averaged over 8 trials; error bars represent twice the standard error. Sparse grid kernel MVM complexity. First, we evaluate the efficiency of MVM algorithms. We compare the basic and efficient implementations of Algorithm 1 to the naive algorithm, which constructs the full kernel matrix and scales quadratically with the sparse grid size. Algorithm 1 has a significant theoretical advantage in terms of both time and memory requirements as the grid size grows. Figure 2 illustrates this for d = 6 by comparing MVM time and memory requirements for sparse grids with resolutions ℓ {2, . . . , 9} (roughly 100 to 1M grid points). The MVM time, preprocessing time, and memory consumption of Algorithm 1 all grow more slowly than the naive algorithm, and the efficient implementation of Algorithm 1 is faster for |G| larger than about 104, after which the naive algorithm also exceeds the 10 GB memory limit. For comparison, at ℓ= 6 (about 40K grid points), Algorithm 1 uses only 0.05 GB of memory. These results indicate that the proposed algorithm is crucial for enabling sparse grid kernel interpolation in higher dimensions. Figure 2 also depicts the typical time to run GP inference (i.e., preprocessing time plus 50 MVM operations). Sparse grid interpolation and GP inference accuracy on synthetic data. A significant advantage of sparse grids over dense rectilinear grids is their ability to perform accurate interpolation in higher dimensions. We demonstrate this for d = 6 by interpolating the function f(x) = cos( x 1) from observation locations on sparse grids of increasing resolution onto 200 random points sampled uniformly from [0, 1]d. Figure 3, left, shows the interpolation error for dense and sparse grids with both cubic and simplicial interpolation. Sparse (cubic) is significantly more accurate than Dense (cubic), and Sparse (simplicial) is more accurate than Dense (simplicial). We next evaluate the accuracy of GP inference in increasing dimensions. We keep the same function f(x) and generate observations as y = f(x) + N(0, 0.05). For all d, we use ℓ= 4 for the sparse grid and compare to the dense grid with the closest possible total number of grid points (i.e., G1/d 4,d points in each dimension)3. For d 8, performance is better with sparse grids than dense grids for both interpolation schemes. Remarkably, our proposal to use simplicial interpolation with dense grids allows SKI to scale to d = 10, which is a significant improvement over prior work, in which SKI is typically infeasible for d 4. 102 103 104 105 |G| Interpolation Error Dense (cubic) Sparse (cubic) Dense (simplicial) Sparse (simplicial) 2 4 6 8 10 d Dense (cubic) Sparse (cubic) Dense (simplicial) Sparse (simplicial) Figure 3: Comparing dense and sparse grids with cubic and simplicial interpolation schemes on synthetic data. Dense (cubic/simplicial) are dense grid methods with cubic/simplicial interpolation schemes; similarly, sparse (cubic/simplicial) are for sparse grids. Left: Function interpolation error vs. the grid size for d = 6. Right: Test root-mean-square error (RMSE) for GP regression for increasing dimensions. For both tasks, sparse grid methods outperform dense grids. For d = 10, both methods with cubic interpolation run out of GPU memory, which is 48 GB for this experiment. GP regression performance on UCI datasets. To evaluate the effectiveness of our proposed methods for scaling GP kernel interpolation to higher dimensions, we consider all UCI [4] data sets with dimension 8 d 10. We compare our proposed methods, Dense-grid (dense SKI with simplicial interpolation) and Sparse-grid (sparse-grid SKI with simplicial interpolation) to SGPR [24], SKIP [10], and Simplex-GP [13]. For SGPR, we report the best results using 256 or 512 inducing points. For SKIP, 100 points per dimension are used. For Simplex-GP [13], the blur stencil order is set to 1. Table 3 shows the root mean squared error (RMSE) for all methods. Our methods have performance comparable to and often better than SGPR and prior methods for scaling kernel interpolation to higher dimensions. This shows that sparse grids and simplicial interpolation can effectively scale SKI to higher dimensions to give a GP regression framework that is competitive with state-of-the-art approaches. We report additional results and analysis in Appendix C. 3We tried to match the sizes of both dense and sparse grids while ensuring that the dense grid always had at least as many points as the sparse grid, to give a fair comparison. Precisely, (d, dense grid size, sparse grid size) tuples are (2, 144, 129) , (4, 1296, 796), (6, 4096, 2561), (8, 6561, 6401), (10, 59049, 13441). Table 2: Test root-mean-square-error (RMSE) on UCI regression datasets with dimensions (i.e., 8 d 10). See text for algorithm descriptions and settings. All mean and standard deviations are computed over 3 trials. Pendulum is abbreviated as Pendu to accommodate table horizontally. Datasets (d) SGPR SKIP Simplex-GP Dense-grid Sparse-grid Energy (8) 1.509 0.003 5.762 0.000 3.076 0.012 1.333 0.009 0.715 0.004 Energy (8) 1.509 0.003 5.762 0.000 3.076 0.012 1.333 0.009 0.715 0.004 Concrete (8) 12.727 0.018 12.727 0.001 12.727 0.000 12.191 0.001 8.655 0.002 Kin40k (8) 0.168 0.009 0.174 0.001 0.287 0.003 0.205 0.003 0.483 0.000 Fertility (9) 0.197 0.016 0.183 0.000 0.187 0.001 0.182 0.000 0.194 0.002 Pendu (9) 1.948 0.021 2.947 0.000 2.577 0.009 2.053 0.010 2.103 0.015 Protein (9) 0.605 0.001 0.778 0.000 0.582 0.018 0.736 0.002 0.595 0.001 Solar (10) 0.790 0.026 0.780 0.002 0.792 0.000 0.775 0.002 0.748 0.006 5 Related Works Beyond SKI and its variants, a number scalable GP approximations have been investigated. Most notable are the different variants of sparse GP approximations [26, 23, 18]. For m inducing points, these methods require either Ω(nm2) time for direct solves or Ω(nm) time for approximate kernel MVMs in iterative solvers. While these methods generally do not leverage structured matrix algebra like the SKI framework, and thus have worse scaling in terms of the number of inducing points m, they may achieve comparable accuracy with smaller m, especially in higher dimensions. Some work utilizes the SKI framework to further boost the performance of sparse GP approximations [12]. A closely related work to ours is on improving the dimension scaling of SKI through the use of low-rank approximation and product structure [10]. In another very closely related work, Kapoor et al. [13] recently proposed to scale SKI to higher dimensions via interpolation on a permutohedral lattice. Like our work, they use simplicial interpolation. Unlike our work, the kernel matrix on the permutohedral lattice does not have special structure that admits fast exact multiplications; they instead use a locality-based approximation that takes into account the length scale of the kernel function by only considering pairs of grid points within a certain distance. Another related direction of research is the adaptation of sparse grid techniques for machine learning problems, e.g., classification and regression [16] and data mining [1]. These methods construct feature representations using sparse grid points [3], often by selecting a subset of grid points relevant for the dataset [6, 3]. 6 Discussion This work demonstrates that two classic numerical techniques, namely, sparse grids and simplicial interpolation, can be used to scale GP kernel interpolation to higher dimensions. SKI with sparse grids and simplicial interpolation has better or competitive regression accuracy compared to state-of-the-art GP regression approaches on several UCI benchmarking datasets with 8 to 10 dimensions. Limitations and future work. Sparse grids and simplicial interpolation address two important bottlenecks when scaling kernel interpolation to higher dimensions. Sparse grids allow scalable matrix-vector multiplications with the grid kernel matrix, and simplicial interpolation allows scalable multiplications with the interpolation matrix W. The relatively large number of rectilinear grids used to form a sparse grid i.e., the factor of ℓ+d 1 d 1 in Proposition 2 is one limiting factor that makes multiplication by W more costly. Future research could investigate methods to mitigate this extra cost, and explore the limits of scaling to even higher dimensions with sparse grids. Acknowledgments This material is based upon work supported by the National Science Foundation under Grant Nos. 1661259, 1749854, 2046235 and 1763618, along with Adobe and Google Research Grants. We thank Javier Burroni, Abhinav Agarwal, Shiv Shankar and Shreyas Chaudhari for their feedback. [1] H-J Bungartz, Dirk Pflüger, and Stefan Zimmer. Adaptive sparse grid techniques for data mining. In Modeling, simulation and optimization of complex processes, pages 121 130. Springer, 2008. [2] Hans-Joachim Bungartz and Michael Griebel. Sparse grids. Acta Numerica, 13:147 269, 2004. [3] Tri Dao, Christopher De Sa, and Christopher Ré. Gaussian quadrature for kernel features. In Advances in Neural Information Processing Systems 30 (Neur IPS), page 6109 6119, 2017. [4] Dheeru Dua and Casey Graff. Uci machine learning repository, 2017. URL http://archive. ics.uci.edu/ml. [5] A. Dutt, Miao Gu, and Vladimir Rokhlin. Fast algorithms for polynomial interpolation, integration, and differentiation. SIAM Journal on Numerical Analysis, 33:1689 1711, 1996. [6] Jochen Garcke. A dimension adaptive sparse grid combination technique for machine learning. Anziam Journal, 48:C725 C740, 2006. [7] Jochen Garcke. Sparse grids in a nutshell. In Sparse grids and applications, pages 57 80. Springer, 2012. [8] Jochen Garcke and Michael Griebel. Classification with sparse grids using simplicial basis functions. Intelligent data analysis, 6(6):483 502, 2002. [9] Jacob R. Gardner, Geoff Pleiss, David Bindel, Kilian Q. Weinberger, and Andrew Gordon Wilson. GPy Torch: Blackbox matrix-matrix Gaussian process inference with GPU acceleration. In Advances in Neural Information Processing Systems 31 (Neur IPS), pages 7576 7586, 2018. [10] Jacob R Gardner, Geoff Pleiss, Ruihan Wu, Kilian Q Weinberger, and Andrew Gordon Wilson. Product kernel interpolation for scalable Gaussian processes. ar Xiv:1802.08903, 2018. [11] John H Halton. Simplicial multivariable linear interpolation. Tehcnical Report 91-002, University of North Carolina at Chapel Hill Department of Computer Science, 1991. [12] Pavel Izmailov, Alexander Novikov, and Dmitry Kropotov. Scalable gaussian processes with billions of inducing inputs via tensor train decomposition. In In Proceedings of the 21st International Conference on Artificial Intelligence and Statistics (AISTATS), volume 84 of Proceedings of Machine Learning Research, pages 726 735. PMLR, 2018. [13] Sanyam Kapoor, Marc Finzi, Ke Alexander Wang, and Andrew Gordon Gordon Wilson. Skiing on simplices: Kernel interpolation on the permutohedral lattice for scalable gaussian processes. In Proceedings of the 38th International Conference on Machine Learning (ICML), pages 5279 5289, 2021. [14] Radford M. Neal. Bayesian Learning for Neural Networks. Springer-Verlag, 1996. [15] Michael Obersteiner and Hans-Joachim Bungartz. A generalized spatially adaptive sparse grid combination technique with dimension-wise refinement. SIAM Journal on Scientific Computing, 43(4):A2381 A2403, 2021. [16] Dirk Michael Pflüger. Spatially adaptive sparse grids for high-dimensional problems. Ph D thesis, Technische Universität München, 2010. [17] George M Phillips and Peter J Taylor. Theory and applications of numerical analysis. Elsevier, 1996. [18] Joaquin Quiñonero-Candela and Carl Edward Rasmussen. A unifying view of sparse approximate Gaussian process regression. Journal of Machine Learning Research, 6(Dec):1939 1959, 2005. [19] Ali Rahimi and Benjamin Recht. Random features for large-scale kernel machines. In Advances in Neural Information Processing Systems 20 (Neur IPS), pages 1177 1184, 2007. [20] Carl Edward Rasmussen. Gaussian Processes in Machine Learning. Springer, 2004. [21] Winfried Sickel and Tino Ullrich. Spline interpolation on sparse grids. Applicable Analysis, 90 (3-4):337 383, 2011. [22] Sergei Abramovich Smolyak. Quadrature and interpolation formulas for tensor products of certain classes of functions. In Doklady Akademii Nauk, volume 148, pages 1042 1045. Russian Academy of Sciences, 1963. [23] Edward Snelson and Zoubin Ghahramani. Sparse Gaussian processes using pseudo-inputs. In Advances in Neural Information Processing Systems 18 (Neur IPS), pages 1257 1264, 2005. [24] Michalis Titsias. Variational learning of inducing variables in sparse gaussian processes. In In Proceedings of the 12th International Conference on Artificial Intelligence and Statistics (AISTATS), pages 567 574. PMLR, 2009. [25] Julian Valentin. B-splines for sparse grids: Algorithms and application to higher-dimensional optimization. ar Xiv:1910.05379, 2019. [26] Christopher KI Williams and Matthias Seeger. Using the Nyström method to speed up kernel machines. In Advances in Neural Information Processing Systems 14 (Neur IPS), pages 682 688, 2001. [27] Andrew Wilson and Hannes Nickisch. Kernel interpolation for scalable structured Gaussian processes (KISS-GP). In Proceedings of the 32nd International Conference on Machine Learning (ICML), pages 1775 1784, 2015. [28] Andreas Zeiser. Fast matrix-vector multiplication in the sparse-grid galerkin method. Journal of Scientific Computing, 47(3):328 346, 2011. 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] (b) Did you describe the limitations of your work? [Yes] See Section 6. (c) Did you discuss any potential negative societal impacts of your work? [N/A] As our method is an improvement on existing methods therefore it doesn t have any direct negative consquences. (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [Yes] (b) Did you include complete proofs of all theoretical results? [Yes] 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] See Section C. (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] See Section C.1. (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [Yes] See Figures 2 and 3. (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] See C. 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [Yes] (b) Did you mention the license of the assets? [Yes] (c) Did you include any new assets either in the supplemental material or as a URL? [Yes] See Section 1. (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [N/A] We use only publicly available datasets. (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [N/A] We use only publicly available datasets. 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]