# target_alignment_in_truncated_kernel_ridge_regression__717ecae3.pdf Target alignment in truncated kernel ridge regression Arash A. Amini1, Richard Baumgartner2, Dai Feng3 1University of California, Los Angeles 2Merck & Co., Inc., Rahway, New Jersey, USA 3Data and Statistical Sciences, Abb Vie Inc. aaamini@ucla.edu, richard_baumgartner@merck.com, dai.feng@abbvie.com Kernel ridge regression (KRR) has recently attracted renewed interest due to its potential for explaining the transient effects, such as double descent, that emerge during neural network training. In this work, we study how the alignment between the target function and the kernel affects the performance of the KRR. We focus on the truncated KRR (TKRR) which utilizes an additional parameter that controls the spectral truncation of the kernel matrix. We show that for polynomial alignment, there is an over-aligned regime, in which TKRR can achieve a faster rate than what is achievable by full KRR. The rate of TKRR can improve all the way to the parametric rate, while that of full KRR is capped at a sub-optimal value. This shows that target alignment can be better leveraged by utilizing spectral truncation in kernel methods. We also consider the bandlimited alignment setting and show that the regularization surface of TKRR can exhibit transient effects including multiple descent and non-monotonic behavior. Our results show that there is a strong and quantifiable relation between the shape of the alignment spectrum and the generalization performance of kernel methods, both in terms of rates and in finite samples. 1 Introduction Kernel methods have become a time-proven popular mainstay in machine learning [23, 27, 22]. Implicit transformation of a learning problem via a suitable kernel with subsequent development of regularized linear models in the associated reproducing kernel Hilbert space (RKHS) is amenable to applications and allow for investigation of theoretical properties of the kernel methods [12]. Recently, there has been a revived interest and increased research into the kernel methods [14, 17, 7, 12, 8]. These efforts have been motivated by the connections between neural network and kernel learning [14]. In particular, the discovery of transient effects such as double descent that emerge in neural networks training [6, 20] has attracted a lot of attention. In this paper, we study the effect of target-model alignment on the performance of kernel ridge regression (KRR) estimators. KRR is a well-known nonparametric regression estimator with desirable theoretical performance guarantees. We focus on a generalization of the KRR, which we refer to as the Truncated Kernel Ridge Regression (TKRR), that introduces an additional parameter, r, for truncating the eigenvalues of the underlying kernel matrix. TKRR is often used as an approximation to the full KRR to scale the computations to large sample sizes. TKRR itself is often approximated in practice by various sampling and sketching schemes [30, 33, 18, 19, 24, 1, 32]. One of our main contributions in this paper is that TKRR can, in fact, improve the statistical rate of convergence in cases where the target function is well-aligned with the underlying RKHS. Roughly speaking, target 36th Conference on Neural Information Processing Systems (Neur IPS 2022). alignment refers to the decay rate of the coefficients of the target function in an orthonormal basis of the eigenfunctions of the kernel operator. In addition to improved rates for TKRR, we also show that target alignment generally predicts the generalization performance of the full KRR as well as TKRR. We also explore the regularization surface of the KRR estimators in both the regularization and truncation parameters, and show some surprising behavior like non-monotonicity, multiple descent and phase transition phenomena. Our main contributions (i) We show that TKRR itself can be viewed as a full KRR estimator in a smaller RKHS embedded in the original RKHS (Propostion 1). This shows that the ad-hoc truncation that is often done for computational reasons can itself be viewed as a proper KRR estimator over a different RKHS. (ii) We introduce target alignment (TA) scores in Section 4 and derive an exact expression for the Mean Sqaured Error (MSE) of TKRR based on the TA scores and eigenvalues of the kernel matrix (Theorem 1). (iii) For the case of polynomially-decaying TA scores, we first derive a simplified expression for the MSE (Eqn. (14)) which is within constant factors of the exact expression. We then identify four target alignment regimes: Under-aligned, just-aligned, weakly-aligned and over-aligned. We show that TKRR can achieve a fast rate of convergence in the over-aligned regime. In contrast, the rate of the full KRR is capped to a sub-optimal level, that of the best achievable in the weakly-aligned regime (Theorem 2). (iv) For bandlimited TA scores, we provide more refined finite-sample analysis. In particular, we show that the regularization curve as a function of the truncation level r is in general non-monotonic. We also show that TA spectra that are narrower and shifted towards lower indices correspond to lower generalization errors (Proposition 2). We provide experiments verifying the multiple-descent and non-monotonic behavior of the regularization curves as well as the improved rate of Theorem 2 (Section 4.2). The MSE in our results, is equivalent to the excess (empirical) generalization error. So all the conclusions, and in particular the rates, are valid for the generalization error. Overall, our results provide strong theoretical and experimental support for the idea that alignment of the target function with the eigenfunctions of the kernel is the key predictor of generalization performance of KRR estimators. It is worth noting that our main theoretical results are non-asymptotic and at times exact. Related work. Kernel ridge regression (KRR) has been widely studied in the literature in the non-parametric statistical framework [29, 26]. Various aspects of the KRR were elucidated. For example, the error rates of the KRR were studied in [9] and more recently benign overfitting [5] and related double descent effects in [13]. The notion of kernel-target alignment was first introduced in [11] for classification. In [8], the generalization error of the KRR was investigated using the tools of statistical physics theory which allows for derivation of the bias-variance decomposition of the generalization error and elucidation of non-monotonic transitional effects reflected by the learning curves. In this work the authors have also demonstrated that the interplay between spectral bias (expressed as kernel eigenvalue decay) and kernel-target alignment or task-model alignment are useful for the characterization of the kernel compatibility with the problem. Building on the [8], the concept of kernel-target alignment and spectral bias was referred to as source and capacity, respectively in [12]. Similarly to [8], the transitions that emerge during a crossover from noiseless to noisy regime corresponding to fast vs slow decay of the learning curve, respectively were studied in [12]. Moreover, in [2] it has been shown that the truncated KRR outperforms full KRR over the unit ball of the RKHS. Another related work is that of [25] where overparametrized KRR is studied. The work [12] is perhaps closest to ours followed by [2] and [25]. We make more detailed comparisons with these after Theorem 2. In particular, we show that the rate obtained in [12] can be improved by using TKRR. In addition, in Appendix D, we provide a more unified view of the settings of [9, 12] which can help clarify how different target alignment assumptions relate to each other. It is worth noting that approximate forms of TKRR via sampling and sketching have been analyzed extensively in the literature [21, 32, 10, 31, 15, 1, 4], but this line of work at best proves that TKRR achieves the same minimax rate achievable with full KRR, over the RKHS ball. What we show in this paper is that TKRR can achieve much faster rates under proper target alignment. 2 Kernel Ridge Regression The general nonparametric regression problem can be stated as yi = f (xi) + εi, i = 1, . . . , n, E(ε) = 0, cov(ε) = σ2In (1) where ε = (εi) Rn is a noise vector and f : X R is the function of interest to be approximated from the noisy observations y = (yi). Here, X is the space to which the covariates x = (xi) belong. We mainly focus on the fixed design regression where the covariates are assumed to be deterministic. A natural estimator is the kernel ridge regression (KRR), defined as the solution of the following optimization problem: bfn,λ := argmin f H i=1 (yi f(xi))2 + λ f 2 H, (2) where λ > 0 is a regularization parameter and H is the underlying RKHS. By the representer theorem [16], this problem can be reduced to a finite-dimensional one: min ω Rn 1 n y n Kω 2 + λωT Kω, where K = 1 n K(xi, xj) Rn n (3) is the (normalized empirical) kernel matrix. Let us define the following operators Sx(f) = 1 n(f(x1), f(x2), . . . , f(xn)), S x(ω) = 1 n j=1 ωj K( , xj) (4) for f H and ω Rn. We refer to Sx as the sampling operator. As the notation suggests, S x is the adjoint of Sx as an operator from H to Rn. We have bfn,λ = S x(bω) where bω is the solution of (3). We note that Sx(S x(ω)) = Kω, i.e., K is the matrix representation of Sx S x : Rn Rn. Notation. We write f, g n = 1 n Pn i=1 f(xi)g(xi) and f n = p f, f n for the empirical inner product and norm, respectively. Note that f n = Sx(f) 2 where 2 is the vector ℓ2 norm. For two functions, f and g, we write f g if there is constant C > 0 such that f Cg, and we write f g if f g and g f. 3 Truncated Kernel Ridge Regression As mentioned earlier, spectral truncation of the kernel matrix has been mainly considered as a computational device in approximating the full KRR estimator. Here, we first argue that it can be reformulated as a KRR estimator over a simpler RKHS. To facilitate future developments, we define this new RKHS based on the given training data x = (x1, . . . , xn). Let K = Pn k=1 µkuku T k be the eigendecomposition of the kernel matrix K where µ1 µ2 µn 0 are the eigenvalues of K and {uk} Rn are the corresponding eigenvectors. We assume for simplicity that µn > 0, that is: Assumption 1. The kernel matrix K = K(x) is invertible. This is a very mild assumption for any infinite-dimensional H. Let uki be the ith coordinate of uk. We would like to find functions {ψk} H such that ψk(xi) = n uki for all k, i [n]. If dim(H) = , there are in general many such functions for each k. We pick the one that minimizes the H-norm: ψk := argmin ψ H : ψ H, Sx(ψ) = uk . Under Assumption 1, the above problem has a unique solution (i.e., the interpolation problem Sx(ψ) = u has a unique minimum-norm solution for any u Rn.). Since {uk} is an orthonormal basis of Rn, by construction, ψk, ψℓ n = 1{k = ℓ}. Consider the set of functions e H := n r X k=1 αkψk | α1, . . . , αr R o equipped with the inner product defined via ψk, ψℓ e H = 1 µk 1{k = ℓ} and extended to the whole of e H by bilinearity. Equivalently, if f = Pr k=1 αkψk and g = Pr k=1 βkψk, define f, g e H = P k αkβk/µk. Then, e H is an r-dimensional RKHS, with reproducing kernel (Appendix A) e K(x, y) := k=1 µkψk(x)ψk(y) (5) and by construction e H H, although e H and H are equipped with different inner products. 3.1 TKRR is itself a valid KRR We define the truncated kernel ridge regression (TKRR) estimator as the usual KRR estimator over e H, that is, efr,λ = argmin f e H i=1 (yi f(xi))2 + λ f 2 e H. (6) We have the following characterization of efr,λ: Proposition 1. Let e K = Pr k=1 µkuku T k = ( 1 n e K(xi, xj)) be the truncated kernel matrix and e S x(ω) := 1 n j=1 ωj e K( , xj), (7) which is the adjoint of Sx as a map from e H to Rn. Let eΩbe the solution set of (3) with K replaced with e K. Then, (6) has a unique solution which is given by efr,λ = e S x(ω) for any ω eΩ. Although the above proposition is not the main contribution of this work, it provides the interesting observation that TKRR is an exact KRR estimator over a smaller RKHS. See also (21) in Appendix B for a closed-form expression for efr,λ. 4 Target alignment We start with the definition of target alignment: Definition 1. The alignment spectrum of a target function f is ξ = U T Sx(f ) Rn where U is the orthogonal matrix of the eigenvectors of K, with columns corresponding to eigenvalues of K in descending order. The elements of ξ are referred to as target alignment (TA) scores. The following theorem gives an exact expression for the expected empirical MSE of TKRR estimate, in terms of the TA scores ξ and the eigenvalues (µi) of the kernel matrix: Theorem 1 (Exact MSE). Let Γλ be an n n diagonal matrix with diagonal elements: (Γλ)ii = µi µi + λ1{1 i r}. (8) For any TKRR solution efr,λ, we have E efr,λ f 2 n = (In Γλ)ξ 2 2 + σ2 (µi + λ)2 (ξ i )2 + i=r+1 (ξ i )2 + σ2 µ2 i (µi + λ)2 (9) 1 (µi + λ)2 h ai(λ)(ξ i )2 + σ2 n µ2 i i (10) where ai(λ) = (µi + λ)2 λ2 and the expectation is w.r.t. the randomness in the noise vector w. Assume that the target function is normalized so that f n = 1. Since U is an orthogonal matrix, this is equivalent to ξ 2 = 1. Since the coefficient of (ξ i )2 in (10) is negative, the more the alignment spectrum peaks near the lower indices, the smaller the expected MSE. This observation is made more precise in Proposition 2 below. 4.1 Bandlimited alignment To see the implications of Theorem 1, let us first consider a bandlimited model as follows: f is randomly generated with alignment scores ξ i satisfying E(ξ i )2 = 1 b 1{ℓ+ 1 i ℓ+ b} (11) for ℓ 0 and b 1 (both integers). Then, we write f Bb,ℓand note that E f 2 n = 1 for such f . We write MSE := E efr,λ f 2 n where the expectation is over the randomness in both f and w the noise vector in (1). One can think of this setting as a Bayesian model for f and of MSE as the Bayes risk w.r.t. to the aforementioned prior on f . We leave the specification of the prior open subject to the condition (11) since this is the only assumption we need. Proposition 2. With f Bb,ℓ, we have ai(λ) (µi + λ)2 + σ2 µ2 i (µi + λ)2 . (12) Assume for simplicity that {µi} are distinct. Then, the following hold: (a) For fixed ℓ, b and λ, let j = max{i [ℓ+ 1, ℓ+ b] | 1 + 2λ/µi σ2b/n} if the set is nonempty, otherwise j = ℓ. Then, the MSE as a function of r is (i) increasing in [1, j ), (ii) decreasing in [j , ℓ+ b], and (iii) increasing in [ℓ+ b, n]. (b) For fixed r, b and λ, the MSE as a function of ℓis increasing in ℓ [0, r b]. (c) For fixed r, ℓand λ, the MSE as a function of b is increasing in b [1, r ℓ]. Part (a) of Proposition 2 shows that MSE as a function of r can be non-monotonic: go up, down and back up. This happens whenever j < ℓ+ b. Similar observations as Proposition 2(a) can be made about bandlimited signals consisting of multiple nonzero bands, e.g., f = 1 K (f 1 + + f K) where f i Bbi,ℓi with non-overlapping bands. In this case, MSE can show multiple descents and ascents (cf. Figure 2a). Part (b) shows that alignment spectra that are concentrated near lower indices are better. Part (c) shows that concentrated alignment spectra are better than diffuse ones. Effect of noise level on optimal rank r Let us explore the implications of Proposition 2(a) for the optimal truncation level r (optimal rank). Fix λ > 0. Then, for small enough noise levels (i.e., small σ2/n), we have 1 + 2λ/µℓ+1 > σ2b/n. Since i 7 µi is (strictly) decreasing, it follows that j = ℓ, and r 7 MSE is decreasing over the entire interval [ℓ, ℓ+ b]. That is, for r < ℓ+ 1, the MSE is increasing, but the moment we enter the signal band (r = ℓ+ 1), the MSE starts to decrease and keeps decreasing till we get to the end of the band (r = ℓ+ b), at which point it starts to increase again. For small enough noise levels, the descent within the band is big enough to bring the MSE below its level at r = 1; hence the optimal truncation level, in this case, is at the end of the band: r = ℓ+ b. If we increase the noise level, at some point, j starts to move to the middle of the band [ℓ, ℓ+ b]. In this case, depending on where in the band j lies, the optimal truncation level could still be r = ℓ+ b. But as we increase the noise level, there comes a point where the noise level is so high that j is very close to the end of the band, in which case, the dip in the MSE is not big enough to bring it below its r = 1 level. Then, the best truncation is just at r = 1 level. This includes the case where j = ℓ+ b in which case the MSE is monotonically increasing (very high noise levels). Similar ideas play out in the multi-band setting as shown in Figure 2a in Section 5. 4.2 Polynomial alignment We now consider the case where the TA scores decay polynomially with a rate potentially faster or slower than what is required by merely belonging to the RKHS. In particular, assume that µi i α, (ξ i )2 i 2γα 1 (13) for some γ > 0 and α 1. Let us justify the choice of decay for (ξ i )2. Asymptotically, both (ξ i )2 and µi should decay similar to the population level TA scores and eigenvalues. Then, f H, if and only if P i(ξ i )2/µi 1 (see Appendix D). Assuming a polynomial decay for (ξ i )2, it follows that that (ξ i )2 should decay faster than i α 1 for the sum to converge, that is i α 1 δ for some δ > 0. Without loss of generality, we are assuming δ = (2γ 1)α. This parameterization is chosen to be consistent with the existing literature. Thus, for a function f H, ξ satisfies (13) with γ > 1/2. The case γ 1/2 then corresponds to a target that does not belong to the RKHS. The following result characterizes the performance of full KRR and TKRR under model (13): Theorem 2. Let η = min(r, λ 1/α). Under the polynomial decay assumption (13), E efr,λ f 2 n λ2 max(1, η 2(γ 1)α) + r 2γα1{r < n} + σ2 (a) Taking λ (σ2/n)γα/(2γα+1) and r (n/σ2)1/(2γα+1), TKRR achieves the following rate E efr,λ f 2 n σ2 2γα/(2γα+1) for γ > 1. (15) (b) Assume n 2α σ2 n, and let δ := min(1, γ). Then, the best rate achievable by the full KRR is obtained for regularization choice λ (σ2/n)α/(2δα+1) and is given by E efn,λ f 2 n σ2 2δα/(2δα+1) for γ > 0. (16) Note that the exponent δ saturates to δ = 1 for all γ > 1 giving the best rate (σ2/n)2α/(2α+1) for the full KRR in those cases. In contrast, the rate in (15) achievable by TKRR improves for γ > 1 without bound, and in the limit of γ , approaches σ2/n, the best parametric rate. It is also well-known that the minimax rate over the unit ball of the RKHS with eigendecay µi i α is (σ2/n)α/(α+1). This can also be seen by letting γ 1/2 in either (15) or (16), removing any target alignment assumption beyond what is provided by the RKHS itself. To summarize, let us define the rate exponent function s(γ) := 2γα/(2γα + 1). (17) There are four regimes of target alignment, implied by Theorem 2: 1. Under-aligned regime, γ (0, 1 2): The target is not in the RKHS. The best achievable rate is (σ2/n)s(γ) which is slower than the minimax rate over the ball of the RKHS, (σ2/n)s( 1 2 ). 2. Just-aligned regime, γ = 1 2 (to be precise γ 1 2): The target is only assumed to be in the RKHS. The best rate achievable is the minimax rate over the ball of the RKHS, (σ2/n)s( 1 2 ). 3. Weakly-aligned regime, γ ( 1 2, 1]: The target is in the RKHS and more aligned with the kernel than what is implied by being in the RKHS. The best achievable rate is (σ2/n)s(γ) which is faster than the minimax rate over the ball of the RKHS, (σ2/n)s( 1 2 ). The rate is achievable by the full KRR and hence TKRR. 4. Over-aligned regime, γ > 1: The target is in the RKHS and is strongly aligned with the kernel. The best achievable rate is (σ2/n)s(γ) which is achieved by TKRR. The full KRR can only achieve the rate (σ2/n)s(1) in this case, which is the best achievable in the weakly-aligned regime. Note that the best achievable rate in the first three regimes is attainable by full KRR (hence TKRR) while for the fourth regime, only TKRR can achieve the given rate. It was shown in [9] (cf. Appendix D.2 for details) that the rate given in the weakly-aligned regime is minimax optimal over the class of targets (13) with γ (1/2, 1]. We conjecture that the minimax optimality of this rate extends to the over-aligned regime, that is, TKRR is minimax optimal for all γ > 1/2. The capped rate (16) for the full KRR is the same as the one obtained in [12, Eqn. (12)]. The rate (15) for the TKRR in the over-aligned regime is new to the best of our knowledge. The work of [2] shows some improvement in TKRR over full KRR, but since the minimax rate was considered only over the unit ball, no difference in rates was shown between TKRR and full KRR. In contrast, by considering the smaller alignment class (13), we can show the rate differences. We also note that the cross-over effects noted in [12] for different regularization regimes follow easily from the simplified formula (14), as can be seen by inspecting the various cases in the proof of part (b) of Theorem 2. To simplify the statement of Theorem 2, we have focused only on the optimal regularization regime. There are interesting parallels but also notable differences between our work and that of [25]. They consider the usual ridge regression while we work with the kernel ridge, although the results can be translated back and forth after some transformation. They use a spectral truncation parameter k, as a proof-device to obtain the tightest upper bound, whereas the truncation level r in our case is a regularization parameter in TKRR that can be tuned in practice. The bound in [25] essentially corresponds to that of the Full KRR in our paper. In fact, we also use a similar proof-device k in the proof of Theorem 2, separate from the r-truncation level of TKRR. As an important corollary, our results about TKRR cannot be deduced from those in [25]. Another point of difference is the focus in [25] which is on how the eigendecay of the kernel (or equivalently the covariance matrix) affects the MSE bound, while our focus is on how the interaction of the target alignment decay and the kernel eigendecay together affect the bound. There are indications of the effect of the target alignment in [25], in the form of the tail energy of their θ parameter, but the implications of this decay seems to have not been explored in detail compared to our work. Generalization error We have considered a fixed design so far, that is, fixed (xi). To see the connections with the generalization error, we need to consider a random design. Let (x, y) be a random test point, and let (x1, y1), . . . , (xn, yn) be i.i.d. training data, all from the same joint distribution P on (x, y). Let PX be the marginal distribution of x under P, and for a function f, write f 2 L2 = R f 2(x)d PX(x) for the corresponding (population) squared L2 norm. Recalling that y = f (x) + ε and E[ε] = 0, the generalization error of a deterministic f is E(y f(x))2 = E(f (x) f(x))2 + σ2 = f f 2 L2 + σ2, where the first expectation is taken w.r.t. the randomness in both x and y (that is, x and ε). The noise variance σ2 is the unimprovable part of the generalization error, i.e., the minimum Bayes risk. So, the excess generalization error is f f 2 L2. For large n, since the L2 norm is an integral, it can be well-approximated by an average based on the training x = (xi)n i=1, that is, 1 n Pn i=1(f(xi) f (xi))2 which is the empirical norm that we have considered in this paper. In fact, letting Bn be the expression in (10) and taking the expectation of both sides of (10) w.r.t. x, by Fubini theorem, we have E efr,λ f 2 L2 = Ex(Bn) where Ex denotes the expectation w.r.t. training x. This controls the expected excess generalization error of efr,λ. More precise high-probability bounds on efr,λ f 2 L2 could be possible, by techniques similar to those in [28, Theorem 14.12], establishing a high probability uniform control of the form f 2 L2 2 f 2 n over the unit ball of an RKHS generated by a Mercer kernel. There are, however, technical difficulties in controlling the so-called critical radius in these arguments at the level needed for the fast rates we present here. Combined with a sub-Gaussian assumption on the noise vector ε, these arguments could lead to high probability bounds on the population excess generation error similar to the results we present here. Whether such arguments are feasible is left for future work. 5 Simulations We present various simulation results to demonstrate the multiple-descent and phase transition behavior of the regularization curves, and corroborate the theoretical results. First, consider the bandlimited alignment (11) where the nonzero entries of ξ are i.i.d. draws from N(0, 1) and ξ is normalized to have unit norm. We use the Gaussian kernel e x x /2h2 in d = 4 dimensions with bandwidth h = p d/2, applied to 200 samples generated from a uniform distribution on [0, 1]d, and let r = l + b. The plots in Figure 1a show the multiple-descent and phase transition behavior of the λ-regularization curves (that is, expected MSE versus log(λ)) for different values of the noise level σ. The corresponding contour plot is shown in Figure 1b. The plots show a transition from monotonically decreasing at σ = 0 to monotonically increasing for large σ, with non-monotonic behavior for moderate values of σ. s = 3 s = 3.4 s = 4 s = 0 s = 2.4 s = 2.8 0 4 8 12 0 4 8 12 0 4 8 12 0 4 8 12 0 4 8 12 0 4 8 12 0.85 MSE vs. - log(l) l=0, b=30, r=30 (a) Panel plot MSE for different lambda and sigma values l=0, b=30, r=30 -log(lambda) (b) Contour plot Figure 1: Multiple-descent and phase transition of for λ-regularization curve: (a) Expected MSE as a function of log(λ) for different values of σ, and (b) overall contour plot of expected MSE for σ vs. log(λ). s n = 0.18 s n = 0.24 s n = 0.3 s n = 0 s n = 0.06 s n = 0.12 0 10 20 30 40 50 0 10 20 30 40 50 0 10 20 30 40 50 0 10 20 30 40 50 0 10 20 30 40 50 0 10 20 30 40 50 MSE vs. r l1 = 0, l2 = 30, b = 10, - log(l) = 11.5 (a) Panel plot 10 20 30 40 50 MSE for different r and sigma values l1=0, l2=30, b=10, -log(lambda)=11.5 sigma/sqrt(n) (b) Contour plot Figure 2: Double-descent and phase transition for r-regularization curves: (a) Expected MSE as a function of the truncation parameter r. Plots are indexed by different values of σ/ n, and the x-axis starts at r = 1. (b) Overall contour plot of expected MSE for σ/ n vs. r. Next, we consider a bandlimited alignment with two non-overlapping bands of length b, starting at indices ℓ1 + 1 and ℓ2 + 1. The nonzero entries of ξ are generated as before, followed by unit norm normalization. We fix λ and consider the r-regularization curve, that is, the plot of MSE versus truncation parameter r. The results are shown in Figure 2a for different values of σ/ n. Figure 2b shows the corresponding contour plot. The plots show the double-descent and phase transition behavior of the r-regularization curves. The double-decent behavior is consistent with Proposition 2(a) that predicts the MSE goes up as r varies between bands. Simulations corroborating Theorem 2 are provided in Appendix C. The code for reproducing the simulations is available at [3]. 6 Conclusion We presented an analysis of the TKRR in the light of recent advances that were made for the KRR. We showed the TKRR equivalence to a KRR on a (smaller) RKHS embedded in the original RKHS. We believe that this insight provides a different perspective on the TKRR and unifies it with the KRR development. We derived an exact MSE decomposition for the TKRR via TA scores and eigenvalues of the kernel matrix. When TA scores exhibit polynomial decay we identified an over-aligned regime, where the TKRR can outperform ordinary KRR. When the TA scores are bandlimited, we demonstrated the non-monotonicity of the regularization curve with respect to the level of truncation and a correspondence of narrower TA spectra that are positioned towards lower indices with lower generalization error. Although we defined the TA spectrum based on the true f , there is a very good plugin estimator for it in practice: Replacing Sx(f ) with the noisy observations y/ n in the definition of the spectrum gives a very good estimate, that is, U T y/ n is a very good estimate of U T Sx(f ). This is because each u T k y/ n will effectively average over the noise in y. Our results can be easily extended to the non-additive and heteroscedastic noise setting, where (x, y) are drawn from a general distribution, f (x) := E[y | x] and ε = y f (x). The extension is possible if we assume var(ε | x) σ2, in which case, Theorem 1, for example, holds as an upper bound on the MSE. However, extensions to more general loss functions, such as logistic loss for classification, requires additional work. Whether the improved rate (15) for TKRR is minimax optimal for γ > 1, appears to be open, although we conjecture it to be the case. Due to the theoretical nature of this contribution, no negative societal impact of our work is anticipated. Here, we only prove Theorem 2. The remaining proofs are given in Appendix B. Proof of Theorem 2. To simplify the argument, we assume that λ 1/α is an integer, without loss of generality. Recall that for a decreasing function f, we have PU i=L f(i) R U L 1 f(x)dx. Let us first consider the sum involved in the variance term. We will use the following inequalities, 1 2 min(a, b) ab a+b min(a, b) which holds for a, b 0. We have µ2 i (µi + λ)2 = 1 (µi + λ)2 1 i=1 min(λ2, µ2 i ). Since µi = i α, the minimum above is λ2 for i [1, λ 1/α], and µ2 i for i > λ 1/α. If r λ 1/α, only the first case happens and the overall bound is 1 λ2 rλ2 = r. If r > λ 1/α =: k, we have i=k+1 i 2α = k + 1 k x 2αdx k + λ 2k 2α+1 λ 1/α To summarize, I1 min(r, λ 1/α) and we note that the variance term in (9) is (σ2/n)I1. For the middle term in (9), we have, assuming r < n, i=r+1 (ξ i )2 = i=r+1 i 2γα 1 Z r x 2γα 1dx r 2γα. Finally, for the first term in (9), we can write (µi + λ)2 (ξ i )2 = λ2µ2 i (µi + λ)2 i=1 min(λ2, µ2 i ) ξ i µi If r λ 1/α, we have I2 λ2 Pr i=1 ξ i µi 2 λ2 Pr i=1 i 2(γ 1)α 1. Letting β = 2(γ 1)α, 1 λ2 I2 1 + i=2 i β 1 1 + Z r 1 x β 1dx = 1 + 1 If γ 1, the sum above is 1. If γ < 1, then β < 0 and the sum is r|β|. The two cases can be compactly written as max(1, r β). If r > λ 1/α := k, we have i=k+1 (ξ i )2 λ2 max(1, k β) + k 2γα using the bounding techniques we have seen before. Plugging in k = λ 1/α, we obtain I2 λ2 max(1, λ2(γ 1)) + λ2γ max(λ2, λ2γ) To summarize, for r λ 1/α, we have I2 λ2 max(1, r β) while for r > λ 1/α, we have I2 λ2 max(1, λ2(γ 1)). Note that these two expressions can be combined into one as follows: I2 λ2 max(1, η β), where η = min(r, λ 1/α). Putting the pieces together gives the desired bound. The above upper bounds are sharp up to constants, since in each step, there is a corresponding matching lower bound. Proof of part (a). With the given choice for λ and r, we have λ 1/α (n/σ2)γ/(2γα+1) (n/σ2)1/(2γα+1) r for γ 1. Thus η = min(r, λ 1/α) r, hence η 2(γ 1)α 1 for γ 1 since r 1. It follows that the E efr,λ f 2 n λ2 + r 2γα + σ2 Plugging in the choices for λ and r, each of the three terms is (σ2/n)2γα/(2γα+1) as desired. Proof of part (b). Here, we have r = n, so the middle term in (14) vanishes. Assume first that γ 1. We consider different regularization regimes. Case 1: λ 1/α [1, n] so that λ [n α, 1] and η = min(n, λ 1/α) = λ 1/α. We have η 2(γ 1)α = λ2(γ 1) 1. It follows that E efr,λ f 2 n λ2 + σ2 The optimal choice of λ is obtained by setting λ2 (σ2/n)λ 1/α, that is, λ = (σ2/n)α/(2α+1) and MSE rate (σ2/n)2α/(2α+1). Note that, assuming n 2α σ2 n, this choice of λ is within the assumed range [n α, 1] up to constants. Case 2: λ 1/α < 1. Then, η = 1 and the MSE λ2 + σ2/n. Since σ2 n by assumption and λ > 1, we have MSE λ2 > 1 which is always worse than the rate in Case 1. Case 3: λ 1/α > n. Then, η = n and MSE λ2 + σ2 σ2 since σ2 n 2α by assumption and λ2 < n 2α. This rate is clearly worst than Case 1. Putting the pieces together the optimal rate is achieved by Case 1. Next, consider the case γ < 1. We again consider the three regularization regimes: Case 1: λ 1/α [1, n] so that λ [n α, 1] and η = min(n, λ 1/α) = λ 1/α. We have η 2(γ 1)α = λ2(γ 1) > 1. It follows from (14) that E efr,λ f 2 n λ2γ + σ2 The optimal choice of λ is obtained by setting λ2γ (σ2/n)λ 1/α, that is, λ = (σ2/n)α/(2γα+1) and MSE rate (σ2/n)2γα/(2γα+1). Note that, as long as n 2γα σ2 n (which holds by the assumption on σ2 since γ < 1), this choice of λ is within the assumed range [n α, 1] up to constants. Case 2: λ 1/α < 1. This is similar to Case 2 when γ 1. Case 3: λ 1/α > n. Then, η = n and η 2(γ 1)α > 1, hence MSE λ2n 2(γ 1)α + σ2 σ2 which is clearly worst than the rate in Case 1. Putting the pieces together the optimal rate is achieved by Case 1. [1] Ahmed Alaoui and Michael W. Mahoney. Fast randomized kernel ridge regression with statistical guarantees. In Advances in Neural Information Processing Systems, pages 775 783, 2015. [2] A. A. Amini. Spectrally-truncated kernel ridge regression and its free lunch. Electronic Journal of Statistics, 15:1935 7524, 2021. [3] Arash A. Amini, Richard Baumgartner, and Dai Feng. Target alignment in truncated kernel ridge regression. https://github.com/aaamini/krr-target-align, 2022. [4] Francis Bach. Sharp analysis of low-rank kernel matrix approximations. In Conference on Learning Theory, pages 185 209, 2013. [5] P. Bartlett, P. Long, G. Lugosi, and A. Tsigler. Benign overfitting in linear regression. Proceedings of the National Academy of Sciences, 117(48):30063 30070, 2020. [6] D. Belkin, D. Hsu, S. Ma, and S. Mandal. Reconciling modern machine-learning practice and the classical bias variance trade-off. Proceedings of the National Academy of Sciences, 116(32):15849 15854, 2019. [7] B. Bordelon, A. Canatar, and C. Pehlevan. Spectrum dependent learning curves in kernel regression and wide neural networks. In International Conference on Machine Learning, pages 1024 1034, 2020. [8] A. Canatar, B. Bordelon, and C. Pehlevan. Spectral bias and task-model alignment explain generalization in kernel regression and infinitely wide neural networks. Nature Communications, 12(2914), 2021. [9] Andrea Caponnetto and Ernesto De Vito. Optimal rates for the regularized least-squares algorithm. Foundations of Computational Mathematics, 7(3):331 368, 2007. [10] Corinna Cortes, Mehryar Mohri, and Ameet Talwalkar. On the impact of kernel approximation on learning accuracy. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, pages 113 120, 2010. [11] Nello Cristianini, John Shawe-Taylor, André Elisseeff, and Jaz Kandola. On kernel-target alignment. In T. Dietterich, S. Becker, and Z. Ghahramani, editors, Advances in Neural Information Processing Systems, volume 14. MIT Press, 2001. [12] Hugo Cui, Bruno Loureiro, Florent Krzakala, and Lenka Zdeborová. Generalization error rates in kernel regression: The crossover from the noiseless to noisy regime. Advances in Neural Information Processing Systems, 34, 2021. [13] Trevor Hastie, Andrea Montanari, Saharon Rosset, and Ryan Tibshirani. Surprises in highdimensional ridgeless least squares interpolation. The Annals of Statistics, 50(2):949 986, 2022. [14] Arthur Jacot, Franck Gabriel, and Clement Hongler. Neural tangent kernel: Convergence and generalization in neural networks. In S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 31. Curran Associates, Inc., 2018. [15] Rong Jin, Tianbao Yang, Mehrdad Mahdavi, Yu-Feng Li, and Zhi-Hua Zhou. Improved bounds for the Nyström method with application to kernel classification. IEEE Transactions on Information Theory, 59(10):6939 6949, 2013. [16] George Kimeldorf and Grace Wahba. Some results on Tchebycheffian spline functions. Journal of Mathematical Analysis and Applications, 33(1):82 95, 1971. [17] Dmitry Kobak, Jonathan Lomond, and Benoit Sanches. The optimal ridge penalty for real-world high-dimensional data can be zero or negative due to the implicit ridge regularization. The Journal of Machine Learning Research, 21(1):1 16, 2020. [18] Sanjiv Kumar, Mehryar Mohri, and Ameet Talwalkar. Ensemble Nyström method. In Advances in Neural Information Processing Systems, pages 1060 1068, 2009. [19] Mu Li, James T. Kwok, and Bao-Liang Lu. Making large-scale Nyström approximation possible. In Proceedings of the 27th International Conference on Machine Learning, pages 631 638, 2010. [20] Marco Loog, Tom Viering, Alexander Mey, and David Tax. A brief prehistory of double descent. Proceedings of the National Academy of Sciences, 117(20):10625 10626, 2020. [21] Alessandro Rudi, Raffaello Camoriano, and Lorenzo Rosasco. Less is more: Nyström computational regularization. In Advances in Neural Information Processing Systems, pages 1657 1665, 2015. [22] B. Schoelkopf and A. Smola. Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond. MIT Press, 2001. [23] Ingo Steinwart and Andreas Christmann. Support vector machines. Springer Science & Business Media, 2008. [24] Ameet Talwalkar and Afshin Rostamizadeh. Matrix coherence and the nyström method. In Proceedings of the Twenty-Sixth Conference on Uncertainty in Artificial Intelligence, pages 572 579, 2010. [25] Alexander Tsigler and Peter L Bartlett. Benign overfitting in ridge regression. ar Xiv preprint ar Xiv:2009.14286, 2020. [26] Alexandre B. Tsybakov. Introduction to Nonparametric Estimation. Springer, 2009. [27] Grace Wahba. Spline Models for Observational Data. SIAM, 1990. [28] Martin J. Wainwright. High-dimensional Statistics: A Non-asymptotic Viewpoint. Cambridge University Press, 2019. [29] Larry Wasserman. All of Nonparametric Statistics. Springer Science & Business Media, 2006. [30] Christopher K. I. Williams and Matthias Seeger. Using the Nyström method to speed up kernel machines. In Advances in Neural Information Processing Systems, pages 682 688, 2001. [31] Tianbao Yang, Yu-Feng Li, Mehrdad Mahdavi, Rong Jin, and Zhi-Hua Zhou. Nyström method vs random Fourier features: A theoretical and empirical comparison. In Advances in Neural Information Processing Systems, pages 476 484, 2012. [32] Yun Yang, Mert Pilanci, and Martin J. Wainwright. Randomized sketches for kernels: Fast and optimal nonparametric regression. The Annals of Statistics, 45(3):991 1023, 2017. [33] Kai Zhang, Ivor W. Tsang, and James T. Kwok. Improved Nyström low-rank approximation and error analysis. In Proceedings of the 25th International Conference on Machine Learning, pages 1232 1239. ACM, 2008. 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] (b) Did you describe the limitations of your work? [Yes] See Section 6. (c) Did you discuss any potential negative societal impacts of your work? [Yes] See Section 6. (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] We adhered to the ethics review guidelines. 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [Yes] See Sections 2, 3, 4. (b) Did you include complete proofs of all theoretical results? [Yes] See Section 7. 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] in the supplemental material. (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [N/A] Our experiments were run to support the theoretical results. See Section 5. (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [N/A] (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [N/A] Our simulation were done on a regular laptop. 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [N/A] (b) Did you mention the license of the assets? [N/A] (c) Did you include any new assets either in the supplemental material or as a URL? [N/A] (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [N/A] (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [N/A] 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]