# ridgeless_regression_with_random_features__885842ee.pdf Ridgeless Regression with Random Features Jian Li1 , Yong Liu2 , Yingying Zhang3 1Institute of Information Engineering, Chinese Academy of Sciences 2Gaoling School of Artificial Intelligence, Renmin University of China 3School of Mathematics and Information Sciences, Yantai University lijian9026@iie.ac.cn, liuyonggsai@ruc.edu.cn, yingyingzhang239@gmail.com Recent theoretical studies illustrated that kernel ridgeless regression can guarantee good generalization ability without an explicit regularization. In this paper, we investigate the statistical properties of ridgeless regression with random features and stochastic gradient descent. We explore the effect of factors in the stochastic gradient and random features, respectively. Specifically, random features error exhibits the double-descent curve. Motivated by the theoretical findings, we propose a tunable kernel algorithm that optimizes the spectral density of kernel during training. Our work bridges the interpolation theory and practical algorithm. 1 Introduction In the view of traditional statistical learning, an explicit regularization should be added to the nonparametric learning objective [Caponnetto and De Vito, 2007; Li et al., 2018], i.e. an 2-norm penalty for the kernelized least-squares problems, known as kernel ridge regression (KRR). To ensure good generalization (out-of-sample) performance, the modern models should choose the regularization hyperparameter λ to balancing the bias and variance, and thus avoid overfitting. However, recent studies empirically observed that neural networks still can interpolate the training data and generalize well when λ = 0 [Zhang et al., 2021; Belkin et al., 2018]. But also, for many modern models including neural networks, random forests and random features, the test error captures a double-descent curve as the increase of features dimensional [Mei and Montanari, 2019; Advani et al., 2020; Nakkiran et al., 2021]. Recent empirical successes of neural networks prompted a surge of theoretical results to understand the mechanism for good generalization performance of interpolation methods without penalty where researchers started with the statistical properties of ridgeless regression based on random matrix theory [Liang and Rakhlin, 2020; Bartlett et al., 2020; Jacot et al., 2020; Ghorbani et al., 2021]. However, there are still some open problems to settle down: 1) In the theoretical front, current studies focused Corresponding author on the direct bias-variance decomposition for the ridgeless regression, but ignore the influence from the optimization problem, i.e. stochastic gradient descent (SGD). Meanwhile, the connection between kernel regression and ridgeless regression with random features are still not well established. Further, while asymptotic behavior in the overparameterized regime is well studied, ridgeless models with a finite number of features are much less understood. 2) In the algorithmic, there is still a great gap between statistical learning for ridgeless regression and algorithms. Although the theories explain the double-descent phenomenon for ridgeless methods well, a natural question is whether the theoretical studies helps to improve the mainstream algorithms or design new ones. In this paper, we consider the Random Features (RF) model [Rahimi and Recht, 2007] that was proposed to approximate kernel methods for easing the computational burdens. In this paper, we investigate the generalization ability of ridgeless random features, of which we explore the effects from stochastic gradient algorithm and random features. And then, motivated by the theoretical findings, we propose a tunable kernel algorithm that optimizes the spectral density of kernel during training, reducing multiple trials for kernel selection to just training once. Our contributions are summarized as: 1) Stochastic gradients error. We first investigate the stochastic gradients error influenced the factors from SGD, i.e. the batch size b, the learning rate γ and the iterations t. The theoretical results illustrate the tradeoffs among these factors to achieve better performance, such that it can guide the set of these factors in practice. 2) Random features error. We then explore the difference between ridgeless kernel predictor and ridgeless RF predictor. In the overparameterized setting M > n, the ridgeless RF converges to the ridgeless kernel as the increase number of random features, where M is the number of random features and n is the number of examples. In the underparameterized regime M n, the error of ridgeless RF also exhibit an interesting double-descent curve in Figure 2 (a), because the variance term explores near the transition point M = n. 3) Random features algorithm with a tunable kernel. Theoretical results illustrate the errors depends on the trace of kernel matrix, motivating us to design a kernel learning algorithm which asynchronously optimizes the spectral density and model weights. The algorithm is friend to random initial- Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence (IJCAI-22) ization, and thus easing the problem of kernel selection. 2 Related Work Statistical properties of random features. The generalization efforts for random features are mainly in reducing the number of random features to achieve the good performance. [Rahimi and Recht, 2007] derived the appropriate error bound between kernel function and the inner product of random features. And then, the authors proved O( n) features to obtain the error bounds with convergence rate O(1/ n) [Rahimi and Recht, 2008]. Rademacher complexity based error bounds have been proved in [Li et al., 2019; Li et al., 2020]. Using the integral operator theory, [Rudi and Rosasco, 2017; Li and Liu, 2022] proved the minimax optimal rates for random features based KRR. In contrast, recent studies [Hastie et al., 2019] make efforts on the overparameterized case for random features to compute the asymptotic risk and revealed the double-descent curve for random Re LU [Mei and Montanari, 2019] features and random Fourier features [Jacot et al., 2020]. Double-descent in ridgeless regression. The doubledescent phenomenon was first observed in multilayer networks on MNIST dataset for ridgeless regression [Advani et al., 2020]. It then been observed in random Fourier features and decision trees [Belkin et al., 2018]. [Nakkiran et al., 2021] extended the double-descent curve to various models on more complicated tasks. The connection between doubledescent curve and random initialization of the Neural Tangent Kernel (NTK) has been established [Geiger et al., 2020]. Recent theoretical work studied the asymptotic error for ridgeless regression with kernel models [Liang and Rakhlin, 2020; Jacot et al., 2020] or linear models [Bartlett et al., 2020; Ghorbani et al., 2021]. 3 Problem Setup In the context of nonparametric supervised learning, given a probability space X Y with an unknown distribution µ(x, y), the regression problem with squared loss is to solve min f E(f), E(f) = Z X Y (f(x) y)2dµ(x, y). (1) However, one can only observe the training set (xi, yi)n i=1 that drawn i.i.d. from X Y according to µ(x, y), where xi Rd are inputs and yi R are the corresponding labels. 3.1 Kernel Ridgeless Regression. Suppose the target regression f (x) = E(y|x = x) lie in a Reproducing Kernel Hilbert Space (RKHS) HK, endowed with the norm K and Mercer kernel K( , ) : X X R. Denote X = [x1, , xn] Rn d the input matrix and y = [y1, , yn] the response vector. We then let K(X, X) = [K(xi, xj)]n i,j=1 Rn n be the kernel matrix and K(x, X) = [K(x, x1), , K(x, xn)] R1 n. Given the data (X, y), the empirical solution to (1) admits a closed-form solution: bf(x) = K(x, X)K(X, X) y, (2) where is the Moore-Penrose pseudo inverse. The above solution is known as kernel ridgeless regression. 3.2 Ridegeless Regression with Random Features The Mercer kernel is the inner product of feature mapping in HK, stated as K(x, x ) = Kx, Kx , x, x X, where Kx = K(x, ) HK is high or infinite dimensional. The integral representation for kernel is K(x, x ) = R X ψ(x, ω)ψ(x , ω)dπ(ω), where π(ω) is the spectral density and ψ : X X R is continuous and bounded function. Random features technique is proposed to approximate the kernel by a finite dimensional feature mapping K(x, x ) φ(x), φ(x ) , with M (ψ(x, ω1), , ψ(x, ωM)) , (3) where φ : X RM and ω1, , ωM are sampled independently according to π. The solution of ridgeless regression with random features can be written as bf M(x) = φ(x) φ(X) φ(X) φ(X) y, (4) where φ(X) Rn M is the feature mapping matrix over X and φ(x) R1 M is the feature mapping over x. 3.3 Random Features with Stochastic Gradients To further accelerate the computational efficiency, we consider the stochastic gradient descent method as bellow bf M,b,t(x) = wt, φ(x) , with wt+1 = wt γt i=b(t 1)+1 wt, φ(xi) yi)φ(xi), (5) where wt RM, w0 = 0, b is the mini-batch size and γt is the learning rate. When b = 1 the algorithm reduces to SGD and b > 1 is the mini-batch version. We assume the examples are drawn uniformly with replacement, by which one pass over the data requires n/b iterations. Before the iteration, the compute of φ(X) consumes O(n M) time. The time complexity is O(Mb) for each iteration and the total complexity is O(n MT) after T passes on the train set. 4 Main Results In this section, we study the statistical properties of estimators bf M,b,t (5), bf M (3) and bf (2). We only present the main theoretical results in this part and the proofs can be find in the full version 1. Denote Eµ[ ] the expectation w.r.t. the marginal x µ and g 2 L2(µ) = Z g2(x)dµ(x) = Eµ[g2(x)], g L2(µ). The squared integral norm over the space L2(µ) = {g : X R| R g2(x)dµ(x) < } and f L2(µ). Combing the above equation with (1), one can prove that E(f) E(f ) = f f 2 L2(µ), f L2(µ). Therefore we can decompose the excess risk of E( bf M,b,t) E(f ) as bellow E( bf M,b,t) E(f ) bf M,b,t bf M + bf M bf + bf f . (6) 1https://arxiv.org/abs/2205.00477 Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence (IJCAI-22) The excess risk bound includes three terms: stochastic gradient error bf M,b,t bf M , random feature error bf M bf , and excess risk of kernel ridgeless regression bf f , which admits the bias-variance form. In this paper, since the ridgeless excess risk bf f has been well-studied [Liang and Rakhlin, 2020], we focus on the first two bounds and explore the factors in them, respectively. Throughout this paper, we assume the true regression f (x) = f , Kx lies in the RKSH of the kernel K, i.e., f HK. Assumption 1 (Random features are continuous and bounded). Assume that ψ is continuous and there is a κ [1, ), such that |ψ(x, ω)| κ, x X, ω Ω. Assumption 2 (Moment assumption). Assume there exists B > 0 and σ > 0, such that for all p 2 with p N, Z R |y|pdρ(y|x) 1 2p!Bp 2σ2. (7) The above two assumptions are standard in statistical learning theory [Smale and Zhou, 2007; Caponnetto and De Vito, 2007; Rudi and Rosasco, 2017]. According to Assumption 1, the kernel K is bounded by K(x, x) κ2. The moment assumption on the output y holds when y is bounded, subgaussian or sub-exponential. Assumptions 1 and 2 are standard in the generalization analysis of KRR, always leading to the learning rate O(1/ N) [Smale and Zhou, 2007]. 4.1 Stochastic Gradients Error We first investigate the approximation ability of stochastic gradient by measuring bf M,b,t bf M 2 L2(µ), and explore the effect of the mini-batch size b, the learning rate γ and the number of iterations t. Theorem 1 (Stochastic gradient error). Under Assumptions 1, 2, let t [T], γ n 9T log n δ 1 8(1+log T ) and n 32 log2 2 δ , the following bounded holds with high probability bf M,b,t bf M γ The first term in the above bound measures the similarity between mini-batch gradient descent estimator and full gradient descent estimator bf M,b,t bf M,t , which depends on the mini-batch size b and the learning rate γ. The second term reflects the approximation between the gradient estimator and the random feature estimator bf M,t bf M,b,t , which is determined by the number of iterations t and the step-size γ, leading to a sublinear convergence O(1/t). Corollary 1. Under the same assumptions of Theorem 1, one of the following cases and the time complexities 1) b = 1, γ 1 n and T = n O(n M) 2) b = n, γ 1 and T = n O(n M) 3) b = n, γ 1 and T = n O(n n M) is sufficient to guarantee with high probability that bf M,b,t bf M 1 n. In the above corollary, we give examples for SGD, minibatch gradient, and full gradient, respectively. It shows the computational efficiency of full gradient descent is usually worse than stochastic gradient methods. The computational complexities O(n M) are much smaller than that of random features O(n M 2 + M 3). All these cases achieve the same learning rate O(1/ n) as the exact KRR. With source condition and capacity assumption in integral operator literature [Caponnetto and De Vito, 2007; Rudi and Rosasco, 2017], the above bound can achieve faster convergence rates. Remark 1. [Carratino et al., 2018] also studied the approximation of mini-batch gradient descent algorithm, but the random features estimator is defined with noise-free labels f (X) and with ridge regularization, which failed to directly capture the effect of stochastic gradients. Note that this work provides empirical estimators bf M,b,t, bf M, and bf with noise labels y and ridgeless, which makes the proof techniques quite different from [Carratino et al., 2018]. The technical difference can be found by comparing the proofs of Theorem 1 in this paper and Lemma 9 in [Carratino et al., 2018]. 4.2 Random Features Error The redgeless RF predictor (3) characterizes different behaviors depending on the relationship between the number of random features M and the number of samples n. Theorem 2 (Overparameterized regime M n). When M n, the random features error can be bounded by E E( bf M) E( bf) (α + c1) f 2 K M , where the constant α M M n. In the overparameterized case, the ridgeless RF predictor is an unbiased estimator of the ridgeless kernel predictor and RF predictors can interpolate the dataset. The error term in the above bound is the variance of the ridgeless RF predictor. As shown in Figure 1 (a), the variance of ridgeless RF estimator explodes near the interpolation as M n, leading to the double descent curve that has been well studied in [Mei and Montanari, 2019; Jacot et al., 2020]. Obviously, a large number of random features in the overparameterized domain can approximate the kernel method well [Rahimi and Recht, 2007; Rudi and Rosasco, 2017], but it introduces computational challenges, i.e. O(n M) time complexity for M > n. To reach good tradeoffs between statistical properties and computational cost, it is rather important to investigate the approximation of ridgeless RF predictor in the underparameterized regime. Theorem 3 (Underparameterized regime M < n). When M < n, under Assumption 1, the ridgeless RF estimator bf M approximates a kernel ridge regression estimator bfλ = K(x, X) (K + λn I) 1 y by E E( bf M) E( bfλ) Tr(K)4 f 2 K M 6 + (α + c1) f 2 K M , where K = K(X, X) is the kernel matrix, α n n M and the regularization parameter λ is the unique positive number satisfying Tr K(K + λn I) 1 = M/n. Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence (IJCAI-22) The above bound estimates the approximation between the ridgeless RF estimator bf M and the ridge regression bfλ, such that the ridgeless RF essentially imposes an implicit regularization. For the shift-invariant kernels K(x, x ) = h( x x ), i.e. Gaussian kernel K(x, x ) = exp( x x 2/(2σ2)), the trace of matrix is a constant Tr(K) = n. The number of random features needs M = Ω(n0.75) to achieve the convergence rate O(1/ n) for Theorem 3. Note that e N(λ) := Tr K(K + λn I) 1 is known as the empirical effective dimension, which has been used to control the capacity of hypothesis space [Caponnetto and De Vito, 2007] and sample points via leverage scores [Rudi et al., 2018]. Theorem 3 states the implicit regularization effect of random features, of which the regularizer parameter is related to the features dimensional M. Together with Theorem 2, Theorem 3 and Figure 1 (a), we discuss the influence from different feature dimensions M for ridgeless RF predictor bf M: 1) In the underparameterized regime M < n, the regularization parameter is inversely proportional to the feature dimensional λ 1/M. The effect of implicit regularity becomes greater as we reduce the features dimensional, while the implicit regularity reduces to zero as M n. As the increase of M, the test error drops at first and then rises due to overfitting (or explored variance). 2) At the interpolation threshold M = n, the condition Tr K(K + λI) 1 = M leads to λ = 0. Thus, M = n not only split the overparameterized regime and the underparameterized regime, but also is the start of implicit regularization for ridgeless RF. At this point, the variance of ridgeless RF predictor explodes, leading to double descent curve. 3) In the overparameterized case M > n, the ridgeless RF predictor is an unbiased estimator ridgeless kernel predictor, but the effective ridge goes to zero. As the number of random features increases, the test error of ridgeless RF drops again. Remark 2 (Excess risk bounds). From (6), one can derive the entire excess risk bound for ridgeless RF-SGD bf M,b,t f by using the existing ridge regression bound bfλ f in [Bartlett et al., 2005; Caponnetto and De Vito, 2007] for Theorem 3, and ridgeless regression bound bf f in [Liang and Rakhlin, 2020] for Theorem 2. These risk bounds usually characterized the bias-variance decomposition with f f 2 L2(µ) B + V , f L2(µ), where the bias usually can be bounded by B 1 n p Tr(K). For the conventional kernel methods, the kernel matrix is fixed and thus Tr(K) is a constant. 5 Random Features with Tunable Kernels By noting that Tr(K) play a dominate role in random features error in Theorem 3 and bias, we thus design a tunable kernel learning algorithm by reducing the trace Tr(K). We first transfer the trace to trainable form by random features Tr(K) Tr φ(X)φ(X) = φ(X) 2 F , where φ : X RM depends on the spectral density π according to (3) and 2 F is the squared Frobenius norm for a Algorithm 1 Ridgeless RF with Tunable Kernels (RFTK) Input: Training data (X, y) and feature mapping φ : Rd RD. Hyparameters σ2, β, T, b, γ, η and s. Output: The ridgeless RF model w T and the learned Ω. 1: for t = 1, 2, , T do 2: Sample a batch examples (xi, yi)b i=1 (X, y). 3: Update RF model weights wt = wt 1 γ L(w,Ω) w 4: if t%s == 0 then 5: Optimize frequency matrix Ωt = Ωt 1 η L(w,Ω) Ω 6: end if 7: end for matrix. Since 2 F is differentiable w.r.t. Ω, we thus can optimize the kernel density with backpropagation. For example, considering the random Fourier features [Rahimi and Recht, 2007], the feature mappings can be written as M cos(Ω x + b), (8) where the frequency matrix Ω= [ω1, , ωM] Rd M composed M vectors drawn i.i.d. from a Gaussian distribution N(0, 1 σ2 I) Rd. The phase vectors b = [b1, , b M] RM are drawn uniformly from [0, 2π]. Theoretical findings illustrate that smaller trace of kernel matrix Tr(K) can lead to better performance. We first propose a bi-level optimization learning problem w φ(xi) yi 2 s.t. Ω = arg min Ω φ(X) 2 F . (9) The above objective includes two steps: 1) given a certain kernel, i.e. the frequency matrix Ω= [ω1, , ωM] π(w), the algorithm train the ridgeless RF model w; 2) given a trained RF model w, the algorithm optimize the spectral density (the kernel) by updating the frequency matrix Ω. To accelerate the solve of (9), we optimize w and Ωjointly by minimize the the following objective L(w, Ω) = 1 i=1 (f(xi) yi)2 + β φ(X) 2 F , (10) where β is a hyperparameter to balance the effect between empirical loss and the trace of kernel matrix. Here, the update of w is still only dependent on the squared loss (thus it is still ridgeless), but the update of Ωis related to both the squared loss and Frobenius norm. Therefore, the time complexity for update w once is O(Mb). However, since the trace is defined on all data, the update of Ωis relevant to all data and time complexity is O(n Md), which is infeasible to update in every iterations. Meanwhile, the kernel needn t to update frequently, and thus we apply an asynchronous strategy for optimizing the spectral density. As shown in Algorithm 1, the frequency matrix Ωis updated after every s iterations for the update of w. The total complexity for the asynchronous strategy is O(n Md T/s). Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence (IJCAI-22) 0 1 2 3 4 5 M/n The value of α (a) The variance factor α w.r.t. M/n 0 20 40 60 80 100 # Epoch Stochastic Error Δt b=1, γ = 0.001 b=10, γ = 0.001 b=100, γ = 0.01 b=1000, γ = 0.1 b=10000, γ = 0.1 (b) Stochastic error vs. factors b, γ and t 0 20 40 60 80 100 # Epoch b=1, γ = 0.001 b=10, γ = 0.001 b=100, γ = 0.01 b=1000, γ = 0.1 b=10000, γ = 0.1 Ridgeless RF (c) MSE vs. the factors b, γ and t Figure 1: (a) The variance factor α w.r.t. ratio M/n, which make the variance explores near the transition point M = n. (b) Empirical stochastic gradient errors t = 1 n Pn i=1 | bf M,b,t(xi) bf M(xi)| of the ridgeless RF-SGD predictors on n = 10000 synthetic examples. (c) The test MSE of all ridgeless RF-SGD estimators. 10 2 10 1 100 101 102 M/n Ridgeless RF RF, λ = 10 4 Ridgeless KRR KRR, λ = 10 4 (a) Test errors v.s. regulerizations 0 50 100 150 200 # Epoch # Accuracy (%) RF RF-SGD RFTK (b) Test accuracies of compared methods 0 50 100 150 200 250 # Epoch 100 150 200 250 RF-SGD Tr(K) of RF-SGD RFTK Tr(K) of RFTK (c) Training loss and the kernel trace Figure 2: (a) Test errors of the RF predictors (solid lines) and kernel predictors (dashed lines) w.r.t. different regularization. Note that, the ridgeless RF predictors exhibit a double descent curve. (b) Test accuracies of the compared methods on the MNIST dataset. (c) Training loss (solid lines) and the trace of kernel Tr(K) (dashed lines) of the RF predictors on the MNIST dataset. 6 Experiments We use random Fourier feature defined in (8) to approximate the Gaussian kernel K(x, x ) = exp( σ2 x x 2/2). Note that, the corresponding random Fourier features (8) is with the frequency matrix Ω N(0, σ2). We implement all code on Pytorch 2 and tune hyperparameters over σ2 {0.01, , 1000}, λ = {0.1, , 10 5} by grid search. 6.1 Numerical Validations Factors of Stochastic Gradient Methods We start with a nonlinear problem y = min( w x, w x)+ ϵ, where ϵ N(0, 0.2) is the label noise, and x N(0, I). Setting d = 10, we generate n = 10000 samples for training and 2500 samples for testing. The optimal kernel hyperparameter is σ2 = 100 after grid search. As stated in Theorem 1, the stochastic gradients error is determined by the batch size b, learning rate γ and the iterations t. To explore effects of these factors, we evaluate the approximation between the ridgeless RF-SGD estimator bf M,b,t and the ridgeless RF estimator bf M on the simulated data. Given a batch size b, we tune the learning rate γ w.r.t. the MSE over γ {101, 100, , 10 4}. As shown in Figure 1 (b), (c), give the b, γ, we estimate the ideal iterations t = # Epoch 10000/b after which the error drops slowly As the step size b increases, the learning rates γ become larger 2https://github.com/superlj666/Ridgeless-Regression-with-Random-Features while the iterations reduces. This coincides with the tradeoffs among these factors b/γ + 1/(γt) in Theorem 1, where the balance between b/γ and 1/γt leads to better performance. Comparing Figure 1 (b) and (c), we find that: 1) After the same passes over the data (same epoch), the stochastic gradient error and MSE of the algorithms with smaller batch sizes are smaller with faster drop speeds. 2) The stochastic error still decreases after the MSE converges, where the other error terms dominates the excess risk, i.e. bias of predictors. Double Descent Curve in Random Features To characterize different behaviors of the random features error w.r.t. M, we fixed the number of training examples n and changes the random features dimensional M. Figure 2 (a) reports test errors in terms of different ratios M/n, illustrating that: 1) Kernel methods with smaller regularization lead to smaller test errors, and the reason may be regularization hurts the performance when the training data is clean . 2) Test errors of RF predictors converges to the corresponding kernel predictors, because a larger M have better approximation to the kernel. This coincides with Theorem 2 and 3 that a larger M reduces the random features error. 3) When the regularization term λ is small, the predictors leads to double descent curves and M = n is the transition point dividing the underparamterized and overparameterized regions. The smaller regularity, the more obvious curve the RF predictor exhibits. In the overparameterized case, the error of RF predictor Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence (IJCAI-22) Kernel Ridge Kernel Ridgeless RF RF-SGD RFTK dna 52.83 1.66 49.67 18.54 52.83 1.66 51.33 1.89 92.92 0.89 letter 96.54 0.25 96.40 0.15 95.33 0.32 91.74 0.46 96.17 0.29 pendigits 97.46 0.42 90.67 4.75 96.91 0.43 46.04 5.62 98.64 0.43 segment 82.99 1.85 56.71 10.71 83.44 1.69 37.75 9.11 94.55 1.52 satimage 90.43 0.48 88.79 0.77 87.67 0.89 90.33 1.36 90.79 1.23 usps 92.49 0.70 87.47 7.38 94.38 0.60 49.81 3.52 97.29 0.61 svmguide2 81.90 2.78 70.13 4.91 66.20 4.64 81.65 4.25 82.78 4.84 vehicle 63.00 2.80 79.35 2.89 75.94 2.88 74.24 3.92 80.06 4.49 wine 39.17 6.27 48.89 13.68 91.11 19.40 43.89 6.19 98.33 1.36 shuttle / / 79.08 26.76 98.94 0.48 99.63 0.19 Sensorless / / 32.92 8.22 17.72 3.84 86.10 0.73 MNIST / / 95.52 0.16 96.61 0.11 98.09 0.07 Table 1: Classification accuracy (%) for classification datasets. We bold the results with the best method and underline the ones that are not significantly worse than the best one. Methods Time Density Regularizer Kernel Ridge O(n3) Assigned Ridge Kernel Ridgeless O(n3) Assigned Ridgeless RF O(n M 2 + M 3) Assigned Ridgeless RF-SGD O(Mb T) Assigned Ridgeless RFTK O(n Mb T/s) Learned φ(X) 2 F Table 2: Compared algorithms. converges to corresponding kernel predictor as M increases, verifying the results in Theorem 2 that the variance term reduces given more features when M > n. In the underparamterized regime, the errors of RF predictors drop first and then increase, and finally explodes at M = n. These empirical results validates theoretical findings in Theorem 3 that the variance term dominates the random features near M = n and it is significantly large as shown in Figure 1 (a). Benefits from Tunable Kernel Motivated by the theoretical findings that the trace of kernel matrix Tr(K) influences the performance, we design a tunable kernel method RFTK that adaptively optimizes the spectral density in the training. To explore the influence of factors upon convergence, we evaluate both test accuracy and training loss on the MNIST dataset. Compared with the exact random features (RF) and random features with stochastic gradients (RF-SGD), we conduct experiments on the entire MNIST datasets. From Figure 2 (b), we find there is a significant accuracy gap between RF-SGD and RFTK. Figure 2 (c) indicates the trace of kernel matrix term takes affects after the current model fully trained near 100 epochs, and it decrease fast. Specifically, since the kernel is continuously optimized, more iterations can still improve the its generalization performance, while the accuracy of RF-SGD decreases after 100 epochs because of the overfitting to the given hypothesis. Figure 2 (b) and (c) explain the benefits from the penalty of Tr(K) that optimizes the kernel during the training, avoiding kernel selection. Empirical studies shows that smaller Tr(K) guarantees better performance, coinciding with the theoretical results in Theorem 1 and Theorem 3 that both stochastic gradients error and random features error depends on Tr(K). 6.2 Empirical Results Compared with related algorithms listed in Table 2, we evaluate the empirical behavior of our proposed algorithm RFTK on several benchmark datasets. Only the kernel ridge regression makes uses of the ridge penalty w 2 2, while the others are ridgeless. For the sake of presentation, we perform regression algorithms on classification datasets with one-hot encoding and cross-entropy loss. We set b = 32 and 100 epochs for the training, and thus the stop iteration is T = 100n/32. Before the training, we tune the hyperparameters σ2, λ, γ, β via grid search for algorithms on each dataset. To obtain stable results, we run methods on each dataset 10 times with randomly partition such that 80% data for training and 20% data for testing. Further, those multiple test errors allow the estimation of the statistical significance among methods. Table 1 reports the test accuracies for compared methods over classification tasks. It demonstrates that: 1) In some cases, the proposed RFTK remarkably outperforms the compared methods, for example dna, segment and Sensorless. That means RFTK can optimize the kernel even with a bad initialization, which makes it more flexible than the schema (kernel selection + learning). 2) For many tasks, the accuracies of RFTK are also significantly higher than kernel predictors, i.e. dna, because the learned kernel is more suitable for these tasks. 3) Compared to kernel predictors, RFTK leads to similar or worse results in some cases. The reason is that kernel hyperparameter σ2 has been tuned and in these cases they are near the optimal one, and thus the spectral density is changed a little by RFTK. 7 Conclusion We study the generalization properties for ridgeless regression with random features, and devise a kernel learning algorithm that asynchronously tune spectral kernel density during the training. Our work filled the gap between the generalization theory and practical algorithms for ridgeless regression with random features. The techniques presented here provides theoretical and algorithmic insights for understanding of neural networks and designing new algorithms. Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence (IJCAI-22) Acknowledgments This work was supported in part by the Excellent Talents Program of Institute of Information Engineering, CAS, the Special Research Assistant Project of CAS, the Beijing Outstanding Young Scientist Program (No. BJJWZYJH012019100020098), Beijing Natural Science Foundation (No. 4222029), and National Natural Science Foundation of China (No. 62076234, No. 62106257). References [Advani et al., 2020] Madhu S Advani, Andrew M Saxe, and Haim Sompolinsky. High-dimensional dynamics of generalization error in neural networks. Neural Networks, 132:428 446, 2020. [Bartlett et al., 2005] Peter L. Bartlett, Olivier Bousquet, and Shahar Mendelson. Local Rademacher complexities. The Annals of Statistics, 33(4):1497 1537, 2005. [Bartlett et al., 2020] Peter L Bartlett, Philip M Long, G abor Lugosi, and Alexander Tsigler. Benign overfitting in linear regression. Proceedings of the National Academy of Sciences, 117(48):30063 30070, 2020. [Belkin et al., 2018] Mikhail Belkin, Siyuan Ma, and Soumik Mandal. To understand deep learning we need to understand kernel learning. ar Xiv preprint ar Xiv:1802.01396, 2018. [Caponnetto and De Vito, 2007] Andrea Caponnetto and Ernesto De Vito. Optimal rates for the regularized least-squares algorithm. Foundations of Computational Mathematics, 7(3):331 368, 2007. [Carratino et al., 2018] Luigi Carratino, Alessandro Rudi, and Lorenzo Rosasco. Learning with sgd and random features. In Advances in Neural Information Processing Systems 31 (Neur IPS), pages 10192 10203, 2018. [Geiger et al., 2020] Mario Geiger, Arthur Jacot, Stefano Spigler, Franck Gabriel, Levent Sagun, St ephane d Ascoli, Giulio Biroli, Cl ement Hongler, and Matthieu Wyart. Scaling description of generalization with number of parameters in deep learning. Journal of Statistical Mechanics: Theory and Experiment, 2020(2):023401, 2020. [Ghorbani et al., 2021] Behrooz Ghorbani, Song Mei, Theodor Misiakiewicz, and Andrea Montanari. Linearized two-layers neural networks in high dimension. The Annals of Statistics, 49(2):1029 1054, 2021. [Hastie et al., 2019] Trevor Hastie, Andrea Montanari, Saharon Rosset, and Ryan J Tibshirani. Surprises in highdimensional ridgeless least squares interpolation. ar Xiv preprint ar Xiv:1903.08560, 2019. [Jacot et al., 2020] Arthur Jacot, Berfin Simsek, Francesco Spadaro, Cl ement Hongler, and Franck Gabriel. Implicit regularization of random feature models. In International Conference on Machine Learning, pages 4631 4640. PMLR, 2020. [Li and Liu, 2022] Jian Li and Yong Liu. Optimal rates for distributed learning with random features. In Proceedings of the 31st International Joint Conference on Artificial Intelligence (IJCAI), 2022. [Li et al., 2018] Jian Li, Yong Liu, Rong Yin, Hua Zhang, Lizhong Ding, and Weiping Wang. Multi-class learning: From theory to algorithm. In Advances in Neural Information Processing Systems 31, pages 1591 1600, 2018. [Li et al., 2019] Jian Li, Yong Liu, Rong Yin, and Weiping Wang. Multi-class learning using unlabeled samples : Theory and algorithm. In Proceedings of the 28th International Joint Conference on Artificial Intelligence (IJCAI), 2019. [Li et al., 2020] Jian Li, Yong Liu, and Weiping Wang. Automated spectral kernel learning. In Thirty-Four AAAI Conference on Artificial Intelligence, 2020. [Liang and Rakhlin, 2020] Tengyuan Liang and Alexander Rakhlin. Just interpolate: Kernel ridgeless regression can generalize. The Annals of Statistics, 48(3):1329 1347, 2020. [Mei and Montanari, 2019] Song Mei and Andrea Montanari. The generalization error of random features regression: Precise asymptotics and the double descent curve. Communications on Pure and Applied Mathematics, 2019. [Nakkiran et al., 2021] Preetum Nakkiran, Gal Kaplun, Yamini Bansal, Tristan Yang, Boaz Barak, and Ilya Sutskever. Deep double descent: Where bigger models and more data hurt. Journal of Statistical Mechanics: Theory and Experiment, 2021(12):124003, 2021. [Rahimi and Recht, 2007] Ali Rahimi and Benjamin Recht. Random features for large-scale kernel machines. In Advances in Neural Information Processing Systems 21 (NIPS), pages 1177 1184, 2007. [Rahimi and Recht, 2008] Ali Rahimi and Benjamin Recht. Weighted sums of random kitchen sinks: Replacing minimization with randomization in learning. In Advances in Neural Information Processing Systems 22 (NIPS), pages 1313 1320, 2008. [Rudi and Rosasco, 2017] Alessandro Rudi and Lorenzo Rosasco. Generalization properties of learning with random features. In Advances in Neural Information Processing Systems 30 (NIPS), pages 3215 3225, 2017. [Rudi et al., 2018] Alessandro Rudi, Daniele Calandriello, Luigi Carratino, and Lorenzo Rosasco. On fast leverage score sampling and optimal learning. In Advances in Neural Information Processing Systems, pages 5672 5682, 2018. [Smale and Zhou, 2007] Steve Smale and Ding-Xuan Zhou. Learning theory estimates via integral operators and their approximations. Constructive approximation, 26(2):153 172, 2007. [Zhang et al., 2021] Chiyuan Zhang, Samy Bengio, Moritz Hardt, Benjamin Recht, and Oriol Vinyals. Understanding deep learning (still) requires rethinking generalization. Communications of the ACM, 64(3):107 115, 2021. Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence (IJCAI-22)