# fast_instrument_learning_with_faster_rates__04d10352.pdf Fast Instrument Learning with Faster Rates Ziyu Wang, Yuhao Zhou, Jun Zhu Dept. of Comp. Sci. and Tech., BNRist Center, State Key Lab for Intell. Tech. & Sys., Institute for AI, Tsinghua-Bosch Joint Center for ML, Tsinghua University {wzy196,yuhaoz.cs}@gmail.com, dcszj@tsinghua.edu.cn We investigate nonlinear instrumental variable (IV) regression given highdimensional instruments. We propose a simple algorithm which combines kernelized IV methods and an arbitrary, adaptive regression algorithm, accessed as a black box. Our algorithm enjoys faster-rate convergence and adapts to the dimensionality of informative latent features, while avoiding an expensive minimax optimization procedure, which has been necessary to establish similar guarantees. It further brings the benefit of flexible machine learning models to quasi-Bayesian uncertainty quantification, likelihood-based model selection, and model averaging. Simulation studies demonstrate the competitive performance of our method. 1 Introduction Instrumental variable (IV) analysis is widely used for causal inference [1 3]. Given confounded observational data, IV analysis identifies the causal effect through the use of instruments. Nonlinear IV regression is typically defined by the following conditional moment restrictions (CMRs): E(y f0(x) | z) = 0 a.s. [P(dz)] (1) where f0 is the causal effect function of interest, and x, y, z denote the observed treatment, response and instrument, respectively. Similar CMR problems also appear in other applications of causal statistics and machine learning [see, e.g., 4, 5, for examples]. Starting from [6], recent works have demonstrated great promise in applying flexible machine learning (ML) methods to IV regression. Modern ML methods are appealing due to their adaptivity to the informative latent structure in data [7]: they may adapt to the low dimensionality of the informative latent features even if the observed input is high-dimensional and its signal-to-noise ratio is low. Sample complexity gaps have been established in such settings, between deep models based on neural networks (NNs) or trees, and linear models such as fixed-form kernels [8 10]. In IV regression, such adaptivity will be highly desirable when the observed instruments are high-dimensional, which is prevalent in applications such as genomics [11], and may also arise from the general desire to use structured data as instrument. Following previous work in the parametric setting [12, 13], we refer to this problem of learning informative latents in instruments as instrument learning. It generalizes the classical problem of instrument selection [14 16]. Comparing with standard supervised learning, IV regression is more challenging, due to the need to estimate a conditional expectation operator which defines (1). Consequently, establishing adaptivity guarantees becomes more difficult. While many recent works have demonstrated promising empirical results using deep models, they are often used as heuristics [e.g., 6], or justified with crude slow-rate analyses, which establish convergence rates that saturate at Ω(n 1/4) [e.g., 17]. This is in contrast to faster rates which approach n 1/2 as the regularity of model improves. The only exception is JZ is the corresponding author. Code available at https://github.com/meta-inf/fil. 36th Conference on Neural Information Processing Systems (Neur IPS 2022). a minimax formulation of IV estimation [5, 18 20]. [19] establish faster rate convergence for this formulation, for models with local Rademacher complexity bounds. Though local Rademacher analysis covers many adaptive ML procedures [e.g., 10, 21], it still does not fully explain the success of modern ML approaches, with prominent alternatives including implicit regularization [22] and PAC-Bayesian analyses [23]. From a practical perspective, minimax optimization is computationally expensive, yet it cannot be avoided in the framework of [19], unless we instantiate their method with the less flexible kernel models. It requires additional hyperparameter tuning, which can be challenging in causal problems where validation is indirect and more difficult. It also prevents the use of such flexible models for uncertainty quantification, or inference, for which reliable methods have only been developed for linear nonparametric models [24 26]. This work bridges the gap between sharp theoretical guarantees and robust, practical implementation. We assume our prior knowledge about the causal effect function f0 is characterized by a reproducing kernel Hilbert space (RKHS) H, and focus on the flexibility of conditional expectation estimation. This is often possible, because the treatment variable is determined by the problem at hand, and thus has a fixed dimensionality.2 We then present a surprisingly simple algorithm, with faster rate guarantees which in many cases match the best known in literature. The algorithm defines the conditional expectation estimates using a learned kernel, the basis of which is defined by applying adaptive regression algorithms to random draws from a Gaussian process (GP) prior. Given this learned first-stage kernel and H, we can estimate f0 using kernelized IV methods [19, 20, 27], which have closed-form solutions and can be efficiently approximated (e.g., with Nyström [19]). Our method allows easy hyperparameter tuning, and exhibits competitive performance in simulations. It accesses the regression algorithm as a black box, thus allowing for the use of any ML methods and benefits from their established theoretical guarantees. It also enables fast quasi-Bayesian uncertainty quantification [24, 28, 29] with improved flexibility. Our algorithm connects to many ideas in literature. Most notably, it can be viewed as an infinitedimensional generalization of [12, 13], which consider a linear outcome model with fixed dimensionality, and use ML methods to learn the optimal instruments [30, 31]. Our setting requires different analyses, for defining an infinite-dimensional estimation target and quantifying errors with its intrinsic complexity. Additionally, analysis of the resultant IV estimator is complicated by the ill-posedness of infinite dimensional IV models [32]. A more subtle distinction is in the choice of basis: while for finite-dimensional function spaces we can pick any set of basis (i.e., features) and apply the black-box regressor separately, in our case seemingly obvious choices of basis lead to inferior results (Appendix C.5.3). Our analysis also connects to the multi-task learning literature, to which we make technical contributions. Section 6 discusses related work in detail. The remaining of the paper is organized as follows: Section 2 reviews background knowledge. Section 3 introduces the instrument learning problem, and reduces it to a general kernel learning problem. We solve the latter problem in Section 4, and return to IV in Section 5 with our main results. We review related work in Section 6, and present numerical experiments in Section 7. 2 Notations and Setup Notations Denote the joint data distribution as P(dz dx dy), its marginal distributions as P(dx), P(dz), etc., and their support as X, Z. For functions of observed variables (e.g., x or z), 2 denotes the L2 norm w.r.t. the respective marginal data distribution. denotes the L norm. We use the notation [m] := {1, . . . , m}. Boldface (x, y, z) emphasizes the denotation of random variables. For any kernel k, GP(0, k) refers to the standard Gaussian process [33] with zero mean, and covariance defined by k. , , represent (in)equalities up to constants; the hidden constants will not depend on any sample size. O( ) denotes inequality up to logarithm factors. Problem Setup Nonparametric IV regression (NPIV) is formulated as (1) [34, 35]. Introduce the conditional expectation operator E : L2(P(dx)) L2(P(dz)), f 7 E(f(x) | z = ), and define g0(z) := E(y | z = z). We can then express (1) as a linear inverse problem: Ef0 = g0, (2) where we observe g0 up to regression error. NPIV deviates from standard inverse problems in its need to estimate both f0 and E. Following conventions in the two stage least square method [1], we 2Our method can also be applied when x contains high-dimensional exogenous covariates; see Appendix G. refer to the modeling of f0 as the second stage, and that of E or equivalently, that of Ef for all f in a hypothesis space as the first stage. The following assumption describes the setup in full detail: Assumption 2.1 (NPIV). (i) The data variables x, y, z satisfy (1), and y is bounded by B. (ii) We observe two sets of i.i.d. samples, with matching (marginal) distributions: D(n2) s2 := {(zi, xi, yi) : i [n2]}, D(n1) s1 := {( zi, xi) : i [n1]}. We impose (ii) since such additional samples are sometimes available, as discussed in e.g. [27]. If only n samples from the joint distribution are available, we can set n1 = n2 = n/2. In the main text, we assume that our prior knowledge about f0 is fully characterized by an RKHS H, in the following sense. Assumption 2.2 (second stage RKHS). (i) X is a bounded subset of Rdx; the reproducing kernel kx of H is bounded and continuous. (ii) The integral operator Tx : f 7 R kx(x, )f(x)P(dx) has eigenvalues λi(Tx) i (b+1), for some b > 0. (iii) One of the following holds true: (iii).a ( kernel scheme ): f0 H. (iii).b ( GP scheme ): b > 1; and for all n, f n H s.t. f n f0 H n 1/2 b+1 , f n f0 2 n b/2 In the above, (i) and (ii) are common technical assumptions: (i) ensures the existence of Mercer s representation, and (ii) is a complexity measure, with a larger value of b indicating a smaller hypothesis space. (iii) requires H is correctly specified for regression; its two cases cover the different assumptions in standard RKHS-based estimation and GP modeling. (iii).a is intuitive. (iii).b is standard in the posterior contraction literature [36]; it roughly requires f0 to be (at least) as regular as typical samples from GP(0, kx), in the sense of [36, Theorem 2.1]. This is different from (iii).a because when H is infinite dimensional, almost all GP samples fall out of H [37, 38]. Our algorithm applies to both settings, but analysis of the GP scheme requires additional effort. It is useful as it allows for quasi-Bayesian uncertainty quantification using a GP(0, kx) prior [26]. The RKHS assumption has been employed in a thread of recent work [27, 20, 39, 26], and generalizes the sieve method in literature [34, 40, 4]. It is most reasonable when x has moderate dimensions; for example, when f0 satisfies certain L2-Sobolev regularity conditions, we can set kx to be a suitable Matérn kernel (Example A.1-A.2). Appendix G studies a more general setting, where x and z include additional, high-dimensional exogenous covariates. Nonetheless, the assumption will be less reasonable when the treatment variable is high-dimensional and variable selection is needed for it. NPIV is typically an ill-posed inverse problem [35]. We now quantify the degree of ill-posedness: Assumption 2.3. The operator E is compact, with singular values si(E) i p, where p > 0. Such mildly ill-posed settings [41] match our polynomial eigendecay assumption for the kernel. In the severely ill-posed setting where the decay of si(E) is exponential, kernels with a similar eigendecay should be used. While the analyses of the two settings share many ideas, the Bayesian inverse problem literature typically restricts to the former for technical reasons [42, 43]. 3 From Instrument to Kernel Learning As we assume H is a correctly specified second-stage model, it remains to determine the first stage. In this section, we show that an ideal first stage model can be defined using another RKHS I, determined by H and E. Although its kernel kz has an unknown form, we demonstrate that we can access noisy samples from GP(0, kz), which, as Section 4 below shows, enable efficient learning of I. This can be viewed as instrument learning, as I will only depend on the informative features in z (Example 3.1). Let us first consider the GP scheme (Assumption 2.2 (iii).b) which roughly requires f0 to be similar to typical samples from GP(0, kx). From a Bayesian perspective, an ideal prior for Ef0 should match the distribution of Ef, for f GP(0, kx). This distribution is almost equivalent to another GP: Lemma 3.1 (proof in Appendix B). Denote by [g] the L2 equivalence class of g.3 Under Assumptions 2.1, 2.2, there exists a kernel kz, with integral operator Tz = ETx E , s.t. for f GP(0, kx), g GP(0, kz), [g] has the same distribution as E[f] . Informally, the lemma shows that GP(0, kz) matches the distribution of Ef. It is thus intuitive that kz could be a good choice for the first stage. The following lemma further motivates its use in the kernel scheme: its (i) shows that I fulfills the conditions in previous work [19, 27]: the restriction of E on H has image contained in I, and is a bounded linear map to I. (ii) shows that I, as a set of functions, cannot be made smaller while maintaining (i) . Lemma 3.2 (proof in Appendix B). Let I be the RKHS defined by kz. Under Assumptions 2.1, 2.2, (i) for any f H, there exists g I s.t. [g] = E[f] ; (ii) for any g I, there exists f H satisfying the above. In both cases, we have f H= g I. We now demonstrate that I only depends on the informative latent features. Example 3.1 (informative latent structure). Let Φ : Z Z be a feature extractor that maps the observed instruments z to latent features z := Φ(z), s.t. E(f(x) | z) = E(f(x) | Φ(z)) for all L2integrable f. Then we can apply Lemma 3.1, with E replaced by E : f 7 E(f(x) | z) L2(P(d z)), leading to a latent-space RKHS I with kernel kz. I induces the input-space RKHS I := {g = g Φ : g I}, g Φ I:= g I; kz(z, z ) = kz(Φ(z), Φ(z )). The above kz satisfies Lemma 3.1-3.2.4 Observe that I perfectly approximates {Ef : f H}, but its complexity only depends on I. In particular, kz has the same Mercer eigenvalues as kz (Claim B.2), the decay of which is a standard complexity measure [e.g., 44, Ch. 7]. While kz has ideal properties, it cannot be used directly as it involves the unknown operator E. Instead, we need to construct an approximation from data. Our main insight is that we can effectively draw noisy samples from GP(0, kz); as we develop in Section 4, such samples enable the approximation of kz. To see how the noisy samples are obtained, consider f GP(0, kx). By Lemma 3.1, g = Ef is L2-equivalent to clean samples from GP(0, kz); and we have f(x) = g(z) + (f(x) (Ef)(z)), where the latter term is unpredictable given z, and from this perspective can be viewed as noise. Thus, if we apply any regression algorithm to f GP(0, kx), with z as input, we will recover a denoised sample from GP(0, kz), up to regression errors. In the informative latent feature setting, optimal regression error can only be achieved by methods that adapt to such structures [8 10]. Approximating I with such denoised samples can then be viewed as a knowledge distillation procedure, which results in a compact representation of the adaptive regression algorithm. This is particularly beneficial in the NPIV setting: as discussed in the introduction, using a learned kernel eliminates the need of minimax optimization in estimation, and allows the (indirect) use of adaptive methods for uncertainty quantification. 4 Black-Box Kernel Learning In this section, we address the problem of kernel learning given noisy GP samples. As our results apply to more general settings, we first state the assumptions with full generality. Assumption 4.1 (RKHS). There exist a continuous function Φ : Z Z, and a reproducing kernel kz over Z, s.t. (i) the random variable z = Φ(z) is supported on a bounded subset of Rdl; kz is bounded. (ii) The eigenvalues of the integral operator T z : g 7 R kz( z, )g( z)P(d z) satisfy λi(T z) i ( b+1), for some b > 0. (iii) g GP(0, kz) have finite sup norm with probability 1. The above assumption applies to a latent-space kernel kz. As shown in Example 3.1, Φ and kz induce an input-space kernel kz, and RKHS I, which inherit the assumed regularity conditions. Our goal is to estimate kz. This is harder than the estimation of kz, as it also involves Φ. 3Recall the L2 space is not a function space, and consists of equivalence classes of functions. Note that for readability, we may occasionally ignore this distinction in the main text, and use (a version of) E to also denote the corresponding map between function spaces. All such denotations can be made unambiguous (Remark B.1), and all null set ambiguities in this section can be removed under mild additional assumptions (Lemma B.5). 4There may be multiple kernels satisfying Lemma 3.1, but they are equivalent up to null sets (Claim B.2); the ambiguity can be removed under mild assumptions (Lemma B.5). All conditions for kz are satisfied by Matérn kernels with a sutiable order; see Appendix A. Appendix B.1 discusses its applicability in the IV setting, where I is defined as in Section 3. Briefly, (ii) always holds for b max{b, 2p 1}, and if H is further correctly specified in the sense of Assumption E.1, b = b + 2p. (i) and (iii) hold under mild technical assumptions. We will denoise noisy GP(0, kz) samples using a regression oracle, which is specified below: Assumption 4.2 (regression oracle). Let D(n1) := {( zi, g( zi) + ei)} be n1 iid replications of the rvs (z, g(z) + e), s.t. E(e | z) = 0 and g(z) + e has a 1-subgaussian distribution. Then the oracle returns estimator ˆgu,n1 s.t. Eg GP(0,kz)ED(n1) ˆgu,n1 g 2 2 ξ2 n1, for some ξn1 0. In the IV setting, we have g = Ef GP(0, kz), and E(e | z) = E(f(x) (Ef)(z) | z) = 0; the subgaussian condition is verified by Lemma A.4. To provide some intuition on adaptivity, we instantiate the assumption with the DNN model in [10], and compare the resulted ξn with fixed-form kernels: Example 4.1 (adaptivity of DNN oracles). Let Z Rdz, Φ : Z Z be β1-Hölder regular, I be a Matérn-β2 RKHS, and β1, β2 1. Let the regression oracle return a ϵ2 opt-approximate empirical risk minimizer for the model in [10]. Then for any ϵ > 0, it holds that (see Appendix C.5.1 for derivations) ξn = O n β1 2β1+dz + n β2 ϵ 2β2+dl + ϵopt =: O ϵfea,n + n β2 ϵ 2β2+dl + ϵopt (3) In the above, ϵfea,n characterizes the hardness of feature learning, i.e., learning Φ. The second term characterizes that of kernelized regression given the optimal features: it matches the optimal regression rate if we had full knowledge about Φ, or equivalently, I, and would be attainable by kernel ridge regression (KRR) using I. As long as ϵopt is small, (3) will match the minimax rate up to logarithms. When β1/dz < β2/dl, the minimax rate is ϵfea,n n β2/(2β2+dl), meaning that the hardness of feature learning cannot be overlooked. Otherwise, the rate ξn nearly matches the rate given full knowledge of the unknown I, up to the infinitesimal ϵ > 0; this is realistic when, e.g., dz dl and Φ is linear (β1 = ). We are interested in the high-dimensional regime where dz dl. In this case, fixed-form Matérn or RBF kernels could only attain the rate of O(n min{β1,β2} 2 min{β1,β2}+dz ), which can always be much worse than (3), regardless of the hardness of feature learning. This comparison suggests that fixed-form kernels cannot adapt to the latent feature structure to avoid the curse of dimensionality.5 We now define the approximate RKHS. Let {g(j) : j [m]} be i.i.d. samples from the GP prior, and ˆg(j) u,n1 be the respective estimate returned by the regression oracle, constructed from the shared dataset D(n1) = {( zi, g(j)( zi) + ϵ(j) i ) : i [n1], j [m]} where ϵ(j) i are subgaussian, meanzero noise. Let ˆg(j) n1 := min{Ck log n, ˆg(j) u,n1( )}, where Ck is a constant determined by I. Define ˆGn(z) := (ˆg(1) n1 (z), . . . , ˆg(m) n1 (z)). Our approximate RKHS is defined as I := {g(z) = θ ˆGn1(z) for some θ Rm}, with norm g I:= m θ 2. (4) As I is a finite-dimensional linear space, it is an RKHS. We can check that g Ck log n1 g I. Theoretical Results Under a given model, regression error is decomposed into approximation and estimation (i.e., generalization) errors. We first present the approximation error bound: Theorem 4.1 (proof in Appendix C.2). Under Assumptions 4.1, 4.2, there exists a universal constant cr > 0, and an event En1 determined by g(1...m) and D(n1) with PD(n1)En1 1, on which for any g L2(P(dz)), there exists g I s.t. g I cr Projm g I, (5) g g 2 cr Projm g I(ξn1 + m ( b+1)/2) p log n1 + g Projm g 2, (6) where m = [m/2], and Projm denotes the projection onto the top m Mercer eigenfunctions of kz. 5[10] establishes formal lower bounds. Also, for small β2, we can replace dz with a manifold dimensionality of Z, but it can still be much larger than dl. We will use I to estimate functions on a separate dataset with n2 samples. For a single regression task, the estimation error can be simply bounded as O( p m/n2) [45]. However, our analysis of IV estimation will require quantifying the intrinsic complexity of I, which will also allow the use of a larger m in practice. The following proposition provides one such result; it will be used in Section 5, to analyze IV estimation in the kernel scheme (Assumption 2.2 (iii).a). Proposition 4.2 (proof in Appendix C.3). Let I, D(n1) be defined as above, and δn2 be the critical radius of the local Rademacher complexity of the norm ball I1 [46, Ch. 14]. On the event defined in Theorem 4.1, we have δn2 = O(n ( b+1)/2( b+2) 2 + m ( b+1)/2 + ξn1). IV estimation in the GP scheme is more delicate, and requires additional analysis of I, which is deferred to App. D. Before we proceed, however, we illustrate the results on a simple regression task: Example 4.2 (Example 4.1, cont d). Let Φ, I be defined as before, and ˆg(j) n1 be estimated by the DNN oracle. Suppose ϵopt is not greater than the other terms. Then i. Let m = n b/( b+1)2 1 . On the event in Theorem 4.1, for any g I, there exists g I s.t. g I cr g I, g g 2= O( g Iξn1). ii. Let g GP(0, kz). A refined analysis, based on Corollary C.6, shows that when m = n1/( b+1) 1 , there exists g s.t. g I n1/2( b+1) 1 , Eg GP(0,kz) g g 2= O(ξn1). (See Appendix C.5.2 for derivations, and another high-probability bound in the GP scheme.) Let ˆg n be the truncated OLS estimate using I, on a dataset {(zi, g (zi) + ei) : i [n2]} where E(ei | zi) = 0, Var(ei) 1, g B. Then E ˆg n g 2= O( g g 2+B p m/n2) [45, Thm. 11.3]. When n1 = n2, the latter term is ξn2, and case (ii) above always matches the DNN rate. Case (i) matches the DNN rate when feature learning becomes harder (ξ2 n1 n b/( b+1) 1 ); otherwise the rate may be slightly inferior, but still approaches n 1/2 1 as the regularity b improves. As discussed in Example 4.1, when dz > dl, the DNN rate can outperform fixed-form kernels by a large margin. The above example demonstrates a similar superiority of the learned kernel. 5 Results for IV Regression We shall use the approximate first stage I for IV regression, by plugging I and H to the kernelized estimators in [19, 26]; see Algorithm 1. We analyze the resulted estimators in this section, while deferring implementation details, including hyperparameter selection, to Appendix F. Algorithm 1 Kernelized IV with learned instruments. Require: D(n1) s1 , D(n2) s2 ; regression algorithm Regress; second-stage kernel kx; m N 1: for j 1 to m do 2: Sample f (j) GP(0, kx) 3: ˆg(j) u,n1 Regress({( zi, f (j)( xi)) : i [n1]}) 4: end for {the m invocations of Regress may be replaced with a single vector-valued regression} 5: Define kz(z, z ) := 1 m Pm j=1 ˆg(j) n1 (z)ˆg(j) n1 (z ),where ˆg(j) n1 := min{ˆg(j) u,n1( ), C log m}. 6: return Kernelized IV(D(n2) s2 , kz, kx) {See (7) below, or Appendix F for the closed-form solution} Both [19] and the posterior mean estimator of [26] have the form arg min f H ℓn2(f)+µ f 2 H:= arg min f H max g I 1 n2 i=1 (yi f(xi) κg(zi))g(zi) λ g 2 I+µ f 2 H. (7) Their difference lies in the regularization scaling, which arises from the different assumptions about f0 and H (Assumption 2.2). Thus, we analyze the resulted two estimators separately, in Section 5.1 and Section 5.2 below. In the setting of [26] we are also able to justify the use of likelihood-based model comparison and (quasi-)Bayesian model averaging (BMA). 5.1 Estimation in the Kernel Scheme [19] establish faster rate convergence of the point estimator under simple assumptions. We now provide corresponding results using our learned I, by plugging in the results in Section 4. Proposition 5.1 (proof in Appendix E.1). Assume Assumptions 2.1, 2.2 (kernel scheme), 4.1 and 4.2. Let I be defined by kz in Algorithm 1, and ˆfn2 be defined by (7), with κ, λ, ν set as in Appendix E.1. On the event defined in Theorem 4.1, we have E( ˆfn2 f0) 2= Op((ξn1 + n b+1 2(b+2) 2 )(1 + f0 2 H)). (8) Let us compare the result with [19] in the setting of Example 4.1. Suppose n1 = n2. [19] establishes the rate of Op((ξ n2 + n (b+1)/(b+2) 2 )(1 + f0 2 H)), where ξ n2 is comparable with a first-stage regression rate established from local Rademacher analysis. For the DNN model in Example 4.1, we have ξ n = Θ(ξn) = O(ϵfea,n+n b/2( b+1)) in the general case,6or ξ n = O(ϵfea,n+n ( b+1)/2( b+2)) assuming additional regularity (Remark C.3). Thus, the two rates are equivalent if ϵfea is sufficiently large, meaning that the difficulty of feature learning cannot be ignored. Otherwise, [19] may be better if b < b + 1; this is a somewhat narrow range, as b max{2p 1, b}. Recall that our method is more appealing computationally: directly instantiating [19] with DNNs requires solving a minimax problem similar to (7), while for our learned kernel (7) can be evaluated in closed form. We can also compare (8) with kernelized IV using a fixed-form first stage. In the above setting, its best rate is also provided by [19], and is dominated by the kernel regression error in Ex. 4.1 which, as we discussed, can be much worse than ξn. Our improved rate has been made possible by the fact that we are approximating a first-stage model with optimal adaptivity (Ex. 3.1), at a rate that is also adaptive to the informative latent structure (Ex. 4.1). In summary, our algorithm combines the best of both worlds: it maintains the sharp guarantees of adaptive models, and the simplicity of kernel methods. 5.2 Quasi-Bayesian Estimation and Uncertainty Quantification Quasi-Bayesian analysis enables efficient uncertainty quantification for NPIV, without introducing extra risks of model misspecification [24, 28]. [26] studies a quasi-Bayesian posterior constructed from (7) and a GP(0, kx) prior. It is defined through the Radon-Nikodym derivative w.r.t. the prior, (dΠ( |D(n2) s2 )/dΠ)(f) e n2ℓn2(f). For kernel first-stage models, the quasi-posterior can be evaluated in closed form (App. F). For general models, however, it is entirely unclear if approximate inference can be possible, since for any parameter f, evaluation of ℓn(f) involves solving a separate optimization problem. Our kernel learning algorithm enables the (indirect) use of such models. Analysis of (quasi-)Bayesian procedures is more challenging, partly because of the weaker regularization. Thus, [26] introduced additional assumptions. Our analysis is further complicated by a different assumption on I, and approximation errors in I, which necessitate further assumptions. App. E.2 discusses these assumptions in detail. For simplicity, we state the result in a rate-optimal case:7 Theorem 5.2 (posterior contraction; proof in App. E.3). Assume Asms. 2.1, 2.2 (GP scheme), 2.3, 4.1, 4.2, D.1, D.2, E.1, E.2. Let n1 be s.t. ξ2 n1 log n2 + n (b+2p)/(b+2p+1) 1 n 1 2 , and m n1/(b+2p+1) 1 . Let Πn1( | D(n2) s2 ) be defined in (52) in appendix. Then, with D(n1) s1 -probability 1, we have ED(n2) s2 Πn1({f : f f0 2 M ϵn2} | D(n2) s2 ) 0, ED(n2) s2 Πn1({f : E(f f0) 2 M δn2} | D(n2) s2 ) 0, where δn2 = O(n (b+2p)/2(b+2p+1) 2 ), ϵn2 = O(n b/2(b+2p+1) 2 ). 6With some abuse of notation, we also use O to hide the infinitesimal deterioration of the polynomial order. 7Classical NPIV lower bounds continue to hold given full knowledge of E [47], so the rate n b/2(b+2p+1) 2 is minimax optimal irrespective of n1. In our setting, it is certainly desirable to improve the dependency on n1, and our restriction is only employed to simplify proof. In simulations we find the choice of n1 = n2 works well. Theorem 5.2 immediately implies Theorems 5, 6 in [26] for our I, with the extra logarithms, as their proofs do not involve the first stage. Those results establish Sobolev norm rates, and justify uncertainty quantification by lower bounding the magnitude of posterior spread. In the nonparametric Bayes literature, contraction results like Theorem 5.2 often lead to the justification of marginal likelihood-based model selection and averaging. This is also the case here. The key ingredient is the following marginal quasi-likelihood bound: Corollary 5.3 (proof in Appendix E.4). In the setting of Theorem 5.2, for some C > 0 we have PD(n2) s2 (C 1n 1 b+2p+1 2 log 6 b n2 log Πn1(D(n2) s2 ) Cn 1 b+2p+1 2 log2 n2) 1. This result allows the comparison of a finite number of second-stage RKHSes. Of particular interest is the comparison between power RKHSes (Defn. A.1), which often have intuitive interpretations: e.g., for a Matérn RKHS H and γ (2/(b + 1), 1), the power RKHS Hγ is equivalent to lower-order Matérn RKHSes ([48]; Ex. A.3). We can verify that such Hγ fulfills the assumptions about H. Thus, provided the other assumptions continue to hold, Corollary 5.3 will hold for all such Hγ, with b + 1 replaced by γ(b + 1), showing the marginal likelihood has a different asymptotics. Consequently, it establishes asymptotically valid comparison between such models, and justifies the use of BMA. Analysis of more general settings requires additional effort: NPIV is an inverse problem, and we anticipate the subtleties of model selection in nonparametric inverse problems. For example, analyses are usually restricted to the selection of γ [43, 49, 50], and the γ > 1 case requires additional assumptions [49].8 In the IV setting, it should also be noted that valid model comparison requires a good approximation to E|H, since otherwise the quasi-likelihood becomes less meaningful at any finite sample size. The same intuition applies to other model selection procedures [18, 20, 27] based on the estimated violation of (1). When the approximation cannot be guaranteed, it could be preferable to stick to the prior knowledge and fix a conservative choice for H. 6 Related Work Multi-Task Learning Our Example 4.2 can also be viewed as quantifying sample efficiency improvements in multi-task learning, if we view the GP prior draws as the labeling functions for a handful of diverse training tasks, which share the representation Φ. This general idea is not new: starting from [51, 52], a line of recent work establishes similar results. Most related is [52, Sec. 5], which assumes a fixed-dimensional linear model for g, and an adaptive Φ with metric entropy bounds. We assume more general models for both components, and do not require different training tasks to have separate inputs. On the flip side, [52] allows for non-iid training tasks. [52, Sec. 6] investigated infinite-dimensional g, but established a slow rate. We are unaware of any work that established fast-rate convergence for infinite-dimensional top-level models, or used ML models as a black box. Both aspects may be interesting for multi-task learning, and are necessary for instrument learning. Causal Statistics The double machine learning framework [53] also uses black-box ML models to estimate certain nuisance parameters in the model. While the operator E can be viewed as a nuisance parameter, the structure of the NPIV problem is quite different: [54, p. 8] noted that it is very unclear if such a view can be helpful for NPIV estimation; consistent with their remarks, we have also been unable to cast our problem into the double ML framework. Note that double ML has been applied to semiparametric estimation and inference for IV [53, 55 57], which are orthogonal to our goal. It has long been known [30, 31] that under a linear outcome model f0(x) = θ x, using E(x | z) as instrument leads to n-consistent estimates. Our Section 3 can be viewed as an infinite-dimensional generalization of this observation.9 Given high-dimensional instruments and a parametric outcome model, there is a large body of literature on efficient inference; see [12] for a review. As we move to nonparametric models, we focus on estimation which becomes much more challenging, in the spirit of [54]. Still, we have provided qualitative characterization for uncertainty estimates in Section 5.2. For the use of ML for nonlinear IV, [6] studied a heuristic application of NNs. We discussed the minimax formulation in introduction. [5, 26] justified the use of NNs with the respective neural 8We do not cover it here for brevity, noting that it is well-understood in inverse problem settings [43, 49]. 9As noted in [27], when f0 H for some RKHS H, the first stage should model E(f(x) | z) for all f H, as opposed to merely modeling E(x | z). Note that [27] did not study the optimal choice of the first stage. step sin sigmoid linear abs 3dpoly 500 2500 5000 500 2500 5000 500 2500 5000 500 2500 5000 500 2500 5000 500 2500 5000 500 2500 5000 AGMM-Tree AGMM-NN AGMM-RBF Proposed Figure 1: Predictive performance: test MSE vs sample size n1 = n2 for all method, and D {2, 100}. Full results are in App. H.2. Table 1: Runtime results for all methods in the predictive experiment, for N = 2500, D = 100. Method AGMM-Tree AGMM-NN AGMM-RBF Proposed Runtime / s 1374 418 303 16 6.7 0.1 25.9 5.6 tangent kernels (NTKs) which, like other fixed-form kernels, cannot adapt to the informative latent structure [8, 9]. [39, 58] investigated the combination of an NN-based second stage and a linear first stage, which could be useful in complementary scenarios. [17] considered feature learning in both stages, but only established a slow rate; as the authors noted, it is also unclear if their algorithm reliably minimizes the empirical risk. For model selection in the setting of Section 5.2, [59] prove the validity of bootstrap-based selection for the sieve estimator [34]. [39, 60] investigate the use of marginal likelihood for two different kernel-based IV estimators: [60] establish a crude 1/4 log n upper bound for the log marginal likelihood, and [39] connect it to the empirical leave-one-out validation error. Neither result fully justifies model selection as our Corollary 5.3. For kernelized IV models, [20, 27, 61] proposed validation statistics for comparing a finite number of first stage models. 7 Simulation Study Our main simulation setup is adapted from [18, 19]; Appendix H.4 presents additional experiment on the demand dataset [6, 17]. In [18, 19], the observed z, x, y are generated by z Unif[ 3, 3] D 2 , z = h( z), u N(0, 1), x := z1 + u + ex, y := (f0(x) + u + ey µ)/σ, where u is the confounder, ex, ey N(0, 0.12) are independent noise, and the constants µ, σ standardize y. We consider three choices for h: (i) D = 2, h is the identity function; this recovers the setup in previous work, and quantifies the hardness of the NPIV problem given true instruments. (ii) dim z = D {40, 100}, h is a three-layer DNN; this simulates a feature learning scenario, and ensures the observation has a low signal-to-noise ratio (O(1/D)). (iii) h maps z1 to a MNIST [62] or CIFAR-10 [63] image with matching label; the MNIST setting also appeared in previous work. We consider two choices for f0: (i) a widely used collection of functions (e.g., sin, abs) in [18]. (ii) f0 GP(0, kx). (ii) ensures the correct specification of H and allows us to focus on the first stage. We use a DNN as the black-box learner, and a RBF kernel for H, with bandwidth determined by marginal likelihood (72). We set N1 = N2 {500, 2500, 5000}. We defer setup details and full results to Appendix H, and summarize the findings below: Hyperparameter Selection (App. H.1) We first study hyperparameter selection in instrument learning. We set f0 GP, D {2, 40, 100}. We find our validation statistics (71) always correlates with the counterfactual MSE ˆfn f0 2 2, and that across a large hyperparameter space, trained DNNs always outperform first-stage models based on RBF kernels, or randomly initialized DNNs. Predictive Performance (App. H.2) For h defined as in (i-ii), we compare our algorithm with [19, AGMM], instantiated with kernel, tree and NN models. As shown in [19], the baselines have Table 2: Test MSE, radius and estimated coverage rate of the 90% L2 credible ball (CB), and the average coverage of pointwise 90% credible interval (CI), for f0 GP, D = 100. For the CB coverage rate estimate, we report its 95% Wilson score interval [64]. Full results are in App. H.3. Method n1 = n2 Test MSE 90% CB. Rad. 90% CB. Cvg. 90% CI. Cvg. Proposed 500 .097 .065 .201 .025 .923 [.888, .948] .915 .123 2500 .035 .024 .074 .008 .917 [.880, .943] .908 .127 5000 .024 .016 .049 .004 .920 [.884, .946] .905 .134 RBF 500 .431 .192 .240 .036 .187 [.147, .235] .640 .191 2500 .176 .089 .175 .023 .517 [.460, .573] .822 .136 5000 .126 .072 .156 .019 .660 [.605, .711] .855 .143 competitive performance on this setup; the latter two models also enjoy adaptivity guarantees. A representative subset of results are plotted in Fig. 1: our method has stable performance as we move to high dimensions, demonstrating excellent adaptivity. In contrast, fixed-form kernels fail to identify the informative features. AGMM-tree and AGMM-NN also have deteriorated performance as D increases, despite their theoretical guarantees, presumably due to the challenges in optimization. Table 1 reports the run time of all methods in this experiment. As we can see, our method is more efficient than both adaptive baselines. For image-based h, we compare with AGMM-NN and [39], which report the best results in the MNIST setting. Our method outperforms both baselines. Uncertainty Quantification (App. H.3) Table 2 presents a subset of results for f0 GP. Comparing with a fixed-form RBF first stage, our method produces sharper credible intervals, which also have better coverage. For f0 specified as in [18], we experiment with BMA over a grid of RBF kernels, and present visualizations in Appendix H.3. We find that when the model is more correctly specified, BMA produces conservative uncertainty estimates which are nonetheless informative. However, when all models are severely misspecified (e.g., when f0 is a step function), we cannot expect model-based uncertainty estimates to have ideal coverage. Exogenous Covariates (App. H.4) We evaluate the extended algorithm in Appendix G on the demand dataset [6], which is a widely used simulation design with high-dimensional exogenous covariates. As shown in the appendix, our extended algorithm has competitive performance. Acknowledgements This work was supported by the National Key Research and Development Program of China (2020AAA0106302); NSFC Projects (Nos. 62061136001, 62076145, 62076147, U19B2034, U1811461, U19A2081, 61972224), Beijing NSF Project (No. JQ19016), BNRist (BNR2022RC01006), Tsinghua Institute for Guo Qiang, and the High Performance Computing Center, Tsinghua University. J.Z is also supported by the XPlorer Prize. [1] J. D. Angrist and J.-S. Pischke, Mostly harmless econometrics. Princeton university press, 2008. [2] M. A. Brookhart, J. A. Rassen, and S. Schneeweiss, Instrumental variable methods in comparative safety and effectiveness research, Pharmacoepidemiology and drug safety, vol. 19, no. 6, pp. 537 554, 2010. [3] S. Burgess, D. S. Small, and S. G. Thompson, A review of instrumental variable estimators for mendelian randomization, Statistical methods in medical research, vol. 26, no. 5, pp. 2333 2355, 2017. [4] X. Chen and D. Pouzo, Estimation of Nonparametric Conditional Moment Models With Possibly Nonsmooth Generalized Residuals, Econometrica, vol. 80, no. 1, pp. 277 321, 2012. [5] L. Liao, Y.-L. Chen, Z. Yang, B. Dai, Z. Wang, and M. Kolar, Provably Efficient Neural Estimation of Structural Equation Model: An Adversarial Approach, ar Xiv:2007.01290 [cs, stat], 2020, ar Xiv: 2007.01290. [6] J. Hartford, G. Lewis, K. Leyton-Brown, and M. Taddy, Deep IV: A flexible approach for counterfactual prediction, in International Conference on Machine Learning, 2017, pp. 1414 1423. [7] F. Bach, Breaking the curse of dimensionality with convex neural networks, The Journal of Machine Learning Research, vol. 18, no. 1, pp. 629 681, 2017. [8] C. Wei, J. D. Lee, Q. Liu, and T. Ma, Regularization matters: Generalization and optimization of neural nets v.s. their induced kernel, in Advances in Neural Information Processing Systems, H. Wallach, H. Larochelle, A. Beygelzimer, F. d Alché Buc, E. Fox, and R. Garnett, Eds., vol. 32, 2019. [9] B. Ghorbani, S. Mei, T. Misiakiewicz, and A. Montanari, Limitations of lazy training of two-layers neural networks, ar Xiv preprint ar Xiv:1906.08899, 2019. [10] J. Schmidt-Hieber, Nonparametric regression using deep neural networks with Re LU activation function, The Annals of Statistics, vol. 48, no. 4, 2020, ar Xiv: 1708.06633. [11] W. Lin, R. Feng, and H. Li, Regularization Methods for High-Dimensional Instrumental Variables Regression With an Application to Genetical Genomics, Journal of the American Statistical Association, vol. 110, no. 509, pp. 270 288, 2015. [12] A. Singh, K. Hosanagar, and A. Gandhi, Machine learning instrument variables for causal inference, in Proceedings of the 21st ACM Conference on Economics and Computation, 2020, pp. 835 836. [13] J. Chen, D. L. Chen, and G. Lewis, Mostly Harmless Machine Learning: Learning Optimal Instruments in Linear IV Models, ar Xiv:2011.06158 [cs, econ, stat], 2021, ar Xiv: 2011.06158. [14] R. Okui, Instrumental variable estimation in the presence of many moment conditions, Journal of Econometrics, vol. 165, no. 1, pp. 70 86, 2011. [15] M. Carrasco, A regularization approach to the many instruments problem, Journal of Econometrics, vol. 170, no. 2, pp. 383 398, 2012. [16] A. Belloni, D. Chen, V. Chernozhukov, and C. Hansen, Sparse models and methods for optimal instruments with an application to eminent domain, Econometrica, vol. 80, no. 6, pp. 2369 2429, 2012. [17] L. Xu, Y. Chen, S. Srinivasan, N. de Freitas, A. Doucet, and A. Gretton, Learning deep features in instrumental variable regression, ar Xiv preprint ar Xiv:2010.07154, 2020. [18] A. Bennett, N. Kallus, and T. Schnabel, Deep generalized method of moments for instrumental variable analysis, in Advances in Neural Information Processing Systems, 2019. [19] N. Dikkala, G. Lewis, L. Mackey, and V. Syrgkanis, Minimax estimation of conditional moment models, in Advances in Neural Information Processing Systems, vol. 33, 2020, pp. 12 248 12 262. [20] K. Muandet, A. Mehrjou, S. K. Lee, and A. Raj, Dual instrumental variable regression, ar Xiv preprint ar Xiv:1910.12358, 2020. [21] V. Syrgkanis and M. Zampetakis, Estimation and Inference with Trees and Forests in High Dimensions, ar Xiv:2007.03210 [cs, math, stat], 2020, ar Xiv: 2007.03210. [22] B. Neyshabur, R. Tomioka, and N. Srebro, In search of the real inductive bias: On the role of implicit regularization in deep learning, ar Xiv preprint ar Xiv:1412.6614, 2014. [23] P. Alquier, User-friendly introduction to PAC-Bayes bounds, ar Xiv preprint ar Xiv:2110.11216, 2021. [24] K. Kato, Quasi-Bayesian analysis of nonparametric instrumental variables models, The Annals of Statistics, vol. 41, no. 5, pp. 2359 2390, 2013. [25] X. Chen and T. M. Christensen, Optimal sup-norm rates and uniform inference on nonlinear functionals of nonparametric IV regression, ar Xiv preprint ar Xiv:1508.03365, 2015. [26] Z. Wang, Y. Zhou, T. Ren, and J. Zhu, Quasi-Bayesian dual instrumental variable regression, ar Xiv preprint ar Xiv:2106.08750, 2021. [27] R. Singh, M. Sahani, and A. Gretton, Kernel Instrumental variable regression, ar Xiv preprint ar Xiv:1906.00232, 2019. [28] V. Chernozhukov and H. Hong, An MCMC approach to classical estimation, Journal of Econometrics, vol. 115, no. 2, pp. 293 346, 2003. [29] J.-P. Florens and A. Simoni, Nonparametric estimation of an instrumental regression: A quasi Bayesian approach based on regularized posterior, Journal of Econometrics, vol. 170, no. 2, pp. 458 475, 2012. [30] H. H. Kelejian, Two-stage least squares and econometric systems linear in parameters but nonlinear in the endogenous variables, Journal of the American Statistical Association, vol. 66, no. 334, pp. 373 374, 1971. [31] G. Chamberlain, Asymptotic efficiency in estimation with conditional moment restrictions, Journal of Econometrics, vol. 34, no. 3, pp. 305 334, 1987. [32] J. L. Horowitz, Ill-Posed Inverse Problems in Economics, Annual Review of Economics, vol. 6, no. 1, pp. 21 51, 2014. [33] J. van Zanten and A. van der Vaart, Reproducing kernel hilbert spaces of gaussian priors, in Pushing the limits of contemporary statistics: contributions in honor of Jayanta K. Ghosh. Institute of Mathematical Statistics, 2008, pp. 200 222. [34] W. K. Newey and J. L. Powell, Instrumental variable estimation of nonparametric models, Econometrica, vol. 71, no. 5, pp. 1565 1578, 2003. [35] J. L. Horowitz, Applied nonparametric instrumental variables estimation, Econometrica, vol. 79, no. 2, pp. 347 394, 2011. [36] A. W. van der Vaart and J. H. van Zanten, Rates of contraction of posterior distributions based on Gaussian process priors, The Annals of Statistics, vol. 36, no. 3, 2008. [37] M. F. Driscoll, The reproducing kernel hilbert space structure of the sample paths of a gaussian process, Zeitschrift für Wahrscheinlichkeitstheorie und verwandte Gebiete, vol. 26, no. 4, pp. 309 316, 1973. [38] A. W. van der Vaart and J. H. van Zanten, Information rates of nonparametric Gaussian process regression, Journal of Machine Learning Research, vol. 12, pp. 2095 2119, 2011. [39] R. Zhang, M. Imaizumi, B. Schölkopf, and K. Muandet, Maximum moment restriction for instrumental variable regression, ar Xiv preprint ar Xiv:2010.07684, 2020. [40] R. Blundell, X. Chen, and D. Kristensen, Semi-Nonparametric IV Estimation of Shape Invariant Engel Curves, Econometrica, vol. 75, no. 6, pp. 1613 1669, 2007, publisher: [Wiley, Econometric Society]. [41] L. Cavalier, Nonparametric statistical inverse problems, Inverse Problems, vol. 24, no. 3, p. 034004, 2008. [42] B. T. Knapik, A. W. van der Vaart, and J. H. van Zanten, Bayesian inverse problems with Gaussian priors, The Annals of Statistics, vol. 39, no. 5, pp. 2626 2657, 2011. [43] B. T. Knapik, B. T. Szabó, A. W. Van Der Vaart, and J. H. van Zanten, Bayes procedures for adaptive inference in inverse problems for the white noise model, Probability Theory and Related Fields, vol. 164, no. 3-4, pp. 771 813, 2016. [44] I. Steinwart and A. Christmann, Support Vector Machines. Springer Science & Business Media, 2008. [45] L. Györfi, M. Kohler, A. Krzyzak, H. Walk et al., A distribution-free theory of nonparametric regression. Springer, 2002, vol. 1. [46] M. J. Wainwright, High-Dimensional Statistics: A Non-Asymptotic Viewpoint. Cambridge University Press, 2019, vol. 48. [47] X. Chen and M. Reiss, On Rate Optimality For Ill-posed Inverse Problems In Econometrics, Econometric Theory, no. 3, pp. 497 521, 2007. [48] S. Fischer and I. Steinwart, Sobolev norm learning rates for regularized least-squares algorithms, Journal of Machine Learning Research, vol. 21, pp. 205 1, 2020. [49] B. Szabó, A. W. Van Der Vaart, and J. van Zanten, Frequentist coverage of adaptive nonparametric Bayesian credible sets, The Annals of Statistics, vol. 43, no. 4, pp. 1391 1428, 2015. [50] J. Jia, J. Peng, and J. Gao, Posterior contraction for empirical bayesian approach to inverse problems under non-diagonal assumption, Inverse Problems & Imaging, vol. 15, no. 2, pp. 201 228, 2021. [51] N. Tripuraneni, M. I. Jordan, and C. Jin, On the Theory of Transfer Learning: The Importance of Task Diversity, ar Xiv:2006.11650 [cs, stat], 2020, ar Xiv: 2006.11650. [52] S. S. Du, W. Hu, S. M. Kakade, J. D. Lee, and Q. Lei, Few-Shot Learning via Learning the Representation, Provably, ar Xiv:2002.09434 [cs, math, stat], 2021, ar Xiv: 2002.09434. [53] V. Chernozhukov, D. Chetverikov, M. Demirer, E. Duflo, C. Hansen, W. Newey, and J. Robins, Double/debiased machine learning for treatment and structural parameters, The Econometrics Journal, vol. 21, no. 1, pp. C1 C68, 2018. [54] D. J. Foster and V. Syrgkanis, Orthogonal statistical learning, ar Xiv preprint ar Xiv:1901.09036, 2019. [55] V. Syrgkanis, V. Lei, M. Oprescu, M. Hei, K. Battocchi, and G. Lewis, Machine learning estimation of heterogeneous treatment effects with instruments, ar Xiv preprint ar Xiv:1905.10176, 2019. [56] R. Singh, L. Xu, and A. Gretton, Generalized kernel ridge regression for nonparametric structural functions and semiparametric treatment effects, ar Xiv preprint ar Xiv:2010.04855, 2020. [57] Y. Jung, J. Tian, and E. Bareinboim, Double machine learning density estimation for local treatment effects with instruments, in Advances in Neural Information Processing Systems, A. Beygelzimer, Y. Dauphin, P. Liang, and J. W. Vaughan, Eds., 2021. [58] J. Chen, X. Chen, and E. Tamer, Efficient Estimation in NPIV Models: A Comparison of Various Neural Networks-Based Estimators, ar Xiv:2110.06763 [cs, econ, math, stat], 2021, ar Xiv: 2110.06763. [59] X. Chen, T. M. Christensen, and S. Kankanala, Adaptive estimation and uniform confidence bands for nonparametric iv, Cowles Foundation for Research in Economics, Yale University, Tech. Rep., 2021. [60] K. Hsu and F. Ramos, Bayesian deconditional kernel mean embeddings, in International Conference on Machine Learning. PMLR, 2019, pp. 2830 2838. [61] R. Zhang, K. Muandet, B. Schölkopf, and M. Imaizumi, Instrument Space Selection for Kernel Maximum Moment Restriction, ar Xiv:2106.03340 [cs], 2021, ar Xiv: 2106.03340. [62] Y. Le Cun, The mnist database of handwritten digits, http://yann. lecun. com/exdb/mnist/, 1998. [63] A. Krizhevsky, G. Hinton et al., Learning multiple layers of features from tiny images, 2009. [64] E. B. Wilson, Probable inference, the law of succession, and statistical inference, Journal of the American Statistical Association, vol. 22, no. 158, pp. 209 212, 1927. [65] I. Steinwart and C. Scovel, Mercer s theorem on general domains: On the interaction between measures, kernels, and RKHSs, Constructive Approximation, vol. 35, no. 3, pp. 363 417, 2012. [66] I. Steinwart, Convergence types and rates in generic Karhunen-Loève expansions with applications to sample path properties, Potential Analysis, vol. 51, no. 3, pp. 361 395, 2019. [67] M. Kanagawa, P. Hennig, D. Sejdinovic, and B. K. Sriperumbudur, Gaussian processes and kernel methods: A review on connections and equivalences, ar Xiv preprint ar Xiv:1807.02582, 2018. [68] I. Steinwart, D. R. Hush, C. Scovel et al., Optimal rates for regularized least squares regression, in COLT, 2009, pp. 79 93. [69] R. A. Adams and J. J. Fournier, Sobolev Spaces. Elsevier, 2003. [70] N. Eldredge, Analysis and probability on infinite-dimensional spaces, ar Xiv preprint ar Xiv:1607.03591, 2016. [71] S. Ghosal and A. Van der Vaart, Fundamentals of Nonparametric Bayesian Inference. Cambridge University Press, 2017, vol. 44. [72] R. Bhatia, Matrix analysis. Springer Science & Business Media, 2013, vol. 169. [73] O. Kallenberg and O. Kallenberg, Foundations of modern probability. Springer, 1997, vol. 2. [74] A. M. Stuart, Inverse problems: a Bayesian perspective, Acta Numerica, vol. 19, pp. 451 559, 2010. [75] B. Simon, Operator theory. American Mathematical Soc., 2015, vol. 4. [76] A. Edelman, Eigenvalues and condition numbers of random matrices, SIAM journal on matrix analysis and applications, vol. 9, no. 4, pp. 543 560, 1988. [77] R. Vershynin, High-Dimensional Probability: An Introduction with Applications in Data Science. Cambridge university press, 2018, vol. 47. [78] H. Wielandt, An extremum property of sums of eigenvalues, Proceedings of the American Mathematical Society, vol. 6, no. 1, pp. 106 110, 1955. [79] A. W. van der Vaart and J. H. van Zanten, Adaptive Bayesian estimation using a Gaussian random field with inverse Gamma bandwidth, The Annals of Statistics, vol. 37, no. 5B, 2009. [80] A. Rudi, R. Camoriano, and L. Rosasco, Less is more: Nyström computational regularization. in NIPS, 2015, pp. 1657 1665. [81] A. Rudi and L. Rosasco, Generalization properties of learning with random features, ar Xiv preprint ar Xiv:1602.04474, 2016. [82] W. V. Li and W. Linde, Approximation, metric entropy and small ball estimates for Gaussian measures, The Annals of Probability, vol. 27, no. 3, 1999. [83] G. Pisier, Sur les espaces de banach k-convexes, Séminaire Analyse fonctionnelle (dit" Maurey-Schwartz"), pp. 1 15, 1979. [84] A. Pajor and N. Tomczak-Jaegermann, Volume ratio and other s-numbers of operators related to local properties of Banach spaces, Journal of Functional Analysis, vol. 87, no. 2, pp. 273 293, 1989. [85] J. Shawe-Taylor, C. Williams, N. Cristianini, and J. Kandola, On the eigenspectrum of the gram matrix and its relationship to the operator eigenspectrum, in Alt 2002, Lnai 2533. Springer Verlag, 2002, pp. 23 40. [86] A. Bhattacharya, D. Pati, and Y. Yang, Bayesian fractional posteriors, The Annals of Statistics, vol. 47, no. 1, 2019. [87] P. Hall and J. L. Horowitz, Nonparametric methods for inference in the presence of instrumental variables, The Annals of Statistics, vol. 33, no. 6, 2005. [88] J. Rousseau and B. Szabo, Asymptotic behaviour of the empirical Bayes posteriors associated to maximum marginal likelihood estimator, The Annals of Statistics, vol. 45, no. 2, 2017. [89] S. Särkkä, State space representation of gaussian processes, Aalto University, Espoo, Finland. Pg, pp. 22 23, 2013. [90] J. L. Horowitz, Specification testing in nonparametric instrumental variable estimation, Journal of Econometrics, vol. 167, no. 2, pp. 383 396, Apr. 2012. [Online]. Available: https://linkinghub.elsevier.com/retrieve/pii/S0304407611002041 [91] F. Liang, Q. Li, and L. Zhou, Bayesian neural networks for selection of drug sensitive genes, Journal of the American Statistical Association, vol. 113, no. 523, pp. 955 972, 2018, p MID: 31354179. [Online]. Available: https://doi.org/10.1080/01621459.2017.1409122 [92] B. Bauer and M. Kohler, On deep learning as a remedy for the curse of dimensionality in nonparametric regression, The Annals of Statistics, vol. 47, no. 4, 2019. [93] N. M. Davies, S. von Hinke Kessler Scholder, H. Farbmacher, S. Burgess, F. Windmeijer, and G. D. Smith, The many weak instruments problem and Mendelian randomization, Statistics in Medicine, vol. 34, no. 3, pp. 454 468, 2015. The checklist follows the references. Please read the checklist guidelines carefully for information on how to answer these questions. For each question, change the default [TODO] to [Yes] , [No] , or [N/A] . You are strongly encouraged to include a justification to your answer, either by referencing the appropriate section of your paper or providing a brief inline description. For example: Did you include the license to the code and datasets? [Yes] See Section ??. Did you include the license to the code and datasets? [No] The code and the data are proprietary. Did you include the license to the code and datasets? [N/A] Please do not modify the questions and only use the provided macros for your answers. Note that the Checklist section does not count towards the page limit. In your paper, please delete this instructions block and only keep the Checklist section heading above along with the questions/answers below. 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] (b) Did you describe the limitations of your work? [Yes] Discussions of limitations are interwoven in the text. E.g., we discussed the limitation of the model selection result at the end of Section 5.2, and cautioned about the limitation of model-based uncertainty quantification at the end of Section 7. (c) Did you discuss any potential negative societal impacts of your work? [No] We do not foresee any potential negative societal impacts, except possible misuses of our method in violation of its core assumptions. As we have cautioned about such issues in the text (see above), we do not believe an additional, dedicated discussion is still necessary. (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [Yes] All theoretical results have their asssumptions in the statement. (b) Did you include complete proofs of all theoretical results? [Yes] Proofs are in appendix and always referenced in the main text. 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] In supplemental material. (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] See Appendix H. (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [Yes] All tables and figures have error bars. (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] See Appendix H, at the end of each experiment setup paragraph. 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [Yes] All assets have creators cited at their first mention. (b) Did you mention the license of the assets? [Yes] In the supplemental material. (c) Did you include any new assets either in the supplemental material or as a URL? [Yes] Our code is included in the suplemental material. (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [N/A] The datasets used are synthetic, and do not contain personal information. (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [N/A] The datasets used are synthetic and do not contain such content. 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] We did not use crowdsourcing, or conduct research with human subjects. (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]