# bayesian_optimization_with_informative_covariance__ecdbe10d.pdf Bayesian Optimization with Informative Covariance Afonso Eduardo afonso.eduardo@ed.ac.uk School of Informatics University of Edinburgh Michael U. Gutmann michael.gutmann@ed.ac.uk School of Informatics University of Edinburgh Bayesian optimization is a methodology for global optimization of unknown and expensive objectives. It combines a surrogate Bayesian regression model with an acquisition function to decide where to evaluate the objective. Typical regression models are given by Gaussian processes with stationary covariance functions. However, these functions are unable to express prior input-dependent information, including possible locations of the optimum. The ubiquity of stationary models has led to the common practice of exploiting prior information via informative mean functions. In this paper, we highlight that these models can perform poorly, especially in high dimensions. We propose novel informative covariance functions for optimization, leveraging nonstationarity to encode preferences for certain regions of the search space and adaptively promote local exploration during optimization. We demonstrate that the proposed functions can increase the sample efficiency of Bayesian optimization in high dimensions, even under weak prior information. 1 Introduction Bayesian optimization (BO) is a methodology for global function optimization that has become popular since the work of Jones et al. (1998). Other influential works date back to Kushner (1962) and Močkus (1975). BO has been applied to a broad range of problems (see e.g. Shahriari et al., 2016), including chemical design (Gómez-Bombarelli et al., 2018), neural architecture search (White et al., 2021) and simulation-based inference (Gutmann et al., 2016; Cranmer et al., 2020). As a Bayesian approach, the idea is to place a prior over the objective function, typically a Gaussian process (GP) prior characterized by a mean and covariance function. The prior, combined with an observation model and past function evaluations (observations), allows the computation of a posterior predictive distribution. The next acquisition is determined by the posterior predictive and a utility function, treating the optimization as a decision problem. Standard BO is generally effective in relatively low-dimensional problems. However, as dimensionality increases, statistical and computational problems become more noticeable. In particular, the curse of dimensionality manifests itself while learning the probabilistic surrogate of the objective function and during the acquisition process. Indeed, as training data become sparser, the estimation of an accurate surrogate becomes more difficult and the reliance on extrapolation increases. In this context, previous work has relied on certain structural assumptions (Binois & Wycoff, 2022), namely low effective dimensionality (Garnett et al., 2014; Li et al., 2016; Wang et al., 2016; Letham et al., 2020; Moriconi et al., 2020; Raponi et al., 2020; Eriksson & Jankowiak, 2021; Grosnit et al., 2021) and additivity (Kandasamy et al., 2015; Gardner et al., 2017; Mutný & Krause, 2018; Rolland et al., 2018; Wang et al., 2018). Local methods based on space-partitioning schemes and trust regions have also been proposed (Assael et al., 2014; Mc Leod et al., 2018; Eriksson et al., 2019; Wang et al., 2020b; Diouane et al., 2021). Notably, these methods employ stationary covariance functions. By construction, stationary covariance functions are translation invariant and thus unable to express spatially-varying information. As shown in Figure 1, stationary models can lead to poor optimization of Ground Truth Predictive Mean Credible Interval (2 SD) Initial Observations Acquisitions Stationary + Quadratic Mean Nonstationary (I+X0) Figure 1: Overconfident models are unable to capture the ground truth, causing BO to yield suboptimal results. The stationary model with an informative quadratic mean fits the trend too well, resulting in small residuals. These small residuals lead to a smaller prior variance and a longer lengthscale that, in turn, decrease epistemic uncertainty. Stationarity posits that these hyperparameters are constant, but this assumption does not hold for functions with irregular bumps. The proposed informative priors, being nonstationary, allow for the specification of spatially-varying properties. relatively simple one-dimensional objectives, suggesting issues in higher-dimensional domains. A common practice is to include input-dependent information via informative mean functions (see e.g. Snoek et al., 2015; Acerbi, 2019; Järvenpää et al., 2021). However, as demonstrated in Figure 1, this practice does not necessarily alleviate the limitations of stationary covariance functions. In contrast, nonstationarity allows for the incorporation of input-dependent information directly in covariance functions. The crux lies in the design of flexible and scalable informative GP priors. In the remainder of this paper, after we introduce the background in Section 2, we lay the foundation for our methodology by discussing the advantages of second-order nonstationary for optimization in Section 3. focusing on spatially-varying prior variance, lengthscales and high-dimensional domains. This includes a discussion of regret bounds for covariance functions with spatially-varying parameters, which we have not previously found in the literature. We then make the following contributions: In Section 4, we design informative covariance functions that leverage nonstationarity to incorporate information about promising points. Our GP priors encode preferences for certain regions and allow for multiscale exploration during optimization. To our knowledge, we are the first to consider informative GP priors for optimization with spatially-varying prior variances and lengthscales, both induced by a shaping function that captures information about promising points. In Section 5, we empirically demonstrate that the proposed methodology can increase the sample efficiency of BO in high-dimensional domains, even under weak prior information. Additionally, we show that it can complement existing methodologies, including GP models with informative mean functions, trust-region optimization and belief-augmented acquisition functions. Section 6 discusses related work and Section 7 concludes the paper. 2 Background 2.1 Bayesian Optimization In global optimization, the goal is to find the global minimizer of an objective function f : X Y R, x = arg min x X f(x), (1) where the domain X RD, for objectives of D variables, and f is a function whose expression and gradient are typically unavailable. Evaluation is possible at any point xi X, yielding the observation yi = f(xi), but its cost is high enough to justify the use of sample-efficient methodologies that aim to minimize the number of evaluations by careful selection. In this context, Bayesian optimization (BO) emerges as a competitive approach. The idea is to train a Bayesian regression model M that explains the evidence collected up to step n, Dn = {(xi, yi)}i n, and provides a distribution over possible functions. Given this probabilistic Algorithm 1 Bayesian Optimization (BO) Input: objective function f, acquisition function α, statistical model M, initial evidence Dn0 repeat xn+1 = arg max α(x | Dn, M) Find best candidate yn+1 = f(xn+1) Evaluate candidate Dn+1 = Dn {(xn+1, yn+1)} Update evidence set until stopping condition is met surrogate, an acquisition function α assigns a utility to each candidate point, guiding the selection of the next acquisition, i.e., the next point to evaluate. The procedure is outlined in Algorithm 1, where the stopping condition may be based on a maximum evaluation budget and a tolerance. For more details on BO, see e.g. Shahriari et al. (2016), Frazier (2018) and Greenhill et al. (2020). 2.2 Gaussian Process Regression BO is traditionally paired with Gaussian process (GP) models. For a given GP prior f GP(mθ, Cθ) with mean function mθ : X Y, covariance function Cθ : X X R+, and an additive Gaussian observation model y = f(x) + ϵ, ϵ N(0, σ2 y), the univariate posterior predictive distribution of the latent f at test input x is Gaussian with mean mn and variance σ2 n, mn(x) = mθ (x) + cn(x) [Cn + σ2 y I] 1(yn mn), (2) σ2 n(x) = Cn(x, x), (3) Cn(xi, xj) = Cθ (xi, xj) cn(xi) [Cn + σ2 y I] 1cn(xj), (4) cn(x) = (Cθ (x, x1), . . . , Cθ (x, xn)) , (5) where the entries in the Gram matrix Cn are [Cn]ij = Cθ (xi, xj), and the mean vector mn has entries [mn]i = mθ (xi), (xi, yi) Dn. Hyperparameters θ can, e.g., be fixed a priori or learned via empirical Bayes, by introducing a hyperprior p(θ) and maximizing the unnormalized posterior p(θ|Dn) p(yn|Xn, θ)p(θ), where the marginal likelihood is p(yn|Xn, θ) = N(yn; mn, Cn + σ2 y I), Xn = {xi}i n. For a comprehensive introduction to GPs, see Rasmussen & Williams (2005). 2.3 Stationary Covariance Functions Most popular covariance functions are translation invariant, depending only on the relative position of the input points. These functions compute Cov(f(xi), f(xj)) based on a translation-invariant distance between xi and xj. Valid covariance functions must be symmetric and positive definite. In this family, we find the squared exponential or Gaussian covariance, CG(xi, xj) = σ2 0 exp( 1/2 d2 M(xi, xj)), d M(xi, xj) = q (xi xj) Λ 1(xi xj), (6) where σ2 0 is the prior variance and d M is the Mahalanobis distance function with matrix Λ. For simplicity and scalability, it is customary to constrain the matrix to be diagonal with entries [Λ]dd = λ2 d R+, d {1, ..., D}, known as (squared) lengthscales. In this case, the Mahalanobis distance turns into the weighted Euclidean distance, which we denote by d WE. These lengthscales control the exponential decay of correlation. For instance, a short lengthscale leads to sample paths, or realizations, that vary more rapidly. The Gaussian covariance, being infinitely differentiable, may generate unrealistically smooth sample paths. A more general class of functions is generated, or reproduced, by the Matérn covariance function which controls the differentiability of the process via a smoothness parameter ν, recovering CG when ν . For optimization purposes, good empirical results have been found with ν = 5/2 (Snoek et al., 2012), for which CM(xi, xj) = σ2 0 5d M(xi, xj) + 5 3d2 M(xi, xj) exp 5d M(xi, xj) . (7) An attractive property of the Gaussian and mixtures thereof, such as Matérn, is their proven universality (Micchelli et al., 2006). Asymptotically, these covariance functions, or kernels, are capable of approximating any continuous function on any compact subset of the domain. In practice, however, other covariance functions may be preferable in the small sample regime, especially if the elicited prior includes relevant information, or structure, that accelerates the learning process. 2.4 Acquisition Functions While many rules to determine promising candidate solutions have been proposed (see e.g. Wilson et al., 2018; Neiswanger et al., 2022), those with an analytical expression remain the most popular. Such acquisition functions can be evaluated without Monte Carlo or quadrature approximations. Given a Bayesian model, myopic acquisition functions α suggest one acquisition at a time, depending only on the univariate posterior predictive distribution with mean mn and variance vn, i.e., α(x | Dn, M) = α(x | mn, vn). In this context, a popular choice is the Lower Confidence Bound (LCB) criterion (Srinivas et al., 2010),1 LCB(x) = mn(x) βnσn(x), (8) where the factor βn is related to a confidence level, balancing exploitation and exploration. Another popular alternative is the Expected Improvement (EI), which has been found to be more effective than LCB when properties of the objective function, e.g. norm bounds, are unknown (Snoek et al., 2012; Chowdhury & Gopalan, 2017; Merrill et al., 2021). This acquisition function is defined as EI(x) = Ep(f(x)|x,Dn)[I(x)] = σn(x)τ(z(x)), (9) τ(z(x)) = z(x)FN (z(x)) + N(z(x); 0, 1), z(x) = (f(xbest) mn(x))/σn(x), (10) where I(x) = max(0, f(xbest) f(x)) is the improvement over the incumbent f(xbest) and FN is the standard Normal cumulative distribution function (CDF). For noisy functions, f(xbest) is replaced by the minimum posterior predictive mean, m n = minx mn(x). In the context of optimization, there are several useful metrics to measure performance. The instantaneous regret is the loss incurred at step n, rn = f(xn) f(x ), where xn is the acquisition and x is the global minimizer of f. The simple regret is the minimum instantaneous regret incurred up to step n, sn = mint n rt, or, equivalently, the regret incurred by the incumbent solution. The cumulative regret is defined as the sum of instantaneous regret over N steps, RN = P n N rn. If the cumulative regret grows at a sublinear rate, then the average regret goes to zero, lim N RN/N = 0, which is also known as no regret. This property implies that simple regret goes to zero, ensuring that the global optimum is asymptotically found. For more details on regret bounds, see e.g. Vakili et al. (2021) and Gupta et al. (2022). 3 Benefits of Second-Order Nonstationarity for Optimization We now turn our attention to GP priors with spatially-varying properties, namely prior variance and lengthscale, which provide the foundations for the informative covariance functions in Section 4. While an analysis in terms of cumulative regret is customary, we focus on the instantaneous regret because we are interested in the relatively small sample regime, for which the asymptotics are not relevant. Furthermore, by demonstrating that the instantaneous regrets rn are smaller (or larger), it follows that the cumulative regret RN is also smaller (or larger). Spatially-varying prior variance The prior variance σ2 0 quantifies the (epistemic) uncertainty of the surrogate model of f before observing data. Relatedly, posterior predictive variances σ2 n(x), defined in Equation 3, determine the mutual information between observations y N and the objective function f, MIN(y N; f) = 1 2 PN n=1 log 1 + σ 2 y σ2 n 1(xn) , see e.g. (Srinivas et al., 2010, Lemma 5.3). Then, for stationary covariance functions, any point xn where cn 1(xn) 0 provides equal information about f because 1The LCB needs to be minimized and thus corresponds to a negative acquisition function α. Uninformative Mean Informative Mean Ground Truth Mean CI (Stationary) CI (Nonstationary) Figure 2: Narrower credible intervals (CI) can be obtained with a spatially-varying prior variance. Typically, GP priors in BO are uninformative, characterized by constant prior mean functions and stationary covariance functions that posit a constant prior variance. Then, for popular acquisition functions, all candidates are considered equally good a priori. Conversely, spatially-varying prior variances can encode preferences. In this example, points near the center are more informative a priori. Constant Lengthscale Spatially-varying Lengthscale Figure 3: Given the same evaluation budget (N = 20), a model with spatially-varying lengthscales can better approximate the function than one with a constant lengthscale. While the latter must set a shorter global lengthscale, a multiscale approach can locally assign shorter lengthscales. the predictive variance is approximately constant, σ2 n 1(xn) σ2 0. Conversely, surrogate models of f with a spatially-varying prior variance are spatially informative, even in the absence of neighboring observations, in which case σ2 n 1(xn) σ2 0(xn). In terms of optimization, a spatially-varying prior variance can express a priori preferences for certain regions, as illustrated in Figure 2. Notably, compared to the stationary case, better optimization performance can be achieved: For popular acquisition functions, the worst-case instantaneous regret is proportional to the predictive standard deviation (Lemma A.1), provided that the interval |f(x) mn(x)| βnσn(x) holds for some factor βn. Then, recall that the predictive variance is upper bounded by the prior variance, σ2 n(x) σ2 0(x). While the latter is constant for stationary covariance functions, covariance functions with spatially-varying prior variances allow for narrower intervals and, in turn, tighter regret bounds. As a caveat, standard regret analysis assumes that 1) intervals include the objective function and 2) global optima of acquisition functions can be found, but in practice these assumptions do not necessarily hold. Spatially-varying lengthscale In terms of sample efficiency, stationary covariance functions are ill-suited for approximating objectives with spatially-varying properties. Intuitively, a function that varies rapidly in one region may be characterized by short lengthscales, but is, otherwise, more efficiently described by longer lengthscales. In order to capture the finest details of such a function, a stationary covariance would require the shortest possible lengthscale, resulting in additional observations to approximate the function and ultimately find the optimum. In contrast, a multiscale approach only requires more neighboring observations where the local lengthscales are shorter (Figure 3). In more detail, let the objective be a realization from a locally stationary process, f(x)= P m M 1x Xmfm(x) with disjoint Xm X, m = {1, . . . , M}, M N+. Each component is a member of a class of functions forming the Hilbert space HCm(Xm) with reproducing stationary kernel Cm and bounded norm f HCm Bm. If f is to be estimated with a stationary covariance function C, then it must reproduce the space HC(X) S m M HCm(Xm) with x X = S m M Xm and norm bound B p P m B2m, by (Aronszajn, 1950, Theorem 6). This corresponds to a search in the space of more complex functions, for which more observations are required because the global lengthscale must be set as the shortest from {Cm}m M. Indeed, for shorter lengthscales, λ λ, we have larger posterior predictive variances, σ2 n,λ (x) σ2 n,λ(x), which lead to greater information gains MIn,λ MIn,λ, and larger norms, f HCλ f HCλ , by (Bull, 2011, Lemma 4), contributing to a larger βn,λ βn,λ because βn = supf HC,MIn f HC+σy p 2(MIn + 1 log δ), δ (0, 1], by (Chowdhury & Gopalan, 2017, Theorem 2). Then, by Lemma A.1, wider intervals |f(x) mn(x)| βnσn(x) lead to larger regret bounds, and hence worse performance. The argument above provides additional justification for spatially-varying prior variances, in which case each Cm is characterized by different but constant σ2 0,m. Then, the stationary covariance function must set σ2 0 = maxm σ2 0,m, leading again to wider intervals, and hence worse performance by Lemma A.1. Finally, as M grows, it becomes more challenging to estimate and maintain M local models. Nonstationary approaches that do not rely on space partitions can avoid these computational and statistical problems. Notably, in the limit, each subset is a singleton, Xm = {xm}, xm X, and the restriction f|Xm is reproduced by Cm(xi, xj) = δimδjm with norm f HCm = |f(xm)|. For an uninformative prior mean, m0(x) = 0, the narrowest interval, before observing data, |f(xm) m0(xm)| β0,mσ0,m, is obtained with a spatiallyvarying β0(x) = |f(x)| and constant σ0 = 1, or, equivalently, with β0 = 1 and spatially-varying prior standard deviation σ0(x) = |f(x)|. Thus, the optimal prior variance depends on the shape of the objective function, as previously suggested by Figure 2. High-dimensional domains In high-dimensional Euclidean spaces, most of the volume is on the boundary, and for stationary GPs, epistemic uncertainty, measured by the posterior predictive variance, increases with the distance to the training data, tilting the acquisition step toward blind exploration of the boundary, typically faces of a hypercube (Binois & Wycoff, 2022). This problem is known as the boundary issue (Swersky, 2017; Oh et al., 2018). Moreover, uninformative priors provide no practical finite-time convergence guarantees. Begin by recalling that the motivation behind BO is the optimization of expensive functions, which puts a practical constraint on the size of the evaluation budget. However, unless the total budget N increases exponentially, the space cannot be covered in such a way that a global contraction of the posterior predictive variance is ensured, see e.g. (Kanagawa et al., 2018, Section 5.2) and (Wüthrich et al., 2021, Appendix C). Indeed, in the relatively small sample regime, there is at least one point xn whose predictive variance is approximately equal to the (constant) prior variance, σ2 n 1(xn) σ2 0. This is explained by the absence of neighboring observations, in which case the training data provides a negligible reduction of uncertainty because cn 1(xn) 0. As a result, the worst-case instantaneous regret, being proportional to the predictive standard deviation (Lemma A.1), remains approximately constant for fixed βn, or alternatively is nondecreasing for nondecreasing βn (Chowdhury & Gopalan, 2017, Theorem 2), and thus there is no guarantee of improvement. These problems can be alleviated by more informative GP priors. In particular, spatially-varying prior variances allow the surrogate to remain spatially informative, even in the absence of neighboring observations, and spatially-varying lengthscales can encourage greedier strategies, in an attempt to improve upon the incumbent solution when given small budgets. 4 Informative Covariance Functions For the construction of our informative covariance functions, we start with a set of anchors {x(l) 0 }, i.e., a set of promising candidate solutions. This prior belief can be interpreted as a point-mass distribution over promising points. A more conservative prior p(x ) is obtained by adding an uninformative slab that ensures the optimal solution is included in its support, and forming a kernel density estimate by associating a non-negative weight wl, distance function dl and kernel kl with each anchor x(l) 0 , p(x ) ϕ(x ), ϕ(x ) = 1 + 1 l L (wl 1) kl dl(x , x(l) 0 ) . (11) Distance functions and kernels characterize the neighborhood of anchors. Larger weights correspond to the belief that better candidates can be found in a given neighborhood. Next, we introduce spatially-varying 1.00 0.75 0.50 0.25 0.00 0.25 0.50 0.75 1.00 0 Prior Variance ( 2 0.1 0.3 0.5 0.8 1.0 1.00 0.75 0.50 0.25 0.00 0.25 0.50 0.75 1.00 0.1 0.3 0.5 0.8 1.0 Figure 4: For a Gaussian kσ and weighted Euclidean distance dσ = d WE, decreasing the ratio rσ increases the prior variance around the anchor x0 = 0. Stationarity is recovered when rσ = 1. The lengthscale λσ in d WE controls the rate at which the prior variance decreases to the stationary baseline. prior variances by rescaling the stationary covariance function CS(xi, xj), CI(xi, xj) = σ2 0(xi, xj)CS(xi, xj), σ2 0(xi, xj) = σ2 p q ϕ(xj), (12) where σp is a scaling constant and CS(x, x) = 1. The modulating σ2 0(xi, xj) is symmetric and separable in xi and xj, and hence a valid covariance function (Genton, 2001). The informative covariance CI(xi, xj), being the product of two covariance functions, is also a valid covariance function. The prior covariance function σ2 0(xi, xj) is a generalization of the constant prior (co)variance σ2 0. By recalling that covariance functions compute Cov(f(xi), f(xj)), the intuition is as follows: For two points xi and xj that are close to anchors, i.e., large ϕ(xi) and ϕ(xj), this prior posits that their values, f(xi) and f(xj), are highly correlated because both should be close to the minimal value and hence similarly small. Then, if the probability for a point, e.g. xj, to be a good candidate decreases, the covariance becomes smaller because f(xj), being less constrained, can take on a greater range of values. In Section 5, we focus on the special case of a single anchor, for which we obtain σ2 0(xi, xj) = σ2 p rσ 1 kσ (dσ(xi, x0)) rσ 1 kσ (dσ(xj, x0)), (13) where kσ k1, dσ d1 and rσ 1/w1 (0, 1]. We find this parametrization more convenient because inverse weights are bounded. The ratio rσ controls the degree of stationarity, with rσ = 1 resulting in a stationary covariance and rσ 0 in increased nonstationary effects, as demonstrated in Figure 4. Next, we introduce spatially-varying lengthscales via input warping. In input warping (Snoek et al., 2014; Risser, 2016), the stationary covariance function CS is computed in a transformed space, i.e., after applying a (possibly nonlinear) warping function hλ to x. Endowed with such a transformation, the informative covariance function CI(xi, xj), in Equation 12, is given by CI(xi, xj) = σ2 0(xi, xj)CS (hλ(xi), hλ(xj)) . (14) Recall that inputs in σ2 0(xi, xj) are already being compared to anchors in the shaping function ϕ (Equation 11). Now, notice that an anisotropic stationary covariance function, equipped with the Mahalanobis distance d M, is equivalent to an isotropic covariance function, equipped with the standard Euclidean distance and linear transformation hλ(x) = Λ 1 2 x. Then, since the warping function can be any real vector-valued function, we propose a nonlinear transformation that shrinks the lengthscales λd around anchors, thereby expanding their neighborhoods. In particular, for a single anchor, we have hλ(x) = u 1 2 x, uλ(x) = 1 + (rλ 1)kλ(dλ(x, x0)), (15) with ratio rλ (0, 1].2 Note that the matrix Λ is diagonal with squared lengthscales λ2 d and the neighborhood of x0 is characterized by the kernel kλ and distance dλ. The shortest lengthscale is rλλd. 2The general case with multiple anchors can be given by the kernel mixture uλ(x) = 1+1/LP l L(1/wl 1)kl(dl(x, x(l) 0 )), an expression similar to that of the shaping function ϕ in Equation 11. The inverse weights 1/wl (0, 1] set shorter lengthscales. 5 Experiments 5.1 Setting Informative covariance functions The shaping function ϕ should capture the existing knowledge about the objective, possibly by being elicited from experts or learned on problems that are known to be similar. This function, based on kernel mixtures, can express arbitrary beliefs, but its optimality crucially depends on its alignment with the shape of the objective function, as discussed in more detail in Section 3. For black-box optimization, in which little is known in advance, a default kernel and distance function may be used, and their hyperparameters learned by empirical Bayes. To avoid overfitting and reduce computational complexity, regularizing hyperpriors and hyperparameter tying are effective techniques that may also be used. In particular, the parameters and functions that control the spatially-varying prior (co)variance and lengthscales can be tied by setting the ratio r0 rσ = rλ, kernel k0 kσ = kλ and distance d0 dσ = dλ. In our experiments, we specify a Gaussian k0, equipped with a weighted Euclidean distance d0 = d WE and a shared lengthscale vector, which also parametrizes Λ. Dirac delta hyperpriors for anchors have also been adopted, but note that none of these choices are strictly required. More implementation details can be found in Appendix B, and we refer to Appendix F for the settings of r0, where we perform a sensitivity analysis. An overview of the main methods using the proposed covariance functions (I) is included in Appendix D. These methods are I+X0 and I+XA, where +X0 denotes a fixed anchor x0 (placed at the center of the search space) and +XA denotes an adaptive anchor (placed at the incumbent solution). More methods can be developed by incorporating complementary methodologies. In this paper, we test I+XA+TR (trust-region optimization, Appendix D.1), I+XA+QM (informative mean functions, Appendix D.2), I+XA+SAAS (sparsityinducing hyperpriors, Appendix H) and I+XA+GKEI (belief-augmented acquisition functions, Appendix I). Baselines Previous work has assumed that optimal solutions are generally closer to the center of the search space rather than its boundary. As dimensionality increases, BO with a stationary covariance (S) may allocate a large portion of the budget toward blind exploration of the boundary, which by assumption does not contain the optimum, resulting in worse performance. To mitigate boundary overexploration, two approaches have been proposed: quadratic mean functions (+QM) (Snoek et al., 2015) and cylindrical covariance functions (C) (Oh et al., 2018). In particular, the latter are nonstationary covariance functions that aim to expand the innermost region, and have been shown to outperform stationary additive models on several high-dimensional additive objectives, without having to infer their structure. Further details can be found in Appendix C. Additionally, the use of trust regions (+TR), despite not specifically intended to address the boundary issue, has become a popular acquisition strategy for high-dimensional BO. This strategy can promote exploitation by constraining acquisitions to an adaptive region centered at the incumbent solution. In our experiments, we refrain from direct comparisons with existing BO packages due to the diverse training and acquisition routines. Instead, we reimplemented the BO baselines to study the effect of different GP priors on optimization performance under controlled conditions. We also investigate the interplay between GP priors and trust regions. Our implementation of +TR follows Tu RBO-1 (Eriksson et al., 2019). The GP models are global, trained on all observations, and the trust region is given by a box whose side lengths are doubled/halved after consecutive successes/failures. The hyperparameter values of +TR are identical to those of Tu RBO-1, except for the failure tolerance, which is originally given by the number of dimensions. On a 100-dimensional problem, this means that the trust region is only shrunk after 100 consecutive failures, half of our evaluation budget. To increase exploitation, this tolerance is set to 10. Other baselines, such as the popular CMA-ES (Hansen, 2016), are described and shown in Appendix D. Performance metric We report the normalized improvement over initial conditions (Hoffman et al., 2011), NIn = f (n0) best f(x(n0+n) best ) f (n0) best f(x ) , n = {0, . . . , N} , (16) where n0 = 16 is the number of initial observations, f (n0) best is the lowest initial value, x(n0+n) best is the incumbent solution, x is the global minimizer and N = 200 is the number of acquisitions.3 3Experiments with a larger evaluation budget can be found in Appendix J. 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement QBranin,20D S S+QM S+TR C I+X0 (Proposed) I+XA (Proposed) I+XA+TR (Proposed) 0 20 40 60 80 100 120 140 160 180 200 QBranin,50D 0 20 40 60 80 100 120 140 160 180 200 QBranin,100D 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement Rosenbrock,20D 0 20 40 60 80 100 120 140 160 180 200 Rosenbrock,50D 0 20 40 60 80 100 120 140 160 180 200 Rosenbrock,100D S S+QM S+TR C I+X0 (Proposed) I+XA (Proposed) I+XA+TR (Proposed) 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement Styblinski-Tang,20D 0 20 40 60 80 100 120 140 160 180 200 Styblinski-Tang,50D 0 20 40 60 80 100 120 140 160 180 200 Styblinski-Tang,100D S S+QM S+TR C I+X0 (Proposed) I+XA (Proposed) I+XA+TR (Proposed) Figure 5: Performance on test functions with 20, 50 and 100 dimensions. Additional results are shown in Appendix D (methods and test functions) and Appendix J (larger evaluation budget). Solid curves and shaded regions represent the mean and standard deviation of the normalized improvement, computed over 10 trials with different initial conditions. Solid vertical lines indicate the interquartile range. Abbreviations: Stationary (S), Cylindrical (C) and Informative (I) covariances; Quadratic Mean (+QM); Origin (+X0) and Adaptive (+XA) anchors; acquisitions within Trust Region (+TR). Unlike simple regret, this performance metric is invariant to initial conditions and to the range of f, always starting from 0 and only attaining the maximum value of 1 if a global optimum is found (zero simple regret). Additionally, our analysis does not hinge on the final best values because these can be deceiving. For instance, a method that rapidly achieves a good, though suboptimal, solution should not be considered inferior to another method that requires more evaluations, only to attain a marginally better value. The former is more sample efficient, which is a desirable feature when functions are expensive to evaluate. Some results are summarized by reporting the mean NI over N acquisitions (or, equivalently, the normalized area under the NI curve), with 0 indicating worst and 1 optimal performance. Respective optimization trajectories are shown in Appendix D. 5.2 Results on Test Functions We first focus on objectives whose landscapes are well-defined and can be easily controlled, allowing us to examine whether more informative priors translate into higher sample efficiency. Our test functions are roughly bowl-shaped, making the quadratic mean prior an appropriate choice for these problems. Test functions, such as Branin and Rosenbrock, have been previously used to test C (Oh et al., 2018). However, the Branin function is known for having multiple global optima, some of which are located close to the boundary. To overcome this, we design the QBranin function, which is obtained by adding a quadratic term, resulting in a bowl-shaped function with a single global optimum. The Styblinsky-Tang function, while multimodal, has a unique global optimum. A more detailed characterization of these test functions, as well as additional benchmarks (Levy and shifted objectives), can be found in Appendix B.3. Table 1: Mean normalized improvement (over 200 acquisitions) on shifted (S35 and S50) Rosenbrock functions (mean standard deviation). Values > 0.5 suggest superlinear improvement rates. Additional methods, test functions and optimization trajectories can be found in Appendix D. Abbreviations: Stationary (S), Cylindrical (C) and Informative (I) covariances; Quadratic Mean (+QM); Origin (+X0) and Adaptive (+XA) anchors; acquisitions within Trust Region (+TR). S35Rosenbrock S50Rosenbrock Method 50D 100D 50D 100D S 0.503 0.078 0.283 0.052 0.690 0.054 0.411 0.048 S+QM 0.507 0.086 0.218 0.083 0.548 0.193 0.104 0.157 S+TR 0.677 0.030 0.486 0.030 0.752 0.020 0.535 0.061 C 0.744 0.077 0.487 0.067 0.700 0.118 0.377 0.142 I+X0 0.803 0.031 0.644 0.026 0.831 0.038 0.729 0.027 I+XA 0.803 0.016 0.659 0.023 0.857 0.030 0.726 0.039 I+XA+TR 0.800 0.038 0.693 0.016 0.844 0.027 0.720 0.031 Fixed anchor The proposed covariance functions are initially tested with a fixed anchor placed at the center (I+X0). Figure 5 shows the performance curves on two objectives (top and middle) where the optimal solution is relatively close to this anchor. These results indicate that second-order nonstationary approaches, such as C and I+X0, can significantly outperform stationary methods, S and S+QM, particularly in higher-dimensional domains. While an informative mean (S+QM) is not necessarily better than an uninformative constant mean (S), prior information introduced via the covariance function leads to a consistent improvement over S. The reason may be that, in certain dimensions, S+QM has relatively large quadratic weights that hinder exploration away from the center, and in other dimensions, very small weights are unable to alleviate the boundary issue. Compared to S, S+QM and C, the proposed I+X0 not only performs better on average but is also more reliable. In Appendix E, an ablation study provides empirical evidence that incorporating both spatially-varying prior variances and lengthscales is important for tackling higher-dimensional problems. In particular, we there observe that the shorter lengthscales near anchors promote local exploration, which is especially effective when anchors are relatively close to optima. Meanwhile, spatially-varying prior variances can mitigate the boundary issue in high-dimensional domains, thereby improving sample efficiency. Next, we discuss the use of an adaptive anchor and the performance on the Styblinsky-Tang function. Adaptive anchor The assumption that optimal solutions are close to the center is strong. As optima deviate from the center, the performance of C decreases markedly, eventually becoming comparable to that of S (Table 1). Despite the lower apparent vulnerability of I+X0 to anchor misspecifications, this motivates the use of informative covariance functions with an adaptive anchor (I+XA). For simplicity, we opt for a greedy strategy where the anchor is given by the incumbent solution. For noise-free objectives, this is the point in the evidence set with the lowest observed value. In addition to QBranin and Rosenbrock, Figure 5 includes the performance on the Styblinski-Tang function, whose optimal solution is relatively distant from the center. Notably, the center is a local maximum, being surrounded by exponentially many local modes. As the assumption that the global minimum is close to the center is no longer valid, we observe that the relative performance of C and I+X0, compared to S and S+QM, deteriorates. In particular, C becomes the worst-performing method and I+X0 is worse than S in the 20-dimensional problem. Conversely, I+XA is consistently among the best-performing methods across all tests shown in the figure, and often outperforms the fixed-anchor variant (I+X0). Trust region The use of an adaptive greedily-chosen anchor is partly motivated by trust-region optimization, wherein regions are centered at the incumbent solution. However, the proposed methodology is complementary to trust regions because it extends a stationary covariance function with spatially-varying parameters that are possibly aligned with prior beliefs about the objective. While S+TR can lead to significant improvements over S, overexploration of boundaries in a trust region can still occur because the model is stationary. The method I+XA+TR shows that the combination of informative covariance functions with trust-region optimization can improve sample efficiency. Efficient Trajectory 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement Rover T,60D S S+QM S+TR C I+X0 (Proposed) I+XA (Proposed) I+XA+TR (Proposed) Figure 6: Map layout of the rover trajectory problem (left) and respective optimization results (right). Additional results and a more detailed analysis can be found in Appendix G. Solid curves and shaded regions represent the mean and standard deviation of the normalized improvement, computed over 10 trials with different initial conditions. Solid vertical lines indicate the interquartile range. Abbreviations: Stationary (S), Cylindrical (C) and Informative (I) covariances; Quadratic Mean (+QM); Origin (+X0) and Adaptive (+XA) anchors; acquisitions within Trust Region (+TR). 0 20 40 60 80 100 120 140 160 180 200 0.0 0.4 MNISTW,100D S S+QM S+TR C I+X0 (Proposed) I+XA (Proposed) I+XA+TR (Proposed) 0 20 40 60 80 100 120 140 160 180 200 0.0 0.4 MNISTW,100D S S+QM I+XA (Proposed) I+XA+QM (Proposed) I+XGood (Proposed) Figure 7: Bayesian optimization of 100 parameters of a neural network trained and tested on MNIST. Solid curves and shaded regions represent the mean and standard deviation of the normalized improvement, computed over 10 trials with different initial conditions. Solid vertical lines indicate the interquartile range. Abbreviations: Stationary (S), Cylindrical (C) and Informative (I) covariances; Quadratic Mean (+QM); Origin (+X0), Good (+XGood) and Adaptive (+XA) anchors; acquisitions within Trust Region (+TR). 5.3 Applications Rover trajectory planning Recent studies have used the rover trajectory (Rover T) problem to showcase BO (Wang et al., 2018; Eriksson et al., 2019; Eriksson & Jankowiak, 2021). The goal is to optimize a trajectory such that the rover satisfies the target endpoints, xstart and xend, while avoiding collisions with objects, as depicted in Figure 6 (left). The trajectory is given by a B-spline that is fitted to 30 2-dimensional points, resulting in a 60-dimensional optimization problem. A more detailed description and analysis is provided in Appendix G. As we show in Figure 6 (right), I+X0 and I+XA, which use an informative covariance, are the best-performing methods. We also observe that the performance of S+TR and I+XA+TR is significantly worse than that of S and I+XA, respectively. The reason may be that points in a trust region do not significantly improve upon the incumbent solution. For optimal performance, the endpoints must match the two targets because deviation incurs a significant loss. However, this may not be feasible when using a small box centered at a suboptimal solution, e.g. xbest = 0. Consecutive failures decrease the size of trust regions, potentially resulting in even slower progress. Neural network optimization In this application, we optimize a neural network layer, as proposed by Oh et al. (2018). In particular, we perform BO of 100 parameters of a two-layer fully-connected network trained and tested on the MNIST dataset (Le Cun, 1998). The architecture is 784 W1,b1 Nhidden W2,b2 10, with weights W, biases b, Nhidden = 10 and Re LU activations. The model is trained for 10 epochs with a batch size of 512 and evaluated on the test set using the negative log-likelihood loss. The weights W2 are found by BO using the test loss, while the remaining parameters are learned on the training set with Adam (Kingma & Ba, 2014). As pointed by Oh et al. (2018), the goal is not to find solutions that generalize well, but rather to evaluate the ability of BO to find solutions that perform well on the test set by minimizing the respective loss. The landscape is multimodal, and points at the boundary can yield relatively good solutions. Results are shown in Figure 7. While S+QM appears to perform slightly better than I+X0 and I+XA, models with an informative covariance function do not preclude the use of an informative mean function. For instance, we observe that I+XA+QM can improve upon I+XA toward the end, achieving similar average performance to that of S+QM. Relatedly, in Appendix D, we show that I+XA+QM outperforms both S+QM and I+XA on Styblinsky-Tang, demonstrating that informative mean and covariance functions are complementary. The best-performing method is I+XGood, which assumes strong prior information by using a fixed anchor placed at a solution found by S+QM. This highlights the value of high-quality information and that it can be incorporated into informative GP priors, without modifying the acquisition process. 5.4 Limitations Despite the overall superior performance shown by methods with informative covariance functions, there is room for improvement that is left for future work. For instance, as dimensionality increases, the weighted Euclidean distance cannot differentiate points that are far only in a few dimensions. This can lead to difficulties in addressing the boundary issue in hypercubes. Potential remedies for this problem may include different parametrizations and distances based on the max-norm. Moreover, there is no mechanism in place that prevents greedily-chosen anchors from being close to the boundary. Since I+XA relies on empirical priors, it can be used without access to reliable prior information. However, as shown in Section 5.3 and Appendix D, methods that include reliable information can significantly outperform I+XA. In Section 4, we provide formulas for covariance functions with multiple anchor points to demonstrate that the proposed methodology can in theory capture knowledge about the objective that goes beyond a single anchor. In our experiments, we focus on methods with a single anchor for several reasons. First, we wanted to show that higher sample efficiency is possible with simple design choices. Furthermore, baselines such as S+QM, S+TR and C implicitly assume the existence of a single anchor. By restricting ourselves to this case, the comparisons are fair, and we can assess the robustness of the different methods to anchor misspecification. Additionally, in sequential myopic BO, which is the setup considered in this paper, it suffices to focus on a region that includes an optimum. The correct identification of such a region may be difficult in multimodal objectives, but as we have shown, the anchor can be dynamically adjusted. One option, without having to specify simultaneous anchors, is to maintain posterior beliefs about promising points and sample them at each step. If the anchors are not known, a similar strategy to that proposed by Eriksson et al. (2019) to maintain multiple trust regions is also a possibility. Other options include setting the anchors after performing cluster analysis and learning them by empirical Bayes. In this context, a different set of baselines would also be more appropriate, including methods with multiple trust regions and quadratic mean functions with multiple centers. Moreover, model selection can be used to automatically determine the number and location of anchors (see e.g. Appendix H). A careful treatment of informative covariance functions with multiple anchors requires substantial analysis and testing, and we believe this is best left for future work. 6 Related Work Second-order nonstationarity Covariance functions with spatially-varying lengthscales were initially explored by Gibbs (1998) and Paciorek & Schervish (2006), and augmented with spatially-varying prior (and noise) variance by Heinonen et al. (2016) for GP regression. These models are derived from basis function expansions or process convolution (Risser, 2016), wherein the covariance function is obtained via integral transform of a base stochastic process with a kernel function, which may also depend on auxiliary mixing variables (e.g. Wang et al., 2020a). In contrast, we introduce spatially-varying lengthscales via input warping, resulting in a simpler expression that does not require extra factors to ensure positive definiteness. Additionally, these other models place GP priors on the spatially-varying parameters and infer the respective latent functions by MCMC, being computationally expensive for high-dimensional parameter spaces. Instead, we use pointwise specifications of nonstationary effects, governed by semiparametric functions whose shape and parameters can be elicited from an expert or learned by empirical Bayes. Importantly, while these works focus on the estimation of the covariance function for regression purposes, we leverage second-order nonstationarity to incorporate information about promising points for optimization. In the context of optimization, input warping was first studied by Snoek et al. (2014), focusing on monotonic input deformations via CDFs, specifically the Beta distribution. The motivation follows from a regression perspective, as the resulting GP is capable of approximating more realistic objective functions. More recently, Oh et al. (2018) proposed cylindrical covariance functions for optimization, outperforming the method by Snoek et al. (2014). In this case, after applying the cylindrical transformation, the radius is warped by the Kumaraswamy distribution, further expanding the innermost region of the search space, which is assumed to contain the optimal solution. A different methodology was explored by Martinez-Cantin (2015), who used linear combinations of local and global stationary covariance functions with input-dependent weights to induce spatially-varying lengthscales (piecewise constant). While the motivation is that the local assignment of shorter lengthscales promotes local exploration, the local component is not guaranteed to learn shorter lengthscales because both are initialized with the same hyperpriors. None of these works provide a regret analysis or consider spatially-varying prior (co)variances in their models. High-dimensional Bayesian optimization To tackle the challenges presented by high-dimensional spaces, previous research has relied on certain structural assumptions (Binois & Wycoff, 2022), namely low effective dimensionality (e.g. Letham et al., 2020; Eriksson & Jankowiak, 2021) and additivity (e.g. Kandasamy et al., 2015; Gardner et al., 2017; Rolland et al., 2018). Space-partitioning schemes and trust regions have also been proposed (e.g. Mc Leod et al., 2018; Eriksson et al., 2019; Wang et al., 2020b; Diouane et al., 2021). Notably, these methods employ stationary covariance functions. Our work complements these by introducing more informative priors (see e.g. Appendices D and H). Cylindrical covariance functions (Oh et al., 2018) can also be augmented with spatially-varying (co)variances and anchors can be defined in the transformed space. Priors over the optimum There have been limited attempts to incorporate beliefs about optimal solutions into BO. Souza et al. (2021) combined predefined prior distributions with the tree-structured Parzen estimator approach by Bergstra et al. (2011), which estimates the input given the output, as opposed to the standard GP approach of modeling the output given the input. By design, the former is restricted to the Expected Improvement acquisition function. Ramachandran et al. (2020) explored input warping via predefined CDFs that encode prior beliefs. More recently, Hvarfner et al. (2022) proposed belief-augmented acquisition functions. As we show in Appendix I, these functions and informative GP priors are complementary. 7 Conclusion This work proposes informative covariance functions for Bayesian optimization, leveraging nonstationarity to incorporate input-dependent information. Our methodology relies on priors with spatially-varying parameters to express a priori preferences for certain regions and to promote multiscale exploration. These priors can more efficiently describe a wider class of objective functions and result in improved worst-case performance. Our experiments showed that the proposed covariance functions can significantly increase sample efficiency, even under weak prior information, challenging the prevalent use of stationary models for optimization. Additionally, we showed that our work complements existing methodologies, including models with informative mean functions, trust-region optimization and belief-augmented acquisition functions. Despite our focus on continuous search spaces, we believe that this work can be extended to other types, such as discrete and mixed spaces, by adopting appropriate priors. Our methodology can also accommodate multiple anchors and be combined with other methods, relying for instance on additivity and low effective dimensionality. We recognize that priors over the possible locations of the optimum can be defined in transformed, possibly lower-dimensional spaces, but these details are left for future research. Overall, we see this work as a contribution toward the wider adoption of scalable nonstationary and informative models for optimization. Luigi Acerbi. An Exploration of Acquisition and Mean Functions in Variational Bayesian Monte Carlo. In Symposium on Advances in Approximate Bayesian Inference, pp. 1 10. PMLR, 2019. Nachman Aronszajn. Theory of Reproducing Kernels. Transactions of the American Mathematical Society, 68(3):337 404, 1950. John-Alexander M. Assael, Ziyu Wang, Bobak Shahriari, and Nando de Freitas. Heteroscedastic Treed Bayesian Optimisation. ar Xiv preprint ar Xiv:1410.7172, 2014. Maximilian Balandat, Brian Karrer, Daniel Jiang, Samuel Daulton, Ben Letham, Andrew G. Wilson, and Eytan Bakshy. Bo Torch: A Framework for Efficient Monte-Carlo Bayesian Optimization. Advances in Neural Information Processing Systems, 33:21524 21538, 2020. James Bergstra, Rémi Bardenet, Yoshua Bengio, and Balázs Kégl. Algorithms for Hyper-Parameter Optimization. Advances in Neural Information Processing Systems, 24, 2011. Mickael Binois and Nathan Wycoff. A survey on high-dimensional Gaussian process modeling with application to Bayesian optimization. ACM Transactions on Evolutionary Learning and Optimization, 2(2):1 26, 2022. Adam D. Bull. Convergence Rates of Efficient Global Optimization Algorithms. Journal of Machine Learning Research, 12(10), 2011. Sayak Ray Chowdhury and Aditya Gopalan. On Kernelized Multi-armed Bandits. In International Conference on Machine Learning, pp. 844 853. PMLR, 2017. Kyle Cranmer, Johann Brehmer, and Gilles Louppe. The frontier of simulation-based inference. Proceedings of the National Academy of Sciences, 117(48):30055 30062, 2020. Y. Diouane, V. Picheny, R. L. Riche, and Alexandre Scotto Di Perrotolo. TREGO: a Trust-Region Framework for Efficient Global Optimization. ar Xiv preprint ar Xiv:2101.06808, 2021. David Eriksson and Martin Jankowiak. High-Dimensional Bayesian Optimization with Sparse Axis-Aligned Subspaces. In Uncertainty in Artificial Intelligence, pp. 493 503. PMLR, 2021. David Eriksson, Michael Pearce, Jacob Gardner, Ryan D Turner, and Matthias Poloczek. Scalable Global Optimization via Local Bayesian Optimization. Advances in Neural Information Processing Systems, 32, 2019. Peter I. Frazier. A Tutorial on Bayesian Optimization. ar Xiv preprint ar Xiv:1807.02811, 2018. Jacob Gardner, Chuan Guo, Kilian Weinberger, Roman Garnett, and Roger Grosse. Discovering and Exploiting Additive Structure for Bayesian Optimization. In International Conference on Artificial Intelligence and Statistics, pp. 1311 1319. PMLR, 2017. Jacob Gardner, Geoff Pleiss, Kilian Q. Weinberger, David Bindel, and Andrew G. Wilson. GPy Torch: Blackbox Matrix-Matrix Gaussian Process Inference with GPU Acceleration. Advances in Neural Information Processing Systems, 31, 2018. Roman Garnett, Michael A. Osborne, and Philipp Hennig. Active Learning of Linear Embeddings for Gaussian Processes. In Proceedings of the Thirtieth Conference on Uncertainty in Artificial Intelligence, pp. 230 239, 2014. Marc G. Genton. Classes of Kernels for Machine Learning: A Statistics Perspective. JMLR, 2:299 312, 2001. Mark N. Gibbs. Bayesian Gaussian processes for Regression and Classification. Ph D thesis, University of Cambridge, 1998. Rafael Gómez-Bombarelli, Jennifer N. Wei, David Duvenaud, José Miguel Hernández-Lobato, Benjamín Sánchez-Lengeling, Dennis Sheberla, Jorge Aguilera-Iparraguirre, Timothy D. Hirzel, Ryan P. Adams, and Alán Aspuru-Guzik. Automatic Chemical Design Using a Data-Driven Continuous Representation of Molecules. ACS Central Science, 4(2):268 276, 2018. Stewart Greenhill, Santu Rana, Sunil Gupta, Pratibha Vellanki, and Svetha Venkatesh. Bayesian Optimization for Adaptive Experimental Design: A Review. IEEE Access, 8:13937 13948, 2020. Antoine Grosnit, Rasul Tutunov, Alexandre Max Maraval, Ryan-Rhys Griffiths, Alexander I. Cowen-Rivers, Lin Yang, Lin Zhu, Wenlong Lyu, Zhitang Chen, Jun Wang, et al. High-Dimensional Bayesian Optimisation with Variational Autoencoders and Deep Metric Learning. ar Xiv preprint ar Xiv:2106.03609, 2021. Sunil Gupta, Santu Rana, Svetha Venkatesh, et al. Regret Bounds for Expected Improvement Algorithms in Gaussian Process Bandit Optimization. In International Conference on Artificial Intelligence and Statistics, pp. 8715 8737. PMLR, 2022. Michael U. Gutmann, Jukka Corander, et al. Bayesian Optimization for Likelihood-Free Inference of Simulator-Based Statistical Models. JMLR, 2016. Nikolaus Hansen. The CMA Evolution Strategy: A Tutorial. ar Xiv preprint ar Xiv:1604.00772, 2016. Nikolaus Hansen, Youhei Akimoto, and Petr Baudis. CMA-ES/pycma. Zenodo, DOI: 10.5281/zenodo.2559634, 2019. Markus Heinonen, Henrik Mannerström, Juho Rousu, Samuel Kaski, and Harri Lähdesmäki. Non-Stationary Gaussian Process Regression with Hamiltonian Monte Carlo. In Artificial Intelligence and Statistics, pp. 732 740. PMLR, 2016. James Hensman, Nicolo Fusi, and Neil D. Lawrence. Gaussian Processes for Big Data. Proceedings of the Twenty-Ninth Conference on Uncertainty in Artificial Intelligence, 2013. Matthew Hoffman, Eric Brochu, Nando de Freitas, et al. Portfolio Allocation for Bayesian Optimization. In UAI, pp. 327 336, 2011. Carl Hvarfner, Danny Stoll, Artur Souza, Marius Lindauer, Frank Hutter, and Luigi Nardi. πBO: Augmenting Acquisition Functions with User Beliefs for Bayesian Optimization. In International Conference on Learning Representations, 2022. Marko Järvenpää, Michael U. Gutmann, Aki Vehtari, and Pekka Marttinen. Parallel Gaussian Process Surrogate Bayesian Inference with Noisy Likelihood Evaluations. Bayesian Analysis, 16(1):147 178, 2021. Stephen Joe and Frances Y. Kuo. Constructing Sobol Sequences with Better Two-Dimensional Projections. SIAM Journal on Scientific Computing, 30(5):2635 2654, 2008. Donald R. Jones, Matthias Schonlau, and William J. Welch. Efficient Global Optimization of Expensive Black-Box Functions. Journal of Global optimization, 13(4):455 492, 1998. M.C. Jones. Kumaraswamy s distribution: A beta-type distribution with some tractability advantages. Statistical Methodology, 6(1):70 81, 2009. Motonobu Kanagawa, Philipp Hennig, Dino Sejdinovic, and Bharath K. Sriperumbudur. Gaussian Processes and Kernel Methods: A Review on Connections and Equivalences. ar Xiv preprint ar Xiv:1807.02582, 2018. Kirthevasan Kandasamy, Jeff Schneider, and Barnabás Poczós. High Dimensional Bayesian Optimisation and Bandits via Additive Models. In International Conference on Machine Learning, pp. 295 304. PMLR, 2015. Diederik P. Kingma and Jimmy Ba. Adam: A Method for Stochastic Optimization. ar Xiv preprint ar Xiv:1412.6980, 2014. Harold J. Kushner. A Versatile Stochastic Model of a Function of Unknown and Time Varying Form. Journal of Mathematical Analysis and Applications, 5(1):150 167, 1962. Yann Le Cun. The MNIST database of handwritten digits. http://yann.lecun.com/exdb/mnist/, 1998. Ben Letham, Roberto Calandra, Akshara Rai, and Eytan Bakshy. Re-Examining Linear Embeddings for High-Dimensional Bayesian Optimization. Advances in Neural Information Processing Systems, 33, 2020. Chun-Liang Li, Kirthevasan Kandasamy, Barnabás Poczós, and Jeff Schneider. High Dimensional Bayesian Optimization via Restricted Projection Pursuit Models. In Artificial Intelligence and Statistics, pp. 884 892. PMLR, 2016. Ruben Martinez-Cantin. Local Nonstationarity for Efficient Bayesian Optimization. ar Xiv preprint ar Xiv:1506.02080, 2015. Mark Mc Leod, Stephen Roberts, and Michael A. Osborne. Optimization, fast and slow: optimally switching between local and Bayesian optimization. In International Conference on Machine Learning, pp. 3443 3452. PMLR, 2018. Erich Merrill, Alan Fern, Xiaoli Z. Fern, and Nima Dolatnia. An Empirical Study of Bayesian Optimization: Acquisition Versus Partition. JMLR, 22:4 1, 2021. Charles A. Micchelli, Yuesheng Xu, and Haizhang Zhang. Universal Kernels. JMLR, 7(Dec):2651 2667, 2006. Jonas Močkus. On Bayesian Methods for Seeking the Extremum. In Optimization Techniques IFIP Technical Conference, pp. 400 404. Springer, 1975. Riccardo Moriconi, Marc Peter Deisenroth, and K. S. Sesh Kumar. High-dimensional Bayesian optimization using low-dimensional feature spaces. Machine Learning, 109(9):1925 1943, 2020. Mojmír Mutný and Andreas Krause. Efficient High Dimensional Bayesian Optimization with Additivity and Quadrature Fourier Features. Advances in Neural Information Processing Systems, 31, 2018. Willie Neiswanger, Lantao Yu, Shengjia Zhao, Chenlin Meng, and Stefano Ermon. Generalizing Bayesian Optimization with Decision-theoretic Entropies. Advances in Neural Information Processing Systems, 35, 2022. Chang Yong Oh, Efstratios Gavves, and Max Welling. BOCK: Bayesian Optimization with Cylindrical Kernels. In International Conference on Machine Learning, pp. 3868 3877. PMLR, 2018. Art B. Owen. Scrambling Sobol and Niederreiter-Xing Points. Journal of Complexity, 14(4):466 489, 1998. Christopher J. Paciorek and Mark J. Schervish. Spatial Modelling Using a New Class of Nonstationary Covariance Functions. Environmetrics, 17(5):483 506, 2006. Anil Ramachandran, Sunil Gupta, Santu Rana, Cheng Li, and Svetha Venkatesh. Incorporating Expert Prior in Bayesian Optimisation via Space Warping. Knowledge-Based Systems, 195:105663, 2020. Elena Raponi, Hao Wang, Mariusz Bujny, Simonetta Boria, and Carola Doerr. High Dimensional Bayesian Optimization Assisted by Principal Component Analysis. In International Conference on Parallel Problem Solving from Nature, pp. 169 183. Springer, 2020. Carl E. Rasmussen and Christopher K. I. Williams. Gaussian Processes for Machine Learning. MIT Press, 2005. Mark D. Risser. Nonstationary Spatial Modeling, with Emphasis on Process Convolution and Covariate Driven Approaches. ar Xiv preprint ar Xiv:1610.02447, 2016. Paul Rolland, Jonathan Scarlett, Ilija Bogunovic, and Volkan Cevher. High-Dimensional Bayesian Optimization via Additive Models with Overlapping Groups. In International Conference on Artificial Intelligence and Statistics, pp. 298 307. PMLR, 2018. Bobak Shahriari, Kevin Swersky, Ziyu Wang, Ryan P. Adams, and Nando De Freitas. Taking the Human Out of the Loop: A Review of Bayesian Optimization. Proceedings of the IEEE, 104(1):148 175, 2016. Jasper Snoek, Hugo Larochelle, and Ryan P. Adams. Practical Bayesian Optimization of Machine Learning Algorithms. In Advances in Neural Information Processing Systems, pp. 2951 2959, 2012. Jasper Snoek, Kevin Swersky, Rich Zemel, and Ryan Adams. Input Warping for Bayesian Optimization of Non-stationary Functions. In International Conference on Machine Learning, pp. 1674 1682. PMLR, 2014. Jasper Snoek, Oren Rippel, Kevin Swersky, Ryan Kiros, Nadathur Satish, Narayanan Sundaram, Mostofa Patwary, Mr Prabhat, and Ryan Adams. Scalable Bayesian Optimization Using Deep Neural Networks. In International Conference on Machine Learning, pp. 2171 2180. PMLR, 2015. Artur Souza, Luigi Nardi, Leonardo B. Oliveira, Kunle Olukotun, Marius Lindauer, and Frank Hutter. Bayesian Optimization with a Prior for the Optimum. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases, pp. 265 296. Springer, 2021. Niranjan Srinivas, Andreas Krause, Sham Kakade, and Matthias Seeger. Gaussian Process Optimization in the Bandit Setting: No Regret and Experimental Design. In Proceedings of the 27th International Conference on International Conference on Machine Learning, pp. 1015 1022, 2010. Kevin Jordan Swersky. Improving Bayesian Optimization for Machine Learning using Expert Priors. Ph D thesis, University of Toronto, 2017. Michalis Titsias. Variational Learning of Inducing Variables in Sparse Gaussian Processes. In Artificial Intelligence and Statistics, pp. 567 574. PMLR, 2009. Sattar Vakili, Kia Khezeli, and Victor Picheny. On Information Gain and Regret Bounds in Gaussian Process Bandits. In International Conference on Artificial Intelligence and Statistics, pp. 82 90. PMLR, 2021. Pauli Virtanen, Ralf Gommers, Travis E. Oliphant, Matt Haberland, Tyler Reddy, David Cournapeau, Evgeni Burovski, Pearu Peterson, Warren Weckesser, Jonathan Bright, Stéfan J. van der Walt, Matthew Brett, Joshua Wilson, K. Jarrod Millman, Nikolay Mayorov, Andrew R. J. Nelson, Eric Jones, Robert Kern, Eric Larson, C. J. Carey, İlhan Polat, Yu Feng, Eric W. Moore, Jake Vander Plas, Denis Laxalde, Josef Perktold, Robert Cimrman, Ian Henriksen, E. A. Quintero, Charles R. Harris, Anne M. Archibald, Antônio H. Ribeiro, Fabian Pedregosa, Paul van Mulbregt, and Sci Py 1.0 Contributors. Sci Py 1.0: Fundamental Algorithms for Scientific Computing in Python. Nature Methods, 17:261 272, 2020. Kangrui Wang, Oliver Hamelijnck, Theodoros Damoulas, and Mark Steel. Nonstationary Nonseparable Random Fields. In International Conference on Machine Learning, pp. 9887 9897. PMLR, 2020a. Linnan Wang, Rodrigo Fonseca, and Yuandong Tian. Learning Search Space Partition for Black-box Optimization using Monte Carlo Tree Search. Advances in Neural Information Processing Systems, 33: 19511 19522, 2020b. Zi Wang, Clement Gehring, Pushmeet Kohli, and Stefanie Jegelka. Batched Large-scale Bayesian Optimization in High-dimensional Spaces. In International Conference on Artificial Intelligence and Statistics, pp. 745 754. PMLR, 2018. Ziyu Wang and Nando de Freitas. Theoretical Analysis of Bayesian Optimisation with Unknown Gaussian Process Hyper-Parameters. ar Xiv preprint ar Xiv:1406.7758, 2014. Ziyu Wang, Frank Hutter, Masrour Zoghi, David Matheson, and Nando de Feitas. Bayesian Optimization in a Billion Dimensions via Random Embeddings. Journal of Artificial Intelligence Research, 55:361 387, 2016. Colin White, Willie Neiswanger, and Yash Savani. BANANAS: Bayesian Optimization with Neural Architectures for Neural Architecture Search. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pp. 10293 10301, 2021. James Wilson, Frank Hutter, and Marc Deisenroth. Maximizing Acquisition Functions for Bayesian Optimization. Advances in Neural Information Processing Systems, 31, 2018. Manuel Wüthrich, Bernhard Schölkopf, and Andreas Krause. Regret Bounds for Gaussian-Process Optimization in Large Domains. Advances in Neural Information Processing Systems, 34, 2021. Ciyou Zhu, Richard H. Byrd, Peihuang Lu, and Jorge Nocedal. Algorithm 778: L-BFGS-B: Fortran Subroutines for Large-Scale Bound-Constrained Optimization. ACM Transactions on Mathematical Software, 23 (4):550 560, 1997. A Instantaneous Regret Bounds Lemma A.1. Assume the interval |f(x) mn(x)| βnσn(x) holds, x X and that global optima of acquisition functions can be found. Then, for popular acquisition functions, namely LCB and EI, and any covariance function, possibly nonstationary, the worst-case instantaneous regret depends on the width of the interval, being proportional to the posterior predictive standard deviation, rn+1 σn(xn+1), where xn+1 is the acquisition at step n + 1. Proof. Let x be the minimizer of f(x). Then, based on (Srinivas et al., 2010, Lemma 5.2), we have the following instantaneous regret bound for LCB, rn+1 = f(xn+1) f(x ) (17) f(xn+1) mn(x ) + βnσn(x ) (18) = f(xn+1) LCB(x ) (19) f(xn+1) LCB(xn+1) (20) = f(xn+1) mn(xn+1) + βnσn(xn+1) (21) 2βnσn(xn+1). (22) Regarding EI, we know Bn In(x) EI(x) In(x) + (βn + 1)σn(x), by (Wang & de Freitas, 2014, Lemma 9), with Bn = τ( βn)/τ(βn) 1, and τ(z) = z FN (z) + N(z; 0, 1), because τ is nondecreasing, τ (z) = FN (z) [0, 1]. Also, recall that the improvement is given by In(x) = max(0, m n f(x)) with m n = minx mn(x). Now, we derive a simpler, tighter upper bound on EI, EI(x) = σn(x)τ(z(x)) (23) σn(x)τ(0) (24) 2π σn(x) < σn(x), (25) where the first inequality follows from z(x) = (m n mn(x))/σn(x) 0 and nondecreasing τ. Let x n = arg minx mn(x). Then, the instantaneous regret bound can be written as rn+1 = f(xn+1) f(x ) (26) = (f(xn+1) m n ) + (m n f(x )) (27) (f(xn+1) m n ) + In(x ) (28) (f(xn+1) m n ) + 1 Bn EI(x ) (29) (f(xn+1) m n ) + 1 Bn EI(xn+1) (30) (f(xn+1) m n ) + 1 Bn σn(xn+1) (31) (mn(xn+1) m n ) + βn + 1 σn(xn+1) (32) Dn(xn+1) + βn + 1 σn(xn+1) (33) D+ n + βn + 1 σn(xn+1), (34) where, in Equation 32, the difference is bounded as |mn(xn+1) m n | Dn(xn+1)σn(xn+1), with Dn(xn+1) = q 2 log(σn(xn+1)/σn(x n )), by (Wang & de Freitas, 2014, Lemma 10). From the same lemma, we also know that σn(x n ) can be lower bounded by a positive value, and that there is a positive constant upper bounding σn(xn+1) by construction. Thus, for any covariance function, possibly nonstationary, there exists a positive constant D+ n such that D+ n Dn(xn+1), which yields the result. B Implementation Details B.1 Gaussian Processes We used GPy Torch (Gardner et al., 2018) to implement all GP models in Python, including the proposed informative (I) covariance functions, cylindrical (C) covariance functions (Oh et al., 2018) and axis-aligned quadratic mean (+QM) (Snoek et al., 2015). Unlike the version provided by GPy Torch, our implementation of C follows the original training and prediction routines, which account for the special treatment of the origin, as discussed in more detail in Appendix C. In order to ensure fair comparisons, all models are optimized by L-BFGS-B (Zhu et al., 1997) with a maximum of 1000 iterations per new acquisition and other options set to default values. In addition, the stationary components CS are Matérn (ν = 5/2), equipped with the weighted Euclidean distance (lengthscales λd). Positivity constraints are handled internally by GPy Torch using the softplus transform, except for C-specific hyperparameters, which use the log transform (Appendix C). General hyperparameters are given uninformative priors with bounds from (Oh et al., 2018). These include the prior variance σ2 0 U(e 12, e20) and lengthscales λd U(e 12, 2 D), d {1 . . . D}, where the upper bound corresponds to the maximum Euclidean distance in the centered hypercube X = [ 1, 1]D. Table 2 summarizes the default settings of I. For noise-free objectives, the noise hyperparameter is set to a fixed value σ2 y = 10 3. By default, models are characterized by an uninformative prior mean function with a uniformly-distributed constant b U(ymin, ymax), where the bounds correspond to the minimum and maximum observed values. Variants +QM use an axis-aligned quadratic mean function m(x) = b + P d ad([x]d [c0]d)2, as suggested by Snoek et al. (2015) and Swersky (2017, Section 4.4). The quadratic weights are independent, distributed as ad HS+(2). The half-Horseshoe is a spike-and-slab hyperprior with a spike at 0 and a slab on the positive real line, only allowing convex quadratic functions. Quadratic mean functions with adaptive, greedily-chosen centers c0 = xbest were tested, but no significant or consistent improvement compared to the default center c0 = 0 was observed. Hence, results using the former (+QM+XA) are only shown in Appendix D. Table 2: Default settings of informative covariance functions. Quantity Setting Stationary corr. CS Matérn (ν = 5/2, σ2 0 = 1) Prior variance σ2 p U(e 12, e20) Lengthscales λd U(e 12, 2 D) Distances dσ, dλ Tied d0 Distance d0 Weighted Euclidean d WE (shared λd) Kernels kσ, kλ Tied k0 Kernel k0 Gaussian Ratios rσ, rλ Tied r0 Ratio r0 Kumar(3.164, 1000), Sensitivity analysis (Appendix F) B.2 Acquisition Prior to acquiring new data, an initial evidence set is formed. The set Dn0 includes the origin and 15 other pseudo-uniformly distributed points, drawn according to a scrambled Sobol sequence (Owen, 1998; Joe & Kuo, 2008). This set provides the initial conditions from which initial models are estimated and their performance measured. For noise-free objectives, the incumbent solution is the point with the lowest observed value x(n0+n) best = arg minx Dn0+n f(x), n0 = 16, n = {0, . . . , N}, N = 200. The default acquisition function is the Expected Improvement (EI), which is implemented in Bo Torch (Balandat et al., 2020). Maximization is performed on the CPU via gradient-based optimization with multiple restarts. The 20 initial points for the optimization are selected from 20010 candidates, including 10 points that are chosen heuristically and 20000 points that are randomly generated. The former are obtained by small Gaussian perturbations of the incumbent solution, and the latter are drawn from a scrambled Sobol sequence. The main difference between this procedure and the one implemented by Oh et al. (2018) is the use of L-BFGS-B via Bo Torch, as opposed to Adam (Kingma & Ba, 2014). B.3 Test Functions Test functions are set up in a way similar to that used by Oh et al. (2018), with their domains adjusted to the centered hypercube X = [ 1, 1]D by applying a linear transformation to the inputs. Additionally, original objectives are shifted and scaled to satisfy f(x ) = 0 and f(0) = 100. Since the resulting functions are nonnegative, models are then estimated on log-transformed data, after adding a very small offset. In general, this leads to improved performance because the distribution of function values becomes more Gaussian-like. Performance metrics are however computed without the log transform. QBranin The Branin function is originally evaluated on [ 5, 10] [0, 15], and has several global optima. This function is modified by adding a quadratic component (last term), transforming it into a bowl-shaped function with only one global optimum, f(x) = [x]2 5.1 4π2 [x]2 1 + 5 π [x]1 6 2 + 10 1 1 cos([x]2 1) + 10 + 5[x]2 1. (35) Higher-dimensional versions are obtained by additive repetition. Shifted QBranin The shifted variants SQBranin and SSQBranin are derived from QBranin by shifting the inputs in the original space by 2 and 3, respectively. Additional experiments with these variants are included in Appendix D. Figure 8 provides a 2D illustration of these functions in the transformed space [ 1, 1]2. Rosenbrock This function is originally evaluated on the hypercube [ 5, 10]D and is given by 100([x]d+1 [x]2 d)2 + ([x]d 1)2 . (36) Similar to QBranin, the global optimum is close to the origin in the transformed space, x = 0.2 1. However, as shown in Figure 9, the Rosenbrock is characterized by narrow banana-shaped valleys, making global optimization more challenging. Shifted Rosenbrock S35Rosenbrock, S50Rosenbrock, and S65Rosenbrock are three variants of the Rosenbrock function, designed to test the robustness of methods that assume good values near the origin. These functions are obtained by shifting the inputs so that the global minima in the transformed space are at x = 0.35 1, x = 0.50 1, and x = 0.65 1. Figure 9 provides a 2D illustration of these functions. Levy The original domain is [ 10, 10]D and the function is defined as f(x) = sin2(πw1) + d=1 (wd 1)2 1 + 10 sin2(πwd + 1) + (wi 1)2 1 + sin2(2πwi) , (37) wd = 1 + [x]d 1 4 , i {1, . . . , D}. (38) As shown in Figure 10 (left), the Levy function is characterized by strong sinusoidal components. The global optimum is again close to the origin, specifically at x = 0.1 1 in the transformed space. Styblinski-Tang The original domain is [ 5, 5]D and the function is given by [x]4 d 16[x]2 d + 5[x]d . (39) Joint optimization of the Styblinski-Tang function is difficult because it has exponentially many local modes 2D 1. In contrast to QBranin, Rosenbrock and Levy, the global optimum is relatively far from the origin, x 0.58 1 in the transformed space. Figure 10 (right) provides an illustration of this multimodal function. 1.00 0.75 0.50 0.25 0.00 0.25 0.50 0.75 1.00 1.00 1.00 QBranin Global Minimum ([-0.333 -0.2 ]) Origin 1.00 0.75 0.50 0.25 0.00 0.25 0.50 0.75 1.00 1.00 1.00 SQBranin Global Minimum ([-0.6 -0.467]) Origin 1.00 0.75 0.50 0.25 0.00 0.25 0.50 0.75 1.00 1.00 1.00 SSQBranin Global Minimum ([-0.733 -0.6 ]) Origin Figure 8: A 2D slice of QBranin and shifted variants. These test functions are unimodal and bowl-shaped. The purpose of SQBranin and SSQBranin is to test the robustness of methods that assume good values near the origin. 1.00 0.75 0.50 0.25 0.00 0.25 0.50 0.75 1.00 1.00 1.00 Rosenbrock Global Minimum ([-0.2 -0.2]) Origin 1.00 0.75 0.50 0.25 0.00 0.25 0.50 0.75 1.00 1.00 1.00 S35Rosenbrock Global Minimum ([0.35 0.35]) Origin 1.00 0.75 0.50 0.25 0.00 0.25 0.50 0.75 1.00 1.00 1.00 S50Rosenbrock Global Minimum ([0.5 0.5]) Origin 1.00 0.75 0.50 0.25 0.00 0.25 0.50 0.75 1.00 1.00 1.00 S65Rosenbrock Global Minimum ([0.65 0.65]) Origin Figure 9: A 2D slice of Rosenbrock and three shifted variants (S35, S50 and S65). 1.00 0.75 0.50 0.25 0.00 0.25 0.50 0.75 1.00 1.00 Global Minimum ([0.1 0.1]) Origin 1.00 0.75 0.50 0.25 0.00 0.25 0.50 0.75 1.00 1.00 1.00 Styblinski-Tang Global Minimum ([-0.581 -0.581]) Origin Figure 10: A 2D slice of Levy and Styblinski-Tang. C Cylindrical Covariance Functions Cylindrical covariance functions (Oh et al., 2018) aim to address the boundary issue that arises in highdimensional Bayesian optimization. The cylindrical transformation T(x) represents a point x X in terms of its radius r and angular components a, (r, a) = T(x) = ( ( x 2, x/ x 2), if x 2 = 0 (0, aarbitrary), if x 2 = 0 . (40) where aarbitrary is a random unit vector. Geometrically, this transformation maps balls of radius R onto the surface of a cylinder of height R. A spherical shell of width δr centered at a point x corresponds to a region in the Euclidean space whose volume increases exponentially with the radius r. As a result, algorithms that aim at equally covering each volume element in the Euclidean space spend exponentially more time at the boundary of the search space. In contrast, Oh et al. (2018) proposed to spend an equal amount of resources on each volume element in the transformed space. Intuitively, the transformation leads to a search in the Euclidean space that expands the region near the center while contracting the regions near the boundary (Oh et al., 2018, Section 3.2). The assumption is that optimal values are more likely to be found near the origin, so it is beneficial to encourage exploration in this region. Remarkably, this transformation poses a challenge when comparing any non-origin point to the origin, since the latter is represented by an infinite set of points with radius 0. Oh et al. (2018) proposed using the point in the set that is closest to the point under comparison x, setting aarbitrary = x/ x 2. However, this solution has significant computational implications, requiring custom inference routines. If the origin is included in the training set, the Gram matrix must be recomputed for each test point. This issue is mitigated by block matrix inversion, where the block containing all non-origin points can be precomputed and reused. Cylindrical covariance functions Ccyl are defined by the product of a 1-dimensional Matérn (ν = 5/2), measuring similarity of radii, and a function Ca that compares angular components, Ca(a1, a2) = p=0 cp(a 1a2)p, (41) where cp is a normalized non-negative weight, given a log-Normal prior log(cp) N(0, 22). To further encourage the expansion of the innermost region, Oh et al. (2018) used input warping on the radius component. This involves transforming the radius according to the Kumaraswamy cumulative distribution, FKuma(r) = 1 (1 rα)β, with α [0.5, 1] and β [1, 2]. Both α and β are given a spike-and-slab hyperprior on the log scale, with spike at log(1). Figure 11 shows possible warpings, including the identity transform obtained with α = 1 and β = 1. 0.0 0.2 0.4 0.6 0.8 1.0 Figure 11: Input warping on the radius. Concave transformations expand regions of small radii. D Additional Results D.1 Main Methods S Bayesian Optimization (BO) with an uninformative Gaussian process (GP) prior, characterized by a constant mean and a stationary Matérn covariance function. S+QM S with an axis-aligned quadratic mean function (Snoek et al., 2015; Swersky, 2017). S+TR S with acquisitions within an adaptive trust region, i.e., a box centered at the incumbent solution that is shrunk/expanded based on consecutive failures/successes (Eriksson et al., 2019). For noise-free objectives, the incumbent solution is the point with the lowest observed value, as described in Appendix B.2. C BO with a GP characterized by a constant mean and a cylindrical covariance function (Oh et al., 2018). I+X0 (Proposed) BO with a GP featuring a constant mean and an informative covariance function. A single, fixed anchor is placed at the center (origin) of the search space. I+XA (Proposed) I+X0 with an adaptive greedily-chosen anchor, given by the incumbent solution. This is the same point as that used in S+TR. I+XA+TR (Proposed) I+XA with a trust region, using the same adaptation scheme as S+TR. D.2 Additional Methods S+QM+XA Quadratic mean function with a greedily-chosen center, given by the incumbent solution. In most tests, including the shifted objectives in Figures 12 and 14, S+QM+XA was outperformed by S+QM. GS+XA Informative covariance functions are based on a search model for the optimum, which is given by a mixture distribution (Equation 11). For a single anchor, Gaussian kernel and weighted Euclidean distance, the equivalent distribution is a mixture of a uniform and a Gaussian distribution with diagonal covariance matrix. In GS+XA, the mean is set to the incumbent solution and the variance is estimated from the data. The standard deviation is the median absolute deviation (MAD) of the best 5% training points (rounded, with a minimum of 2), assuming that the coordinates are independently and identically distributed. The Gaussian mixing weight is set to 0.9. Other combinations involving a smaller mixing weight (0.5), a larger training set (10%) and a scale correction of the MAD estimate were also tested, but overall performed no better than GS+XA. In turn, GS+XA typically performed worse than S or C, as shown in Figures 12, 13 and 14. CMA-ES Covariance Matrix Adaptation Evolution Strategy (Hansen, 2016) is a popular global optimization algorithm and is related to GS+XA, in that it samples from a Gaussian whose mean and covariance matrix are adaptive. In terms of implementation, we use pycma (Hansen et al., 2019) with a population size of 10. In general, CMA-ES was outperformed by S, as shown in Figures 12 and 14. On Styblinski-Tang (Figure 13), however, it performed well, but no better than I+XA or I+XA+QM. I+XA+QM (Proposed) The uninformative constant mean in I+XA is replaced by the quadratic mean from S+QM. In tests where S+QM performed comparatively well, e.g. Styblinski-Tang (Figure 13), the combination I+XA+QM proved to be effective. I+XA+F (Proposed) I+XA uses empirical priors, optimizing the hyperparameters via marginal likelihood. However, if additional information about the objective is available, better performance can be achieved. For instance, in cases where the center is already a good solution, e.g. Levy (Figure 13) and Rosenbrock (Figure 14), it can be difficult to improve upon initial conditions. In such cases, we can commit to a more exploitative search around XA. Based on this intuition, the focused method I+XA+F uses relatively short fixed lengthscales λ0 = 0.1 D to compute distances d0, and a fixed r0 = 0.1. The remaining hyperparameters are still learned from data. As future work, several priors can be used in tandem. A portfolio can then be managed using strategies similar to those proposed by Hoffman et al. (2011) and Mc Leod et al. (2018). I+XA+F+TR (Proposed) In order to further encourage the search around XA, I+XA+F can also be combined with trust regions, achieving better performance in certain cases, e.g. Levy (Figure 13) and Rosenbrock (Figure 14). D.3 Results 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement QBranin,20D S S+QM S+TR S+QM+XA C GS+XA CMA-ES I+X0 (Proposed) I+XA (Proposed) I+XA+TR (Proposed) I+XA+QM (Proposed) I+XA+F (Proposed) I+XA+F+TR (Proposed) 0 20 40 60 80 100 120 140 160 180 200 QBranin,50D 0 20 40 60 80 100 120 140 160 180 200 QBranin,100D 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement SQBranin,20D S S+QM S+TR S+QM+XA C GS+XA CMA-ES I+X0 (Proposed) I+XA (Proposed) I+XA+TR (Proposed) I+XA+QM (Proposed) I+XA+F (Proposed) I+XA+F+TR (Proposed) 0 20 40 60 80 100 120 140 160 180 200 SQBranin,50D 0 20 40 60 80 100 120 140 160 180 200 SQBranin,100D 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement SSQBranin,20D S S+QM S+TR S+QM+XA C GS+XA CMA-ES I+X0 (Proposed) I+XA (Proposed) I+XA+TR (Proposed) I+XA+QM (Proposed) I+XA+F (Proposed) I+XA+F+TR (Proposed) 0 20 40 60 80 100 120 140 160 180 200 SSQBranin,50D 0 20 40 60 80 100 120 140 160 180 200 SSQBranin,100D Figure 12: Performance on QBranin and variants, ranging from 20 to 100 dimensions. Solid curves and shaded regions represent the mean and standard deviation of the normalized improvement, computed over 10 trials with different initial conditions. Solid vertical lines indicate the interquartile range. 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement 0 20 40 60 80 100 120 140 160 180 200 0 20 40 60 80 100 120 140 160 180 200 S S+QM S+TR S+QM+XA C GS+XA CMA-ES I+X0 (Proposed) I+XA (Proposed) I+XA+TR (Proposed) I+XA+QM (Proposed) I+XA+F (Proposed) I+XA+F+TR (Proposed) 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement Styblinski-Tang,20D 0 20 40 60 80 100 120 140 160 180 200 Styblinski-Tang,50D 0 20 40 60 80 100 120 140 160 180 200 Styblinski-Tang,100D S S+QM S+TR S+QM+XA C GS+XA CMA-ES I+X0 (Proposed) I+XA (Proposed) I+XA+TR (Proposed) I+XA+QM (Proposed) I+XA+F (Proposed) I+XA+F+TR (Proposed) Figure 13: Performance on Levy and Styblinski-Tang, ranging from 20 to 100 dimensions. Solid curves and shaded regions represent the mean and standard deviation of the normalized improvement, computed over 10 trials with different initial conditions. Solid vertical lines indicate the interquartile range. 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement Rosenbrock,20D 0 20 40 60 80 100 120 140 160 180 200 Rosenbrock,50D 0 20 40 60 80 100 120 140 160 180 200 Rosenbrock,100D S S+QM S+TR S+QM+XA C GS+XA CMA-ES I+X0 (Proposed) I+XA (Proposed) I+XA+TR (Proposed) I+XA+QM (Proposed) I+XA+F (Proposed) I+XA+F+TR (Proposed) 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement S35Rosenbrock,20D S S+QM S+TR S+QM+XA C GS+XA CMA-ES I+X0 (Proposed) I+XA (Proposed) I+XA+TR (Proposed) I+XA+QM (Proposed) I+XA+F (Proposed) I+XA+F+TR (Proposed) 0 20 40 60 80 100 120 140 160 180 200 S35Rosenbrock,50D 0 20 40 60 80 100 120 140 160 180 200 S35Rosenbrock,100D 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement S50Rosenbrock,20D S S+QM S+TR S+QM+XA C GS+XA CMA-ES I+X0 (Proposed) I+XA (Proposed) I+XA+TR (Proposed) I+XA+QM (Proposed) I+XA+F (Proposed) I+XA+F+TR (Proposed) 0 20 40 60 80 100 120 140 160 180 200 S50Rosenbrock,50D 0 20 40 60 80 100 120 140 160 180 200 S50Rosenbrock,100D 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement S65Rosenbrock,20D S S+QM S+TR S+QM+XA C GS+XA CMA-ES I+X0 (Proposed) I+XA (Proposed) I+XA+TR (Proposed) I+XA+QM (Proposed) I+XA+F (Proposed) I+XA+F+TR (Proposed) 0 20 40 60 80 100 120 140 160 180 200 S65Rosenbrock,50D 0 20 40 60 80 100 120 140 160 180 200 S65Rosenbrock,100D Figure 14: Performance on Rosenbrock objectives with 20, 50 and 100 dimensions. Solid curves and shaded regions represent the mean and standard deviation of the normalized improvement, computed over 10 trials with different initial conditions. Solid vertical lines indicate the interquartile range. E Ablation Study In this section, we investigate the contribution of spatially-varying prior variances (V) and lengthscales (L) toward more efficient high-dimensional optimization. For this purpose, we test these spatially-varying properties individually, referring to them as I(V) and I(L). Overall, the combination of spatially-varying prior variances and lengthscales is most effective in higherdimensional spaces (50 and 100), where I+X0 and I+XA generally outperform their I(V) and I(L) counterparts. In comparison to S, I(V) and I(L) prove to be effective on their own, but I(L) seems to account for most of the performance increase on Rosenbrock (Figure 15), where anchors are relatively close to the optimum. The shorter lengthscales around anchors promote local exploration, even after having acquired data in their neighborhood. Interestingly, when the anchor is misspecified (Styblinski-Tang), the performance of I(L)+X0 is better than that of I+X0 and similar to that of S in 20 dimensions. However, despite anchor misspecification, we observe the usefulness of spatially-varying prior variances as the dimensionality increases to 100, likely because these mitigate the boundary issue. For adaptive anchors, it is less clear whether the performance of I+XA on Styblinski-Tang is mostly due to spatially-varying prior variances or lengthscales. Both seem to be important because I+XA consistently outperforms I(V)+XA and I(L)+XA. 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement Rosenbrock,20D 0 20 40 60 80 100 120 140 160 180 200 Rosenbrock,50D 0 20 40 60 80 100 120 140 160 180 200 Rosenbrock,100D S S+QM C I+X0 (Proposed) I(V)+X0 (Proposed) I(L)+X0 (Proposed) 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement Rosenbrock,20D 0 20 40 60 80 100 120 140 160 180 200 Rosenbrock,50D 0 20 40 60 80 100 120 140 160 180 200 Rosenbrock,100D S S+QM C I+XA (Proposed) I(V)+XA (Proposed) I(L)+XA (Proposed) 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement Styblinski-Tang,20D 0 20 40 60 80 100 120 140 160 180 200 Styblinski-Tang,50D 0 20 40 60 80 100 120 140 160 180 200 Styblinski-Tang,100D S S+QM C I+X0 (Proposed) I(V)+X0 (Proposed) I(L)+X0 (Proposed) 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement Styblinski-Tang,20D 0 20 40 60 80 100 120 140 160 180 200 Styblinski-Tang,50D 0 20 40 60 80 100 120 140 160 180 200 Styblinski-Tang,100D S S+QM C I+XA (Proposed) I(V)+XA (Proposed) I(L)+XA (Proposed) Figure 15: Ablation study. Solid curves and shaded regions represent the mean and standard deviation of the normalized improvement, computed over 10 trials with different initial conditions. Solid vertical lines indicate the interquartile range. Abbreviations: Stationary (S), Cylindrical (C) and Informative (I) covariances; Quadratic Mean (+QM); Origin (+X0) and Adaptive, greedily-chosen (+XA) anchors; spatially-varying prior Variances (V) and Lengthscales (L). F Sensitivity Analysis The informative covariance functions used in our experiments introduce a ratio r0 (0, 1], which leads to an uninformative stationary GP prior as it approaches 1. In terms of optimization, this hyperparameter balances global and local exploration under the informative search model (Equation 11) by adjusting the magnitude of spatially-varying prior (co)variances via rσ (Equation 13) and lengthscales through rλ (Equation 15). For example, compared to the stationary case, a ratio of r0 = 0.1 indicates that the prior variance is up to 10 larger and the squared lengthscales are 10 shorter in the neighborhood of anchor x0. By default, the ratio r0 is learned by empirical Bayes and its prior is an informative Kumaraswamy density function with a peak at 0.1, given by Kumar(3.164, 1000) and referred to as K3.4 The Kumaraswamy distribution is related to the beta distribution and is similarly parametrized by two positive shape parameters, a and b, with probability density function p Kumar(r; a, b) = abra 1(1 ra)b 1, r (0, 1). (42) Notably, both Beta( ; α, β) and Kumar( ; a, b) are special cases of the generalized (p, γ, δ)-beta distribution (Jones, 2009, Equation 2.1), given by (1, α, β) and (a, 1, b), respectively. However, the latter is more tractable because the beta function in the denominator simplifies as B(1, b) = b 1. For a list of similarities and other advantages over the beta distribution, see (Jones, 2009, Section 7). In the remainder of this section, we examine the performance of I+XA under different priors for the hyperparameter r0. We tested uniform (U) and Dirac delta (D) priors, as well as several Kumaraswamy priors, with parameters chosen to yield densities that become increasingly narrower, while maintaining a peak at 0.1. In particular, we tested Kumar(1.467, 10) denoted by K1, Kumar(2.253, 100) denoted by K2 and Kumar(3.164, 1000) denoted by K3. Among these, K1 is the broadest prior, as shown in Figure 16. The use of priors that favor smaller values of r0 is motivated by the belief that the informative search model is useful. In this case, strategies that promote local exploration under this search model are more likely to improve upon initial conditions than those using an uninformative, uniform search model (stationary GP prior). Figure 17 shows the performance of I+XA variants on QBranin, Rosenbrock and Styblinski-Tang. We begin by observing that, on QBranin, I+XA is relatively robust to the choice of prior over r0 because the performance of all variants is similar and significantly better than that of S, S+QM and C. However, the differences become more noticeable on Rosenbrock and Styblinski-Tang. The uniform prior leads to the worst-performing variant, but I+XA+U can still outperform the other baselines on Rosenbrock. Conversely, narrower priors, such as K2 and K3, which promote exploration under the informative search model, lead to the best-performing methods. Interestingly, I+XA with a delta prior (I+XA+D) can outperform other variants on Rosenbrock. This result suggests that learning all parameters via marginal likelihood may not be the best approach. Alternative training objectives that also consider optimization performance may provide better results, but this study is left for future work. 0.0 0.2 0.4 0.6 0.8 1.0 0 U K1 K2 K3 D Figure 16: Probability density functions of the priors over r0. 4The suffix +K3 is omitted in other sections because it is the default prior over r0. As discussed in Appendix D, the only exception is I+XA+F that uses r0 = 0.1, corresponding to the Dirac delta (D) prior that is tested in this section. 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement QBranin,20D S S+QM C I+XA+U (Proposed) I+XA+K1 (Proposed) I+XA+K2 (Proposed) I+XA+K3 (Proposed) I+XA+D (Proposed) 0 20 40 60 80 100 120 140 160 180 200 QBranin,50D 0 20 40 60 80 100 120 140 160 180 200 QBranin,100D 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement Rosenbrock,20D 0 20 40 60 80 100 120 140 160 180 200 Rosenbrock,50D 0 20 40 60 80 100 120 140 160 180 200 Rosenbrock,100D S S+QM C I+XA+U (Proposed) I+XA+K1 (Proposed) I+XA+K2 (Proposed) I+XA+K3 (Proposed) I+XA+D (Proposed) 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement Styblinski-Tang,20D 0 20 40 60 80 100 120 140 160 180 200 Styblinski-Tang,50D 0 20 40 60 80 100 120 140 160 180 200 Styblinski-Tang,100D S S+QM C I+XA+U (Proposed) I+XA+K1 (Proposed) I+XA+K2 (Proposed) I+XA+K3 (Proposed) I+XA+D (Proposed) Figure 17: Performance of I+XA under different priors over r0. Solid curves and shaded regions represent the mean and standard deviation of the normalized improvement, computed over 10 trials with different initial conditions. Solid vertical lines indicate the interquartile range. Abbreviations: Stationary (S), Cylindrical (C) and Informative (I) covariances; Quadratic Mean (+QM); Adaptive, greedily-chosen (+XA) anchors; Uniform (+U), Kumaraswamy (+K) and Dirac delta (+D) priors. G Rover Trajectory Planning In this section, we examine the rover trajectory problem proposed by Wang et al. (2018), which has also been adopted by more recent studies (Eriksson et al., 2019; Eriksson & Jankowiak, 2021). In brief, the goal is to optimize a 2D trajectory such that the rover starts as close as possible to xstart and stops near xend, while avoiding collisions with objects. The trajectory is given by a B-spline that is fitted to 30 2-dimensional points, resulting in a 60D optimization problem. The original reward function is defined as f(x) = c(x) + γ ( [x]1,2 xstart 1 + [x]59,60 xend 1) + b, x [0, 1]60, (43) where γ = 10, b = 5 and c(x) is a nonpositive function that penalizes collisions. For minimization, we turn the reward into a loss function. In particular, we take the negative reward and apply a +b shift, yielding a non-negative objective to be minimized. Models are again estimated on log-transformed data. The domain is also adjusted to the centered hypercube X = [ 1, 1]D by applying a linear transformation to the inputs. Figure 18 shows the original map layout with two overlaid trajectories found by S and I+XA (left), as well as the corresponding performance curves (right). In this problem, we immediately observe that the trajectories do not pass through the 2D points. This outcome is due to oversmoothed splines, which explains the performance of S. Intuitively, given the map layout, we would expect points near the boundary to yield penalized trajectories, but this is only true if the trajectories pass through the specified points. Instead, points near the boundary have a significant effect on the oversmoothed trajectories. The overexploration of boundaries that is characteristic of S in higher-dimensional problems confers an advantage in this situation. However, we observe that I+XA is eventually able to find trajectories of similar cost. In order to align the problem formulation with our expectations, we modify the procedure that fits B-splines to the collection of 2D points. In terms of implementation, the function call that performs interpolation, scipy.interpolate.splprep (Virtanen et al., 2020), is extended with a smoothing factor s that is set to 0, which forces the trajectories to pass through all points. An example is shown in Figure 19 (left). Additionally, we clamp all trajectories into the admissible range and increase the cost for deviating from the endpoints, γ = 50. The latter promotes unfolded trajectories away from the initial incumbent solution x(n0) best = 0, as the relative cost of collisions is too steep otherwise. Performance curves for all rover tests are shown in Figure 20. Notably, we observe that S is no longer among the best-performing methods because the specification of points near the boundary correctly yields penalized trajectories. S (cost=4.898) I+XA (cost=3.580) 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement Oversmoothed Rover T,60D S S+QM S+TR C GS+XA CMA-ES I+X0 (Proposed) I+XA (Proposed) I+XA+TR (Proposed) I+XA+QM (Proposed) I+XA+F (Proposed) I+XA+F+TR (Proposed) Figure 18: Original map layout with overlaid rover trajectories (left). Trajectories are given by a B-spline that is fitted to the 30 2-dimensional points. Performance on the original rover trajectory problem (right). Solid curves and shaded regions represent the mean and standard deviation of the normalized improvement, computed over 10 trials with different initial conditions. Solid vertical lines indicate the interquartile range. Efficient Trajectory Shifted Goal Rover T Figure 19: Original map layout with an overlaid rover trajectory (left). In Rover T, trajectories must pass through the specified collection of 2D points. The handcrafted trajectory demonstrates that it is not necessary to specify 30 different points. However, the reward function does not penalize less efficient trajectories, which leads to an infinite number of equally-good trajectories. As an additional test, we consider the same map layout with a shifted endpoint xend (right). 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement Oversmoothed Rover T,60D 0 20 40 60 80 100 120 140 160 180 200 Rover T,60D 0 20 40 60 80 100 120 140 160 180 200 Shifted Goal Rover T,60D S S+QM S+TR C GS+XA CMA-ES I+X0 (Proposed) I+XA (Proposed) I+XA+TR (Proposed) I+XA+QM (Proposed) I+XA+F (Proposed) I+XA+F+TR (Proposed) Figure 20: Performance on the rover trajectory problem. Solid curves and shaded regions represent the mean and standard deviation of the normalized improvement, computed over 10 trials with different initial conditions. Solid vertical lines indicate the interquartile range. As shown in Figure 19 (left), an optimal trajectory does not require the specification of 30 different points. However, the reward function does not take into account the distance traveled, which means that the rover is free to roam the environment, provided that it starts at xstart and ends at xend without colliding with other objects. For this reason, trajectories of similar cost can be significantly different and the choice of GP priors becomes even more important. The importance of GP priors is further demonstrated in Figure 21, which shows an example of the best Rover T solutions found by BO. In particular, the stationary covariance in S (top-left) does not promote exploration around any particular region. In this case, the best trajectories tend to cover the entire space, while trying to avoid collisions. In constrast, the quadratic mean prior (top-middle) and the cylindrical covariance function (top-right) promote exploration near the center. As quadratic weights can be nearly zero, some points along the trajectory can also be near the boundary. In both cases, the prior dominates to such an extent that the endpoints of the best trajectories are comparatively far from the target endpoints. Notably, I+X0 (bottom-left) also encourages exploration around the center, but to a lesser degree than S+QM and C, leading to better solutions. Finally, I+XA (bottom-middle) and I+XA+F (bottom-right) find the most cost-effective and efficient trajectories because both promote exploration around the incumbent solution, emphasizing the search for fine-tuned trajectories. S (acq=0, cost=90.000) S (acq=50, cost=66.962) S (acq=100, cost=50.592) S (acq=150, cost=43.883) S (acq=200, cost=38.560) S+QM (acq=0, cost=90.000) S+QM (acq=50, cost=90.000) S+QM (acq=100, cost=48.690) S+QM (acq=150, cost=41.183) S+QM (acq=200, cost=39.241) C (acq=0, cost=90.000) C (acq=50, cost=88.912) C (acq=100, cost=88.485) C (acq=150, cost=88.316) C (acq=200, cost=84.644) I+X0 (acq=0, cost=90.000) I+X0 (acq=50, cost=60.298) I+X0 (acq=100, cost=51.344) I+X0 (acq=150, cost=31.491) I+X0 (acq=200, cost=31.491) I+XA (acq=0, cost=90.000) I+XA (acq=50, cost=52.781) I+XA (acq=100, cost=39.914) I+XA (acq=150, cost=27.089) I+XA (acq=200, cost=26.934) I+XA+F (acq=0, cost=90.000) I+XA+F (acq=50, cost=78.327) I+XA+F (acq=100, cost=29.854) I+XA+F (acq=150, cost=27.741) I+XA+F (acq=200, cost=26.915) Figure 21: Example of the best rover trajectories found by BO using different GP priors. Top row: baselines S, S+QM and C. Bottom row: proposed methods I+X0, I+XA and I+XA+F. H Sparse Axis-Aligned Subspaces (SAAS) Hyperprior The Sparse Axis-Aligned Subspaces (SAAS) hyperprior, introduced by Eriksson & Jankowiak (2021), assumes that only a subset of dimensions affects the objective function. This sparsity-inducing hyperprior posits that the inverse squared lengthscales are distributed according to a half-Cauchy, 1/λ2 d HC(τ), where τ is a global shrinkage hyperparameter. At each step, multiple GP models are trained by empirical Bayes with τ {1, 10 1, 10 2, 10 3}. The GP model with the highest leave-one-out cross-validation likelihood is then used for acquisition. We evaluated a modified version of S, which incorporates SAAS and follows the original training procedure described above. As shown in Figures 22 and 23, S+SAAS generally performs worse than S, indicating that this hyperprior may adversely affect performance when the low effective dimensionality assumption does not hold. To further demonstrate the complementary nature of our proposed methodology, we tested I+XA+SAAS. This combination proved to be beneficial on QBranin, Rosenbrock and Levy. Recall that, in I+XA, the lengthscales also govern the nonstationary effects, and incorporating SAAS allows for longer lengthscales (c.f. Table 2). On the Styblinski-Tang function, however, this hyperprior continued to negatively impact performance, renewing the question of its suitability when the objective does not exhibit low effective dimensionality. 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement QBranin,20D 0 20 40 60 80 100 120 140 160 180 200 QBranin,50D 0 20 40 60 80 100 120 140 160 180 200 QBranin,100D S S+QM S+SAAS S+TR C I+X0 (Proposed) I+XA (Proposed) I+XA+SAAS (Proposed) I+XA+TR (Proposed) 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement Rosenbrock,20D 0 20 40 60 80 100 120 140 160 180 200 Rosenbrock,50D 0 20 40 60 80 100 120 140 160 180 200 Rosenbrock,100D S S+QM S+SAAS S+TR C I+X0 (Proposed) I+XA (Proposed) I+XA+SAAS (Proposed) I+XA+TR (Proposed) 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement 0 20 40 60 80 100 120 140 160 180 200 0 20 40 60 80 100 120 140 160 180 200 S S+QM S+SAAS S+TR C I+X0 (Proposed) I+XA (Proposed) I+XA+SAAS (Proposed) I+XA+TR (Proposed) Figure 22: Performance on test functions, ranging from 20 to 100 dimensions. Solid curves and shaded regions represent the mean and standard deviation of the normalized improvement, computed over 10 trials with different initial conditions. Solid vertical lines indicate the interquartile range. Abbreviations: Stationary (S), Cylindrical (C) and Informative (I) covariances; Quadratic Mean (+QM); Origin (+X0) and Adaptive (+XA) anchors; acquisitions within Trust Region (+TR); Sparse Axis-Aligned Subspace (+SAAS) hyperprior. 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement Styblinski-Tang,20D 0 20 40 60 80 100 120 140 160 180 200 Styblinski-Tang,50D 0 20 40 60 80 100 120 140 160 180 200 Styblinski-Tang,100D S S+QM S+SAAS S+TR C I+X0 (Proposed) I+XA (Proposed) I+XA+SAAS (Proposed) I+XA+TR (Proposed) Figure 23: Performance on the Styblinski-Tang function with 20, 50 and 100 dimensions. Solid curves and shaded regions represent the mean and standard deviation of the normalized improvement, computed over 10 trials with different initial conditions. Solid vertical lines indicate the interquartile range. Abbreviations: Stationary (S), Cylindrical (C) and Informative (I) covariances; Quadratic Mean (+QM); Origin (+X0) and Adaptive (+XA) anchors; acquisitions within Trust Region (+TR); Sparse Axis-Aligned Subspace (+SAAS) hyperprior. I Belief-Augmented Acquisition Functions In order to incorporate beliefs about optimal solutions into BO, Hvarfner et al. (2022) proposed priorweighted acquisition functions, which are obtained by multiplying an acquisition function α with a prior distribution over possible locations of the optimum. As optimization progresses, the influence of the fixed prior π(x) decays according to απ,n(x) α(x | Dn, M)π(x)ζ/n, where the hyperparameter ζ controls the decay. As more acquisitions are made, the prior approaches a uniform distribution, only tilting the acquisition step initially, when the exponent ζ/n is not small. By default, Hvarfner et al. (2022) used Gaussian-Weighted EI (GWEI) and ζ = N/10, where N is the evaluation budget. The default Gaussian prior is characterized by a diagonal covariance matrix, with standard deviation set to 25% of the domain. We evaluated two additional baselines, S+GWEIX0 and S+GWEIXA, which use Gaussian-Weighted EI with fixed and adaptive locations. However, the results on high-dimensional test functions indicate that these baselines perform poorly when compared to S, as shown in Figures 24 and 25. We found that poor performance was also linked to machine precision because EI, whose values are already small, is multiplied by small factors, resulting in an acquisition function that is difficult to optimize in high dimensions. In order to improve performance, we removed the normalizing constant in the Gaussian prior, obtaining a Gaussian Kernelweighted EI (GKEI). The method I+XA+GKEI shows that combining informative covariance functions and belief-augmented acquisition functions can be an effective strategy. 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement QBranin,20D 0 20 40 60 80 100 120 140 160 180 200 QBranin,50D 0 20 40 60 80 100 120 140 160 180 200 QBranin,100D S S+QM S+GWEIX0 S+GWEIXA S+TR C I+X0 (Proposed) I+XA (Proposed) I+XA+GKEIX0 (Proposed) I+XA+GKEIXA (Proposed) I+XA+TR (Proposed) 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement Rosenbrock,20D 0 20 40 60 80 100 120 140 160 180 200 Rosenbrock,50D 0 20 40 60 80 100 120 140 160 180 200 Rosenbrock,100D S S+QM S+GWEIX0 S+GWEIXA S+TR C I+X0 (Proposed) I+XA (Proposed) I+XA+GKEIX0 (Proposed) I+XA+GKEIXA (Proposed) I+XA+TR (Proposed) 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement 0 20 40 60 80 100 120 140 160 180 200 0 20 40 60 80 100 120 140 160 180 200 S S+QM S+GWEIX0 S+GWEIXA S+TR C I+X0 (Proposed) I+XA (Proposed) I+XA+GKEIX0 (Proposed) I+XA+GKEIXA (Proposed) I+XA+TR (Proposed) Figure 24: Performance on test functions, ranging from 20 to 100 dimensions. Solid curves and shaded regions represent the mean and standard deviation of the normalized improvement, computed over 10 trials with different initial conditions. Solid vertical lines indicate the interquartile range. Abbreviations: Stationary (S), Cylindrical (C) and Informative (I) covariances; Quadratic Mean (+QM); Origin (+X0) and Adaptive (+XA) anchors; acquisitions within Trust Region (+TR); Gaussian-Weighted Expected Improvement with fixed (+GWEIX0) and adaptive (+GWEIXA) location; Gaussian Kernel-weighted Expected Improvement with fixed (+GKEIX0) and adaptive (+GKEIXA) location. 0 20 40 60 80 100 120 140 160 180 200 0.0 Normalized Improvement Styblinski-Tang,20D 0 20 40 60 80 100 120 140 160 180 200 Styblinski-Tang,50D 0 20 40 60 80 100 120 140 160 180 200 Styblinski-Tang,100D S S+QM S+GWEIX0 S+GWEIXA S+TR C I+X0 (Proposed) I+XA (Proposed) I+XA+GKEIX0 (Proposed) I+XA+GKEIXA (Proposed) I+XA+TR (Proposed) Figure 25: Performance on the Styblinski-Tang function with 20, 50 and 100 dimensions. Solid curves and shaded regions represent the mean and standard deviation of the normalized improvement, computed over 10 trials with different initial conditions. Solid vertical lines indicate the interquartile range. Abbreviations: Stationary (S), Cylindrical (C) and Informative (I) covariances; Quadratic Mean (+QM); Origin (+X0) and Adaptive (+XA) anchors; acquisitions within Trust Region (+TR); Gaussian-Weighted Expected Improvement with fixed (+GWEIX0) and adaptive (+GWEIXA) location; Gaussian Kernel-weighted Expected Improvement with fixed (+GKEIX0) and adaptive (+GKEIXA) location. J High-dimensional Experiments with Larger Evaluation Budgets Thus far, we have considered a budget of 200 acquisitions. As highlighted in Sections 2.1 and 3, the rationale is that BO is competitive when function evaluations are expensive, putting a practical constraint on the size of the budget, even for high-dimensional problems.5 Additionally, we have favored an analysis in terms of sample efficiency, demonstrating that the proposed methodology, which employs informative covariance functions, offers methods that discover superior solutions with fewer function evaluations. Despite the rationale above, we now increase the evaluation budget to 600 acquisitions. Figure 26 shows the results for two test functions that we consider representative (see Appendix B.3). The Rosenbrock function features a banana-shaped valley, with the global optimum relatively close to the center of the search space. Conversely, the Styblinski-Tang is a multimodal function, whose optimal solution is relatively distant from the center (local maximum). Once again, we observe that the proposed covariance functions can be combined with other methods, including quadratic mean functions and trust regions, to achieve higher sample efficiency and to discover superior solutions. As a final note, while our methodology provides scalable methods in terms of dimensionality (e.g., I+X0, I+XA), these still share some of the limitations of standard GPs. In particular, when dealing with large training sets, standard GPs become impractical because they require O(n3) computation and O(n2) memory, where n is the number of training points. To address this limitation, one option is to use sparse GPs with inducing variables, as proposed by e.g. Titsias (2009) and Hensman et al. (2013). This allows for a significant reduction in computational and memory requirements due to low-rank matrix approximations. 0 50 100 150 200 250 300 350 400 450 500 550 600 0.0 Normalized Improvement Rosenbrock,100D 0 50 100 150 200 250 300 350 400 450 500 550 600 Styblinski-Tang,100D S S+QM S+TR C I+X0 (Proposed) I+XA (Proposed) I+XA+QM (Proposed) I+XA+TR (Proposed) Figure 26: Performance on the 100-dimensional Rosenbrock and Styblinski-Tang functions. A value of 1 indicates that the global optimum has been found (zero simple regret). Solid curves and shaded regions represent the mean and standard deviation of the normalized improvement, computed over 10 trials with different initial conditions. Solid vertical lines indicate the interquartile range. Abbreviations: Stationary (S), Cylindrical (C) and Informative (I) covariances; Quadratic Mean (+QM); Origin (+X0) and Adaptive (+XA) anchors; acquisitions within Trust Region (+TR). 5Based on a similar reasoning, Eriksson & Jankowiak (2021) conducted 100-dimensional experiments with a maximum budget of 100 acquisitions.