# bayesian_optimization_using_pseudopoints__c1014bc9.pdf Bayesian Optimization using Pseudo-Points Chao Qian1 , Hang Xiong2 and Ke Xue1 1National Key Laboratory for Novel Software Technology, Nanjing University, Nanjing 210023, China 2University of Science and Technology of China, Hefei 230027, China {qianc, xuek}@lamda.nju.edu.cn, flybear@mail.ustc.edu.cn Bayesian optimization (BO) is a popular approach for expensive black-box optimization, with applications including parameter tuning, experimental design, and robotics. BO usually models the objective function by a Gaussian process (GP), and iteratively samples the next data point by maximizing an acquisition function. In this paper, we propose a new general framework for BO by generating pseudo-points (i.e., data points whose objective values are not evaluated) to improve the GP model. With the classic acquisition function, i.e., upper confidence bound (UCB), we prove that the cumulative regret can be generally upper bounded. Experiments using UCB and other acquisition functions, i.e., probability of improvement (PI) and expectation of improvement (EI), on synthetic as well as real-world problems clearly show the advantage of generating pseudo-points. 1 Introduction One often needs to solve an optimization problem: x arg maxx X f(x), where X Rd is the solution space, f : X R is the objective function, and x is an optimal solution. Usually, it is assumed that f has a known mathematical expression, is convex, or cheap to evaluate at least. Increasing evidences, however, show that f may not satisfy these assumptions, but is an expensive black-box model [Brochu et al., 2010]. That is, f can be non-convex, or even the closedform expression of f is unknown; meanwhile, evaluating f can be noisy and computationally very expensive. Expensive black-box optimization is involved in many real-world decision making problems. For example, in machine learning, one has to tune hyper-parameters to maximize the performance of a learning algorithm [Snoek et al., 2012]; in physical experiments, one needs to set proper parameters of the experimental environment to obtain an ideal product [Brochu et al., 2010]. More applications can been This work was supported by the Fundamental Research Funds for the Central Universities (14380004) and the project of HUAWEILAMDA Joint Laboratory of Artificial Intelligence. found in robotic control [Martinez-Cantin et al., 2007], computer vision [Denil et al., 2012], sensor placing [Garnett et al., 2010], and analog circuit design [Lyu et al., 2018]. BO [Mockus, 1994] has been a type of powerful algorithm to solve expensive black-box optimization problems. The main idea is to build a model, usually by a GP, for the objective f based on the observation data, and then sample the next data point by maximizing an acquisition function. Many BO algorithms have been proposed, with the goal of reaching the optima using as few objective evaluations as possible. Most existing works focus on designing effective acquisition functions, e.g., PI [Kushner, 1964], EI [Jones et al., 1998], and UCB [Srinivas et al., 2012]. Recently, Wang et al. [2016] proposed the EST function by directly estimating x , which automatically and adaptively trades off exploration and exploitation in PI and UCB. Another major type of acquisition function is based on information entropy, including entropy search (ES) [Hennig and Schuler, 2012], predictive ES [Hern andez-Lobato et al., 2014], max-value ES [Wang and Jegelka, 2017], FITBO [Ru et al., 2018], etc. As BO is a sequential algorithm, some parallelization techniques have been introduced for acceleration, e.g., [Azimi et al., 2010; Desautels et al., 2014; Shah and Ghahramani, 2015; Gonz alez et al., 2016]. There is also a sequence of works addressing the difficulty of BO for high-dimensional optimization, e.g., [Wang et al., 2013; Kandasamy et al., 2015; Wang et al., 2017; Mutny and Krause, 2018]. For any BO algorithm with a specific acquisition function, the GP model becomes increasingly accurate with the observation data augmenting. However, the number of data points to be evaluated is often limited due to the expensive objective evaluation. In this paper, we propose a general framework for BO by generating pseudo-points to improve the GP model. That is, before maximizing the acquisition function to select the next point in each iteration, some pseudo-points are generated and added to update the GP model. The pseudo-points are neighbors of the observed data points, and take the same function values as the observed ones. Without increasing the evaluation cost, the generation of pseudo-points can reduce the variance of the GP model, while introducing little accuracy loss under the Lipschitz assumption. This framework is briefly called BO-PP. Theoretically, we study the performance of BO-PP w.r.t. the acquisition function UCB, called UCB-PP. We prove a Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence (IJCAI-20) Algorithm 1 BO Framework Input: iteration budget T Process: 1: let D0 = ; 2: for t = 1 : T do 3: xt = arg maxx X acq(x); 4: evaluate f at xt to obtain yt; 5: augment the data Dt = Dt 1 {(xt, yt)} and update the GP model 6: end for general upper bound of UCB-PP on the cumulative regret, i.e., PT t=1(f(x ) f(xt)), where xt denotes the sampled point in the t-th iteration. It is shown to be a generalization of the known bound [Srinivas et al., 2012] of UCB. Empirically, we compare BO-PP with BO on synthetic benchmark functions as well as real-world optimization problems. The acquisition functions UCB, PI and EI are used. The results clearly show the superior performance of BO-PP. 2 Background The general framework of BO is shown in Algorithm 1. It sequentially optimizes a given objective function f(x) with assumptions on a prior distribution, i.e., a probabilistic model, over f(x). In each iteration, BO selects a point x by maximizing an acquisition function acq( ), evaluates its objective value f(x), and updates the prior distribution with the new data point. 2.1 GPs A GP [Rasmussen and Williams, 2006] is commonly used as the prior distribution, which regards the f value at each data point as a random variable, and assumes that all of them satisfy a joint Gaussian distribution specified by the mean value function m( ) and the covariance function k( , ). For convenience, m( ) is set to zero. Assume that the objective evaluation is subject to i.i.d. additive Gaussian noise, i.e., y = f(x) + ϵ, where ϵ N(0, σ2). Let [t] denote the set {1, 2, . . . , t}. Given an observation data Dt = {(xi, yi)}t i=1, we can obtain the posterior mean µt(x) = kt(x)T(Kt + σ2I) 1y1:t, (1) and the posterior variance σ2 t (x) = k(x, x) kt(x)T(Kt + σ2I) 1kt(x), (2) where kt(x) = [k(xi, x)]t i=1, Kt = [k(xi, xj)]i,j [t] and y1:t = [y1; y2; . . . ; yt]. For a GP, the log likelihood of observed data Dt is log Pr(y1:t | {xi}t i=1, θ) = (1/2)y T 1:t(Kt + σ2I) 1y1:t (1/2) log det(Kt + σ2I) (t/2) log 2π, where θ denote the hyper-parameters of k( , ), and det( ) denotes the determinant of a matrix. When updating the GP model in line 5 of Algorithm 1, the hyper-parameters θ can be updated by maximizing the log likelihood of the augmented data, or treated to be fully Bayesian. 2.2 Acquisition Functions The data point to be evaluated in each iteration is selected by maximizing an acquisition function, which needs to trade off exploration, i.e., large posterior variances, and exploitation, i.e., large posterior means. Many acquisition functions have been proposed, and we introduce three typical ones, i.e., PI [Kushner, 1964], EI [Jones et al., 1998] and UCB [Srinivas et al., 2012], which will be examined in this paper. Let x+ be the best point generated in the first (t 1) iterations, and Z = (µt 1(x) f(x+))/σt 1(x). Let Φ and φ denote the cumulative distribution and probability density functions of standard Gaussian distribution, respectively. PI selects the point by maximizing the probability of improvement, i.e., PI(x) = Pr(f(x) > f(x+)) = Φ(Z). (3) EI selects the data point by maximizing the expectation of improvement, i.e., (µt 1(x) f(x+))Φ(Z) + σt 1(x)φ(Z) if σt 1(x) > 0, 0 if σt 1(x) = 0. (4) UCB integrates the posterior mean and variance via a tradeoff parameter βt, i.e., UCB(x) = µt 1(x) + β1/2 t σt 1(x), (5) and selects the data point by maximizing this measure. 2.3 Regrets To evaluate the performance of BO algorithms, regrets are often used. The instantaneous regret rt = f(x ) f(xt) measures the gap of function values between an optimal solution x and the currently selected point xt. The simple regret ST = mini [T ] ri measures the gap between x and the best point found in the first T iterations. The cumulative regret RT = PT i=1 ri is the sum of instantaneous regrets in the first T iterations. A BO algorithm is said to be no-regret if lim T + RT /T = 0. 3 The BO-PP Framework In BO, a GP is used to characterize the unknown objective function. The posterior variance of a GP describes the uncertainty about the unknown objective, while the posterior mean provides a closed form of the unknown objective. As the observation data augments, the posterior variance decreases and the posterior mean gets close to the unknown objective, making the GP express the unknown objective better. Thus, a straightforward way to improve the GP model is collecting more data points, which is, however, impractical, because the objective evaluation is expensive. In this section, we propose a general framework BO-PP by generating pseudo-points to improve the GP model. As shown in Eq. (2), the posterior variance of f does not depend on the objective values, and will be decreased by adding new data points. As shown in Eq. (1), the posterior mean of f can be regarded as a linear combination of the observed objective values, and will be influenced by the error on Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence (IJCAI-20) Algorithm 2 BO-PP Framework Input: iteration budget T Parameter: {li}T 1 i=0 , {τi}T 1 i=0 Process: 1: let D0 = and l0 = 0; 2: for t = 1 : T do 3: generate lt 1 pseudo-points {(x i, ˆy i)}lt 1 i=1 ; 4: re-compute ˆµt 1 and ˆσt 1 by Dt 1 {(x i, ˆy i)}lt 1 i=1 ; 5: xt = arg maxx X acq(x); 6: evaluate f at xt to obtain yt; 7: augment the data Dt = Dt 1 {(xt, yt)} and update the GP model 8: end for where each pseudo-point in the t-th iteration has distance τt 1 to some observed data point in Dt 1, and takes the same objective value as the observed one. the objective values of new data points. Inspired by the Lipschitz assumption, i.e., close data points have close objective values, the pseudo-points are selected to be neighbors of the observed data points, and take the same objective values as the observed ones. The BO-PP framework is described in Algorithm 2. Before selecting the next data point in line 5, BO-PP generates a few pseudo-points to re-compute the posterior mean and variance of the GP model in lines 3-4, rather than directly using the GP model updated in the last iteration. After evaluating a new data point in line 6, the hyper-parameters of the covariance function employed by the GP model will be updated in line 7 using the truly observed data points by far. Note that the pseudo-points are only used to re-compute the posterior mean and variance. The way of generating pseudo-points can be diverse, e.g., randomly sampling a point with distance τ from some observed data point. The only requirement is that the pseudopoint takes the same objective value as the corresponding observed data point, which does not increase the evaluation cost. The number lt of pseudo-points and the distance τt employed in each iteration could affect the performance of the algorithm. For example, as τt decreases, the error on the objective values of pseudo-points will decrease, whereas the reduction on the posterior variance will also decrease. Their relationship will be analyzed in theoretical analysis, and we will provide an effective way of setting lt and τt in experiments. Note that BO-PP can be equipped with any acquisition function. 4 Theoretical Analysis In this section, we theoretically analyze the performance of BO-PP w.r.t. the acquisition function UCB, called UCB-PP. Specifically, we prove that the cumulative regret RT of UCBPP can be generally upper bounded. We first give some notations that will be used in the following analysis. Let µt and σt denote the posterior Here, two data points x, x Rd have distance τ means that i [d] : |xi x i| = τ. mean and variance after obtaining Dt; let ˆµt and ˆσt denote the posterior mean and variance after adding pseudopoints {(x i, ˆy i)}lt i=1 into Dt; let µt and σt denote the posterior mean and variance after adding pseudo-points with true observed objective values, i.e., {(x i, y i))}lt i=1, where y i = f(x i)+ϵ i with ϵ i N(0, σ2). Some notations about pseudopoints: ˆy 1:lt = [ˆy 1; ˆy 2; . . . ; ˆy lt]; y 1:lt = [y 1; y 2; . . . ; y lt]; k lt(x) = [k(x i, x)]lt i=1; K lt = [k(x i, x j)]i,j [lt]; e Kt,lt = [k(xi, x j)]i [t],j [lt]; p(x) = e KT t,lt(Kt + σ2I) 1kt(x) k lt(x); M = (K lt e KT t,lt(Kt +σ2I) 1 e Kt,lt +σ2I) 1. For convenience of analysis, assume k(x, x) = 1. Let A be a finite subset of X, f A denote their true objective values (which are actually random variables satisfying the posterior Gaussian distribution over the true objective values), and y A denote the noisy observations. Let PP denote all generated pseudo-points, and ˆy P P denote their selected objective values. Note that ˆy P P are random variables, as they are actually the noisy observations of the objective values of PP s neighbor observed points. Let γ T = max A:|A|=T I(y A; f A) min A:|A|=T,P P I(y A; ˆy P P ), where I( ; ) denotes the mutual information. Theorem 1 gives an upper bound of UCB-PP on the cumulative regret RT . As the analysis of UCB in [Srinivas et al., 2012], Assumption 1 is required, implying Pr( x, x : |f(x) f(x )| L x x 1) 1 dae (L/b)2. (6) Assumption 1. Suppose the kernel k( , ) satisfies the following probability bound on the derivatives of f: for some constants a, b > 0, j [d] : Pr(supx X | f/ xj| > L) ae (L/b)2. Theorem 1. Let X [0, r]d, δ (0, 1), and set βt in Eq. (5) as βt= 2 log(2π2t2/(3δ)) + 2d log(t2dbr p log(4da/δ)). Running UCB-PP for T iterations, it holds that CTβT γ T + 2 + 2PT t=1 m(lt 1, τt 1) where C = 8/ log(1 + σ 2), and m(lt, τt) = 1 + σ 2 bdτt p log(4da/δ)/σ + 2 q log 4 PT 1 t=0 lt Lemma 1 bounds the error on the posterior mean led by the incorrect objective values of pseudo-points, which will be used in the proof of Theorem 1. Its proof is provided in the supplementary material due to space limitation. Lemma 1. After obtaining Dt in UCB-PP, the difference on the posterior mean by adding pseudo-points, i.e., {(x i, ˆy i)}lt i=1, and that with true observed objective values, i.e., {(x i, y i))}lt i=1, is ˆµt(x) µt(x) = p(x)TM(ˆy 1:lt y 1:lt). Furthermore, it holds that Pr 0 t T 1, x X : |ˆµt(x) µt(x)| m(L, lt, τt) 1 dae (L/b)2 δ/4, where m(L, lt, τt)=l2 t log 4 PT 1 t=0 lt Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence (IJCAI-20) The proof of Theorem 1 is inspired by that of Theorem 2 in [Srinivas et al., 2012], which gives an upper bound of UCB on the cumulative regret RT . Their proof intuition is mainly that the instantaneous regret rt can be upper bounded by the width of confidence interval of f(xt), relating to the posterior variance. The generation of pseudo-points will introduce another quantity into the upper bound on rt, characterized by the error on the posterior mean in Lemma 1. Proof of Theorem 1. According to Assumption 1 and βt = 2 log(2π2t2(dt2r L)d/(3δ)), where L = b p log(4da/δ), we can apply Lemma 5.7 in [Srinivas et al., 2012] to derive that Pr t 1 : |f(x ) µt 1([x ]t)| β1/2 t σt 1([x ]t) + 1/t2 1 δ/2, (8) where [x ]t denotes the discretized data point closest to x in the t-th iteration. Note that m(lt, τt) is just m(L, lt, τt) with L = b p log(4da/δ) in Lemma 1. By the definition of rt, we have, t 1: rt = f(x ) f(xt) β1/2 t σt 1([x ]t) + µt 1([x ]t) f(xt) + 1/t2 β1/2 t σt 1([x ]t) + ˆµt 1([x ]t) f(xt) + 1/t2 + m(lt 1, τt 1) = β1/2 t ˆσt 1([x ]t) + ˆµt 1([x ]t) f(xt) + 1/t2 + m(lt 1, τt 1) β1/2 t ˆσt 1(xt) + ˆµt 1(xt) f(xt) + 1/t2 + m(lt 1, τt 1) β1/2 t ˆσt 1(xt) + µt 1(xt) f(xt) + 1/t2 + 2 m(lt 1, τt 1) 2β1/2 t ˆσt 1(xt) + 1/t2 + 2 m(lt 1, τt 1), where the first inequality holds with probability at least 1 δ/2 by Eq. (8), the second and fourth inequalities hold with probability at least 1 dae (L/b)2 δ/4 = 1 δ/2 by Lemma 1, the equality holds because the posterior variance in Eq. (2) does not depend on the objective values, leading to x : ˆσt 1(x) = σt 1(x), the third inequality holds because xt is selected by maximizing ˆµt 1(x) + β1/2 t ˆσt 1(x) in Eq. (5), and the last inequality holds with probability at least 1 δ/4 by Lemma 5.5 in [Srinivas et al., 2012]. Note that to prove Lemma 5.7 in [Srinivas et al., 2012] and Lemma 1, Assumption 1, i.e., Eq. (6), is both used; thus, the probability dae (L/b)2 = δ/4 has been repeated. By the union bound, we have Pr t 1 : rt 2β1/2 t ˆσt 1(xt)+1/t2+2 m(lt 1, τt 1) 1 δ/2 δ/4 δ/4 = 1 δ, Pr RT = PT t=1rt PT t=1 2β1/2 t ˆσt 1(xt) + 1/t2 + 2 m(lt 1, τt 1) 1 δ. By the Cauchy Schwarz inequality, C = 8/ log(1 + σ 2) and t T : βt βT , we have PT t=12β1/2 t ˆσt 1(xt) q TPT t=14βtˆσ2 t 1(xt) 2 PT t=1 log(1 + σ 2ˆσ2 t 1(xt)). Let PPt denote the pseudo-points generated in the t-th iteration, and ˆy P Pt denote their selected objective values. Let H( ) denote the entropy. We have 1 2 PT t=1 log(1 + σ 2ˆσ2 t 1(xt)) + H(y1:T |f1:T ) 2 PT t=1 log(1 + σ 2ˆσ2 t 1(xt)) + 1 2 log(det(2πeσ2I)) 2 PT t=1 log(2πe(σ2 + ˆσ2 t 1(xt))) = H(y1 | ˆy P P1) + H(y2 | y1, ˆy P P2) + + H(y T | y1:T 1, ˆy P PT ) = H(y1 | ˆy P P ) + H(y2 | y1, ˆy P P ) + + H(y T | y1:T 1, ˆy P P ) = H(y1:T | ˆy P P ), where the first equality holds because f(x) is subject to additive Gaussian noise N(0, σ2). Thus, 1 2 PT t=1 log(1 + σ 2ˆσ2 t 1(xt)) = H(y1:T | ˆy P P ) H(y1:T | f1:T ) = H(ˆy P P | y1:T ) H(ˆy P P ) + H(y1:T ) H(y1:T | f1:T ) = I(y1:T ; f1:T ) I(y1:T ; ˆy P P ) γ T . Considering P t 1 1/t2 = π2/6 < 2, Eq. (7) holds. Thus, the theorem holds. Adding pseudo-points is to improve the GP model when the number of observed data points is not large. After UCBPP runs many iterations, there are already enough observed points, and thus pseudo-points are not needed. That is, UCBPP will only add pseudo-points in a finite number of iterations, denoted by T0. This implies that t T0, lt = 0, leading to m(lt, τt) = 0. Thus, lim T + RT /T = 0, implying that UCB-PP is no-regret. Under the same assumption, it has been proved [Srinivas et al., 2012] that the cumulative regret RT of UCB satisfies CTβT γT + 2 1 δ, (9) where γT = max A:|A|=T I(y A; f A), and the other parameters have the same meaning as that in Theorem 1. Without generating pseudo-points, t 0 : lt = 0 I(y A; ˆy P P ) = 0, and thus, γ T = γT m(lt, τt) = 0, implying that Eq. (7) specializes to Eq. (9). Thus, we have: Remark 1. Our bound on RT of UCB-PP is a generalization of the bound on RT of UCB in [Srinivas et al., 2012]. As γ T γT , the comparison between Eqs. (7) and (9) suggests that the generation of pseudo-points can be helpful if the negative influence of introducing the error on the posterior mean, i.e., introducing the term m(lt, τt), can be compensated by the positive influence of reducing the posterior variance, i.e., introducing the term I(y1:T ; ˆy P P ). Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence (IJCAI-20) Function UCB UCB-PP01 UCB-PP001 UCB-PP0001 Dropwave 0.2710 0.1311 0.2232 0.1053 0.1630 0.1014 0.2121 0.1038 Griewank 0.2357 0.2125 0.2272 0.1644 0.2350 0.1690 0.2085 0.1177 Hart6 1.0256 0.3498 1.0565 0.3620 1.0868 0.3153 0.9276 0.3307 Rastrigin 3.3492 3.2602 3.6975 2.7991 3.5124 2.4124 3.0077 2.3245 SVM wine 0.6182 0.0029 0.6186 0.0030 0.6186 0.0036 0.6189 0.0042 NN wine 0.9149 0.0004 0.9151 0.0005 0.9151 0.0004 0.9151 0.0004 NN cancer 0.9585 0.0006 0.9589 0.0006 0.9589 0.0006 0.9590 0.0006 NN housing 8.6733 1.6916 8.6776 1.7149 9.0691 1.8656 8.7216 1.8076 Function PI PI-PP01 PI-PP001 PI-PP0001 Dropwave 0.1526 0.1534 0.1221 0.1462 0.1251 0.1355 0.1457 0.1539 Griewank 0 0 0 0 0 0 0 0 Hart6 0.5795 0.2959 0.4558 0.1048 0.5599 0.0982 0.5500 0.2529 Rastrigin 0.0524 0.2285 0.0524 0.2285 0 0 0.0524 0.2285 SVM wine 0.6192 0.0037 0.6176 0.0025 0.6204 0.0050 0.6208 0.0053 NN wine 0.9140 0.0011 0.9143 0.0006 0.9140 0.0008 0.9138 0.0008 NN cancer 0.9571 0.0024 0.9578 0.0020 0.9576 0.0024 0.9574 0.0024 NN housing 7.5570 1.4822 7.9702 1.4000 7.8585 1.5515 7.6147 1.3592 Function EI EI-PP01 EI-PP001 EI-PP0001 Dropwave 0.2557 0.1720 0.1924 0.0818 0.2307 0.1461 0.2276 0.1752 Griewank 0.3098 0.1722 0.3028 0.1005 0.3187 0.1594 0.2729 0.1471 Hart6 0.6652 0.2685 0.6050 0.2328 0.6028 0.1656 0.6828 0.3081 Rastrigin 3.3069 2.4955 2.6602 2.2063 3.0492 1.5602 3.1987 2.3818 SVM wine 0.6189 0.0037 0.6182 0.0035 0.6198 0.0037 0.6179 0.0034 NN wine 0.9149 0.0006 0.9150 0.0005 0.9150 0.0004 0.9148 0.0005 NN cancer 0.9587 0.0006 0.9589 0.0006 0.9588 0.0006 0.9587 0.0006 NN housing 8.1780 1.8382 8.0277 1.3844 8.0471 1.5024 8.1992 1.7971 Table 1: The results (mean std.) of BO-PP and BO on synthetic benchmark functions and real-world optimization problems, when reaching the iteration budget. ST : the smaller, the better; f: the larger, the better. The bolded values denote that BO-PP is no worse than BO. UCB, PI and EI are tested. 5 Empirical Study In this section, we empirically compare BO-PP with BO. Three common acquisition functions, i.e., UCB, PI and EI, are used. The ARD squared exponential kernel is employed, whose hyper-parameters are tuned by maximum likelihood estimation (MLE), and the acquisition function is maximized via the DIRECT algorithm [Jones et al., 1993]. To alleviate the cold start issue, each algorithm starts with five random initial points. To compare BO-PP with BO on each problem, we repeat their running 20 times independently and report the average results; in each running, BO-PP and BO use the same five random initial points. The noise level is set to σ2 = 0.0001, and the iteration budget is set to 100. In the (t + 1)-th iteration of BO-PP, for each point in Dt, one pseudo-point is generated by randomly sampling within its distance τt and taking the same function value; thus, lt = |Dt|. To control the error of objective values with pseudo-points increasing, τt is set to rτ0/(dlt), which decreases with lt. Note that r corresponds to the width of each dimension of the search domain. τ0 is set to a small value. We will use 0.01, 0.001 and 0.0001 to explore its influence, and the corresponding algorithms are denoted as BO-PP01, BO-PP001 and BO-PP0001, respectively. We use four common synthetic benchmark functions: Dropwave, Griewank, Hart6 and Rastrigin, whose dimensions are 2, 2, 6 and 2, respectively. Their search domains are scaled to [ 1, 1]d. As the minima are known, the sim- ple regret ST is used as the metric. We also employ four real-world optimization problems, widely used in BO experiments [Springenberg et al., 2016; Wang and Jegelka, 2017; Ru et al., 2018]. The first is to tune the hyper-parameters, i.e., box constraint C [0.001, 1000] and kernel scale l [0.0001, 1], of SVM for classification on the data set Wine quality (1,599 #inst, 11 #feat). The second is to tune the hyper-parameters of 1-hidden-layer neural network (NN) for this task. The NN is trained by backpropagation, and the hyper-parameters are the number of neurons n [1, 100] and the learning rate lr [0.000001, 1]. The last two problems are to tune the hyper-parameters of 1-hidden-layer NN for classification on Breast cancer (699 #inst, 9 #feat) and regression on Boston housing (506 #inst, 13 #feat), respectively. The NN is trained by Levenberg-Marquardt optimization, and there are four hyper-parameters: n [1, 100], the damping factor µ [0.000001, 100], the µ-decrease and µincrease factors µdec [0.01, 1], µinc [1.01, 20]. All data sets are randomly split into training/validation/test sets with ratio 0.7/0.2/0.1, and the performance on validation sets is used as the objective f. For classification, f is the classification accuracy; for regression, f equals 20 minus the regression L2-loss. For UCB, βt in Eq. (5) is set to 2 log(td/2+2π2/3δ) where δ = 0.1, as suggested in [Brochu et al., 2010; Srinivas et al., 2012]. For PI and EI, the best observed function value by far is used as f(x+) in Eqs. (3) and (4). The results are summarized in Table 1. We can observe that UCB-PP0001 Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence (IJCAI-20) UCB UCB-PP01 UCB-PP001 UCB-PP0001 20 40 60 80 100 Iterations (T) 20 40 60 80 100 Iterations (T) 20 40 60 80 100 Iterations (T) 20 40 60 80 100 Iterations (T) (a) SVM wine (b) NN wine (c) NN cancer (d) NN housing Figure 1: The results (mean (1/4)std.) of UCB-PP and UCB on real-world optimization problems. f: the larger, the better. UCB UCB-PP01 UCB-PP001 UCB-PP0001 20 40 60 80 100 Iterations (T) 20 40 60 80 100 Iterations (T) 20 40 60 80 100 Iterations (T) 20 40 60 80 100 Iterations (T) (a) SVM wine (b) NN wine (c) NN cancer (d) NN housing Figure 2: The results (mean (1/4)std.) of UCB-PP and UCB with the Gaussian kernel on real-world optimization problems. f: the larger, the better. is always better than UCB, and UCB-PP01/UCB-PP001 surpasses UCB in most cases, disclosing that the performance of UCB-PP is not very sensitive to the distance τt. Also, PI-PP and EI-PP perform better than PI and EI, respectively, in most cases, showing the applicability of generating pseudo-points. Furthermore, we plot the curves of the simple regret ST or the objective f over iterations for each algorithm on each problem. Figure 1 shows the curves of UCB-PP and UCB on real-world problems. It can be observed that on each problem, there is at least one curve of UCB-PP almost always above that of UCB, implying that UCB-PP can consistently outperform UCB during the running process. The other five figures, showing similar observations, are provided in the supplementary material due to space limitation. To examine the robustness of BO-PP against kernels, we use the Gaussian kernel with hyper-parameters tuned by MLE. We compare UCB-PP with UCB on real-world optimization problems. Figure 2 shows that UCB-PP can be better than UCB except UCB-PP01/UCB-PP001 on SVM wine. 6 Conclusion In this paper, we propose a general framework BO-PP by generating pseudo-points to improve the GP model of BO. BO-PP can be implemented with any acquisition function. Equipped with UCB, we prove that the cumulative regret of BO-PP can be well bounded. This bound generalizes the well-known bound of UCB. Experiments with UCB, PI and EI on synthetic as well as real-world optimization problems show the superior performance of BO-PP over BO. It is expected that the generation of pseudo-points can be helpful for more BO algorithms. Note that the dimensionality of problems tested in our experiments is low. Thus, studying the effectiveness of BO-PP on high-dimensional optimization problems is an interesting topic. It is also noted that the added pseudo-points take the same function values with their neighbor observed data points, requiring the function to vary smoothly locally. Thus, it is interesting to study strategies of improving BO when the function can fluctuate widely in the future. Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence (IJCAI-20) [Azimi et al., 2010] J. Azimi, A. Fern, and X. Z. Fern. Batch Bayesian optimization via simulation matching. In Advances in Neural Information Processing Systems 23 (NIPS 10), pages 109 117, Vancouver, Canada, 2010. [Brochu et al., 2010] E. Brochu, V. M. Cora, and N. de Freitas. A tutorial on Bayesian optimization of expensive cost functions, with application to active user modeling and hierarchical reinforcement learning. Co RR abs/1012.2599, 2010. [Denil et al., 2012] M. Denil, L. Bazzani, H. Larochelle, and N. de Freitas. Learning where to attend with deep architectures for image tracking. Neural Computation, 24(8):2151 2184, 2012. [Desautels et al., 2014] T. Desautels, A. Krause, and J. W. Burdick. Parallelizing exploration-exploitation tradeoffs in Gaussian process bandit optimization. Journal of Machine Learning Research, 15:3873 3923, 2014. [Garnett et al., 2010] R. Garnett, M. A. Osborne, and S. J. Roberts. Bayesian optimization for sensor set selection. In Proceedings of the 9th ACM/IEEE International Conference on Information Processing in Sensor Networks (IPSN 10), pages 209 219, Stockholm, Sweden, 2010. [Gonz alez et al., 2016] J. Gonz alez, Z. Dai, P. Hennig, and N. D. Lawrence. Batch Bayesian optimization via local penalization. In Proceedings of the 19th International Conference on Artificial Intelligence and Statistics (AISTATS 16), pages 648 657, Cadiz, Spain, 2016. [Hennig and Schuler, 2012] P. Hennig and C. J. Schuler. Entropy search for information-efficient global optimization. Journal of Machine Learning Research, 13:1809 1837, 2012. [Hern andez-Lobato et al., 2014] J. M. Hern andez-Lobato, M. W. Hoffman, and Z. Ghahramani. Predictive entropy search for efficient global optimization of black-box functions. In Advances in Neural Information Processing Systems 27 (NIPS 14), pages 918 926, Montreal, Canada, 2014. [Jones et al., 1993] D. R. Jones, C. D. Perttunen, and B. E. Stuckman. Lipschitzian optimization without the Lipschitz constant. Journal of Optimization Theory and Applications, 79(1):157 181, 1993. [Jones et al., 1998] D. R. Jones, M. Schonlau, and W. J. Welch. Efficient global optimization of expensive black-box functions. Journal of Global Optimization, 13(4):455 492, 1998. [Kandasamy et al., 2015] K. Kandasamy, J. G. Schneider, and B. P oczos. High dimensional Bayesian optimisation and bandits via additive models. In Proceedings of the 32nd International Conference on Machine Learning (ICML 15), pages 295 304, Lille, France, 2015. [Kushner, 1964] H. J. Kushner. A new method of locating the maximum point of an arbitrary multipeak curve in the presence of noise. Journal of Basic Engineering, 86(1):97 106, 1964. [Lyu et al., 2018] W. Lyu, F. Yang, C. Yan, D. Zhou, and X. Zeng. Batch Bayesian optimization via multi-objective acquisition ensemble for automated analog circuit design. In Proceedings of the 35th International Conference on Machine Learning (ICML 18), pages 3306 3314, Stockholm, Sweden, 2018. [Martinez-Cantin et al., 2007] R. Martinez-Cantin, N. de Freitas, A. Doucet, and J. A. Castellanos. Active policy learning for robot planning and exploration under uncertainty. In Robotics: Science and Systems III (RSS 07), pages 321 328, Atlanta, GA, 2007. [Mockus, 1994] J. Mockus. Application of Bayesian approach to numerical methods of global and stochastic optimization. Journal of Global Optimization, 4(4):347 365, 1994. [Mutny and Krause, 2018] M. Mutny and A. Krause. Efficient high dimensional Bayesian optimization with additivity and quadrature fourier features. In Advances in Neural Information Processing Systems 31 (Neur IPS 18), pages 9005 9016, Montreal, Canada, 2018. [Rasmussen and Williams, 2006] C. E. Rasmussen and C. Williams. Gaussian Processes for Machine Learning. The MIT Press, Cambridge, MA, 2006. [Ru et al., 2018] B. X. Ru, M. Mc Leod, D. Granziol, and M. A. Osborne. Fast information-theoretic Bayesian optimisation. In Proceedings of the 35th International Conference on Machine Learning (ICML 18), pages 4381 4389, Stockholm, Sweden, 2018. [Shah and Ghahramani, 2015] A. Shah and Z. Ghahramani. Parallel predictive entropy search for batch global optimization of expensive objective functions. In Advances in Neural Information Processing Systems 28 (NIPS 15), pages 3330 3338, Montreal, Canada, 2015. [Snoek et al., 2012] J. Snoek, H. Larochelle, and R. P. Adams. Practical Bayesian optimization of machine learning algorithms. In Advances in Neural Information Processing Systems 25 (NIPS 12), pages 2951 2959, Lake Tahoe, NV, 2012. [Springenberg et al., 2016] J. T. Springenberg, A. Klein, S. Falkner, and F. Hutter. Bayesian optimization with robust Bayesian neural networks. In Advances in Neural Information Processing Systems 29 (NIPS 16), pages 4134 4142, Barcelona, Spain, 2016. [Srinivas et al., 2012] N. Srinivas, A. Krause, S. M. Kakade, and M. W. Seeger. Information-theoretic regret bounds for Gaussian process optimization in the bandit setting. IEEE Transactions on Information Theory, 58(5):3250 3265, 2012. [Wang and Jegelka, 2017] Z. Wang and S. Jegelka. Max-value entropy search for efficient Bayesian optimization. In Proceedings of the 34th International Conference on Machine Learning (ICML 17), pages 3627 3635, Sydney, Australia, 2017. [Wang et al., 2013] Z. Wang, M. Zoghi, F. Hutter, D. Matheson, and N. de Freitas. Bayesian optimization in high dimensions via random embeddings. In Proceedings of the 23rd International Joint Conference on Artificial Intelligence (IJCAI 13), pages 1778 1784, Beijing, China, 2013. [Wang et al., 2016] Z. Wang, B. Zhou, and S. Jegelka. Optimization as estimation with Gaussian processes in bandit settings. In Proceedings of the 19th International Conference on Artificial Intelligence and Statistics (AISTATS 16), pages 1022 1031, Cadiz, Spain, 2016. [Wang et al., 2017] Z. Wang, C. Li, S. Jegelka, and P. Kohli. Batched high-dimensional Bayesian optimization via structural kernel learning. In Proceedings of the 34th International Conference on Machine Learning (ICML 17), pages 3656 3664, Sydney, Australia, 2017. Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence (IJCAI-20)