# eventtriggered_timevarying_bayesian_optimization__9ac09dff.pdf Published in Transactions on Machine Learning Research (01/2025) Event-Triggered Time-Varying Bayesian Optimization Paul Brunzema paul.brunzema@dsme.rwth-aachen.de Institute for Data Science in Mechanical Engineering RWTH Aachen University Alexander von Rohr vonrohr@dsme.rwth-aachen.de Institute for Data Science in Mechanical Engineering RWTH Aachen University Friedrich Solowjow friedrich.solowjow@dsme.rwth-aachen.de Institute for Data Science in Mechanical Engineering RWTH Aachen University Sebastian Trimpe trimpe@dsme.rwth-aachen.de Institute for Data Science in Mechanical Engineering RWTH Aachen University Reviewed on Open Review: https: // openreview. net/ forum? id= WEYMCLu8u7 We consider the problem of sequentially optimizing a time-varying objective function using time-varying Bayesian optimization (TVBO). Current approaches to TVBO require prior knowledge of a constant rate of change to cope with stale data arising from time variations. However, in practice, the rate of change is usually unknown. We propose an event-triggered algorithm, ET-GP-UCB, that treats the optimization problem as static until it detects changes in the objective function and then resets the dataset. This allows the algorithm to adapt online to realized temporal changes without the need for exact prior knowledge. The event trigger is based on probabilistic uniform error bounds used in Gaussian process regression. We derive regret bounds for adaptive resets without exact prior knowledge of the temporal changes and show in numerical experiments that ET-GP-UCB outperforms competing GP-UCB algorithms on both synthetic and real-world data. The results demonstrate that ET-GP-UCB is readily applicable without extensive hyperparameter tuning. 1 Introduction Over the last two decades, Bayesian optimization (BO) has emerged as a powerful method for sequential decision-making and design of experiments under uncertainty (Garnett et al., 2010; Snoek et al., 2012; Calandra et al., 2016; Frazier & Wang, 2016; Chen et al., 2018; Neumann-Brosig et al., 2019; Griffiths & Hernández-Lobato, 2020; Colliandre & Muller, 2023). These problems are typically formalized as optimization problems of a static unknown objective function. At the core of BO is the exploration-exploitation trade-off, where the decision maker weights the potential benefit of a new unexplored query against the known reward of the currently assumed optimum. The performance of BO algorithms is typically measured in terms of regret, i.e., the difference between the actual decision taken and the (unknown) optimal one. If one assumes a static objective function, there exist algorithms that achieve a desirable sub-linear regret (see Garnett (2023) for an overview), meaning they will efficiently converge to the globally optimal solution. In the real world, however, objective functions are often time-varying. Time-varying objective functions are widespread across various domains, including economics due to fluctuations in price or supply and demand (Primiceri, 2005; Dangl & Halling, 2012), environmental science (Marchant & Ramos, 2012; Marchant et al., 2014), and control systems due to changing system dynamics (Åström & Wittenmark, 1995; Ellis & Christofides, 2014; Colombino et al., 2019; Simonetto et al., 2020; König et al., 2021; Brunzema et al., 2022). Published in Transactions on Machine Learning Research (01/2025) Event Trigger Data set Acquisition function Objective Event trigger bound GP model next time step detected change Figure 1: Illustration of our proposed concept. After the standard BO steps of optimizing the acquisition function and obtaining an observation, an event trigger decides whether to reset the data set (γreset = 1) if a change is detected (see red cross), or augment it with the observed data point in the usual way (γreset = 0). This event trigger allows our algorithm ET-GP-UCB to be adaptive to changes in the objective function without relying on exact prior knowledge of the rate of change. A time-varying objective function alters the nature of the exploration-exploitation trade-off. Time affects the optimal decision and the information a decision maker gains from each query, and any collected data set becomes stale. Generally, in a setting of ongoing changes and without further assumptions, we cannot expect to find the optimal query, and conversely, no algorithm can achieve sub-linear regret (Besbes et al., 2015, Proposition 1). Instead, our goal is to design algorithms with an explicit relation between regret and the rate of change ε, which quantifies how fast the objective function varies. In their seminal work, Bogunovic et al. (2016) proposed two algorithms for the time-varying Bayesian optimization (TVBO) problem; however, the proposed algorithms and their subsequent variants rely on prior knowledge of the rate of change ε as a hyperparameter. In typical BO applications, where the objective is unknown, knowledge of its rate of change ε is hard to obtain. Moreover, in these algorithms, ε is specified at the start and assumed to remain constant over time, which is rarely the case in practice. To address these limitations, we propose the novel concept of event-triggered TVBO (see Figure 1). We model the objective function as time-invariant until an event trigger detects a significant deviation of the observed data from the current model. The algorithm then reacts to this change by resetting the data set. This way, the time-varying nature of the problem is considered in the algorithm s exploration-exploitation trade-off only when necessary. The event trigger replaces previously required exact prior knowledge of a rate of change, ET-GP-UCB exhibits strong empirical performance for different time variations, including gradual, sudden, and no change, and desirable theoretical properties under common regularity assumptions and upper and lower bounds on the rate of change. In summary, our contributions are as follows: (i) A conceptually new event-triggered BO algorithm for TVBO; (ii) Bayesian regret bounds for event-triggered BO for an unknown but bounded rate of change; (iii) Empirical evaluations of ET-GP-UCB on synthetic data and real-world benchmarks. 2 Problem Setting We aim to sequentially optimize an unknown time-varying objective function ft : X R on a compact set X Rd over the discrete time steps t I1:T := {1, . . . , T} up to the time horizon T as x t = arg max x X ft(x). (1) Here and in the following, subscript t denotes time dependency. At each time step, an algorithm chooses a query xt X and obtains noisy observations following the standard assumption in BO. Assumption 1 Observations yt = ft(xt)+wt are perturbed by independent and identically distributed (i.i.d.) Gaussian noise with noise variance σ2 n as wt N(0, σ2 n). Without assumptions on the temporal change, the problem is intractable. We impose the Markov chain model proposed by Bogunovic et al. (2016) and used in various prior applications (see Sec. 3). Published in Transactions on Machine Learning Research (01/2025) Assumption 2 (Bogunovic et al. (2016)) Let GP (µ, k(x, x )) be a Gaussian process (GP) with mean function µ : X R and kernel k: X X R. Given i.i.d. samples g1, g2, . . . with gi GP (0, k(x, x )), where k(x, x ) is a stationary kernel with k(x, x ) 1 for all x, x R, and a rate of change ε [0, 1], the time-varying objective function ft(x) follows the Markov chain: ( gt(x) , t = 1 1 εft 1(x) + εgt(x) , t 2. (2) Remark 1 Due to Assumption 2, any choice of ε [0, 1] yields that ft GP (0, k) for all t I1:T . We further use standard smoothness assumptions on the kernel choice introduced in Srinivas et al. (2010) as well as Bogunovic et al. (2016). Assumption 3 The kernel k(x, x ) is such that, given f GP (0, k(x, x )), L, Lf 1, and a0, b0, a1, b1 0, the following holds: P sup x X |f(x)| > Lf a0e (Lf /b0)2 and P sup x X > L a1e (L/b1)2, j = 1, ..., d. (3) This assumption limits the choice of kernel to those that are at least four times differentiable (Ghosal & Roy, 2006, Theorem 5), such as the popular squared exponential (SE) kernel. The performance of an algorithm optimizing (1) is measured in terms of the cumulative regret RT . Definition 1 Let x t be the maximizer of ft(x), and xt X be the query at time step t. Then, the instantaneous regret at t is rt = ft(x t ) ft(xt), and the regret after T time steps is RT = PT t=1 rt. Sub-linear regret is defined as lim T RT /T = 0. In the time-invariant setting (ε = 0), algorithms can achieve such sub-linear regret, e.g., GP-UCB (Srinivas et al., 2010, Theorem 2) with regret of order O( T)1. For a time-varying objective function following Assumption 2, Bogunovic et al. (2016, Theorem 4.1) shows that the lower bound on the regret of an algorithm is linear as Ω(Tε). Problem Statement We consider the time-varying optimization problem (1) under Assumptions 1 3. To develop a practical algorithm and in contrast to the state-of-the-art in TVBO we assume the exact rate of change ε to be unknown. According to prior work (Bogunovic et al., 2016), it is reasonable to demand regret bounds that are no more than linear in T and explicitly dependent on the true ε. 3 Related Work and Background We build on prior work in BO and TVBO. This section gives an overview of these two fields. Bayesian Optimization Our event trigger design and algorithm extends prior work by Srinivas et al. (2010) on the BO algorithm GP-UCB and transfers some of their concepts to the time-varying setting. Srinivas et al. (2010) proved sub-linear regret bounds when using an upper confidence bound (UCB) acquisition function of the form µDt(x) + βtσDt(x), with standard posterior updates of the GP model, given a dataset Dt := {(xi, yi)}t 1 i=1, as µDt(x) = kt(x)T Kt + σ2 n I 1 yt time-invariant posterior mean and σ2 Dt(x) = k(x, x) kt(x)T Kt + σ2 n I 1 kt(x) time-invariant posterior variance (4) where Kt = [k(xi, xj)]t 1 i,j=1 is the Gram matrix, kt(x) = [k(xi, x)]t 1 i=1, and noisy measurements yt = [y1, . . . , yt 1]T follow Assumption 1. Our algorithm uses these standard posterior updates only as long as all the measurements comply with the current posterior model; it resets the dataset if this is no longer the case. Consequently, in contrast to GP-UCB, our algorithm can adapt to changes in the objective function by discarding stale data. For a detailed overview of other standard BO algorithms as well as the standard BO framework, we refer to Garnett (2023). 1As in Bogunovic et al. (2016) and Srinivas et al. (2010), we denote asymptotics up to log factors as O( ). Published in Transactions on Machine Learning Research (01/2025) Time-Varying Bayesian Optimization TVBO can be considered a special case of contextual BO (Krause & Ong, 2011), with the key difference that time is treated as a strictly increasing variable. The inclusion of time as an input for spatial-temporal monitoring within the BO framework was explored by Marchant & Ramos (2012) and Marchant et al. (2014). Building on spatio-temporal GPs, Bogunovic et al. (2016) introduced two time-varying UCB algorithms tailored for TVBO, where they suggest that changes in the objective function can be effectively captured using the model outlined in Assumption 2. Embedding this assumption into the GP surrogate model yields the following posterior updates µDt(x) = kt(x)T Kt + σ2 n I 1 yt time-varying posterior mean and σ2 Dt(x) = k(x, x) kt(x)T Kt + σ2 n I 1 kt(x) time-varying posterior variance (5) where Kt = Kt Ktime t with Ktime t = [(1 ε)|i j|/2]t 1 i,j=1, and kt(x) = kt(x) ktime t with ktime t = [(1 ε)|t+1 i|/2]t 1 i=1, and is the Hadamard product. Using updates following (5) ensures that the posterior distribution follows the same dynamics as ft. We will utilize this in the proof of regret bounds for algorithms without exact knowledge of ε under the mentioned regularity assumptions. Since the introduction of this setting by Bogunovic et al. (2016), their algorithm TV-GP-UCB with posterior updates as in (5) has been used in various applications, such as controller learning (Su et al., 2018), safe adaptive control (König et al., 2021), and hyperparameter optimization for reinforcement learning (Parker Holder et al., 2020; 2021). Similarly, Brunzema et al. (2022) use time-varying posterior updates in their algorithm UI-TVBO but with a different temporal kernel. Both approaches rely on a priori estimation of the rate of change and assume that it stays constant over time. In contrast, the event trigger in ET-GP-UCB does not need an estimate of the rate of change and enables our algorithm to adapt to changes in the objective. We note that if the rate of change is known beforehand and the structural assumptions are satisfied, the TV-GP-UCB is an efficient choice since these properties are explicitly embedded in the algorithm. Bogunovic et al. (2016) further introduced R-GP-UCB, which resets the data set periodically to account for data becoming uninformative over time. In essence, ET-GP-UCB is an adaptive version of R-GP-UCB and resets only if an obtained measurement is no longer consistent with the current dataset (cf. Sec. 5). As in R-GP-UCB, our algorithm uses time-invariant posterior updates within time intervals, but our time intervals do not have to be specified a priori based on ε; rather, they are determined online by the event trigger. Under frequentist regularity assumption, Zhou & Shroff (2021) also propose an algorithm with a constant reset time, BR-GP-UCB2, as well as a sliding window algorithm. The crucial difference to our problem setting is their assumption of a fixed variational budget, i.e., PT t=1 ||ft+1 ft||H < BT , where H is a Reproducing Kernel Hilbert Space (RKHS) and BT a scalar. Leveraging these same assumptions, Deng et al. (2022) propose an algorithm, W-GP-UCB, which utilizes a similar weighted GP model to TV-GP-UCB. In this setting of a fixed variational budget, sub-linear regret is possible (Besbes et al., 2015). In contrast, we are interested in the setting given by Assumption 2, which corresponds to a variational budget that increases linearly in time; hence, the best possible regret is also linear (Besbes et al., 2015, Proposition 1). Nevertheless, in Appendix A.5, we empirically compare ET-GP-UCB against the algorithms of Zhou & Shroff (2021) and Deng et al. (2022) and display improved performance on their benchmarks with less prior information. Event-Triggered Learning The idea of resetting a dataset only if an important change occurs as detected by a trigger event is similar to the concepts of event-triggered learning (Solowjow & Trimpe, 2020) and eventtriggered online learning (Umlauft & Hirche, 2019). In the multi-armed bandit setting, Wei & Luo (2021) discuss the idea of including tests into optimization algorithms to detect and adapt to non-stationarity. Our algorithm ET-GP-UCB is the first to introduce event triggers to BO and specifically TVBO. Since then, this idea has been taken up for a different BO problem, namely safe BO, in recent work (Holzapfel et al., 2024). The event-triggered concepts aim to be efficient with the available data, only updating a model when necessary or only adding new data that is relevant. ET-GP-UCB also aims for data efficiency; as long as we have a good model of our objective function, we want to use already available data to minimize regret. If the event trigger detects a significant model mismatch, we reset the dataset and thereby delete stale data. 2We label it BR-GP-UCB to avoid confusion with R-GP-UCB of Bogunovic et al. (2016). Published in Transactions on Machine Learning Research (01/2025) Dynamic and Non-Stationary Multi-Armed Bandits Related problems to (1) have also been considered in the non-stationary multi-armed bandit (MAB) literature. The key distinction to our problem setting is that for MABs the number of arms is finite and they are not correlated. Similar to our setting with ongoing changes, Whittle (1988) introduce restless MABs; here, the state of an arm can change in each step according to a known (yet arbitrary) stochastic transition function (Slivkins & Upfal, 2008). Such restless MABs are known to be intractable (Papadimitriou & Tsitsiklis, 1994; Slivkins & Upfal, 2008). For restless bandits with a time-variation parameter greater zero linear regret as Ω(T) is inevitable (Slivkins & Upfal, 2008, Theorem 1.3). The setting in Slivkins & Upfal (2008) is, in its effects, similar to the regularity assumption on the temporal changes in Assumption 2 with a corresponding lower bound on the regret of Ω(Tε) (Bogunovic et al., 2016, Theorem 4.1). Also related are MABs under a fixed number of abrupt distribution shifts (Auer et al., 2019; Abbasi-Yadkori et al., 2023). For these problems, adaptive resetting algorithms have been proposed (Chen et al., 2019; Wei & Luo, 2021; Suk & Kpotufe, 2022). In this more structured problem setting, similar to the frequentist regularity assumptions discussed above, sub-linear regret is possible. 4 Regret Bounds for TVBO with Adaptive Resets The core goal of this work is to develop an algorithm for an unknown rate of change. To achieve this goal, we first generalize the results by Bogunovic et al. (2016) and derive regret bounds for a BO algorithm with variable resets. We analyze the general problem for all strategies that reset anytime between a lower ( N) and upper ( N) reset threshold. Our analysis reveals that for any resetting strategy, given reasonable N and N, linear regret with explicit dependency on the true ε is guaranteed, even if the exact ε is unknown to the resetting algorithm. This is in contrast to previous work on TVBO, where the true ε does not only appear in the regret bounds but needs to be known to the algorithm or GP model. We first introduce a theorem applicable to all algorithms with such a general resetting strategy, and then apply it to specific kernels to derive scaling laws. In Sec. 5, we will propose a specific event trigger as a resetting strategy. Theorem 1 Let the domain X [0, r]d Rd be convex and compact with d N+ and let ft follow Assumption 2 with a kernel k(x, x ) such that Assumption 3 is satisfied. Pick δ (0, 1) and set βt = 2 ln 2π2t2/(3δ) + 2d ln (t2db1r p ln(2da1π2t2/(3δ))). Pick a lower block length N N+ and a upper block length N N+ with N N T. Let γ N be the maximum information gain over N points. Then, running any algorithm with block sizes between N and N satisfies N + 1 γ N + 2 + TϕT (ε, N) (6) with probability at least 1 δ, where C1 = 8/ ln(1 + σ 2 n ) and we defined ϕT (ε, N) := 2 q βT (3σ 2 n + σ 4 n ) N 3ε + 2(σ 2 n + σ 4 n ) N 3ε(b0 + ln 4(a0+1)π2T 2 Sketch of proof: We extend the analysis of R-GP-UCB to a setting in which resets are not enforced periodically but can occur at any time step between N and N. Building on Assumptions 1 to 2, we use the upper threshold N to bound the difference between the model used in the algorithm (time-invariant model, see (4)) and the time-varying model in (5). We use the lower threshold N to bound the maximum number of resets over the time horizon. The full proof is in Appendix F. We next apply this theorem to specific kernels. We specify how N and N scale with upper and lower bounds on the rate of change to obtain the desirable scaling of the regret. Corollary 1 Let the assumptions in Theorem 1 hold. Given a lower and upper bound on the true ε such that ε [ ε, ε] (0, 1) there exist λ1 (0, 1] and λ2 [1, 1 ε) such that ε = λ1 ε and ε = λ2 ε. For the SE kernel, set N = Θ(min{T, ε 1/4}) and N = Θ(min{T, ε 1/4}). Then, with probability 1 δ, we have RT = O max n T, λ1/8 2 ε1/8T, λ 3/8 1 ε1/8T, λ 3/4 1 ε1/4T o . (8) Published in Transactions on Machine Learning Research (01/2025) For the Matèrn kernel (ν > 2), define ξ = d(d+1) 2ν+d(d+1) and set N = Θ(min{T, ε 1/(4 ξ)}) and N = Θ(min{T, ε 1/(4 ξ)}). Then, with probability 1 δ, we have T 1 ξ, λ ξ 2(4 ξ) 1 λ 1 2(4 ξ) 2 ε 1 ξ 2(4 ξ) T, λ 3 2(4 ξ) 1 ε 1 ξ 2(4 ξ) T, λ 3 4 ξ 1 ε 1 ξ 4 ξ T . (9) From Corollary 1, we can derive the following insights: (i) It is possible to design algorithms with regret explicitly depending on ε, without knowing its exact value. (ii) A larger interval for ε, i.e., decreasing λ1 or increasing λ2, results in a worse scaling of the regret bound. It essentially is the price one has to pay for not knowing the exact value of ε. (iii) For ε = ε = ε, hence λ1 = λ2 = 1, the scaling of R-GP-UCB is recovered. From Theorem 1, we can further observe that for ε = 0 the bounds recover the sub-linear scaling of GP-UCB iff an appropriate lower bound on ε is specified such that N = T which follows intuition as no resets should should take place. The above results demonstrate that it is possible to design algorithms that do not require the exact rate of change; as long as lower and upper bounds on ε are known and resets occur within [ N, N] the algorithm has bounded regret. Crucially, we did not actually specify when to reset, thus our analysis includes random and worst case resets. Next, we utilize the flexibility in choosing the reset timing within [ N, N] to achieve better empirical regret, e.g., as compared to periodic resets. 5 An Event Trigger for Adaptive Resets in TVBO We will now introduce our algorithm ET-GP-UCB. It will decide when to reset through its event trigger (see Figure 1). We build on established uniform error bounds for GP regression to decide if new data is consistent with the posterior. When new data deviates significantly, we reset the dataset. This idea constitutes a new approach compared to previous work in TVBO, as the resulting algorithm does not rely on the true rate of change ε as a hyperparameter. The core component of our algorithm is the event trigger: Definition 2 (Event trigger in TVBO) Given a test function ψt and a threshold function κt, both of which can depend on the current dataset Dt and the latest query location and measurement pair (xt, yt), the event trigger at time t is γreset = 1 ψt (Dt, (xt, yt)) > κt (Dt, (xt, yt)) (10) where γreset is the binary indicator for whether to reset the dataset (γreset = 1) or not (γreset = 0). We next design and develop ψt and κt for the specific event trigger of ET-GP-UCB. We base these functions on Bayesian uniform error bounds for GP regression (Srinivas et al., 2010). Other design options would also be possible, e.g., likelihood-based or performance-based event triggers. In the following, we denote the time step of a reset as tr(1), tr(2), . . . and define tr tr(i), tr(i+1) N+ as the time in between resets. We have tr = t if no reset has yet occurred. Lemma 1 (Srinivas et al. (2010, Lemma 5.5)) Pick δ (0, 1) and set ρtr = 2 ln πtr δ , where P tr 1 π 1 tr = 1, πtr > 0. Let f GP(0, k(x, x )). Then, |f(xt) µDt(xt)| ρtrσDt(xt) holds for all tr 1 with probability at least 1 δ. Remark 2 For P tr 1 π 1 tr = 1 to hold, one can choose πtr = π2tr2/6 as in Srinivas et al. (2010). The result gives a high probability bound for all time steps on the absolute deviation between the posterior mean and a time-invariant objective function at the query location xt based on the posterior variance and parameter ρtr. If a time-varying objective function satisfies this bound for all time steps, the time variations are negligible with high probability (i.e., ε 0). On the other hand, if the bound in Lemma 1 is violated at a chosen query location xt, the objective function has likely changed. The algorithm needs to evaluate the trigger event at every time step. While it cannot directly evaluate |ft(xt) µDt(xt)| in Lemma 1 (cf. Assumption 1), we can evaluate |yt µDt(xt)|. Thus, we adapt Lemma 1 to design a probabilistic bound that includes only yt instead of ft(xt). For this, we first bound the noise in probability. All proofs of the following Lemmas are in Appendix G. Published in Transactions on Machine Learning Research (01/2025) Algorithm 1 Event-triggered GP-UCB (ET-GP-UCB). 1: Define: GP(0, k), X Rd, δB (0, 1), D1 = , lower reset bound N, upper reset bound N, tr = 1 2: for t = 1, 2, . . . , T do 3: Train GP model with the current data set Dt 4: Choose βt (e.g., according to Theorem 1) 5: Select xt = arg maxx X µDt(x) + βtσDt(x) time-invariant posterior using (4) 6: Sample next observation yt = ft(xt) + wt 7: γreset Event Trigger(Dt, (yt, xt), δB) evaluate (10) with (11) 8: if (γreset and tr [ N, N]) or tr = N then reset only if within reset window 9: Reset dataset Dt+1 = {(yt, xt)} and set tr = 1 10: else 11: Update dataset Dt+1 = Dt {(yt, xt)} and set tr = tr + 1 Lemma 2 Pick δ (0, 1) and set w2 tr = 2σ2 n ln πtr δ , where P tr 1 π 1 tr = 1, πtr > 0. Then, the noise sequence in Assumption 1 obtained since the last reset satisfies |wtr| wtr for all tr 1 with probability at least 1 δ. With Lemma 2, we can extend Lemma 1 to build a probabilistic error bound that can act as the test and corresponding threshold functions of our event trigger. Lemma 3 Let ft(x) follow Assumption 2 with ε = 0. Pick δB (0, 1) and set ρtr = 2 ln 2πtr δB , where P tr 1 π 1 tr = 1, πtr > 0. Also set w2 tr = 2σ2 n ln 2πtr δB . Then, observations yt under Assumption 1 satisfy |yt µDt(xt)| ρtrσDt(xt) + wtr for all tr 1 with probability at least 1 δB. We can now formally define the event trigger for ET-GP-UCB following the framework outlined in Definition 2. Leveraging the result from Lemma 3, we choose as the test and threshold function ψt (Dt, (xt, yt)) = |yt µDt(xt)| := test function ψt in Def. 2 and κt (Dt, (xt, yt)) = ρtrσDt(xt) + wtr := threshold function κt in Def. 2 , (11) respectively. With this, we have designed a principled event trigger to detect changes in the objective function. Note that our event trigger monitors for changes in the objective function only at the current query location and cannot detect changes anywhere else. Our problem setting restricts an algorithm to querying the objective only once per time step. The resulting algorithm ET-GP-UCB is then as follows: We perform standard GP-UCB as long as the event trigger evaluates to γreset = 0. When the event trigger is activated the objective function ft has likely changed, and the data has become stale. If tr [ N, N], ETGP-UCB resets the dataset and restarts the optimization. This is summarized in Algorithm 1. The single design parameter of our event trigger is δB in Lemma 3, which has a well-understood meaning as defining the tightness of the probabilistic uniform error bound. Since resets are confined to the range between the lower and upper reset thresholds, ET-GP-UCB guarantees a regret no worse than that specified in Corollary 1. 6 Empirical Evaluations In the following, we will first investigate the empirical impact of N and N on the reset times of the algorithm as well as on the regret. In Section 6.2, we will benchmark our algorithm against other baselines on various synthetic and real-world problems and end with practical extensions for our proposed algorithm. 6.1 Empirical Impact of N and N on Reset Times and Regret Impact on the Reset Times To analyze the impact of N and N on the behavior and performance of ET-GP-UCB, we first define the window size in which the event trigger can reset the data set as W := N N. Further, note that the time of reset using (11) in the event trigger framework is a random variable as it depends on the realization of ft as well as the observation model. The ridgeline plot in Figure 2 (left) shows the empirical distribution of this reset time as (scaled) histograms; starting from the reset time of R-GP-UCB, the window size increases. For this, we ran ET-GP-UCB on 30 000 different objective functions following Assumption 2, so-called within-model comparisons (cf. Section 6.2.1), for each window size. Small window sizes cut off the probability mass in the tails of the histograms, while for larger window sizes, the histogram resembles an inverse gamma distribution. Published in Transactions on Machine Learning Research (01/2025) 0 50 100 150 200 Reset time (ε = 0.03) Window size W Constant reset time of R-GP-UCB 27 55 82 110 137 165 193 0 50 100 150 200 Window size W Mean empirical RT ( ) ε = 0.01 ε = 0.03 ε = 0.05 R-GP-UCB Reset at N Figure 2: Left: Influence of the window size on the empirical PDF of reset times. For small window sizes, the probability mass in the tails is cut off through N and N. As the window size W increases, hence the flexibility of our event trigger increases, the histogram resembles an inverse-gamma-like distribution. Right: Influence of the window size on the empirical regret for different ε. As the window size increases, the regret of ET-GP-UCB (markers) decreases always staying below R-GP-UCB (dash-dotted lines). Increasing the flexibility of our trigger empirically reduces regret. Always resetting at N (dashed line) shows increased empirical regret. This indicates that Corollary 1 is a relatively loose bound for our trigger. Impact on the Regret Figure 2 (right) demonstrates how increasing the window size affects regret for different ε. As the window size grows, ET-GP-UCB s performance (lines with markers) consistently remains below the regret of R-GP-UCB with known ε (dash-dotted lines). This suggests that even coarse bounds lead to robust performance, which is desirable from a practical viewpoint. The improved performance is contrary to the regret bounds in Corollary 1, which gets larger as the window size grows. However, Corollary 1 gives regret bounds for all triggers, including the worst-case. Empirically, the proposed event trigger performs much better than the worst-case and even improves on R-GP-UCB despite less prior knowledge of the rate of chance. ET-GP-UCB performs best for larger window sizes. This suggests that prior knowledge of the rate of change is less important than the regret bounds in Corollary 1 imply. The proposed event trigger uses the gained flexibility from the theoretical investigation to perform resets when necessary within the reset window. Other event triggers may not show the same behavior: resetting, e.g., always at the lower bound N (dashed lines in Figure 2) exhibits increased regret with growing window sizes highlighting the importance of choosing a suitable resetting strategy for good empirical performance. 6.2 Empirical Evaluation on Synthetic and Real-World Data Using both synthetic data and standard BO benchmarks, we compare ET-GP-UCB to state-of-the-art GPUCB baseline for TVBO. Specifically, we compare against R-GP-UCB, which resets periodically, TV-GPUCB (Bogunovic et al., 2016), which gradually forgets past data points, and UI-TVBO (Brunzema et al., 2022), which continuously injects uncertainty at past data points. We further compare these time-varying algorithms to the time-invariant GP-UCB (Srinivas et al., 2010). In our experiments, we find that: (i) The adaptive resetting of ET-GP-UCB always outperforms periodic resetting; (ii) ET-GP-UCB outperforms all baselines on synthetic data if the rate of change is misspecified; (iii) ET-GP-UCB outperforms all baselines on real-world data (where the rate of change is unknown). We reiterate that ET-GP-UCB does not rely on knowledge about the true rate of change. To highlight the generalization capabilities of ET-GP-UCB to different scenarios, we use the same parameter for the event trigger (δB = 0.1 in Lemma 3) in all experiments. A sensitivity analysis of the hyperparameter δB on the regret is in Appendix B and shows that ET-GP-UCB performs similarly for a wide range of δB. As in (Bogunovic et al., 2016) and (Srinivas et al., 2010), we utilize a logarithmic scaling βt as βt = O(d ln(t)), where βt = c1 ln(c2t). This approximates βt in Theorem 1 and allows for a direct comparison to Bogunovic et al. (2016), as they suggest to set c1 = 0.8 and c2 = 4 for a suitable exploration-exploitation trade-off. All experiments in this section are conducted using Bo Torch (Balandat et al., 2020) and GPy Torch (Gardner Published in Transactions on Machine Learning Research (01/2025) Table 1: Within-model comparisons. Top: Algorithms with time-invariant updates following (4). Bottom: Algorithms with more compute intensive time-varying updates similar to (5). The best median performance for each section is highlighted in bold, and in green for the best performance across top and bottom. ε = 0.01 ε = 0.03 ε = 0.05 Miss. ε = 0.001 Miss. ε = 0.2 Algorithm RT /T ( ) RT /T ( ) RT /T ( ) RT /T ( ) RT /T ( ) GP-UCB 0.748 0.904 0.610 1.051 1.243 0.924 1.276 1.377 1.088 1.276 1.377 1.088 1.276 1.377 1.088 R-GP-UCB 0.622 0.681 0.571 0.831 0.894 0.770 0.985 1.035 0.899 0.902 0.985 0.829 1.054 1.119 0.995 ET-GP-UCB0.05 0.01 0.604 0.682 0.531 0.778 0.841 0.720 0.879 0.966 0.824 0.879 0.966 0.824 0.879 0.966 0.824 ET-GP-UCB0.1 0.001 0.507 0.581 0.448 0.688 0.782 0.654 0.866 0.927 0.806 0.866 0.927 0.806 0.866 0.927 0.806 ET-GP-UCB1.0 0.0 0.483 0.571 0.407 0.686 0.738 0.639 0.849 0.899 0.748 0.849 0.899 0.748 0.849 0.899 0.748 TV-GP-UCB 0.299 0.365 0.258 0.500 0.565 0.447 0.622 0.686 0.582 0.960 1.036 0.854 1.223 1.289 1.126 UI-TVBO 0.351 0.376 0.311 0.647 0.678 0.600 0.868 0.904 0.833 0.953 1.072 0.839 1.381 1.417 1.340 * Our algorithms with formatting ET-GP-UCB upper bound on ε lower bound on ε . True rate of change is ε = 0.05. et al., 2018), and full details are in Appendix A. All results show the median and interquartile performance. In the following, we indicate the direction of better performance with upward ( ) and downward ( ) arrows. 6.2.1 Within-Model Comparison First, we consider the same two-dimensional setting for synthetic data as in Bogunovic et al. (2016). We choose the compact set to be X = [0, 1]2 and generate 50 different objective functions according to Assumption 2 with a known ε, a SE kernel with fixed lengthscales l = 0.2, time horizon T = 400, and a noise variance of σ2 n = 0.02. The setting in which all the hyperparameters are known to the algorithm is referred to as within-model comparison (Hennig & Schuler, 2012); such comparisons are used to evaluate an algorithm s performance given all assumptions are fulfilled (cf. Appendix A.1 for further details on the creation of the within-model objective functions). Here, TV-GP-UCB acts as the reference case with full information. It shows the best performance an UCB algorithm can obtain in this setting as it takes Assumption 2 explicitly into account by using the time-varying posterior updates in (5). R-GP-UCB is also parameterized using the true ε for the periodic reset time as Nconst = min{T, 12ε 1/4} (Bogunovic et al., 2016, Corollary 4.1). For ET-GP-UCB, we compare three different variants with different lower and upper bounds on ε and set N = min{T, 12 ε 1/4} and N = min{T, 12 ε 1/4} according to Corollary 1. For UI-TVBO, we set ˆσ2 w = ε (cf. Brunzema et al. (2022)). Table 1 shows the normalized regret (medianq75% q25%) for different ε. We can observe that in all cases the ET-GP-UCB variant with the largest flexibility of the event trigger outperforms all baselines with a time-invariant GP model (top). As expected, TV-GP-UCB is the best-performing algorithm if the rate of change is correct. However, if the rate of change is misspecified (right columns), ET-GP-UCB also outperforms the baselines with a time-varying GP model as it does not rely on exact knowledge of ε and can react to the realized changes of ft. 6.2.2 Real-World Temperature Data To benchmark the algorithms on real-world data, we use the temperature dataset collected from 46 sensors deployed at Intel Research Berkeley over eight days at 10 minute intervals.3 This dataset was also used as a benchmark in previous work on GP-UCB and TVBO (Srinivas et al., 2010; Krause & Ong, 2011; Bogunovic et al., 2016). The objective is to activate the sensor with the highest temperature at each time step resulting in a spatio-temporal monitoring problem. In our TVBO setting, ft consists of all sensor readings at time step t, and an algorithm can query only a single sensor every 10 minutes. Hence, the regret following Definition 1 is the difference between the maximum and the recorded temperature by the activated sensor. For each experiment, two test days are selected, while the preceding days serve as the training set. Using this training set, all data is normalized to have a mean of zero and a standard deviation of one. The empirical covariance matrix is computed based on the training set and employed as the kernel for all algorithms as in Bogunovic et al. (2016). The sensor data of the days 7 and 8 is shown in Figure 3 (a) and an overview of the remaining days is given in Appendix E. For TV-GP-UCB and UI-TVBO, we set ε = ˆσ2 w = 0.03 and the 3We thank the researchers at Intel for the publically available data set under: https://db.csail.mit.edu/labdata/labdata.html. Published in Transactions on Machine Learning Research (01/2025) 0 12 24 36 48 Time t [h] (a) Temperature data Norm. Temperature [ C] 0 12 24 36 48 Time t [h] (b) Days 7 & 8 0 12 24 36 48 Time t [h] (c) Days 5 & 6 GP-UCB R-GP-UCB TV-GP-UCB TV-GP-UCB (MLE) UI-TVBO ET-GP-UCB (ours) 100 200 300 Time steps t (d) Policy Search Resets of R-GP-UCB Resets of ET-GP-UCB Resets of R-GP-UCB Resets of ET-GP-UCB Figure 3: Real-world examples. Subfigure (a) shows the temperature data from the test days used in Bogunovic et al. (2016). In (b), the performance of the algorithms on these two days (top) and the resets of R-GP-UCB and ET-GP-UCB (bottom) are displayed. Subfigure (c) shows the performance on other test days. ET-GP-UCB displays superior performance improvement compared to all other algorithms. For these experiments on real-world data, we use the same parameter for the event trigger as in all previous experiments whereas TV-GP-UCB and R-GP-UCB relied on estimating ε. Lastly, in subfigure (d), ET-GPUCB also outperforms all competing baselines on a policy search benchmark. For all methods, the median performance and the interquartile (25% and 75%) are displayed. periodic reset time of R-GP-UCB to Nconst = 15 according to Bogunovic et al. (2016). For this benchmark, we further compare against TV-GP-UCB with online maximum likelihood estimation (MLE) of ε as one might choose to do in practice if the true rate of change is unknown. Note that this significantly increases the computational overhead. For ET-GP-UCB, we use the best-performing variant form Sec. 6.2.1 as well as the same parameter for the event trigger as in all the previous experiments. Figure 3 (b) presents the normalized regret (top) for each algorithm over 20 runs and the resets of R-GPUCB and ET-GP-UCB (bottom). In this real-world benchmark, ET-GP-UCB outperforms all the competing baselines. It highlights that using a constant rate of change in the GP model as in TV-GP-UCB can be problematic for real-world problems. ET-GP-UCB also shows the best performance on other test days as shown in Figure 3 (c). considering the larger interquartile of TV-GP-UCB. While ET-GP-UCB and TV-GPUCB have similar median performance, ET-GP-UCB displayed a significantly smaller interquartile range, indicating more consistent and reliable results. From the bottom figure, we can observe the adaptive nature of ET-GP-UCB. Considering the timing of resets of ET-GP-UCB, it appears that especially when changes in the objective function are more distinct (e.g., changes between night and day), the proposed event trigger framework is empirically beneficial compared to using an estimated constant rate of change. 6.2.3 Policy Search in Time-Varying Environments We also compare all algorithms on a four-dimensional policy search benchmark of a cart-pole system as proposed in (Brunzema et al., 2022). Given the system dynamics of the cart-pole system, it is possible to calculate the optimal policy and thus calculate the regret following Definition 1. To induce a change in the objective function, we increase the friction in the bearing at t = 100 by a factor of 2.5. We expect the event trigger of ET-GP-UCB to detect this change in the objective function and converge to the new optimum after resetting the dataset. In contrast to previous examples, we also perform online hyperparameter optimization of the spatial lengthscales using gamma priors to guide the search (all parameters are listed in Appendix H). For TV-GP-UCB, we choose ε = 0.03 and we pick the constant reset timing of R-GP-UCB according to Bogunovic et al. (2016, Corollary 4.1). For ET-GP-UCB, we again choose the same parameter for the event trigger as in all previous examples. Published in Transactions on Machine Learning Research (01/2025) 0.01 0.02 0.03 0.04 0.05 Rate of change ε R-GP-UCB R-GP-UCB (N = mean reset time of ET-GP-UCB) ET-GP-UCB ET-GP-UCB Figure 4: Impact of Reset Frequency vs. Timing. Table 2: Online MLE hyperparameter optimization of the GP when using ET-GP-UCB. ε = 0.01 ε = 0.03 ε = 0.05 Algorithm RT /T ( ) RT /T ( ) RT /T ( ) GP-UCB 0.835 1.260 1.405 R-GP-UCB 0.776 0.998 1.123 ET-GP-UCB 0.612 0.870 1.057 *with learn-then-monitor approach The results over 20 runs are displayed in Figure 3 (d). We can observe that ET-GP-UCB is the best performing algorithm since it can detect the change in the objective function and converge to the new optimal solution. We can further see that TV-GP-UCB and R-GP-UCB, which rely on an ongoing and smooth change in the objective function, empirically struggle to cope with such sudden changes resulting in significantly increased regret. While UI-TVBO can also recover from the change, ET-GP-UCB still outperforms this baseline. The example also highlights that a constant reset time that is too short can yield divergent behavior in higher-dimensional settings and non-smooth changes. 6.2.4 Ablation on the Impact of Reset Frequency vs. Reset Timing To better understand the source of empirical improvement of ET-GP-UCB, we conducted an ablation study to test whether the performance gains are due to resetting at the correct frequency or at the correct time steps. Specifically, we investigated the effect of using the average resetting time of ET-GP-UCB, derived from Monte-Carlo simulations of the event trigger, as the reset interval N for R-GP-UCB. Figure 4 shows within-model experiments across multiple rate of changes for 50 different seeds. The results demonstrate that resetting at the average frequency of the trigger improves regret performance at low rates of change. However, as the rate of change increases, adaptively resetting at the correct time steps, i.e., considering the realized changes in the objective function, becomes crucial for minimizing regret. 6.3 Practical Extensions of ET-GP-UCB Next, we discuss practical extensions to ET-GP-UCB that can enhance its performance in practice. Upper Bound on the Noise Imposing a fixed upper bound on the noise term wtr after a certain number of time steps can be practical. While the gradual increase of wtr is crucial for Lemma 3 to hold, it results in large thresholds in (11) over time. This might lead to undetected changes, especially when combined with a large N and rare changes in the objective. Considering bounded noise in Assumption 1 instead of Gaussian noise might also be more suitable for some applications. Online Hyperparameter Optimization Practitioners should be cautious when using naïve MLE for all hyperparameters, as our algorithm fits a time-invariant GP to time-varying data. This can inflate the noise variance, again resulting in large thresholds in (11). The core issue lies in the dual use of the model mismatch: the event-trigger mechanism is designed to detect model mismatches, while hyperparameter optimization aims to minimize this mismatch. These two objectives conflict. We suggest estimating the noise on hold-out data, as demonstrated in Section 6.2 with the temperature dataset. If this is not feasible, we recommend using reasonable priors and constraints for all hyperparameters and then performing maximum-a-posteriori estimation, as is common practice for most GP and BO libraries (GPy, 2012; Gardner et al., 2018; Balandat et al., 2020) and demonstrated in the policy search example. In the case of no prior knowledge, a simple learn-then-monitor approach, i.e., learning the parameters and freezing them afterward while monitoring for changes, can be helpful. To demonstrate this, we conducted experiments with online hyperparameter optimization for all baselines using only very coarse constraints. Specifically, we use a lower bound of 0.001 and an upper bound of 0.1 for the noise (true noise value is 0.02) and a lower bound of 0.01 and an upper bound of 1 for the lengthscales (true lengthscales are 0.2). For ET-GP-UCB, we use the first 2 d time steps Published in Transactions on Machine Learning Research (01/2025) Norm. cumulative value P t ft(xt)/t ( ) Input dimension: d = 8 Forgetting factor: ε = 0.01 R-GP-UCB TV-GP-UCB ET-GP-UCB ET-GP-UCB w. backtracking 100 200 300 400 Time t Resets of respective algorithm Resets of respective algorithm 330 340 0 Consecutive 2 4 6 8 Input dimensions d (ε = 0.01) t ft(xt)/t ( ) R-GP-UCB TV-GP-UCB ET-GP-UCB ET-GP-UCB w. backtracking Figure 5: Left: Performance on an eight-dimensional within-model objective function. Retaining a portion of past data using Algorithm 2 leads to a significant performance boost, nearly matching the results of TV-GPUCB, even without prior knowledge of ε. Right: Final performance for different dimensions (d {2, 4, 6, 8}) with a fixed ε. As dimensionality increases, the benefit of retaining data becomes more pronounced. after a trigger event to learn the noise variance and individual lengthscales, and then freeze the parameters and start monitoring for changes using our event trigger. Table 2 shows that this approach decouples the two conflicting objectives, yielding reasonable performance without prior knowledge of ε, the spatial hyperparameters, nor the observation model. Retaining Past Information Exploring alternative strategies to augment the data set instead of performing a full reset every time the event trigger activates is interesting to preserve the most recent information. In this paper, we demonstrated that even with a full reset we obtain superior performance on commonly used real-world benchmarks but preserving information can become crucial for higher-dimensional problems. One idea that arises directly from our event-triggered setup is to implement a backtracking mechanism when a change is detected: after detecting a change, the algorithm backtracks to check how many of the most recent data points still satisfy the trigger bounds. These relevant observations are then retained for further optimization. By re-using the event trigger to determine which data to keep, we ensure that, in the event of sudden changes, there are no inconsistent data points in the data set. We cap the re-initialization size of the data set by 2 d; similar data set sizes are commonly used to initialize BO algorithms (Eriksson et al., 2019; Balandat et al., 2020; Müller et al., 2021). The backtracking procedure is summarized in Algorithm 2. Algorithm 2 Backtracking procedure to retain data 1: procedure Backtracking(D, Event Trigger) 2: ˆDkeep {} Initialize an empty data set 3: for i |D| 1 downto 0 do 4: (yi, xi) D[i] Extract input-output pair 5: γreset Event Trigger( ˆDkeep, (yi, xi)) 6: if γreset or | ˆDkeep| 2 d then 7: break Stop backtracking 8: ˆDkeep ˆDkeep {(yi, xi)} 9: return ˆDkeep To test this approach, we conduct withinmodel comparisons for different dimensions summarized in Figure 5. Retaining some of the past data proves to be very beneficial. In Figure 5 (left), we can see that by using Algorithm 2, ET-GP-UCB is almost able to match the performance of TV-GP-UCB even without prior knowledge of ε. Notably, combining data reuse and smooth changes can yield consecutive resets, occasionally transforming the method into an adaptive sliding window approach. From a computational standpoint, this poses no significant overhead, as backtracking through the dataset and evaluating the event trigger are computationally efficient. Figure 5 (right) shows that as the dimensions increase, using a strategy that reuses some of the past data becomes more beneficial. The results for more forgetting factors are in Appendix A. Published in Transactions on Machine Learning Research (01/2025) 7 Conclusion and Future Work This paper presented the novel concept of event-triggered resets to optimize time-varying objective functions with TVBO. Our theoretical results show that algorithms that reset between lower and upper thresholds yield meaningful regret even if the rate of change is unknown. This allowed us to optimize the actual reset timing for increased empirical performance using probabilistic error bounds. The resulting algorithm, ETGP-UCB, adapts to the realized changes in the objective function without relying on exact prior knowledge of these temporal changes, making it especially useful in real-world scenarios. We empirically showed that the adaptive resetting of ET-GP-UCB consistently outperforms periodic resetting and further demonstrated ET-GP-UCB s increased performance on real-world benchmarks. Lastly, we presented practical extensions to our algorithm. These extensions included an approach for practical online hyperparameter estimation within the event trigger framework and an approach to retaining past information, which further increases performance in higher-dimensional settings. The latter leverages the event trigger not only to detect changes but also to determine how much recent data can be reused, thereby improving sample efficiency. Overall, ET-GP-UCB offers an adaptive solution for optimizing time-varying objectives with minimal hyperparameter tuning and prior knowledge required, making it a valuable tool for many time-varying optimization problems. An interesting direction for future work on ET-GP-UCB is its further theoretical analysis. While Corollary 1 ensures the desirable scaling of the regret for worst-case resets, it does not reflect the improved empirical regret when increasing the window size as the event trigger is not explicitly included in the analysis. Regret bounds that explicitly consider the choice of event trigger are a promising direction for future research and would bridge the current gap between theoretical analysis and empirical performance. For further discussion, see Appendix C. Another research direction would be to move away from a setting solely focused on dynamic regret and consider e.g., a setting with multiple comparators that are fixed for certain window lengths (Zhang et al., 2020). Here, we believe that adaptive approaches as proposed herein can be especially beneficial. Lastly, investigating different event triggers within the framework of Definition 2 is interesting, such as event triggers that consider multiple time steps in their test and threshold function to, e.g., identify slow trends in the objective. New and tailored data augmentation strategies may also arise with new event triggers or event triggers tailored to specific applications. Acknowledgments The authors thank Pierre-François Massiani, Dominik Baumann, Christian Fiedler, and Noel Brindise for their helpful comments and discussions, as well as the reviewers and action editor for their constructive feedback, which helped improve the final version of the paper. This work is partially funded by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) RTG 2236/2 (Un RAVe L). Simulations were performed with computing resources granted by RWTH Aachen University under project rwth1579. Yasin Abbasi-Yadkori, András György, and Nevena Lazić. A new look at dynamic regret for non-stationary stochastic bandits. Journal of Machine Learning Research, 24(288), 2023. Karl Johan Åström and Björn Wittenmark. Adaptive control. In Mathematical System Theory: The Influence of RE Kalman. Springer, 1995. Peter Auer, Pratik Gajane, and Ronald Ortner. Adaptively tracking the best bandit arm with an unknown number of distribution changes. In Conference on Learning Theory, pp. 138 158. PMLR, 2019. Maximilian Balandat, Brian Karrer, Daniel R. Jiang, Samuel Daulton, Benjamin Letham, Andrew Gordon Wilson, and Eytan Bakshy. Bo Torch: A framework for efficient Monte-Carlo Bayesian optimization. In Advances in Neural Information Processing Systems, 2020. Omar Besbes, Yonatan Gur, and Assaf Zeevi. Non-stationary stochastic optimization. Operations Research, 63(5), 2015. Ilija Bogunovic, Jonathan Scarlett, and Volkan Cevher. Time-varying Gaussian process bandit optimization. In International Conference on Artificial Intelligence and Statistics. PMLR, 2016. Published in Transactions on Machine Learning Research (01/2025) Paul Brunzema, Alexander von Rohr, and Sebastian Trimpe. On controller tuning with time-varying Bayesian optimization. In IEEE Conference on Decision and Control, 2022. Roberto Calandra, André Seyfarth, Jan Peters, and Marc Peter Deisenroth. Bayesian optimization for learning gaits under uncertainty: An experimental comparison on a dynamic bipedal walker. Annals of Mathematics and Artificial Intelligence, 76:5 23, 2016. Yifang Chen, Chung-Wei Lee, Haipeng Luo, and Chen-Yu Wei. A new algorithm for non-stationary contextual bandits: Efficient, optimal and parameter-free. In Conference on Learning Theory, pp. 696 726. PMLR, 2019. Yutian Chen, Aja Huang, Ziyu Wang, Ioannis Antonoglou, Julian Schrittwieser, David Silver, and Nando de Freitas. Bayesian optimization in Alpha Go. ar Xiv:1812.06855, 2018. Sayak Ray Chowdhury and Aditya Gopalan. On kernelized multi-armed bandits. In International Conference on Machine Learning. PMLR, 2017. Lionel Colliandre and Christophe Muller. Bayesian optimization in drug discovery. High Performance Computing for Drug Discovery and Biomedicine, 2023. Marcello Colombino, Emiliano Dall Anese, and Andrey Bernstein. Online optimization as a feedback controller: Stability and tracking. IEEE Transactions on Control of Network Systems, 7(1), 2019. Thomas Dangl and Michael Halling. Predictive regressions with time-varying coefficients. Journal of Financial Economics, 106(1):157 181, 2012. Yuntian Deng, Xingyu Zhou, Baekjin Kim, Ambuj Tewari, Abhishek Gupta, and Ness Shroff. Weighted Gaussian process bandits for non-stationary environments. In Proceedings of The 25th International Conference on Artificial Intelligence and Statistics, volume 151. PMLR, 2022. Matthew Ellis and Panagiotis D Christofides. Economic model predictive control with time-varying objective function for nonlinear process systems. AICh E Journal, 60(2), 2014. David Eriksson, Michael Pearce, Jacob Gardner, Ryan D Turner, and Matthias Poloczek. Scalable global optimization via local Bayesian optimization. Advances in Neural Information Processing Systems, 2019. Peter I Frazier and Jialei Wang. Bayesian optimization for materials design. Information Science for Materials Discovery and Design, 2016. Jacob Gardner, Geoff Pleiss, Kilian Q Weinberger, David Bindel, and Andrew G Wilson. GPy Torch: Blackbox matrix-matrix Gaussian process inference with GPU acceleration. In Advances in Neural Information Processing Systems, 2018. Roman Garnett. Bayesian Optimization. Cambridge University Press, 2023. Roman Garnett, Michael A Osborne, and Stephen J Roberts. Bayesian optimization for sensor set selection. International Conference on Information Processing in Sensor Networks, 2010. Subhashis Ghosal and Anindya Roy. Posterior consistency of Gaussian process prior for nonparametric binary regression. The Annals of Statistics, 34(5):2413 2429, 2006. GPy. GPy: A gaussian process framework in python. http://github.com/Sheffield ML/GPy, 2012. Ryan-Rhys Griffiths and José Miguel Hernández-Lobato. Constrained Bayesian optimization for automatic chemical design using variational autoencoders. Chemical Science, 11(2), 2020. Philipp Hennig and Christian J. Schuler. Entropy search for information-efficient global optimization. Journal of Machine Learning Research, 13, 2012. Antonia Holzapfel, Paul Brunzema, and Sebastian Trimpe. Event-triggered safe Bayesian optimization on quadcopters. In Learning for Dynamics and Control Conference. PMLR, 2024. Published in Transactions on Machine Learning Research (01/2025) Andreas Krause and Cheng Ong. Contextual Gaussian process bandit optimization. In Advances in Neural Information Processing Systems. Curran Associates, Inc., 2011. Christopher König, Matteo Turchetta, John Lygeros, Alisa Rupenyan, and Andreas Krause. Safe and efficient model-free adaptive control via Bayesian optimization. In IEEE International Conference on Robotics and Automation, 2021. Roman Marchant and Fabio Ramos. Bayesian optimisation for intelligent environmental monitoring. In IEEE/RSJ International Conference on Intelligent Robots and Systems. IEEE, 2012. Roman Marchant, Fabio Ramos, and Scott Sanner. Sequential Bayesian optimisation for spatial-temporal monitoring. In Conference on Uncertainty in Artificial Intelligence. Citeseer, 2014. Sarah Müller, Alexander von Rohr, and Sebastian Trimpe. Local policy search with Bayesian optimization. Advances in Neural Information Processing Systems, 34:20708 20720, 2021. Matthias Neumann-Brosig, Alonso Marco, Dieter Schwarzmann, and Sebastian Trimpe. Data-efficient autotuning with Bayesian optimization: An industrial control study. IEEE Transactions on Control Systems Technology, 28(3), 2019. Christos H Papadimitriou and John N Tsitsiklis. The complexity of optimal queueing network control. In Conference on Structure in Complexity Theory. IEEE, 1994. Jack Parker-Holder, Vu Nguyen, and Stephen J Roberts. Provably efficient online hyperparameter optimization with population-based bandits. In Advances in Neural Information Processing Systems, 2020. Jack Parker-Holder, Vu Nguyen, Shaan Desai, and S Roberts. Tuning mixed input hyperparameters on the fly for efficient population based Auto RL. In Advances in Neural Information Processing Systems, volume 34, pp. 15513 15528. Curran Associates, Inc., 2021. Giorgio E Primiceri. Time varying structural vector autoregressions and monetary policy. The Review of Economic Studies, 72(3):821 852, 2005. Ali Rahimi and Benjamin Recht. Random features for large-scale kernel machines. Advances in Neural Information Processing Systems, 20, 2007. Andrea Simonetto, Emiliano Dall Anese, Santiago Paternain, Geert Leus, and Georgios B Giannakis. Timevarying convex optimization: Time-structured algorithms and applications. Proceedings of the IEEE, 108 (11), 2020. Aleksandrs Slivkins and Eli Upfal. Adapting to a changing environment: the Brownian restless bandits. In Conference on Learning Theory, 2008. Jasper Snoek, Hugo Larochelle, and Ryan P Adams. Practical Bayesian optimization of machine learning algorithms. Advances in Neural Information Processing Systems, 2012. Friedrich Solowjow and Sebastian Trimpe. Event-triggered learning. Automatica, 117, 2020. Niranjan Srinivas, Andreas Krause, Sham Kakade, and Matthias Seeger. Gaussian process optimization in the bandit setting: No regret and experimental design. In International Conference on Machine Learning, 2010. Jie Su, Junfeng Wu, Peng Cheng, and Jiming Chen. Autonomous vehicle control through the dynamics and controller learning. IEEE Transactions on Vehicular Technology, 67(7):5650 5657, 2018. Joe Suk and Samory Kpotufe. Tracking most significant arm switches in bandits. In Conference on Learning Theory. PMLR, 2022. Jonas Umlauft and Sandra Hirche. Feedback linearization based on Gaussian processes with event-triggered online learning. IEEE Transactions on Automatic Control, 65(10), 2019. Published in Transactions on Machine Learning Research (01/2025) Chen-Yu Wei and Haipeng Luo. Non-stationary reinforcement learning without prior knowledge: An optimal black-box approach. In Conference on Learning Theory, pp. 4300 4354. PMLR, 2021. Peter Whittle. Restless bandits: Activity allocation in a changing world. Journal of Applied Probability, 25 (A):287 298, 1988. Lijun Zhang, Shiyin Lu, and Tianbao Yang. Minimizing dynamic regret and adaptive regret simultaneously. In International Conference on Artificial Intelligence and Statistics, pp. 309 319. PMLR, 2020. Xingyu Zhou and Ness Shroff. No-regret algorithms for time-varying Bayesian optimization. In Annual Conference on Information Sciences and Systems, pp. 1 6. IEEE, 2021. Published in Transactions on Machine Learning Research (01/2025) Supplementary Material for Event-Triggered Time-Varying Bayesian Optimization A Additional Results and Details on Experimental Design 13 A.1 Creating Within-Model Objective Functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13 A.2 Additional Results on Within-Model Objective Functions . . . . . . . . . . . . . . . . . . . . 13 A.3 Comparison to Performance-based Triggers . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13 A.4 Performance on Standard Optimization Benchmarks . . . . . . . . . . . . . . . . . . . . . . . 15 A.5 Comparisons to Algorithms in the Finite Variational Budget Setting . . . . . . . . . . . . . . 16 B Sensitivity of ET-GP-UCB to the Event Trigger Hyperparameter δB 19 C Discussion on the Theoretical Analysis and Future Directions 20 D Influence of the window size on the distribution of resets for different ε 21 E Overview of the Temperature Data 22 F Proof of Theorem 1 and Corollary 1 from Section 4 22 F.1 Proof of Theorem 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22 F.2 Proof of Corollary 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25 G Proof of Lemma 2 and Lemma 3 27 H Hyperparameters 27 Following is the technical appendix. Note that all citations here are in the bibliography of the main document, and similarly for many of the cross-references. Published in Transactions on Machine Learning Research (01/2025) A Additional Results and Details on Experimental Design This section contains details about the design of the experiment for the results in Section 6.2 for reproducibility. A.1 Creating Within-Model Objective Functions To create the within-model objective functions, we take two different approaches. For the two-dimensional objectives in Section 6.2.1, we draw sample paths gi from a GP with the corresponding hyperparameters on a dense equidistant grid of the compact set X = [0, 1]2 with 100 100 input locations. To create the time-varying objective function ft we then combine the sample paths following Assumption 2 with a specified rate of change ε and perform linear interpolation in the dense grid. For the higher-dimensional within-model objective functions in Section 6.3, this procedure is not feasible. Therefore, we approximate prior samples gi with a M random Fourier features (RFFs) following Rahimi & Recht (2007). This yields a set of parametric functions m=1 wmϕm(x) with ϕm(x) = 2 M cos(θ mx + τm). (12) Here, θm are sampled proportional to the kernel s spectral density and τm U(0, 2π). These approximate sample paths again can be combined following Assumption 2 to yield the objective function ft(x) given ε. In all experiments in Section 6.3, we use M = 1028 RFFs. This approach does not require interpolation in the input space since the approximate sample paths are parametric. A.2 Additional Results on Within-Model Objective Functions ε = 0.002 ε = 0.1 GP-UCB R-GP-UCB TV-GP-UCB ET-GP-UCB 100 200 300 400 500 Time t Resets of respective algorithm Resets of respective algorithm Figure 6: Within-model with change in ε. Here, we list some of the additional results on within-model objective functions. On the right in Figure 6, we conducted an experiment with a change in the rate of change after 175 time steps (segment highlighted in gray). We initialize all algorithms that require epsilon as an input to their algorithm with the rate of change at time step 0. In this experiment, the adaptive nature of our algorithm becomes evident. After the change in the rate of change ε, ET-GP-UCB starts to increase the reset frequency adjusting to the higher rate of change in the unknown objective. We can observe how TV-GP-UCB with the too small rate of change starts to diverge over time. While the difference for 500 time steps is not as big, there is a clear diverging trend. We next further visualize the results from Table 1 in Figure 7. Here, we only list the best-performing variant of ET-GP-UCB to increase the readability of the plot. Lastly, Figure 8 shows the results of using Algorithm 2 to keep some of the past information for different rate of changes. We can observe that using the backtracking procedure improves the performance for all dimensions for the given rate of changes. A.3 Comparison to Performance-based Triggers Another idea that arises naturally for an event trigger is to consider a reset strategy that is solely based on an improvement in performance. This strategy is inspired by trust-region BO, Tu RBO (Eriksson et al., 2019), where a lack of improvement triggers a refinement of the search domain. We apply a variant of this mechanism to the time-varying setting: If an algorithm is not able to improve its performance for τfail N+ time steps in a row, the event trigger resets the data set. To set this threshold, we use the same heuristic proposed in the Tu RBO paper and set τfail = D/q = D (c.f. Appendix D in Eriksson et al. (2019) and our batch size is q = 1 given our problem setting). We refer to this event trigger as Tu RBO Trigger. To test its performance, we ran the same benchmark from the previous section, i.e., multiple within-model functions generated using RFFs (cf. Section A.1) with different rates of change and different dimensions. The results are in Figure 9. Published in Transactions on Machine Learning Research (01/2025) ε = 0.03 ε = 0.05 εtrue = 0.05 εmissp. = 0.2 GP-UCB R-GP-UCB TV-GP-UCB UI-TVBO ET-GP-UCB 0 200 400 Time steps t (a) Constant ε 0 200 400 Time steps t (b) Constant ε 0 200 400 Time steps t (c) Constant ε 0 200 400 Time steps t (d) Misspecified ε Resets of R-GP-UCB Resets of ET-GP-UCB Resets of R-GP-UCB Resets of ET-GP-UCB Figure 7: Top row: performance of the different algorithms in terms of normalized regret. The shaded areas show one standard deviation of the respective algorithms. Bottom row: reset timings of R-GP-UCB and ET-GP-UCB. In (a) to (c), rate of change is fixed for all time steps, and in (d), it is misspecified for TV-GPUCB, UI-TVBO, and R-GP-UCB. In (a) to (c), TV-GP-UCB is the full information reference, showing the best possible UCB performance with known constant ε. We compare ET-GP-UCB to R-GP-UCB, the latter of which uses the true ε to determine its periodic reset time. For ET-GP-UCB, we use the same parameter δB in the event trigger for the test cases (a) to (d). In all cases, ET-GP-UCB outperforms R-GP-UCB. It also outperforms TV-GP-UCB and UI-TVBO if ε is misspecified, as shown in (d), since ET-GP-UCB does not rely on prior information on ε. The plots show the median performance and the interquartile. 2 4 6 8 Input dimensions d (ε = 0.01) t ft(xt)/t ( ) 2 4 6 8 Input dimensions d (ε = 0.03) R-GP-UCB TV-GP-UCB ET-GP-UCB ET-GP-UCB w. backtracking 2 4 6 8 Input dimensions d (ε = 0.05) Figure 8: Benefit of reusing some of the past information through Algorithm 2 for different ε. We can observe that for all rates of change, the Tu RBO trigger performs worse than the other baselines. Unlike in Tu RBO, a trigger event significantly impacts performance since all data is discarded (while shrinking the trust region for Tu RBO is less drastic). This is especially the case for lower-dimensional objective functions where we only allow for a few failures before the dataset is reset. An additional factor is that the objective function is noisy. Therefore, even consecutive evaluations at the optimum of a non-changing function could yield a reset, which, in contrast to Tu RBO, requires significant new data to recover from. While performance-based resets hold promise as they are not as tightly coupled to the prior assumptions of the GP, they require further in-depth investigation and algorithm design for the dynamic settings to yield well-performing approaches by further including heuristics. Also, further tuning of τfail for each problem could improve its performance. Published in Transactions on Machine Learning Research (01/2025) 2 4 6 8 Input dimensions d (ε = 0.01) t ft(xt)/t ( ) 2 4 6 8 Input dimensions d (ε = 0.03) R-GP-UCB TV-GP-UCB ET-GP-UCB ET-GP-UCB w. backtracking Tu RBO Trigger 2 4 6 8 Input dimensions d (ε = 0.05) Figure 9: Comparison to a performance-based event trigger (Tu RBO Trigger). A.4 Performance on Standard Optimization Benchmarks Next, we will test our event trigger on some standard benchmarks, specifically Ackley (d = 2) and Hartmann (d = 6), which are popular for testing BO algorithms. The event trigger in ET-GP-UCB is tightly coupled to the GP prior. We, therefore, expect some false positives of our event trigger on these benchmarks. Similar to Section 6.2.2 and Section 6.2.3, the comparisons in this section aim to test to what extent resets affect the empirical performance. For Ackley, we set the time horizon to T = 60 and induce a shift in the optimum after T/2 time steps. Similarly, we set Hartmann s time horizon to T = 200 and induce the shift after T/2 time steps. To induce this shift, we perturb the input of the objective function by a constant factor. We use the same UCB acquisition function stated in Section 6, run ten different seeds, and learn the lengthscales ℓi online for all baselines. For the lengthscales, we use corse bounds as ℓi [0.005, 4.] (Eriksson et al., 2019) and standardize the output data at each time step given the current data set which follows best practices in BO (Balandat et al., 2020). Since we learn the lengthscales online, we apply the learnthen-monitor approach introduced Section 6.3 for the ET-GP-UCB baselines and use the same δB as in all previous experiments. For TV-GP-UCB, we set the rate of change to ε = 0.01 to yield good performance in the static setting but still allow for some adaptation. Figure 10 shows the performance of all baselines. For Ackley, we can observe that at the beginning of optimization, there are some resets of both ET-GP-UCB variants. However, we can also observe that these resets do not significantly impact the final performance (before the change) when considering simple regret. After the shift, we can observe that all the time-varying baselines can adjust to the change in the objective, and as expected, GP-UCB struggles to cope with the stale data in the data set. On Hartmann, the event-triggered baseline without backtracking struggles and is outperformed by GP-UCB. Similar to Ackley, we can observe early resets of both ET-GP-UCB variants. In this higher-dimensional example, these resets are more costly for vanilla ET-GP-UCB. For the backtracking variant, we can observe consecutive resets resulting in an event-triggered sliding window (cf. Section 6.3). With this backtracking mechanism, the ET-GP-UCB variant outperforms GP-UCB significantly in terms of simple and cumulative regret. Similar to the Ackley example, most of the time-varying baselines can better adjust to the shift in the optimum compared to GP-UCB. Noteworthy is also that ET-GP-UCB with backtracking significantly outperforms GP-UCB in terms of simple regret even before the change. Removing some initial data allows this variant to achieve better GP hyperparameters and converge to the optimum. This result raises further questions on how classic BO algorithms should deal with past data for better convergence. In summary, we observed that ET-GP-UCB does yield additional resets when optimizing an objective function that is not well explained by the GP prior, but the impact on final performance is relatively small for the considered test functions when leveraging the proposed backtracking mechanism. We further observed that this mismatch in prior and objective functions can also harm static algorithms. Additionally, we observe Published in Transactions on Machine Learning Research (01/2025) Ackley (2D) Best value ( ) GP-UCB TV-GP-UCB ET-GP-UCB ET-GP-UCB w. backtracking Hartmann (6D) Best value ( ) GP-UCB TV-GP-UCB ET-GP-UCB ET-GP-UCB w. backtracking Ackley (2D) Hartmann (6D) 10 20 30 40 50 60 Time t Resets of respective algorithm Resets of respective algorithm 50 100 150 200 Time t Resets of respective algorithm Resets of respective algorithm Figure 10: Comparison on the standard benchmarks Ackley (left) and Hartmann (right). We show the best obtained value on top and the normalized cumulative regret on the bottom. We reset the best value an algorithm as observed after the shift (indicated by the dashed line) to test how the algorithms cope with the shift in the optimum. All plots show the median performance as well as the interquantiles. that the same algorithms developed to tackle time-varying problems may also be helpful for static problems to overcome some problems with past data points, which is an exciting research direction. A.5 Comparisons to Algorithms in the Finite Variational Budget Setting In this section of the appendix, we compare ET-GP-UCB to W-GP-UCB (Deng et al., 2022), and BRGP-UCB and SW-GP-UCB (Zhou & Shroff, 2021). As discussed in Section 3, their underlying assumptions differ from ours in that they assume a fixed variational budget. Therefore, a theoretical comparison is not reasonable, but an empirical comparison, especially on real-world data, is interesting. As the baseline that does not consider time-variations, we also compare against IGP-UCB (Chowdhury & Gopalan, 2017), which is an improved form of the agnostic GP-UCB of Srinivas et al. (2010). For the comparison, we use the code base submitted by Deng et al. (2022) in the supplementary material of https://proceedings.mlr.press/v151/deng22b. Their experiments consist of: (a) an one-dimensional experiment with sudden changes at time steps t = 100 and t = 200, (b) an one-dimensional experiment with a slowly changing objective function, (c) an experiment on real-world stock market data.4 We refer to Deng et al. (2022) for the details on how the objective function for (a) and (b) is composed. We implemented our algorithm ET-GP-UCB in their Matlab code base and added random search as a baseline. While going through the implementation details, we noted that in the stock market data benchmark, the data was not normalized and values of the objective function ranged from 19 to 250. This is problematic as all algorithms use a prior mean of zero and a prior variance of one. Thus, all observations will have a very small likelihood under the given model. Secondly, the noise variance was set to 300, which seems high for stock data where observation noise should be low. The high noise variance combined with the prior signal variance of one also explains the very high standard deviations of all the algorithms in Deng et al. (2022, 4Note that stock market data does not fit the fixed variational budget setting as stock prices constantly change. However, since this example was used in prior work, we treat it as a benchmark but wanted to highlight that ongoing change is the more suitable assumption. Published in Transactions on Machine Learning Research (01/2025) 0 100 200 300 400 500 (a) Abruptly-change, SE kernel 0 100 200 300 400 500 (b) Slowly-change, SE kernel 0 200 400 600 800 Time t (c) Stock market data Random IGP-UCB BR-GP-UCB SW-GP-UCB W-GP-UCB ET-GP-UCB (ours) Figure 11: Performance using the refactored GP prior in the experiments of Deng et al. (2022). Figure 1 (c)). Overall, the stock market data has a very low marginal likelihood for the model chosen in Deng et al. (2022). Therefore, we decided to improve the model and add a preprocessing step to the stock market data example by normalizing the data to a mean of zero and a variance of one to fit the prior model assumptions of all algorithms. Next, we chose a noise variance of 0.01 to account for smaller short-term fluctuations in stock prices. We also changed the noise variance of the synthetic experiments (a) and (b) to 0.1. It was previously set to 1 (same as the signal variance), amounting to a signal to noise ratio of 1, which is very low for practical optimization problems. We then re-ran their experiments using 20 runs for the synthetic experiments and 5 runs for the stock market data as in Deng et al. (2022). For ET-GP-UCB, we use the same parameters for the event trigger as in all previous experiments, as well as no hard upper and lower bounds, as this resulted in the best empirical performance in Section 6.2. Note that all other time-varying algorithms rely on specifying the amount of change BT a priori, which is hard to estimate in practice. The results are shown in Figure 11 and regret at the last time step is listed in Table 3. We can see that ET-GP-UCB can outperform all baselines in experiments Figure 11 (a) to (c). Especially in Figure 11 (a), ET-GP-UCB can detect the sudden changes in the objective function and then quickly find the new optimum due to the simple one-dimensional objective function. In Figure 11 (b), ET-GP-UCB is the best performing algorithm, but W-GP-UCB is competitive. Considering the good performance of IGP-UCB, we can also conclude that the changes in the objective function over the time horizon are relatively small. In Figure 11 (c), ET-GP-UCB is again the best-performing algorithm. We also observe that in Figure 11 (c) all algorithms significantly improve in performance and yield a smaller standard deviation compared to Deng et al. (2022, Figure 1 (c)) because the GP prior model is more suitable for modeling the objective function. After approx. 500 time steps ET-GP-UCB, W-GP-UCB, SW-GP-UCB, and IGP-UCB (see Table 3), no longer accumulate regret. Looking at the objective function, this is expected as the optimal bandit arm no longer changes and the optimal values do not change significantly. Table 3: Comparison on experiments in Deng et al. (2022). (a) Abruptly-change (b) Slowly-change (c) Stock market Algorithm Regret (µ σ) Regret (µ σ) Regret (µ σ) IGP-UCB 704.6 238.9 136.0 32.2 11 462.3 56.0 BR-GP-UCB 312.5 4.3 153.9 11.0 45 362.2 430.1 SW-GP-UCB 233.2 9.1 111.6 15.7 17 854.9 27.9 W-GP-UCB 298.8 27.3 197.0 29.7 11 461.3 32.9 ET-GP-UCB (ours) 55.8 2.7 86.3 42.8 2 483.8 39.8 Published in Transactions on Machine Learning Research (01/2025) Results with the Original Gaussian Process Prior Below in Figure 12 are the results using the original GP prior from Deng et al. (2022). We can observe that for Figure 12 (a), ET-GP-UCB can reliably detect the first change in the objective function at t = 100. However, due to the low signal-to-noise ratio, ET-GP-UCB can only detect the second change approximately 50% of the time. It still outperforms periodic resetting and the sliding window approach. In Figure 12 (b), ET-GP-UCB shows a similar performance as W-GP-UCB. Both examples, however, highlight that ET-GPUCB is not as dominant as in previous examples if the signal-to-noise ratio is too low since wt will be large relative to |yt µDt(xt)|. For the stock market data, we observe the discussed large standard deviations. All algorithms outperform the random baseline. However, we can observe that the ordering of algorithms has changed compared to Deng et al. (2022, Figure 1 (c)) even though we used their implementations and only added ET-GP-UCB and the random baseline to the run script. Further, changing the random seed again changed this ordering of the algorithms, which indicates that the prior model assumptions yield random-like behavior as the prior does not allow the GP model to make meaningful decisions. 0 100 200 300 400 500 (a) Abruptly-change, SE kernel 0 100 200 300 400 500 (b) Slowly-change, SE kernel 250 500 750 Time t (c) Stock market data Random IGP-UCB BR-GP-UCB SW-GP-UCB W-GP-UCB ET-GP-UCB (ours) Figure 12: Performance using the original GP prior for the experiments of Deng et al. (2022). Published in Transactions on Machine Learning Research (01/2025) 0 200 400 Time t 0 200 400 Time t δB = 0.005 δB = 0.01 δB = 0.05 δB = 0.1 δB = 0.5 0 200 400 Time t Figure 13: Sensitivity of our event trigger from (11) with respect to its hyperparameter δB. B Sensitivity of ET-GP-UCB to the Event Trigger Hyperparameter δB Table 4 and Figure 13 show the sensitivity of ET-GP-UCB to the event trigger parameter δB in Lemma 3. To solely focus on the influence of δB with use the maximum window size as W = T. It should be noted, that for different window sizes, different δB may be optimal for the best possible empirical performance as both, the location of lower and upper reset bounds, and the hyperparameter δB, influence how much (and where) probability mass from the distribution over reset times is cut off. For each δB, we ran 50 different seeds on within-model objective functions (cf. Section A.1). All hyperparameters are listed in Table 12. As δB is increased, the average number of resets increases as the probabilistic uniform error bound is softened. We observe that for different rates of changes ε different δB may be optimal to obtain the lowest possible empirical regret. From Figure 13, we can observe that for higher rates of change a smaller δB can yield a smaller variance in the final regret. In summary, ET-GP-UCB displays a low sensitivity to δB highlighting that our algorithm is readily applicable without the need for excessive hyperparameter tuning. In practice, we found δB = 0.1 to yield good empirical performance in all of our experiments. Table 4: Sensitivity of ET-GP-UCB to δB in Lemma 3. Rate of Change ε ET-GP-UCB Hyperparameter δB Avg. No. of Resets Avg. Regret RT 0.005 2.66 191.39 0.01 2.96 193.05 ε = 0.01 0.05 3.20 200.16 0.1 3.38 200.33 0.5 3.98 196.03 0.005 6.42 276.84 0.01 6.82 273.37 ε = 0.03 0.05 7.72 269.01 0.1 8.04 271.59 0.5 10.32 280.05 0.005 9.60 331.69 0.01 9.96 328.74 ε = 0.05 0.05 11.40 329.47 0.1 11.88 332.04 0.5 14.44 334.80 Published in Transactions on Machine Learning Research (01/2025) C Discussion on the Theoretical Analysis and Future Directions The current theoretical analysis does not explicitly include the event trigger (11). As stated in Section 7, we are interested in tighter regret bounds, tailored to the specific event trigger. This has the potential for additional insights into the design of new triggers. In the following, we list some of the challenges that may arise when conducting this theoretical investigation and give some pointers that may be helpful to derive tighter bounds. Reset times are stopping times. First, note that the activation time of the event trigger is a random variable as τ := min{t N : ψt(xt) > κt(xt)}. This random variable τ is a stopping time, as ψt(xt) is a discrete-time stochastic process and the trigger event is completely determined by the information known up to t. The stochastic process depends recursively on the acquisition function, which makes an analytical study of the stopping time difficult; even for simple stochastic processes, calculating expected values of stopping times is typically intractable. The regret analysis should be specific for the event trigger. There is some information available on τ: As long as T is finite, τ is a bounded random variable as τ [1, T]. This also means, that τ is a sub-Gaussian random variable. One can use concentration inequalities of sub-Gaussian random variables to bound the maximum of τ with high probability and then use this bound in the theoretical analysis. For the analysis to be specific for the event trigger, this concentration inequality has to depend on the test and threshold functions of the event trigger which, as stated above, is in itself very challenging: Random sampling block lengths within [1, T] have the same sub-Gaussian property, i.e., with the same sub-Gaussian constant. However, the empirical performance would be very different compared to a sophisticated event trigger. Using standard bounds for sub-Gaussian random variables would therefore likely still result in a mismatch between empirical performance and theoretical analysis. Bounding the random block sizes and number of block simultaneously is promising. Once a event trigger specific bound on the stopping time is established, the theoretical analysis would still require a fundamentally new way of quantifying the model mismatch compared to Theorem 1. The current analysis bounds the maximum model mismatch by considering the maximum block length at every time step. Here, one could substitute in the bound on the stopping time τ (which depends on T). However, the bound only holds with high probability and, therefore, for a valid probabilistic statement, a union bound over all time steps is required. This union bound can result in a super-linear scaling of the regret bound. An interesting direction for future work would be instead of considering the maximum model mismatch at each time step to bound the composition of the number of blocks and size of the blocks simultaneously. Published in Transactions on Machine Learning Research (01/2025) D Influence of the window size on the distribution of resets for different ε In Figure 2 (left), we displayed the influence of the window size on the distribution of reset times. In Figure 14, we show this influence also for different ε. We further add the empirical cumulative distribution in the right column. 0 50 100 150 200 Reset time (ε = 0.03) Window size W Constant reset time of R-GP-UCB 27 55 82 110 137 165 193 0 50 100 150 200 Cum. reset time (ε = 0.03) Window size W 0 27 55 82 110 137 165 193 Constant reset time of R-GP-UCB 0 50 100 150 200 Reset time (ε = 0.05) Window size W Constant reset time of R-GP-UCB 28 55 83 110 138 165 193 0 50 100 150 200 Cum. reset time (ε = 0.05) Window size W 0 28 55 83 110 138 165 193 Constant reset time of R-GP-UCB 0 50 100 150 200 Reset time (ε = 0.10) Window size W Constant reset time of R-GP-UCB 27 55 82 110 137 165 192 0 50 100 150 200 Cum. reset time (ε = 0.10) Window size W 0 27 55 82 110 137 165 192 Constant reset time of R-GP-UCB Figure 14: Left column: Influence of the window size on the empirical PDF of reset times. Right column: Influence of the window size on the empirical CDF. Published in Transactions on Machine Learning Research (01/2025) E Overview of the Temperature Data In Figure 15, we show the normalized temperature data used for the empirical kernel in the results in Figure 3. The full pre-processing of the temperature data is described in the corresponding jupyter-notebook in the provided code base. 0 24 48 72 96 Time t [h] Temperature data Norm. Temperature [ C] Day 3 Day 4 Day 5 Day 6 Figure 15: Normalized temperature data used for the results in Figure 3. Each color resembles the measurement data from one sensor. F Proof of Theorem 1 and Corollary 1 from Section 4 F.1 Proof of Theorem 1 Restated Theorem 1 Let the domain X [0, r]d Rd be convex and compact with d N+ and let ft follow Assumption 2 with a kernel k(x, x ) such that Assumption 3 is satisfied. Pick δ (0, 1) and set βt = 2 ln 2π2t2/(3δ) + 2d ln (t2db1r p ln(2da1π2t2/(3δ))). Pick a lower block length N N+ and a upper block length N N+ with N N T. Let γ N be the maximum information gain over N points. Then, running any algorithm with block sizes between N and N satisfies N + 1 γ N + 2 + TϕT (ε, N) (13) with probability at least 1 δ, where C1 = 8/ ln(1 + σ 2 n ) and we defined ϕT (ε, N) := 2 q βT (3σ 2 n + σ 4 n ) N 3ε + 2(σ 2 n + σ 4 n ) N 3ε(b0 + ln 4(a0+1)π2T 2 Proof: The proof of Theorem 1 is mainly based on the proof of Bogunovic et al. (2016, Theorem 4.2) which builds on arguments from Srinivas et al. (2010). The key distinction of our analysis is that we will consider the block length not to be fixed but within an interval [ N, N]. For completeness, we list the full proof here. As in Bogunovic et al. (2016) and Srinivas et al. (2010), we first fix a discretization Xt X [0, r]d of size (ϑt)d satisfying x [x]t rd/ϑt, x Xt (15) where [x]t denotes the closest point in Xt to x. Now fix a constant δ > 0 and an increasing sequence of positive constants {πt} t=1 satisfying P t 1 π 1 t = 1 (cf. Remark 2). Next, we introduce some Lemmas needed to bound the regret. Published in Transactions on Machine Learning Research (01/2025) Lemma 4 (Bogunovic et al. (2016, Proof of Theorem 4.2, cf. Appendix C.1)) Pick δ (0, 1) and set βt = 2 ln πt δ , where P t 1 π 1 t = 1, πt > 0. Then we have P n |ft(xt) µDt(xt)| p βt σDt(xt), t 1 o 1 δ. (16) Lemma 5 (Bogunovic et al. (2016, Proof of Theorem 4.2, cf. Appendix C.1)) Pick δ (0, 1) and set βt = 2 ln |Xt|πt δ , where P t 1 π 1 t = 1, πt > 0. Then we have P n |ft(xt) µDt(xt)| p βt σDt(xt), t 1, x Xt o 1 δ. (17) Lemma 6 (Bogunovic et al. (2016, Proof of Theorem 4.2, cf. Appendix D)) Pick δ (0, 1) and set Lt = b1 q δ , where P t 1 π 1 t = 1, πt > 0. Then, with Assumption 3 we have > Lt, t 1, x Xt, j {1, . . . , d} 1 δ. (18) Lemma 7 Pick δ (0, 1) and set yt = (b0 + ln 2(a0+1)πt δ , where P t 1 π 1 t = 1, πt > 0. Then we have P {|yt| yt, t 1} 1 δ. (19) Proof: We have that |yt| |ft(xt)| + |wt| given Assumption 1. We can use Lemma 2 (with t instead of tr) with δ/2 to bound |wt| for all time steps. Combined with Assumption 3 we have P {|yt| Lf + wt} 1 a0 exp{ L2 f/b2 0} δ/2. (20) Solving the remaining term with δ/2 and using the union bound, we get y t = Lf + wt := b0 q wt. Here, we can further bound b0 q δ with b0 q ln 2(a0+1)πt δ using the monotonicity of the logarithm (recall that a0 > 0). Since we have wt = δ , we can use the same trick to state that wt ln 2(a0+1)πt δ . To end, we can concatenate the terms obtaining y t yt := (b0 + ln 2(a0+1)πt δ proving the claim. Remark 3 Note that Lemma 7 slightly differs from the bound used in Bogunovic et al. (2016). We explicitly consider the variance of the measurement noise to bound the observations in probability. For Theorem 1, we condition on four high probability events i.e., Lemma 4, Lemma 5, Lemma 6, and Lemma 7 each with with δ/4. Conditioned on these events, the following Lemma holds. Lemma 8 (Bogunovic et al. (2016, Appendix D)) Let Lemma 4, Lemma 5, and Lemma 6 each hold with probability 1 δ/4. Pick δ (0, 1) and set βt = 2 ln 4πt + 2d ln rdt2b1 p ln(4a1dπt/δ) , t 1 π 1 t = 1, πt > 0. Let ϑt = rdt2b1 p ln(2a1dπt/δ) and let [x ]t be the closest point in Xt to x . Then we have P |ft(x t ) µDt([x t ]t)| p βt σDt([x t ]t) + 1 t2 , t 1 1 δ. (21) In the following, we will explicitly highlight the usage of each Lemma. We first bound the instantaneous regret at a time step t as rt = ft(x t ) ft(xt) (22) µDt([x t ]t) + p βt σDt([x t ]t) + 1 t2 µDt(xt) + p βt σDt(xt). (Lemma 4, Lemma 8) (23) Published in Transactions on Machine Learning Research (01/2025) This is identical to Bogunovic et al. (2016, Proof of Theorem 4.2) for R-GP-UCB. We now account for model mismatch between using the time-invariant model in the algorithm and the true" time-varying posterior as µDt(x) µDt(x) + (µ) Dt with (µ) Dt := sup x X {| µDt(x) µDt(x)|} (24) σDt(x) σDt(x) + (σ) Dt with (σ) Dt := sup x X {| σDt(x) σDt(x)|} . (25) We can thus further bound the regret as rt µDt([x t ]t) + (µ) Dt ([x t ]t) + p βt σDt([x t ]t) + (σ) Dt ([x t ]t) (26) µDt(xt) + (µ) Dt (xt) + p βt σDt(xt) + (σ) Dt (xt) + 1 Per definition of Algorithm 1, we have that through the optimization of its UCB-type acquisition function µDt([x t ]t) + βtσDt([x t ]t) µDt(xt) + βtσDt(xt), and, therefore, βtσDt(xt) + 1 t2 + (µ) Dt ([x t ]t) + (µ) Dt (xt) + p βt (σ) Dt ([x t ]t) + (σ) Dt (xt) . (28) What is left to bound is the model mismatch.5 For this, we slightly adjust the bounds from Bogunovic et al. (2016, Lemma D.1.) to the setting where the block size can vary but is upper bounded by N. Lemma 9 (Adjusted from Bogunovic et al. (2016, Lemma D.1.)) Conditioned on the event in Lemma 7, we have for any block size N N that the maximum model mismatch in the mean and variance between the time-varying GP model and the time-invariant GP model is a.s. bounded as (σ) Dt q 3σ 2 n + σ 4 n N 3ε (29) (µ) Dt σ 2 n + σ 4 n N 3ε yt. (30) Proof: For establishing the bounds on the model mismatch, we refer to the proof of Bogunovic et al. (2016, Lemma D.1.). If focuses on bounding (24) and (25) by explicitly bounding the difference in posterior distributions of the time-invariant updates in (4) and time-varying updates (5). The bounds in (29) and (30) are monotonic functions in the block length. Hence, the model mismatch is upper bounded by the maximum model mismatch with N = N. With Lemma 9, we can summarize the bound on the model mismatch at each time step as 2 (µ) Dt + 2 p βt (σ) Dt 2 σ 2 n + σ 4 n N 3ε yt + 2 q βt 3σ 2 n + σ 4 n N 3ε =: ϕt(ε, N). (31) For completeness, we follow the steps in Srinivas et al. (2010, Proof of Theorem 2) and Bogunovic et al. (2016, Proof of Theorem 4.2 (R-GP-UCB)) to bound the regret following Definition 1 as βtσDt(xt) + 1 t2 + ϕt(ε, N). (32) With P t=1 1/t2 = π2/6 2 and using Cauchy-Schwarz with PT t=1 xt 2 T PT t=1 x2 t we obtain t=1 4βtσ2 Dt(xt) + 2 + 2 σ 2 n + σ 4 n N 3ε yt + 2 q βt 3σ 2 n + σ 4 n N 3ε . (33) 5We noticed a minor mistake in Bogunovic et al. (2016, Theorem 4.2). There, it should be 2 TψT (N, ϵ) in Eq. (17) as model differences in mean and variance have to be bounded twice as it is the case for the variance in our proof (cf. Bogunovic et al. (2016, Proof of Theorem 4.2), Eq. (72)). There are no consequences from the missing constant but it might be helpful for future work to state it here explicitly. Published in Transactions on Machine Learning Research (01/2025) Furthermore, using the same arguments as in Bogunovic et al. (2016, Proof of Theorem 4.2 (R-GP-UCB)) and Srinivas et al. (2010, Proof of Lemma 5.3 (GP-UCB)) on the information gain within a block of size M is I(y M; f M) = 1 2 PM t=1 ln 1 + σ 2 n σ2 Dt(xt) γM with γM = maxx1,...x M I(y M; f M) yields t=1 4βtσ2 Dt(xt) 4βT σ2 n σ 2 n σ2 Dt(xt) (34) 4βT σ2 n C2 t=1 ln 1 + σ 2 n σ2 Dt(xt) (35) with C2 = σ 2 n / ln(1 + σ 2 n ) 1, since s2 C2 ln(1 + s2) for s [0, σ2 n], and σ 2 n σ2 Dt(xt) σ 2 n k(xt, xt) σ 2 n due to the bounded kernel function in Assumption 2 (also see Srinivas et al. (2010, Proof of Lemma 5.4), we get t=1 4βtσ2 Dt(xt) 8βT σ2 n C2 T N I(y N; f N) 8βT σ2 n C2 T N γ N. (36) Here, the lower reset bound N becomes relevant with the following argument. With block lengths as N [ N, N], the maximum number of blocks one could obtain until T is bounded by T N (assuming this is an integer for now). In each of the blocks, the maximum information gain is bounded by γ N since the information gain is non decreasing. We assumed that T N is an integer as the number of block in T and we can upper this with T N + 1 . We therefore have N + 1 γ N + 2 + t=1 2 σ 2 n + σ 4 n N 3ε yt + 2 q βt 3σ 2 n + σ 4 n N 3ε (37) with C1 = 8/ ln(1 + σ 2 n ). Further bounding the sum over maximum model mismatch yields N + 1 γ N + 2 + T 2 σ 2 n + σ 4 n N 3ε y T + 2 q βT 3σ 2 n + σ 4 n N 3ε | {z } ϕT (ε, N) in Theorem 1 F.2 Proof of Corollary 1 Restated Corollary 1 Let the assumptions in Theorem 1 hold. Given a lower and upper bound on the true ε such that ε [ ε, ε] (0, 1), there exist λ1 (0, 1] and λ2 [1, 1 ε) such that ε = λ1 ε and ε = λ2 ε. For the SE kernel, set N = Θ(min{T, ε 1/4}) and N = Θ(min{T, ε 1/4}). Then, RT = O max n T, λ1/8 2 ε1/8T, λ 3/8 1 ε1/8T, λ 3/4 1 ε1/4T o . (38) For the Matèrn kernel (ν > 2), define ξ = d(d+1) 2ν+d(d+1) and set N = Θ(min{T, ε 1/(4 ξ)}) and N = Θ(min{T, ε 1/(4 ξ)}). Then, T 1 ξ, λ ξ 2(4 ξ) 1 λ 1 2(4 ξ) 2 ε 1 ξ 2(4 ξ) T, λ 3 2(4 ξ) 1 ε 1 ξ 2(4 ξ) T, λ 3 4 ξ 1 ε 1 ξ 4 ξ T . (39) Proof: First recall the bound from Theorem 1 and divide it into three terms to increase readability. N + 1 γ N | {z } Term 1 βT 3σ 2 n + σ 4 n N 3ε | {z } Term 2 + 2T σ 2 n + σ 4 n N 3ε y T | {z } Term 3 Published in Transactions on Machine Learning Research (01/2025) In the following, we will first consider the SE kernel and afterwards the general Matérn kernel. We will bound each term separately for both kernels. The proof builds on Bogunovic et al. (2016, Appendix E) but extends it to the setting where only lower and upper bounds on ε are known. In the following, let IT (z) be the closest integer in {1, . . . , T} which is closest to z R. Squared Exponential Kernel Term 1 For the SE kernel, Srinivas et al. (2010, Theorem 5) shows that γ N = O(ln N) = O(1). Setting N = IT ( ε 1/4), we obtain upon substitution s N + 1 γ N = O = O λ1/8 2 ε1/8T . (40) Term 2 For the second term, we have upon substitution of N = IT ( ε 1/4) βT 3σ 2 n + σ 4 n N 3ε = O T p ε 3/4ε = O T q (λ1ε) 3/4ε = O λ 3/8 1 ε1/8T . (41) Term 3 For the third term, we have upon substitution of N = IT ( ε 1/4) 2T σ 2 n + σ 4 n N 3ε y T = O T ε 3/4ε = O T(λ1ε) 3/4ε = O λ 3/4 1 ε1/4T . (42) From these results, we can further see that for the scaling to be sub-linear ε has to be sufficiently small relative to the time horizon T and the lower and upper bounds on ε have to be sufficiently close. Specifically, if ε < min n 1 λ2T 4 , λ3 1 T 4 o we obtain sub-linear scaling, i.e., O( Matérn Kernel (ν > 2) Term 1 For the Matérn Kernel kernel, Srinivas et al. (2010, Theorem 5) shows that γ N = O( N ξ ln N) = O( N ξ) where ξ = d(d+1) 2ν+d(d+1). Setting N = IT ( ε 1/(4 ξ)) and N = IT ( ε 1/(4 ξ)), we have upon substitution s N + 1 γ N = O T T ε 1/(4 ξ) ε ξ/(4 ξ) ! (λ2ε) 1/(4 ξ) (λ1ε) ξ/(4 ξ) ! = O λ ξ 2(4 ξ) 1 λ 1 2(4 ξ) 2 ε 1 ξ 2(4 ξ) T . (45) Term 2 For the second term, we have upon substitution of N = IT ( ε 1/(4 ξ)) βT 3σ 2 n + σ 4 n N 3ε = O T p ε 3/(4 ξ)ε (46) (λ1ε) 3/(4 ξ)ε (47) = O λ 3 2(4 ξ) 1 ε 1 ξ 2(4 ξ) T . (48) Term 3 For the third term, we have upon substitution of N = IT ( ε 1/(4 ξ)) 2T σ 2 n + σ 4 n N 3ε y T = O T ε 3/(4 ξ)ε (49) = O T(λ1ε) 3/(4 ξ)ε (50) = O λ 3 4 ξ 1 ε 1 ξ 4 ξ T . (51) Published in Transactions on Machine Learning Research (01/2025) Similar to the SE kernel, we can also identify the sub-linear region for the Matérn kernel. Specifically, if ε < min λξ 1 λ1 ξ 2 T 4 ξ , λ3(1 ξ) 1 T 4 ξ we obtain sub-linear scaling, i.e., O( T 1 ξ). We can furthermore nicely observe that for ν (therefore ξ 0), the bounds for Terms 1-3 as well as the sub-linear region converges to the bounds of the SE kernel. G Proof of Lemma 2 and Lemma 3 Restated Lemma 2 Pick δ (0, 1) and set w2 tr = 2σ2 n ln πtr δ , where P tr 1 π 1 tr = 1, πtr > 0. Then, the noise sequence in Assumption 1 obtained since the last reset satisfies |wtr| wtr for all tr 1 with probability at least 1 δ. Proof: Since wtr N(0, σ2 n), we can use a standard bound on the tail probability of the normal distribution P {|r| > s} exp s2/(2σ2) with r N(0, σ2) to obtain P {|wtr| wtr} 1 exp w2 tr/(2σ2 n) . Solving for wt with δ/πtr using the same arguments as in the proof of Srinivas et al. (2010, Lemma 5.5) and taking the union bound over all time steps yields the results. Restated Lemma 3 Let ft(x) follow Assumption 2 with ε = 0. Pick δB (0, 1) and set ρtr = 2 ln 2πtr tr 1 π 1 tr = 1, πtr > 0. Also set w2 tr = 2σ2 n ln 2πtr δB . Then, observations yt under Assumption 1 satisfy |yt µDt(xt)| ρtrσDt(xt) + wtr for all tr 1 with probability at least 1 δB. Proof: From Lemma 1 with δB/2 and incorporating the noise, we have |ft(xt) + wtr µDt(xt)| ρtrσDt(xt) + |wtr| |yt µDt(xt)| ρtrσDt(xt) + |wtr|. (52) with ρtr = 2 ln 2πtr δB . Using Lemma 2 with δB/2 and taking the union bound yields the result. H Hyperparameters To ensure the reproducibility of our results, we give a list of all hyperparameters used in any simulation performed for this paper. All experiments in Section 6.2 were conducted on a 2021 Mac Book Pro with an Apple M1 Pro chip and 16GB RAM. The Monte Carlo simulations in Section 5 were conducted over 3 days on a compute cluster. The temperature data set is publicly available at http://db.csail.mit.edu/ labdata/labdata.html. The code will be published upon acceptance and is also part of the supplementary material of the submission. Table 5: Hyperparameters of the Monte Carlo Simulation for ET-GP-UCB (cf. Figure 2). Hyperparameters Monte Carlo Simulation dimension d 2 compact set X [0, 1]2 lengthscales l 0.2 time horizon T 200 noise variance σ2 n 0.02 number of i.i.d. runs 30 000 event trigger parameter in Lemma 3 δB 0.1 rate of change ε {0.03, 0.1} Published in Transactions on Machine Learning Research (01/2025) Table 6: Hyperparameters for the within-model comparisons (cf. Figure 7). Algorithm Hyperparameters Correct rate of change Miss. rate of change dimension d 2 2 compact set X [0, 1]2 [0, 1]2 lengthscales l 0.2 0.2 all time horizon T 400 400 noise variance σ2 n 0.02 0.02 βt approximation parameters c1 = 0.4, c2 = 4 c1 = 0.4, c2 = 4 number of i.i.d. runs 50 50 UI-TVBO rate of change ˆσ2 w {0.01, 0.03, 0.05} {0.001, 0.2} TV-GP-UCB rate of change ε {0.01, 0.03, 0.05} {0.001, 0.2} R-GP-UCB reset time Nconst {38, 29, 26} {68, 17} ET-GP-UCB event trigger parameter δB 0.1 0.1 Table 7: Hyperparameters for the temperature dataset (cf. Figure 3). Algorithm Hyperparameters Temperature Dataset dimension d 1 compact set X 46 arms kernel k empirical kernel all time horizon T 286 noise variance σ2 n 0.01 βt approximation parameters c1 = 0.8, c2 = 0.4 number of i.i.d. runs 50 UI-TVBO rate of change ˆσ2 w 0.03 TV-GP-UCB rate of change ε (see Bogunovic et al. (2016)) 0.03 R-GP-UCB reset time Nconst (see Bogunovic et al. (2016)) 15 ET-GP-UCB event trigger parameter in Lemma 3 δB 0.1 Table 8: Hyperparameters for the policy search experiment (cf. Figure 3 (c)). Algorithm Hyperparameters Policy search dimension d 4 lower bound to [ 3, 6, 50, 4] compact set X upper bound to [ 2, 4, 25, 2] scaling factors [1/8, 1/4, 3, 1/8] lengthscales l gamma prior G(15, 10/3) all time horizon T 300 noise variance σ2 n 0.02 βt approximation parameters c1 = 0.8, c2 = 4 number of i.i.d. runs 50 UI-TVBO rate of change ˆσ2 w 0.03 TV-GP-UCB rate of change ε 0.03 R-GP-UCB reset time Nconst induced by rate of change ε 29 UI-TVBO rate of change ˆσ2 w 0.03 ET-GP-UCB event trigger parameter in Lemma 3 δB 0.1 Published in Transactions on Machine Learning Research (01/2025) Table 9: Ablation on Impact of Reset Frequency vs. Timing (cf. Figure 4). Algorithm Hyperparameters Sensitivity Analysis dimension d 2 compact set X [0, 1]2 kernel k fixed, l = 0.2 all time horizon T 400 noise variance σ2 n 0.02 βt approximation parameters c1 = 0.4, c2 = 4 number of i.i.d. runs 50 R-GP-UCB reset time Nconst {38, 29, 26} R-GP-UCB* reset time Nconst as average from ET-GP-UCB {100, 45, 32} ET-GP-UCB event trigger parameter in Lemma 3 δB {0.1, 0.1, 0.1} Table 10: Hyperparameters for the synthetic experiments in Deng et al. (2022). Algorithm Hyperparameters Synthetic Experiments dimension d 1 compact set X [0, 1]2 kernel k fixed, l = 0.2 time horizon T 500 noise variance σ2 n {0.01, 1} number of i.i.d. runs 20 ET-GP-UCB event trigger parameter in Lemma 3 δB 0.1 Table 11: Hyperparameters for the stock market data experiment in Deng et al. (2022). Algorithm Hyperparameters Stock Market Data dimension d 1 compact set X 48 arms kernel k empirical kernel time horizon T 823 noise variance σ2 n {0.01, 300} number of i.i.d. runs 20 ET-GP-UCB event trigger parameter in Lemma 3 δB 0.1 Table 12: Hyperparameters for the sensitivity analysis (cf. Table 4). Algorithm Hyperparameters Sensitivity Analysis dimension d 2 compact set X [0, 1]2 kernel k fixed, l = 0.2 all time horizon T 400 noise variance σ2 n 0.02 βt approximation parameters c1 = 0.4, c2 = 4 number of i.i.d. runs 50 ET-GP-UCB event trigger parameter in Lemma 3 δB {0.005, 0.01, 0.05, 0.1, 0.5}