# optimal_scalarizations_for_sublinear_hypervolume_regret__17a84d3f.pdf Optimal Scalarizations for Sublinear Hypervolume Regret Qiuyi (Richard) Zhang Google Deepmind qiuyiz@google.com Scalarization is a general, parallizable technique that can be deployed in any multiobjective setting to reduce multiple objectives into one, yet some have dismissed this versatile approach because linear scalarizations cannot explore concave regions of the Pareto frontier. To that end, we aim to find simple non-linear scalarizations that provably explore a diverse set of k objectives on the Pareto frontier, as measured by the dominated hypervolume. We show that hypervolume scalarizations with uniformly random weights achieves an optimal sublinear hypervolume regret bound of O(T 1/k), with matching lower bounds that preclude any algorithm from doing better asymptotically. For the setting of multiobjective stochastic linear bandits, we utilize properties of hypervolume scalarizations to derive a novel non-Euclidean analysis to get regret bounds of e O(d T 1/2 + T 1/k), removing unnecessary poly(k) dependencies. We support our theory with strong empirical performance of using non-linear scalarizations that outperforms both their linear counterparts and other standard multiobjective algorithms in a variety of natural settings. 1 Introduction Optimization objectives in modern AI systems are becoming more complex with many different components that must be combined to perform precise tradeoffs in machine learning models. Starting from standard ℓp regularization objectives in regression problems [Kutner et al., 2005] to increasingly multi-component losses used in reinforcement learning [Sutton et al., 1998] and deep learning [Le Cun et al., 2015], many of these single-objective problems are phrased as a scalarized form of an inherently k-objective problem. Scalarization Method. Practitioners often vary the weights of the scalarization method, with the main goal of exploring the entire Pareto frontier, which is the set of optimal objectives that cannot be simultaneously improved. First, one chooses some weights λ Rk and scalarization functions sλ(y) : Rk R that convert k multiple objectives F(a) := (f1(a), ..., fk(a)) over some parameter space a A Rd into a single-objective scalar. Optimization is then applied to this family of single-objective functions sλ(F(x)) for various λ and since we often choose sλ to be monotonically increasing in all coordinates, xλ = arg maxa A sλ(F(a)) is on the Pareto frontier and the various choices of λ recovers an approximation to the Pareto frontier [Paria et al., 2018]. Due to its simplicity of use, many have turned to a heuristic-based scalarization strategy to pick the family of scalarizer and weights, which efficiently splits the multi-objective optimization into numerous single "scalarized" optimizations [Roijers et al., 2013]. Linear scalarizations with varying weights are often used in multi-objective optimization problems, such as in multi-objective reinforcement learning to combine task reward with the negative action norm [Abdolmaleki et al., 2021] or in RLHF to align responses with human preferences Ouyang et al. [2022]. However, the appeal of 38th Conference on Neural Information Processing Systems (Neur IPS 2024). using scalarizations in multiobjective optimization declined as linear scalarizations are shown to be provably incapable of exploring the full Pareto frontier [Boyd and Vandenberghe, 2004]. Beyond Linear Scalarization. To address this, some works have proposed piecewise linear scalarizations inspired by economics [Busa-Fekete et al., 2017], while for multi-armed bandits, scalarized knowledge gradient methods empirically perform better with non-linear scalarizations [Yahyaa et al., 2014]. The classical Chebyshev scalarization has been shown to be effective in many settings, such as evolutionary strategies Qi et al. [2014], Li et al. [2016], general blackbox optimization [Kasimbeyli et al., 2019] or reinforcement learning Van Moffaert et al. [2013]. Other works have come up with novel scalarizations that perform better empirically in some settings [Aliano Filho et al., 2019, Schmidt et al., 2019]. There also have been specific multi-objective algorithms tailored to specific settings such as Par Ego [Knowles, 2006] and MOEAD [Zhang and Li, 2007] for black-box optimization or multivariate iteration for reinforcement learning [Yang et al., 2019]. Furthermore, many adaptive reweighting strategies have been proposed in order to target or explore the full Pareto frontier, which have connections to gradient-based multi-objective optimization [Lin et al., 2019, Abdolmaleki et al., 2021]. However these adaptive strategies are heuristic-driven and hard to compare, while understanding simple oblivious scalarizations remain very important especially in batched settings where optimizations are done heavily in parallel Gergel and Kozinov [2019]. Hypervolume Regret. To determine optimality, a natural and widely used metric to measure progress of an optimizer is the hypervolume indicator, which is the volume of the dominated portion of the Pareto set [Zitzler and Thiele, 1999]. The hypervolume metric has become a gold standard because it has strict Pareto compliance meaning that if set A is a subset of B and B has at least one Pareto point not in A, then the hypervolume of B is greater than that of A. Unsurprisingly, multiobjective optimizers often use hypervolume related metrics for progress tracking or acquisition function, such as the Expected Hypervolume Improvement (EHVI) or its differentiable counterpart [Daulton et al., 2020, Hupkens et al., 2015]. Only recently has some works provide sub-linear hypervolume regret bounds which guarantees convergence to the full Pareto frontier; however, they are exponential in k and its analysis only applies to a specially tailored algorithm that requires an unrealistic classification step [Zuluaga et al., 2013]. We draw inspiration on work by [Golovin and Zhang, 2020] that introduces random hypervolume scalarizations but their work does not consider the finite-sample regime and their regret bounds are in the blackbox setting, where the convergence rate quantifies the statistical convergence of the Pareto frontier estimation. 1.1 Our Contributions We show, perhaps surprisingly, that a simple ensemble of nonlinear scalarizations, known as Hypervolume scalarizations, are theoretically optimal to minimize hypervolume regret and are empirically competitive for general multiobjective optimization. Intuitively, we quantify how fast scalarizations can approximate the Pareto frontier under finite samples even with perfect knowledge of the Pareto frontier, which is the white-box setting. Specifically, as the hypervolume scalarization has sharp level curves, they naturally allow for the targeting of a specific part of the Pareto frontier, without any assumptions on the Pareto set or the need for adaptively changing weights. Additionally, we can oblivously explore the Pareto frontier by choosing T maximizers of randomly weighted Hypervolume scalarizations and achieve a sublinear hypervolume regret rate of O(T 1/k), where T is the number of points sampled. Sublinear Hypervolume Regret. Given any set of objectives Y that are explicitly provided, our goal is to choose points on the Pareto frontier in a way to maximize the hypervolume. In this white-box setting, we introduce the notion of the hypervolume regret convergence rate, which is both a function of both the scalarization and the weight distribution, and show that the maxima of Hypervolume scalarization with uniform i.i.d. weights enjoy O(T 1/k) hypervolume regret (see Theorem 7). In fact, our derived regret rate of the Hypervolume scalarization holds for all frontiers, regardless of the inherent multiobjective function F or the underlying optimizer. Therefore, we emphasize that analyzing these model-agnostic rates can be a general theoretical tool to compare and analyze the effectiveness of proposed multiobjective algorithms. Although many scalarizers will search the entire Pareto frontier as T , the rate at which this convergence occurs can differ significantly, implying that this framework paves the road for a theoretical standard by which to judge the effectiveness of advanced strategies, such as adaptively weighted scalarizations. Lower bounds. On the other hand, we show surprisingly that no multiobjective algorithm, whether oblivious or adaptive, can beat the optimal hypervolume regret rates of applying single-objective optimization with the hypervolume scalarization. To accomplish this, we prove novel lower bounds showing one cannot hope for a better convergence rate due to the exponential nature of our regret, for any set of T points. Specifically, we show that the hypervolume regret of any algorithm after T actions is at least Ω(T 1/k), demonstrating the necessity of the O(T 1/k) term up to small constants in the denominator. As a corollary, we leverage the sublinear regret properties of hypervolume scalarization to transfer our lower bounds to the more general setting of scalarized Bayes regret. Together, we demonstrate that for general multiobjective optimzation, finding maximas of the hypervolume scalarizations with a uniform weight distribution optimally finds the Pareto frontier asymptotically. Theorem 1 (Informal Restatement of Theorem 7 and Theorem 8). Let YT = {y1, ..., y T } be a set of T points in Rk such that yi arg max y Y s HV λi (y) with λi S+ randomly drawn i.i.d. from an uniform distribution and s HV are hypervolume scalarizations. Then, the hypervolume regret satisfies HV(Y ) HV(YT ) = O(T 1 where Y is the Pareto frontier and HV is the hypervolume function. Furthermore, any algorithm for choosing these T points must suffer hypervolume regret of at least Ω(T 1 Scalarized Algorithm for Linear Bandit. Next, we use a novel non-Euclidean analysis to prove improved hypervolume regret bounds for our theoretical toy model: the classic stochastic linear bandit setting. For any scalarization and weight distribution, we propose a new scalarized algorithm (Algorithm 1) for multiobjective stochastic linear bandit that combines uniform exploration and exploitation via an UCB approach to provably obtain scalarized Bayes regret bounds, which we then combine with the hypervolume scalarization to derive optimal hypervolume regret bounds. Specifically, for any scalarization sλ, we show that our algorithm in the linear bandit setting has a scalarized Bayes regret bound of e O(Lpk1/pd T 1/2 + T 1/k), where Lp is the Lipschitz constant of the sλ( ) in the ℓp norm. Finally, by using hypervolume scalarizations and exploiting their ℓ - smoothness, we completely remove the dependence on the number of objectives, k, which had a polynomial dependence in previous regret bound given by Golovin and Zhang [2020]. Theorem 2 (Informal Restatement of Theorem 11). For the multiobjective linear bandit problem, let AT A be the actions generated by T rounds of Algorithm 1, then our hypervolume regret is bounded by: HVz(Θ A) HVz(Θ AT ) O(d T 1 Experiments. Guided by our theoretical analysis, we empirically evaluate a diverse combination of scalarizations and weight distributions in multiple natural settings. First, we consider a synthetic optimization that is inspired by our theoretical derivation of hypervolume regret when the entire Pareto front is known and analytically given. Thus we can explicitly calculate hypervolume regret as a function of the number of Pareto points chosen and compare the regret convergence rates of multiple scalarizations with the uniform S+ distribution with k = 3 objectives, for easy visualization. This is important since when k = 2, we show that the Chebyshev and Hypervolume scalarizations are in fact equivalent. In our setup, we consider synthetic combinations of concave, convex, and concave/convex Pareto frontiers in each dimension. As expected, non-linear Hypervolume and Chebyshev scalarizations enjoy fast sublinear convergence, while the performance of the Linear scalarization consistently lacks behind, surprisingly even for convex Pareto frontiers. We observe that generally the Hypervolume scalarization does better on concave Pareto frontiers and on some convex-concave frontiers, while the Chebyshev distribution have faster hypervolume convergence in certain convex regimes. For multiobjective linear bandits, our experiments show that for many settings, despite having a convex Pareto frontier, applying linear or Chebyshev scalarizations naively with various weight distributions leads to suboptimal hypervolume progress, especially when the number of objective increase to exceed k 5. This is because the non-uniform curvature of the Pareto frontier, exaggerated by the curse of dimensionality and combined with a stationary weight distribution, hinders uniform progress in exploring the frontier. Although one can possibly adapt the weight distribution to the varying curvature of the Pareto frontier when it is convex, the use of simple non-linear scalarizations allow for fast parallelization and are theoretically sound. For general multiobjective optimization, we perform empirical comparisons on BBOB benchmarks for bi-ojective functions in a bayesian optimization setting, using classical Gaussian Process models [Williams and Rasmussen, 2006]. When comparing EHVI with hypervolume scalarization approaches, we find that EHVI tends to limit its hypervolume gain by over-focusing on the central portion of the Pareto frontier, whereas the hypervolume scalarization encourages a diverse exploration of the extreme ends. We summarize our contributions as follows: Show that hypervolume scalarizations optimizes for the full Pareto frontier and enjoys sublinear hypervolume regret bounds of O(T 1/k), a theoretical measure of characterizing scalarization effectiveness in the white-box setting. Establish a tight lower bound on the hypervolume regret for any algorithm and Bayes regret of Ω(T 1/k) by a packing argument on the Pareto set. Introduce optimization algorithm for multiobjective linear bandits that achieves improved e O(d T 1/2 + T 1/k) hypervolume regret via a novel non-Euclidean regret analysis and metric entropy bounds. Empirically justify the adoption of hypervolume scalarizations for finding a diverse Pareto frontier in general multiobjective optimization via synthetic, linear bandit, and blackbox optimization benchmarks. 2 Problem Setting and Notation For a scalarization function sλ(x), sλ is Lp-Lipschitz with respect to x in the ℓp norm on X Rd if for x1, x2 X, |sλ(x1) sλ(x2)| Lp x1 x2 p, and Lλ analogously for λ in the Euclidean norm. We let Sk 1 + = {y Rk | y = 1, y > 0} be the sphere in the positive orthant and by abuse of notation, we also let y Sk 1 + denote that y is drawn uniformly on Sk 1 + . Our usual settings of the weight distribution D = Sk 1 + will be uniform, unless otherwise stated. For two outputs y, z Y Rk, we say that y is Pareto-dominated by z if y z and there exists j such that yj < zj, where y z is defined for vectors element-wise. A point is Paretooptimal if no point in the output space Y can dominates it. Let Y denote the set of Pareto-optimal points (objectives) in Y, which is also known as the Pareto frontier. Our main progress metrics for multiobjective optimization is given by the standard hypervolume indicator. For S Rk compact, let vol(S) be the regular hypervolume of S with respect to the standard Lebesgue measure. Definition 3. For a set Y of points in Rk, we define the (dominated) hypervolume indicator of Y with respect to reference point z as: HVz(Y) = vol({x | x z, x is dominated by some y Y}) We can formally phrase our optimization objective as trying to rapidly minimize the hypervolume (psuedo-)regret. Let A be our action space and for some general multi-objective function F, let Y be the image of A under F. Let AT RT d be any matrix of T actions and let YT = F(AT ) RT k be the k objectives corresponding. Then, the hypervolume regret of actions AT , with respect to the reference point z, is given by: Hypervolume-Regret(AT ) = HVz(Y ) HVz(YT ) which is 0 for all z if and only if YT contains all unique points in Y . For various scalarizations and weight distributions, an related measure of progress that attempts to capture the requirement of diversity in the Pareto front is scalarized Bayes regret for some scalarization function sλ. For some fixed scalarization with weights λ, sλ : Rk R, we can define the instantaneous scalarized (psuedo-)regret as r(sλ, at) = maxa A sλ(F(a)) sλ(F(at)). To capture diversity and progress, we will vary λ D according to some distribution of non-negative weight vectors and define regret with respect with respect to all past actions At. Specifically, we define the (scalarized) Bayes regret with respect to a set of actions At to be: BR(sλ, At) = Eλ D[maxa A sλ(F(a)) maxa At sλ(F(a))] = Eλ D[ min a Atr(sλ, a)] Unlike previous notions of Bayes regret in literature, we are actually calculating the Bayes regret of a reward function that is maximized with respect to an entire set of actions At. Specifically, by maximizing over all previous actions, this captures the notion that during multi-objective optimization our Pareto set is always expanding. We will see later that this novel definition is the right one, as it generalizes to the multi-objective setting in the form of hypervolume regret. 2.1 Scalarizations for Multiobjective Optimization For multiobjective optimization, we generally only consider monotone scalarizers that have the property that if y > z, then sλ(y) > sλ(z) for all λ. Note this guarantees that an optimal solution to the scalarized optimization is on the Pareto frontier. A common scalarization used widely in practice is the linear scalarization: s LIN λ (y) = λ y for some chosen positive weights λ Rk. By Lagrange duality and hyperplane separation of convex sets, one can show that any convex Pareto frontier can be characterized fully by an optimal solution for some weight settings. However, linear scalarizations cannot recover the non-convex regions of Pareto fronts since the linear level curves can only be tangent to the Pareto front in the protruding convex regions (see Figure 1). To overcome this drawback, another scalarization that is proposed is the Chebyshev scalarization: s CS λ (y) = min i λiyi. Indeed, one can show that the sharpness of the scalarization, due to its minimum operator, can discover non-convex Pareto frontiers. Proposition 4 (Emmerich and Deutz [2018]). For a point y Y for convex Y, there exists λ > 0 such that y = arg max y Y s LIN λ (y). For a point y Y for any possibly non-convex set Y that lies in the positive orthant, there exists λ > 0 such that y = arg max y Y s CS λ (y). Figure 1: Left: Comparisons of the scalarized minimization solutions with various weights with convex and non-convex Pareto fronts. The colors represent different weights; the dots are scalarized optima and the dotted lines represent level curves. Linear scalarization does not have an optima in the concave region of the Pareto front for any set of weights, but the non-linear scalarization, with its sharper level curves, can discover the whole Pareto front (Figure from [Emmerich and Deutz, 2018]). Right: The dotted red lines represent the level curves of the hypervolume scalarization with λ = v, discovering b, whereas the linear scalarization would prefer a or c. Furthermore, the optima is exactly the Pareto point that is in the direction of v. 3 Hypervolume Scalarizations In this section, we show the utility and optimality of a related scalarization known as the Hypervolume scalarization, s HV λ (y) = min i (yi/λi)k that was introduced in Golovin and Zhang [2020]. To gain intuition, we visualize the non-linear level curves of the scalarization, which shows that our scalarization targets the portion of the Pareto frontier in the direction of λ for any λ > 0 (see Figure 1), since the tangent point of the level curves of the scalarization is always on the vector in the direction of λ. This implies that with an uniform distribution on λ, we are guaranteed to have a uniform spread of Pareto points in terms of its angular direction. Metric 1 (higher is better) Metric 2 (higher is better) Dominated Hypervolume (grey) Metric 1 (higher is better) Metric 2 (higher is better) Dominated Hypervolume (grey) Figure 2: The hypervolume scalarization taken with respect to a direction λ = w corresponds to a differential area element within dominated hypervolume and averaging over random directions is analogous to integrating over the dominated hypervolume in polar coordinates. We exploit this fact to show that by choosing the maximizers of T random hypervolume scalarizers, we quickly converge to the hypervolume of the Pareto frontier at an optimal rate of O(T 1/k). Figure from [Song et al., 2024] Lemma 5. For any point y on the Pareto frontier of any set Y that lies in the positive orthant, there exists λ > 0 such that y = arg max y Y s HV λ (y). Furthermore, for any α, λ > 0 such that αλ is on the Pareto frontier, then αλ arg max y Y s HV λ (y). This scalarization additionally has the special property that the expected maximized scalarized value under a uniform weight distribution on Sk 1 + gives the dominated hypervolume, up to a constant scaling factor. Thus, the optima of the hypervolume scalarization over some static uniform distribution will be provably sufficiently diverse for any Pareto set in expectation. Lemma 6 (Hypervolume in Expectation [Golovin and Zhang, 2020]). Let YT = {y1, ..., y T } be a set of T points in Rk. Then, the hypervolume with respect to a reference point z is given by: HVz(YT ) = ck E λ Sk 1 + max y YT s HV λ (y z) where ck = πk/2 2k Γ(k/2+1) is a constant depending only on k. While this lemma is useful in the infinite limit, we supplement it by showing that finite-sample bounds on the strongly sublinear hypervolume convergence rate. In fact, many scalarizations will eventually explore the whole Pareto frontier in the infinite limit, but the rate at which the exploration improves the hypervolume is not known, and may be exponentially slow. We show that the optimizing hypervolume scalarizations with a uniform weight distribution enjoys sublinear hypervolume regret, specifically O(T 1/k) hypervolume regret convergence rates for any Pareto set Y. Note that this rate is agnostic of the underlying optimization algorithm or Pareto set, meaning this is a general property of the scalarization. Our novel proof of convergence uses a generalization argument to connect hypervolume-scalarized Bayes regret and its finite sample form, exploiting the Lipschitz properties of s HV λ to derive metric entropy bounds. Proving smoothness properties of our hypervolume scalarizations for any λ > 0 with λ normalized on the unit sphere is non-obvious as s HV λ (y) depends inversely on λi so when λi is small, s HV λ might change wildly. Theorem 7 (Sublinear Hypervolume Regret). Let YT = {y1, ..., y T } be a set of T points in Rk such that yi arg max y Y s HV λi (y z) with respect to a reference point z and Bl yi z Bu. Then, with probability at least 1 δ over λi S+ i.i.d., the hypervolume of YT with respect to z satisfies sublinear hypervolume regret in T: HVz(Y ) HVz(YT ) = O(T 1 k+1 + p where O hides constant factors in k, Bu, Bl. For k = 2, this also holds when Chebyshev scalarization is used: yi arg max y Y s CS λi (y z). We note that the distribution of Pareto points selected by the Chebyshev scalarizer is quite similar to the Hypervolume scalarizer as the formulas are almost identical, except for the inverse weights of the latter. In fact, we show that when k = 2, both scalarizers behave the same and enjoy strong convergence rates; however for k > 2, the Pareto distributions are in fact different under λ S+ and we can empirically observe the suboptimality of the Chebyshev scalarizer. By definition, using the inverse weight distribution with the Chebyshev scalarizer will be equivalent to applying the Hypervolume scalarizer. 3.1 Lower Bounds and Optimality The dominating factor in our derived convergence rate is the O(T 1/(k+1)) term and we show that this cannot be improved. Over all subsets YT Y of size T, note that our optimal convergence rate is given by the the subset that maximizes the dominated hypervolume of YT , although finding this is in fact a NP-hard problem due to reduction to set cover. By constructing a lower bound via a novel packing argument, we show that even this optimal set would incur at least Ω(T 1/k) regret, implying that our convergence rates, derived from generalization rates when empirically approximating the hypervolume, are optimal. Specifically, we show that for hypervolume regret, any algorithm cannot achieve better than O(T 1/(k 1)) regret even when using linear objectives, and this matches the dominating factor in our algorithm up to a small constant in the denominator. By using hypervolume scalarizations and its connection to hypervolume regret, we conclude that this also implies a Ω(T 1/(k 1)) lower bound on the scalarized Bayes regret. Theorem 8 (Hypervolume Regret Lower Bound). There exists a setting of linear objective parameters Θ and A = {a : a = 1} such that for any actions AT , the hypervolume regret at z = 0 after T rounds is HVz(Θ A) HVz(Θ AT ) = Ω(T 1 k 1 ) Corollary 9. There is a setting of objectives Θ and A = {a : a = 1} such that for any actions AT , the scalarized Bayes regret after T rounds is BR(s HV λ , AT ) = Ω(T 1 k 1 ) 4 Multiobjective Stochastic Linear Bandits We propose a simple scalarized algorithm for linear bandits and provide a novel ℓp analysis of the hypervolume regret that removes the polynomial dependence on k in the scalarized regret bounds. When combined with the ℓ sharpness of the hypervolume scalarization, this analysis gives an optimal O(d/ T) bound on the scalarized regret, up to log(k) factors. This log dependence on k is perhaps surprising but is justified information theoretically since each objective is observed separately. Note that our scalarized algorithm works despite of noise in the observations, which makes it difficult to even statistically infer measures of hypervolume progress. Our setup and algorithm is given in the Appendix (see Section A). By using the confidence ellipsoids given by the UCB algorithm, we can determine each objective parameter Θ i , up to a small error. To bound the scalarized regret, we utilize the ℓp smoothness of sλ, Lp, to reduce the dependence on k to be O(k1/p), which effectively removes the polynomial dependence on k when p . This is perhaps not surprising, since each objective is observed independently and fully, so the information gain scales with the number of objectives. Lemma 10. Consider running EXPLOREUCB (Algorithm 1) for T > max(k, d) iterations and for T even, let a T be the action that maximizes the scalarized UCB in iteration T/2. Then, with probability at least 1 δ, the instantaneous scalarized regret can be bounded by r(sλ, a T ) 10k 1 p Lpd log(k/δ) + log(T) where Lp is the ℓp-Lipschitz constant for sλ( ). Finally, we connect the expected Bayes regret with the empirical average of scalarized regret via uniform convergence properties of all functions of the form f(λ) = max a Asλ(Θ a). By using s HV λ and setting p = , we derive our final fast hypervolume regret rates for stochastic linear bandits, which is the combination of the scalarized regret rates and the hypervolume regret rates. Theorem 11 (Hyper Volume Regret of EXPLOREUCB). Let z Rk be a reference point such that over all a A, Bl Θ a z Bu. Then, with constant probability, running Algorithm 1 with s HV λ (y) and D = S+ gives hypervolume regret bound, for k constant, HVz(Θ A) HVz(Θ AT ) O 5 Experiments In this section, we empirically justify our theoretical results by comparing hypervolume convergence curves for multiobjective optimization in synthetic, linear bandit and blackbox optimization environments with multiple scalarizations and weight distributions. Our empirical results highlight the advantage of the hypervolume scalarization with uniform weights in maximizing the diversity and hypervolume of the resulting Pareto front when compared with other scalarizations and weight distributions, especially when there are a mild number of output objectives k. Our experiments are not meant to show that scalarizations is the best way to solve multiobjective optimization; rather, it is a simple yet competitive baseline that is easily parallelized in a variety of settings. Also, we use slightly altered form of our hypervolume scalarization as sλ(y) = min i yi/λi, which is a simply a monotone transform and does not inherently affect the optimization. All error bars are given between the 30 to 70 percentile over independent repeats. 5.1 Synthetic Optimization Our synthetic optimization benchmarks assume the knowledge of Y and the Pareto frontier and thus we can compute the total hypervolume and compare the hypervolume regret of multiple scalarizations with the uniform S+ distribution. For our experiments, we fix our weight distribution and compare the three widely types of scalarizations that were previously mentioned: the Linear, Chebyshev, and the Hypervolume scalarization. We focus on the k = 3 setting and apply optimization for a diverse set of Pareto frontiers in the region x, y [0, 1], z > 0. We discretize our region into 30 points per dimension to form a discrete Pareto frontier and set our reference point to be at z = [ ϵ, ϵ, ϵ] for ϵ = 1e-4 and run for 10 repeats. The synthetic Pareto frontiers that we are consider are a product of 1-dimensional frontiers and specifically, z = g(x) g(y), where g(x) = exp( x), 3 exp(x), cos(πx) + 1, which form a concave, convex, and concave/convex Pareto frontiers respectively. Now that since the derivatives of these functions are all negative in our feasible region, z is a valid Pareto front. From our combination of functions, we observe that both Hypervolume and Chebyshev scalarizations enjoy fast convergence, while the performance of the Linear scalarization consistently lacks behind, surprisingly even for convex Pareto frontiers (see full plots in C.1). Interestingly, we observe that generally the Hypervolume scalarization does better on concave Pareto frontiers and on some convex-concave frontiers; however since these are static algorithms, it is not surprising that the Chebyshev distribution can perform better in certain convex regimes. However, we still advocate the usage of the hypervolume indicator as it is guaranteed to have a uniform spread especially when the number of objectives is increased, as shown by our later experiments. Figure 3: Comparisons of multiple scalarizations for the synthetic concave Pareto frontier given by z = exp( x y). The hypervolume regret for Linear is constant, and the Hypervolume enjoys a faster regret convergence rate than the Chebyshev. 5.2 Stochastic Linear Bandits We run Algorithm 1 and compare scalarization effects for the multiobjective linear bandit setting. We set our reference point to be z = 2 in k dimension space, since our action set of A = {a : a = 1} and our norm bound on Θ ensures that our rewards are in [ 1, 1]. In conjunction with the scalarizer, we use our weight distribution D = S+, which samples vectors uniformly across the unit sphere. In addition, we also compare this with the bounding box distribution methods that were suggested by [Paria et al., 2018], which samples from the uniform distribution from the min to the max each objective and requires some prior knowledge of the range of each objective [Hakanen and Knowles, 2017]. Given our reward bounds, we use the bounding box of [ 1, 1] for each of the k objectives. Following their prescription for weight sampling, we draw our weights for the linear and hypervolume scalarization uniformly in [1, 3] and take an inverse for the Chebyshev scalarization. We name this the boxed distribution for each scalarization, respectively. To highlight the differences between the multiple scalarizations, we configure our linear bandits parameters to be anti-correlated, which creates a convex Pareto front with non-uniform curvature. Note that a perfect anti-correlated Pareto front would be linear, which would cause linear scalarizations to always optimize at the end points. We start with simple k = 2 case and let θ0 be random and θ1 = θ0 + η, where η is some small random Gaussian perturbation (we set the standard deviation to be about 0.1 times the norm of θi). We renormalize after the anti-correlation to ensure Θ = 1. We run our algorithm with inherent dimension d = 4 for T = 100, 200 rounds with k = 2, 6, 10. As expected, we find the hypervolume scalarization consistently outperforms the Chebyshev and linear scalarizations, with linear scalarization as the worst performing (see Figure 4). Note that when we increase the output dimension of the problem by setting k = 10, the hypervolume scalarization shows a more distinct advantage. The boxed distribution approach of [Paria et al., 2018] does not seem to fare well and consistently performs worse than its uniform counterpart. While linear scalarization provides relatively good performance when the number of objective k 5, it appears that as the number of objectives increase in multi-objective optimization, more care needs to be put into the design of scalarization and their weights due to the curse of dimensionality, since the regions of non-uniformity will exponentially increase. We suggest that as more and more objectives are being added to modern machine learning systems, using smart scalarizations is critical to an uniform exploration of the Pareto frontier. 5.3 BBOB Functions We empirically demonstrate the competitiveness of hypervolume scalarizations for Bayesian Optimization by comparing them to the popular BO method of EHVI [Hupkens et al., 2015]. Running our proposed multiobjective algorithms on the Black-Box Optimization Benchmark (BBOB) functions, Figure 4: Comparisons of the cumulative hypervolume plots with some anti-correlated θ. When the output dimension increase, there is a clearer advantage to using the hypervolume scalarization over the linear and Chebyshev scalarization. We find that the boxed weight distribution does consistently worse than the uniform distribution. which can be paired up into multiple bi-objective optimization problems [Tušar et al., 2016]. Our goal is to use a wide set of non-convex benchmarks to supplement our experiments on our simple toy example of linear bandits. For scalarized approaches, we use hypervolume scalarizations with the scalarized UCB algorithm with a constant standard deviation multiplier of 1.8 and all algorithms with use a Gaussian Process as the underlying model with a standard Matérn kernel that is tuned via ARD Williams and Rasmussen [2006]. Our objectives are given by BBOB functions, which are usually non-negative and are minimized. The input space is always a compact hypercube [ 5, 5]n and the global minima is often at the origin. For bi-objective optimization, given two different BBOB functions f1, f2, we attempt to maximize the hypervolume spanned by ( f1(xi), f2(xi)) over choices of inputs xi with respect to the reference point ( 5, 5). We normalize each function due to the drastically different ranges and add random observation noise. We also apply vectorized shifts of the input space of 2, 2 on the two functions respectively, so that the optima of each function do not co-locate at the origin. We run each of our algorithms in dimensions d = 8, 16, 24 and optimize for 160 iterations with 5 repeats. From our results, we see that both EHVI and UCB with hypervolume scalarizations are competitive on the BBOB problems but the scalarized UCB algorithm seems to be able to explore the extreme ends of the Pareto frontier, whereas EHVI tends to favor points in the middle (see Appendix and Figure 5). From our experiments, this trend appears to be consistent across different functions and is more prominent as the input dimensions d increase. Michael H Kutner, Christopher J Nachtsheim, John Neter, William Li, et al. Applied linear statistical models, volume 5. Mc Graw-Hill Irwin Boston, 2005. Richard S Sutton, Andrew G Barto, et al. Introduction to reinforcement learning, volume 2. MIT press Cambridge, 1998. Yann Le Cun, Yoshua Bengio, and Geoffrey Hinton. Deep learning. nature, 521(7553):436 444, 2015. Biswajit Paria, Kirthevasan Kandasamy, and Barnabás Póczos. A flexible framework for multiobjective bayesian optimization using random scalarizations. ar Xiv preprint ar Xiv:1805.12168, 2018. Diederik M Roijers, Peter Vamplew, Shimon Whiteson, and Richard Dazeley. A survey of multiobjective sequential decision-making. Journal of Artificial Intelligence Research, 48:67 113, 2013. Abbas Abdolmaleki, Sandy H Huang, Giulia Vezzani, Bobak Shahriari, Jost Tobias Springenberg, Shruti Mishra, Dhruva TB, Arunkumar Byravan, Konstantinos Bousmalis, Andras Gyorgy, et al. On multi-objective policy optimization as a tool for reinforcement learning. ar Xiv preprint ar Xiv:2106.08199, 2021. Long Ouyang, Jeffrey Wu, Xu Jiang, Diogo Almeida, Carroll Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, et al. Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35: 27730 27744, 2022. Stephen Poythress Boyd and Lieven Vandenberghe. Convex optimization. Cambridge University Press, Cambridge, 2004. ISBN 0-521-83378-7. Róbert Busa-Fekete, Balázs Szörényi, Paul Weng, and Shie Mannor. Multi-objective bandits: Optimizing the generalized gini index. In International Conference on Machine Learning, pages 625 634. PMLR, 2017. Saba Q Yahyaa, Madalina M Drugan, and Bernard Manderick. The scalarized multi-objective multiarmed bandit problem: An empirical study of its exploration vs. exploitation tradeoff. In 2014 International Joint Conference on Neural Networks (IJCNN), pages 2290 2297. IEEE, 2014. Yutao Qi, Xiaoliang Ma, Fang Liu, Licheng Jiao, Jianyong Sun, and Jianshe Wu. Moea/d with adaptive weight adjustment. Evolutionary computation, 22(2):231 264, 2014. Hui Li, Qingfu Zhang, and Jingda Deng. Biased multiobjective optimization and decomposition algorithm. IEEE transactions on cybernetics, 47(1):52 66, 2016. Refail Kasimbeyli, Zehra Kamisli Ozturk, Nergiz Kasimbeyli, Gulcin Dinc Yalcin, and Banu Icmen Erdem. Comparison of some scalarization methods in multiobjective optimization. Bulletin of the Malaysian Mathematical Sciences Society, 42(5):1875 1905, 2019. Kristof Van Moffaert, Madalina M Drugan, and Ann Nowé. Scalarized multi-objective reinforcement learning: Novel design techniques. In 2013 IEEE symposium on adaptive dynamic programming and reinforcement learning (ADPRL), pages 191 199. IEEE, 2013. Angelo Aliano Filho, Antonio Carlos Moretti, Margarida Vaz Pato, and Washington Alves de Oliveira. An exact scalarization method with multiple reference points for bi-objective integer linear optimization problems. Annals of Operations Research, pages 1 35, 2019. Marie Schmidt, Anita Schöbel, and Lisa Thom. Min-ordering and max-ordering scalarization methods for multi-objective robust optimization. European Journal of Operational Research, 275 (2):446 459, 2019. Joshua Knowles. Parego: a hybrid algorithm with on-line landscape approximation for expensive multiobjective optimization problems. IEEE Transactions on Evolutionary Computation, 10(1): 50 66, 2006. Qingfu Zhang and Hui Li. Moea/d: A multiobjective evolutionary algorithm based on decomposition. IEEE Transactions on evolutionary computation, 11(6):712 731, 2007. Runzhe Yang, Xingyuan Sun, and Karthik Narasimhan. A generalized algorithm for multi-objective reinforcement learning and policy adaptation. Advances in Neural Information Processing Systems, 32, 2019. Xi Lin, Hui-Ling Zhen, Zhenhua Li, Qing-Fu Zhang, and Sam Kwong. Pareto multitask learning. In Advances in Neural Information Processing Systems 32, pages 12037 12047. Curran Associates, Inc., 2019. URL http://papers.nips.cc/paper/ 9374-pareto-multi-task-learning.pdf. Victor Gergel and Evgeny Kozinov. Parallel computations for various scalarization schemes in multicriteria optimization problems. In International Conference on Parallel Processing and Applied Mathematics, pages 174 184. Springer, 2019. Eckart Zitzler and Lothar Thiele. Multiobjective evolutionary algorithms: a comparative case study and the strength pareto approach. IEEE transactions on Evolutionary Computation, 3(4):257 271, 1999. Samuel Daulton, Maximilian Balandat, and Eytan Bakshy. Differentiable expected hypervolume improvement for parallel multi-objective bayesian optimization. Advances in Neural Information Processing Systems, 33:9851 9864, 2020. Iris Hupkens, André Deutz, Kaifeng Yang, and Michael Emmerich. Faster exact algorithms for computing expected hypervolume improvement. In international conference on evolutionary multi-criterion optimization, pages 65 79. Springer, 2015. Marcela Zuluaga, Guillaume Sergent, Andreas Krause, and Markus Püschel. Active learning for multi-objective optimization. In International Conference on Machine Learning, pages 462 470, 2013. Daniel Golovin and Qiuyi Zhang. Random hypervolume scalarizations for provable multi-objective black box optimization. ar Xiv preprint ar Xiv:2006.04655, 2020. Christopher KI Williams and Carl Edward Rasmussen. Gaussian processes for machine learning, volume 2. MIT press Cambridge, MA, 2006. Michael TM Emmerich and André H Deutz. A tutorial on multiobjective optimization: fundamentals and evolutionary methods. Natural computing, 17(3):585 609, 2018. Xingyou Song, Qiuyi Zhang, Chansoo Lee, Emily Fertig, Tzu-Kuo Huang, Lior Belenki, Greg Kochanski, Setareh Ariafar, Srinivas Vasudevan, Sagi Perel, et al. The vizier gaussian process bandit algorithm. ar Xiv preprint ar Xiv:2408.11527, 2024. Jussi Hakanen and Joshua D Knowles. On using decision maker preferences with parego. In International Conference on Evolutionary Multi-Criterion Optimization, pages 282 297. Springer, 2017. Tea Tušar, Dimo Brockhoff, Nikolaus Hansen, and Anne Auger. Coco: the bi-objective black box optimization benchmarking (bbob-biobj) test suite. Ar Xiv e-prints, 2016. Tor Lattimore and Csaba Szepesvári. Bandit algorithms. Cambridge University Press, 2020. Jack Kiefer and Jacob Wolfowitz. The equivalence of two extremum problems. Canadian Journal of Mathematics, 12:363 366, 1960. Peter L Bartlett and Shahar Mendelson. Rademacher and gaussian complexities: Risk bounds and structural results. Journal of Machine Learning Research, 3(Nov):463 482, 2002. Lee-Ad Gottlieb, Aryeh Kontorovich, and Robert Krauthgamer. Adaptive metric dimensionality reduction. Theoretical Computer Science, 620:105 118, 2016. Yasin Abbasi-Yadkori, Dávid Pál, and Csaba Szepesvári. Improved algorithms for linear stochastic bandits. Advances in neural information processing systems, 24:2312 2320, 2011. A Linear Bandit Setup and Algorithm For theory, we use the classic stochastic linear bandit setting. For the single-objective setting, in round t = 1, 2, ..., T, the and receives a reward yt = θ , at + ξt where ξt is i.i.d. 1-sub-Gaussian noise and θ Rd is the unknown true parameter vector. In the multi-objective stochastic linear bandit setting, the learner chooses an action at Rd from the action set A and receives a vectorized reward yt = Θ at + ξt, where Θ Rk d is now a matrix of k unknown true parameters and ξt Rk is a vector of independent 1-sub-Gaussian noise. We assume, for sake of normalization, that Θ i 1 and that at 1, where denotes the ℓ2 norm unless otherwise stated. Other norms that are used include the classical ℓp norms p and matrix norms x M = x Mx for a positive semi-definite matrix M. We also denote At Rd t to be the history action matrix, whose i-th column is ai, the action taken in round i. Similarly, yt is defined analogously. Finally, for sake of analysis, we assume that A contains an isotropic set of actions and specifically, there is E A with size |E| = O(d) such that P 2I, where denotes the PSD ordering on symmetric matrices. This assumption is not restrictive, as it can be relaxed by using optimal design for least squares estimators [Lattimore and Szepesvári, 2020] and the Kiefer-Wolfrowitz Theorem [Kiefer and Wolfowitz, 1960], which guarantees the existence and construction of an uniform exploration basis of size O(poly(d)). Algorithm 1: EXPLOREUCB(T, D, sλ) Input :number of maximum actions T, weight distribution D , scalarization sλ 1 repeat 2 Play action en E for n i mod d 3 Let Cti be the confidence ellipsoid for Θi and let UCBi(a) = maxθ Ci θ a 4 Draw λ D, play a = argmaxa A sλ(UCBi(a)) 5 until number of rounds i exceed T/2 B Missing Proofs Proof of Lemma 5. Let λ = y / y . Note that λ > 0 since y is in the positive orthant and for the sake of contradiction, assume there exists z such that sλ(z) > sλ(y ). However, note that for any i, zi λi > mini y i λi = y i λi , where the last line follows since y i /λi = y for all i by construction. Therefore, we conclude that y < z, contradicting that y is Pareto optimal. Finally, note that if αλ is on the Pareto frontier, then we see that mini αλi/λi = α and furthermore, this min value is achieved for all i. Therefore, since αλ is on the Pareto frontier, any other point y Y has some coordinate j such that yj < αλj, which implies that mini yi/λi < α. Proof of Theorem 7. Let us denote sλ = s HV λ is the hypervolume scalarization and WLOG let z = 0. Note that we can first decompose our regret as HVz(Y ) HVz(YT ) |HVz(Y ) ck T PT i=1 maxy Y sλi(y)| + | ck T PT i=1 maxy Y sλi(y) HVz(YT )| T PT i=1 sλi(y)| + | ck T PT i=1 maxy YT sλi(y) HVz(YT )| where the second inequality uses the fact that yi arg max y Y sλi(y). We proceed to bound both parts separately and we note that it suffices to prove uniform concentration of the empirical sum to the expectation, as by our choice of scalarization is the hypervolume by Lemma 6. Let f Y(λi) = maxy Y sλi(y). We let F = {f Y : Y A} be our class of functions over all possible output sets Y. We will first demonstrate uniform convergence by bounding the complexity of F. Specifically, by generalization bounds from Rademacher complexities Bartlett and Mendelson [2002], over choices of λi D, we know that with probability 1 δ, for all Y, we have the bound E λ D[f Y] 1 i=1 f Y(λi) where Rm(F) = Eλi D,σi , where σi are i.i.d. 1 Rademacher variables. To bound Rm(F), we appeal to Dudley s integral formulation that allows us to use the metric entropy of F to bound Rm(F) inf α>0 log(N(ϵ, F, 2)) where N denotes the standard covering number for F under the ℓ2 function norm metric over λ D. Since D is the uniform distribution over S+, this induces a natural ℓ function norm metric on F that is f = supλ S+ |f(λ)|. By Lemma 15, since yi z is bounded below and above by Bu, Bl respectively, sλ(y) is Lλ Lipschitz with respect to the Euclidean norm in λ. Note that since the maximal operator preserves Lipschitzness, f Y(λ) is also Lλ-Lipschitz with respect to λ Rk for any Y. Since F contains Lλ-Lipschitz functions in Rk, we can bound the metric entropy via a covering of λ via a Lipschitz covering argument (see Lemma 4.2 of Gottlieb et al. [2016]), so we have log(N(ϵ, F, 2)) log(N(ϵ, F, )) (4Lλ/ϵ)k log(8/k) Finally, we follow the same Dudley integral computation of Theorem 4.3 of Gottlieb et al. [2016] to get that Rm(F) inf α>0 4α + 12 Z 2 (4Lλ/ϵ)k log(8/k) = O(Lλ/m1/(k+1)) Therefore, we conclude that with probability at least 1 δ over the independent choices of λi D, for all Y and setting m = T max a Y sλ(Θ a) 1 i=1 max a Y sλi(Θ a) O Lλ T 1/(k+1) where B is some ℓ upper bound on the output set Y. Finally, we conclude by using Lemma 6 to replace the expectation by the hypervolume and by setting Y = Y, YT respectively. Now consider the case when k = 2 and we are using the Chebyshev scalarizer s CS λ . We claim that the distribution YT = {y1, ..., y T } is the same and specifically let Y denote that random variable over the choice of λ S+ that corresponds to y arg max y Y s CS λ (y). Then, note that a standard way to draw a random weight on S+ is to draw random absolute Gaussians and then renormalize, so if R2 = P i |Xi|2, where Xi are i.i.d. Gaussian, then Y = arg max y Y min( |X1| R y2). Note that the optimization of the arg-max is unaffected by positive scalar multiples of the objective, so multiplying by R2/(|X1||X2|) gives Y = arg max y Y min( R |X2|y1, R |X1|y2). Note since X1, X2 are i.i.d., we conclude that Y is the same distribution as the random variable that is given by y arg max y Y s HV λ (y). Since monotone transforms of the objective does not change the arg-max distribution, we conclude. B.1 Proofs of Lipschitz Properties Again, recall that for a scalarization function sλ(x), sλ is Lp-Lipschitz with respect to x in the ℓp norm on X if for x1, x2 X, |sλ(x1) sλ(x2)| Lp x1 x2 p, and analogously for define Lλ for p = 2 so |sλ1(x) sλ2(x)| Lλ λ1 λ2 2. We utilize the fact that if sλ is differentiable everywhere except for a finite set, bounding Lipschitz constants is equivalent to bounding the dual norm sλ q, where 1/p + 1/q = 1, which follows from mean value theorem, which we state as Proposition 12. Proposition 12. Let f : X R be a continuous function that is differentiable everywhere except on a finite set, then if f(x) q Lp for all x X, f(x) is Lp-Lipshitz with respect to the ℓp norm. Lemma 13. Let sλ(y) = λ y be the linear scalarization with λ 1 and y 1. Then, we may bound Lp max(1, k1/2 1/p) and Lλ Proof of Lemma 13. Since λsλ(x) = y, we use Proposition 12 to bound Lλ maxy y k. Similarly, since ysλ(y) = λ, we may bound for p 2, Lp λ q λ 1 for 1/p + 1/q = 1 and for p 2, we may use Holder s inequality to bound Lp λ q k1/q 1/2 λ k1/2 1/p. To bound the absolute value of sλ, note sλ(y) = λ y k for all since y 2 Lemma 14. Let sλ(y) = mini λiyi be the Chebyshev scalarization with λ 1 and y 1. Then, we may bound Lp 1 and Lλ 1 and |sλ| 1/ Proof of Lemma 14. For a specific λ, y, let i be the optimal index of the minimization. Then, the gradient λsλ(x) is simply zero in every coordinate except at i , where it is yi . Therefore, since we can only have a finite number of discontinuities due to monotonicity, we use Proposition 12 to bound Lλ yi 1. Similarly, since ysλ(y) has only one non-zero coordinate except at i , which is λi , we may bound for Lq λi 1. To bound the absolute value of sλ, note that there must exists λi < 1/ k as λ 1. Thus, mini λiyi < 1/ Lemma 15. Let sλ(y) = mini(yi/λi)k be the hypervolume scalarization with λ = 1 and 0 < Bl yi Bu. Then, we may bound Lp Bk u Blkk/2 1 and Lλ Bk+1 u Blk(k 1)/2 and |sλ| Bk u kk/2 . Proof of Lemma 15. For a specific λ, y, let i be the optimal index of the minimization. Then, the gradient λsλ(x) is simply zero in every coordinate except at i , which in absolute value is k(yi /λi )k(1/λi ). Let j be the index such that λj is maximized and since λ = 1, we know that λj 1/ k. Therefore, we see that since yi /λi yj/λj yj/ k, we conclude that 1/λi (Bu/Bl)/ Therefore, using Proposition 12, we have Lλ k(yi /λi )k(1/λi ) k(Bu/ k)k (Bu/Bl) Bk+1 u Blk(k 1)/2 And similarly, since ysλ(y) has only one non-zero coordinate except at i , which is k(yi /λi )k 1(1/λi ), we may bound for Lq k(yi /λi )k 1(1/λi ) k(Bu/ k)k 1 (Bu/Bl) k Bk u Blkk/2 1 To bound the absolute value of sλ, note that sλ(y) ( yj λj )k Bk u kk/2 . B.2 Proofs for Linear Bandits The following lemma about the UCB ellipsoid is borrowed from the original analysis of linear bandits. Lemma 16 (Abbasi-Yadkori et al. [2011]). Consider the least squares estimator bθt = (Mt) 1A t yt, where the covariance matrix of the action matrix is Mt = A t At + λI, then with probability 1 δ, δ ) + d log(T/λ) Proof of Lemma 10. Let bΘT be the least squares estimate of the true parameters after observing (AT , y T ). Since the noise ξt in each objective is independent and 1-sub-Gaussian, by Lemma 16, if we let MT = A T AT + λI, then with regularization λ = 1 bΘT i Θ i MT 1 + p 2 log(k/δ) + d log(T) := DT holds with probability at least 1 δ/k. Note that this describes the confidence ellipsoid, CT i = {θ Rd : bΘT i θi MT DT } for ΘT i. By the definition of the UCB maximization of at, we see at, eΘt = argmaxa A maxθi CT i sλ(Θ i a). Note that since Θ CT , we can bound the instantaneous scalarized regret as: r(sλ, at) = max a A sλ(Θ a) sλ(Θ at) sλ(eΘtat) sλ(Θ at) By the Lipschitz smoothness condition, we conclude that r(sλ, at) Lp (f Θt Θ )at p. To bound the desired ℓp norm, first note that by triangle inequality, f Θt Θ MT 2DT . Since we apply uniform exploration every other step and P 2I for ei E with size |E| = d, we conclude that MT T 5d I. Therefore, we conclude that bΘT i Θ i 5 p d/TDT := ET with probability at least 1 δ/k. Since at 1, we conclude by Cauchy-Schwarz, that |(bΘT i Θ i )at| ET . Together with our Lipschitz condition, we conclude that r(sλ, at) k1/p Lp ET 10k1/p Lpd p (log(k/δ) + log(T))/T Theorem 17. Assume that for any a A, |sλ(Θ a)| B for some B and sλ is Lλ-Lipschitz with respect to the ℓ2 norm in λ. With constant probability, the Bayes regret of running Algorithm 1 at round T can be bounded by BR(sλ, AT ) O Proof of Theorem 17. For any set of actions A A, we define f A(λ) = maxa A sλ(Θ a). We let F = {f A : A A} be our class of functions over all possible action sets and for any Bayes regret bounds, we will first demonstrate uniform convergence by bounding the complexity of F. Specifically, by generalization bounds from Rademacher complexities Bartlett and Mendelson [2002], over choices of λi D, we know that with probability 1 δ, for all A, we have the bound E λ D[f A] 1 i=1 f A(λi) where Rm(F) = Eλi D,σi , where σi are i.i.d. 1 Rademacher variables. To bound Rm(F), we appeal to Dudley s integral formulation that allows us to use the metric entropy of F to bound Rm(F) inf α>0 log(N(ϵ, F, 2)) where N denotes the standard covering number for F under the ℓ2 function norm metric over λ D. Since D is the uniform distribution over S+, this induces a natural ℓ function norm metric on F that is f = supλ S+ |f(λ)|. Since sλ(Θ a) is Lλ Lipschitz with respect to the Euclidean norm in λ. Note that since the maximal operator preserves Lipschitzness, f A(λ) is also Lλ-Lipschitz with respect to λ Rk. Since F contains Lλ-Lipschitz functions in Rk, we can bound the metric entropy via a covering of λ via a Lipschitz covering argument (see Lemma 4.2 of Gottlieb et al. [2016]), so we have log(N(ϵ, F, 2)) log(N(ϵ, F, )) (4Lλ/ϵ)k log(8/k) Finally, we follow the same Dudley integral computation of Theorem 4.3 of Gottlieb et al. [2016] to get that Rm(F) inf α>0 4α + 12 Z 2 (4Lλ/ϵ)k log(8/k) = O(Lλ/m1/(k+1)) Therefore, we conclude that with probability at least 1 δ over the independent choices of λi D, for all A, max a A sλ(Θ a) 1 i=1 max a A sλi(Θ a) O BLλ m1/(k+1) Finally, note that for T even, with constant probability, BR(sλ, At) = E λ D[r(sλ, At)] = E λ D[max a A sλ(Θ a) max a AT sλ(Θ a)] max a A sλi(Θ a) max a AT sλi(Θ a) + O BLλ T 1/(k+1) max a A sλi(Θ a) sλi(Θ a2i) + O BLλ T 1/(k+1) i=1 r(sλi, a2i) + O BLλ T 1/(k+1) T + BLλ T 1/(k+1) where the last line used Lemma 10 with δ = 1/T 2 and applied a union bound over all O(T) iterations. Proof of Theorem 11. Note that by Lemma 6, we connect the Bayes regret to the hypervolume regret for D : HVz(Θ A) HVz(Θ At) = ck E λ D[max a A sλ(Θ a) max a AT sλ(Θ a)] where sλ(y) = mini(yi zi/λi)k. Note that since Θ a 1 for all a A and B is maximal, we have B Θ a z B + 2. Therefore, we conclude by Lemma 15 that sλ is Lipschitz with Lp (B + 2)k Bkk/2 1 , Lλ (B + 2)k+1 Bk(k 1)/2 , |sλ| (B + 2)k Finally, we combine this with Theorem 17 with p = as the optimal choice of p (since Lp does not depend on p) to get our desired bound on hypervolume regret. Proof of Theorem 8. We let A = {a : a 1} be the unit sphere and Θ i = ei be the unit vector directions. Note that in this case the Pareto frontier is exactly Sk 1 + . Consider a uniform discretization of the Pareto front by taking an ϵ grid with respect to each angular component with respect to the polar coordinates. Let p1, ..., pm be the center (in terms of each of the k 1 angular dimensions) in the m = Θ((1/ϵ)k 1) grid elements. We consider the output y T = Θ AT and assume that for some grid element i, it contains none of the T outputs y T . Since our radial component r = 1, by construction of our grid in the angular component, we deduce that mint yt pi > ϵ/10 by translating polar to axis-aligned coordinates. Let ϵ = ϵ/10. Assume also that 1 k < pi < 1 1 k. Next, we claim that the hypercube from pi ϵ /k2 to pi is not dominated by any points in YT . Assume otherwise that there exists yt such that yt pi ϵ /k2. Now, this combined with the fact that since mint yt pi > ϵ implies that there must exist a coordinate such that ytj pj + ϵ . However, this implies that i=1 y2 ti X i =j (pi ϵ /k2)2 + (pj + ϵ )2 i p2 i 2(ϵ /k2) X i =j pi + 2ϵ pj > 1 where the last inequality follows since P k and pj > 1 k by assumption. However, this contradicts that yt 1, so it follows that pi ϵ is not dominated. Therefore, for any grid element such that pi > 1/k, if there is no yt in the grid, we must have a hypervolume regret of at least Ω(ϵ k) = Ω(ϵk) be simply consider the undominated hypervolume from pi to pi ϵ , which lies entirely within the grid element. In fact, since there are Θ((1/ϵ)k 1) such grid elements satisfying pi > 1/k, we see that if T < O((1/ϵ)k 1), by pigeonhole, there must be a hypervolume regret of at least Ω((1/ϵ)k 1ϵk) = Ω(ϵ) Therefore, for any 1/2 > ϵ > 0, HVz(Θ A) HVz(Θ AT ) < ϵ implies that T = Ω((1/ϵ)k 1). Rearranging shows that HVz(Θ A) HVz(Θ AT ) = Ω(T 1/(k 1)) C.1 Synthetic Figures Here are the relevant full plots from synthetic optimization of Pareto frontiers. Note that the title of each plots explicitly mention the optimized function. Generally, we observe that the hypervolume scalarizer has better performance on more concave curves. We also include a plot of a concave function multiplied by a linear combination of a convex and concave function, given by z = exp( x)(x exp( y) + (1 x)(3 exp(y))), which demonstrates that hypervolume scalarization performs competitively even with convex frontier regions. C.2 BBOB Figures Figure 5: Comparisons of the hypervolume indicator and the optimization fronts with BBOB functions. The left plot tracks the dominated hypervolume as a function of trials that were evaluated. The blue/orange dots represent the frontier points of the UCB-HV/EHVI algorithms respectively, over 5 repeats. Especially in high dimensions, EHVI tends overly concentrate on points in the middle of the frontier, limiting its hypervolume gain, while hypervolume scalarizations produce more diverse points. 0 20 40 60 80 100 120 140 160 # of Trials 18.007 18.407 18.808 19.208 19.608 20.008 20.408 20.808 21.208 21.609 22.009 22.409 22.809 23.209 23.609 24.009 24.410 24.810 25.210 25.610 hypervolume Dim: 16 Exp: ELLIPSOID_SEPARABLE EHVI UCB_HV 3.0 2.5 2.0 1.5 1.0 0.5 0.0 3.340 3.143 2.946 2.749 2.552 2.355 2.158 1.961 1.764 1.567 1.370 1.173 0.976 0.779 0.582 0.385 0.188 0.009 0.206 0.403 pareto_plot Dim: 16 Exp: ELLIPSOID_SEPARABLE EHVI UCB_HV 0 20 40 60 80 100 120 140 160 # of Trials 18.142 18.498 18.854 19.210 19.566 19.922 20.278 20.634 20.990 21.346 21.702 22.058 22.414 22.770 23.126 23.482 23.838 24.193 24.549 24.905 hypervolume Dim: 24 Exp: ELLIPSOID_SEPARABLE EHVI UCB_HV 3.0 2.5 2.0 1.5 1.0 0.5 0.0 2.626 2.472 2.318 2.165 2.011 1.858 1.704 1.551 1.397 1.244 1.090 0.937 0.783 0.630 0.476 0.323 0.169 0.015 0.138 0.292 pareto_plot Dim: 24 Exp: ELLIPSOID_SEPARABLE EHVI UCB_HV 0 20 40 60 80 100 120 140 160 # of Trials 18.495 18.856 19.218 19.579 19.940 20.302 20.663 21.025 21.386 21.748 22.109 22.471 22.832 23.193 23.555 23.916 24.278 24.639 25.001 25.362 hypervolume Dim: 8 Exp: ELLIPSOID_SEPARABLE EHVI UCB_HV 2.5 2.0 1.5 1.0 0.5 0.0 3.032 2.848 2.663 2.479 2.295 2.111 1.926 1.742 1.558 1.374 1.190 1.005 0.821 0.637 0.453 0.268 0.084 0.100 0.284 0.468 pareto_plot Dim: 8 Exp: ELLIPSOID_SEPARABLE EHVI UCB_HV 0 20 40 60 80 100 120 140 160 # of Trials 16.976 17.337 17.699 18.061 18.423 18.785 19.146 19.508 19.870 20.232 20.593 20.955 21.317 21.679 22.040 22.402 22.764 23.126 23.487 23.849 hypervolume Dim: 16 Exp: ELLIPSOID_SEPARABLE:RASTRIGIN_SEPARABLE EHVI UCB_HV 2.5 2.0 1.5 1.0 0.5 2.128 1.998 1.868 1.739 1.609 1.479 1.350 1.220 1.091 0.961 0.831 0.702 0.572 0.442 0.313 0.183 0.053 0.076 0.206 0.336 pareto_plot Dim: 16 Exp: ELLIPSOID_SEPARABLE:RASTRIGIN_SEPARABLE EHVI UCB_HV 0 20 40 60 80 100 120 140 160 # of Trials 16.609 16.971 17.333 17.695 18.057 18.419 18.780 19.142 19.504 19.866 20.228 20.590 20.952 21.313 21.675 22.037 22.399 22.761 23.123 23.485 hypervolume Dim: 24 Exp: ELLIPSOID_SEPARABLE:RASTRIGIN_SEPARABLE EHVI UCB_HV 3.0 2.5 2.0 1.5 1.0 0.5 2.182 2.051 1.921 1.790 1.660 1.529 1.399 1.269 1.138 1.008 0.877 0.747 0.616 0.486 0.355 0.225 0.094 0.036 0.167 0.297 pareto_plot Dim: 24 Exp: ELLIPSOID_SEPARABLE:RASTRIGIN_SEPARABLE EHVI UCB_HV 0 20 40 60 80 100 120 140 160 # of Trials 17.565 17.950 18.335 18.720 19.105 19.491 19.876 20.261 20.646 21.031 21.416 21.801 22.186 22.571 22.956 23.341 23.726 24.111 24.496 24.881 hypervolume Dim: 8 Exp: ELLIPSOID_SEPARABLE:RASTRIGIN_SEPARABLE EHVI UCB_HV 2.5 2.0 1.5 1.0 0.5 0.0 2.230 2.091 1.952 1.813 1.674 1.535 1.396 1.257 1.118 0.979 0.840 0.701 0.562 0.423 0.284 0.145 0.006 0.133 0.272 0.411 pareto_plot Dim: 8 Exp: ELLIPSOID_SEPARABLE:RASTRIGIN_SEPARABLE EHVI UCB_HV 21.230 21.719 22.207 22.696 23.184 23.672 24.161 24.649 25.138 25.626 Dim: 16 Exp: ELLIPSOID_SEPARABLE:SCHWEFEL EHVI UCB_HV 1.087 0.928 0.770 0.611 0.452 0.294 0.135 0.024 0.182 0.341 Dim: 16 Exp: ELLIPSOID_SEPARABLE:SCHWEFEL EHVI UCB_HV 21 0 20 40 60 80 100 120 140 160 # of Trials 16.346 16.834 17.323 17.811 18.300 18.788 19.276 19.765 20.253 20.742 4 3 2 1 0 2.673 2.515 2.356 2.197 2.039 1.880 1.721 1.563 1.404 1.245 0 20 40 60 80 100 120 140 160 # of Trials 16.447 16.929 17.411 17.893 18.375 18.857 19.338 19.820 20.302 20.784 21.266 21.748 22.230 22.712 23.194 23.676 24.158 24.639 25.121 25.603 hypervolume Dim: 24 Exp: ELLIPSOID_SEPARABLE:SCHWEFEL EHVI UCB_HV 4 3 2 1 0 3.227 3.039 2.851 2.662 2.474 2.286 2.097 1.909 1.721 1.533 1.344 1.156 0.968 0.780 0.591 0.403 0.215 0.027 0.162 0.350 pareto_plot Dim: 24 Exp: ELLIPSOID_SEPARABLE:SCHWEFEL EHVI UCB_HV 0 20 40 60 80 100 120 140 160 # of Trials 15.641 16.156 16.670 17.185 17.699 18.214 18.729 19.243 19.758 20.272 20.787 21.301 21.816 22.330 22.845 23.359 23.874 24.389 24.903 25.418 hypervolume Dim: 8 Exp: ELLIPSOID_SEPARABLE:SCHWEFEL EHVI UCB_HV 4 3 2 1 0 4.338 4.084 3.831 3.578 3.324 3.071 2.818 2.565 2.311 2.058 1.805 1.552 1.298 1.045 0.792 0.538 0.285 0.032 0.221 0.475 pareto_plot Dim: 8 Exp: ELLIPSOID_SEPARABLE:SCHWEFEL EHVI UCB_HV 0 20 40 60 80 100 120 140 160 # of Trials 18.428 18.813 19.197 19.582 19.967 20.351 20.736 21.121 21.505 21.890 22.274 22.659 23.044 23.428 23.813 24.198 24.582 24.967 25.351 25.736 hypervolume Dim: 16 Exp: ELLIPSOID_SEPARABLE:SPHERE EHVI UCB_HV 2.0 1.5 1.0 0.5 0.0 1.975 1.861 1.748 1.635 1.521 1.408 1.295 1.182 1.068 0.955 0.842 0.728 0.615 0.502 0.388 0.275 0.162 0.048 0.065 0.178 pareto_plot Dim: 16 Exp: ELLIPSOID_SEPARABLE:SPHERE EHVI UCB_HV 0 20 40 60 80 100 120 140 160 # of Trials 17.818 18.197 18.575 18.953 19.332 19.710 20.089 20.467 20.845 21.224 21.602 21.981 22.359 22.737 23.116 23.494 23.872 24.251 24.629 25.008 hypervolume Dim: 24 Exp: ELLIPSOID_SEPARABLE:SPHERE EHVI UCB_HV 1.75 1.50 1.25 1.00 0.75 0.50 0.25 0.00 0.25 1.728 1.635 1.542 1.449 1.356 1.263 1.170 1.077 0.984 0.891 0.798 0.705 0.612 0.519 0.426 0.333 0.240 0.147 0.054 0.039 pareto_plot Dim: 24 Exp: ELLIPSOID_SEPARABLE:SPHERE EHVI UCB_HV 0 20 40 60 80 100 120 140 160 # of Trials 17.978 18.431 18.883 19.335 19.788 20.240 20.692 21.145 21.597 22.050 22.502 22.954 23.407 23.859 24.311 24.764 25.216 25.668 26.121 26.573 hypervolume Dim: 8 Exp: ELLIPSOID_SEPARABLE:SPHERE EHVI UCB_HV 2.0 1.5 1.0 0.5 0.0 2.604 2.451 2.297 2.144 1.991 1.838 1.685 1.532 1.378 1.225 1.072 0.919 0.766 0.613 0.459 0.306 0.153 0.000 0.153 0.306 pareto_plot Dim: 8 Exp: ELLIPSOID_SEPARABLE:SPHERE EHVI UCB_HV 0 20 40 60 80 100 120 140 160 # of Trials 15.824 16.089 16.354 16.619 16.884 17.149 17.414 17.678 17.943 18.208 18.473 18.738 19.003 19.268 19.533 19.797 20.062 20.327 20.592 20.857 hypervolume Dim: 16 Exp: RASTRIGIN_SEPARABLE EHVI UCB_HV 2.5 2.0 1.5 1.0 0.5 0.0 5.241 4.960 4.680 4.399 4.119 3.838 3.558 3.277 2.997 2.716 2.436 2.155 1.875 1.594 1.314 1.034 0.753 0.473 0.192 0.088 pareto_plot Dim: 16 Exp: RASTRIGIN_SEPARABLE EHVI UCB_HV 15 769 16.016 16.262 16.509 16.756 17.003 17.250 17.497 17.743 17.990 18.237 18.484 18.731 18.978 19.224 19.471 19.718 19.965 20.212 hypervolume Dim: 24 Exp: RASTRIGIN_SEPARABLE EHVI UCB_HV 4 842 4.577 4.311 4.046 3.781 3.516 3.251 2.986 2.721 2.456 2.191 1.925 1.660 1.395 1.130 0.865 0.600 0.335 0.070 pareto_plot Dim: 24 Exp: RASTRIGIN_SEPARABLE EHVI UCB HV # -*- coding: utf-8 -*- """3D Hypervolume Experiments Automatically generated by Colab. Original file is located at https://colab.corp.google.com/drive/1Xbxab Rf_aqg E0n RY-c X6Mcie Uc3Xn HSG ### Imports """ import numpy as np import matplotlib.pyplot as plt from vizier.pyvizier import multimetric from vizier.pyvizier.multimetric import xla_pareto eps = 1e-4 origin = np.zeros(shape=3) - eps front = multimetric.Pareto Frontier( points=np.zeros(shape=(2,3)) - eps, origin=origin, num_vectors=10000, cum_hypervolume_base=xla_pareto.jax_cum_hypervolume_origin, ) hv_curve = front.hypervolume(is_cumulative=True) import numpy as np import matplotlib.pyplot as plt pi = np.pi + 1e-4 # Define the function to plot threshold = 0.1 def g(x): if x <= threshold: return np.cos(x*pi/(2*threshold)) + 1 else: return np.exp(-3*(x-threshold)) g = lambda x : np.cos(x*pi) + 1 # Generate some data x = np.linspace(0, 1, 500) y = np.vectorize(g)(x) # Plot the data plt.plot(x, y, bo ) plt.xlabel( x ) plt.ylabel( y ) plt.title( 1D Function Plot ) plt.show() def f(x, y): # return np.vectorize(g)(x) * np.vectorize(g)(y) # return (3 - np.exp(x)) * (x * np.exp(-y) + (1-x)*(3-np.exp(y))) # return (3 - np.exp(x))*(3np.exp(y)) # return np.exp(-x) * (3-np.exp(y)) return (3-np.exp(x)) * (np.cos(y * pi) + 1) return np.exp(-x) * (x * np.exp(-y) + (1-x) * (3 - np.exp(y))) # return np.exp(-x - y) x = np.linspace(0, 1, 30) y = np.linspace(0, 1, 30) X, Y = np.meshgrid(x, y) Z = f(X, Y) fig = plt.figure(figsize=(4, 4)) ax = plt.axes(projection= 3d ) ax.contour3D(X, Y, Z, 500) ax.set_xlabel( x ) ax.set_ylabel( y ) ax.set_zlabel( z ) ax.set_title( z = (3-exp(x))(3 - exp(y)) ) x = X.flatten() y = Y.flatten() z = Z.flatten() xla_pareto.is_frontier(np.array([x, y, z]).T) total_hypervolume = front.hypervolume(additional_points=np.array([x, y, z]).T) total_hypervolume #@title Hypervolume code def get_points(scalarizer_generator, num_points = 50): index_maxes = [] for _ in range(num_points): outputs = scalarizer_generator(np.array([x, y, z])) index_max = np.argmax(outputs) index_maxes.append(index_max) return index_maxes def linear_gaussian_generator(array): weights = abs(np.random.normal(size=(array.shape[0], 1))) return np.sum(weights * array, axis=0) def cheby_gaussian_generator(array): weights = abs(np.random.normal(size=(array.shape[0], 1))) return np.min(weights * array, axis=0) def hv_gaussian_generator(array): weights = abs(np.random.normal(size=(array.shape[0], 1))) return np.min((1/weights) * array, axis=0) generators = { linear : linear_gaussian_generator, cheby : cheby_gaussian_generator, hv : hv_gaussian_generator, } index_maxes = get_points(hv_gaussian_generator) from vizier.benchmarks.analyzers import plot_median_convergence num_repeats = 10 generator_curves = {} for _ in range(num_repeats): for key, generator in generators.items(): index_maxes = get_points(generator, num_points=500) new_points = np.vstack([x[index_maxes], y[index_maxes], z[index_maxes]]).T hv_curve = front.hypervolume(is_cumulative=True, additional_points=new_points) regret_curve = total_hypervolume - hv_curve if key in generator_curves: generator_curves[key] = np.vstack([generator_curves[key], regret_curve]) else: generator_curves[key] = regret_curve fig, ax = plt.subplots(1, 1, figsize=(4,4)) ax.set_xlabel( # of Pareto Points ) ax.set_ylabel( Hypervolume Regret ) ax.set_title( z = (3exp(x)) * (cos(pi * y) + 1) ) ax.set_yscale( log ) for key, curves in generator_curves.items(): print(curves.shape) plot_median_convergence(ax, curves, label=key, percentiles=((25, 75),)) plt.legend() # -*- coding: utf-8 -*- """Multiobjective Linear Bandits Automatically generated by Colab. Original file is located at https://colab.corp.google.com/drive/1CD7ek1DV4f3FNzo O7H1rmb0k Ok Mvx80Z """ import dataclasses import itertools import math import os import re from absl import flags import matplotlib.pyplot as plt import numpy as np # @title Simple Lin UCB implementation { display-mode: "form" } import numpy as np import pandas as pd import seaborn as sns from google3.file.recordio.python import recordio from google3.learning.vizier import benchmark_v2 from google3.learning.vizier.benchmark_v2 import analysis from google3.learning.vizier.benchmark_v2 import config_pb2 import google3.pyglib.gfile as gfile class Lin UCB: """Simple Lin UCB implementation.""" def __init__( self, dimension: int, max_inst_regret: float = 2.0, regularizer: float = 1.0, var_noise: float = 1.0, max_parameter_norm: float = 1.0, failure_probabilty: float = 0.1, ): assert regularizer >= 0.0 self.dimension = dimension self._regularizer = regularizer self._var_noise = var_noise self._max_parameter_norm = max_parameter_norm self._max_inst_regret = max_inst_regret self._failure_probability = failure_probabilty self.reset() def reset(self): self._covariance_inv = np.eye(self.dimension) * self._regularizer self._reward_scaled_features = np.zeros(self.dimension) self._parameter_estimate = np.zeros(self.dimension) self._num_observations = 0 self._last_conf_radius = None self._conf_ellipsoid_width_raw = self._conf_ellipsoid_rhs() def add_observation(self, action, reward: float, context=None): """add an observation (action, reward) to the learner. This function updates the covariance matrix and parameter estimate. Args: action: action that was taken reward: achieved reward context: context for this observation blamed: whether this algorithm is to be blamed for this observation. """ self._num_observations += 1 # Sherman Morrison update of covariance matrix y = np.dot(self._covariance_inv, action) self._covariance_inv -= np.outer(y, y) / (1 + np.inner(action, y)) # regression target self._reward_scaled_features += reward * action # parameter estimate self._parameter_estimate = np.dot( self._covariance_inv, self._reward_scaled_features ) self._conf_ellipsoid_width_raw = self._conf_ellipsoid_rhs() def ucb(self, actions, confidence_scale=1.0): """computes the upper-confidence bound for a batch of actions. Args: actions: A x d matrix confidence_scale: scaling factor in front of the bonus prescribed by theory Returns: upper-confidence bound for each A action, A vector """ rewards = np.dot(actions, self._parameter_estimate) var = np.sum(actions * np.dot(actions, self._covariance_inv), axis=1) beta = self._conf_ellipsoid_width_raw * confidence_scale return rewards, beta * np.sqrt(var) def choose_action(self, actions, confidence_scale=1.0): """chooses the action that maximizes the upper confidence bound. Args: actions: A x d matrix, possible actions to choose from confidence_scale: scaling factor in front of the bonus prescribed by theory Returns: chosen action (d vector) """ rewards, conf_b = self.ucb(actions, confidence_scale) action_index = randargmax(rewards + conf_b) return actions[action_index, :] def _conf_ellipsoid_rhs(self): # from Abbassi-Yadkori et al. 2011 beta_t = np.sqrt(self._regularizer) * self._max_parameter_norm log_dets = -np.log(self._failure_probability) log_dets -= self.dimension / 2 * np.log(self._regularizer) log_dets -= np.linalg.slogdet(self._covariance_inv)[1] / 2 beta_t += np.sqrt(2 * self._var_noise * log_dets) return beta_t def randargmax(b): """takes the argmax but randomly picks from the set of maximizers.""" return np.random.choice(np.flatnonzero(b == b.max())) def linear_scalarizer(acquisitions, weights): sum = 0.0 for acquisition, weight in zip(acquisitions, weights): sum += weight * acquisition return sum def chebyshev_scalarizer(acquisitions, weights): min = np.inf for acquisition, weight in zip(acquisitions, weights): min = np.minimum((acquisition - reference) * weight, min ) return min def hypervolume_scalarizer(acquisitions, weights): min = np.inf for acquisition, weight in zip(acquisitions, weights): min = np.minimum((acquisition - reference)/ weight, min ) return min def uniform_weights(): weights = np.random.normal(size=num_metrics) return abs(weights)/np.linalg.norm(weights) def boxed_linear_weights(): # Use bounding box. u = np.random.uniform(low=1.0, high=3.0, size=num_metrics) return u / np.linalg.norm(u, ord=1, keepdims=True) def boxed_chebyshev_weights(): lmda = boxed_linear_weights() lmda_prime = 1.0/lmda return lmda_prime/np.linalg.norm(lmda_prime, ord=1, keepdims=True) dim = 5 num_metrics = 16 thetas = np.random.normal(size=(num_metrics, dim)) # Anti-correlate thetas. thetas[0, :] = -thetas[1, :] + np.random.normal( size=thetas[1, :].shape, scale=0.01 ) for i in range(int(num_metrics / 2)): index = int(2 * i) thetas[index, :] = -thetas[index + 1, :] + np.random.normal( size=thetas[index + 1, :].shape, scale=0.01 ) # Renormalize to make sure |theta| = 1 thetas = thetas / np.linalg.norm(thetas, axis=1, keepdims=True) reference = -2 def run_linear_bandit(thetas, scalarizer, weight_generator, num_rounds=100): algs = [Lin UCB(dimension=dim) for theta in thetas] expected_rewards = np.empty(shape=(num_rounds, num_metrics)) actions = np.random.normal(size=(1000, dim)) actions = actions / np.linalg.norm(actions, axis=1)[..., np.newaxis] for t in range(num_rounds): index = np.random.choice(len(actions)) action = actions[index] expected_reward = np.inner(thetas, action) rewards = expected_reward + np.random.normal(size=expected_reward.shape) for alg, reward in zip(algs, rewards): alg.add_observation(action, reward) weights = weight_generator() acquisitions = scalarizer([sum(alg.ucb(actions)) for alg in algs], weights) assert len(acquisitions) == len(actions) index = randargmax(acquisitions) action = actions[index] expected_reward = np.inner(thetas, action) rewards = expected_reward + np.random.normal(size=expected_reward.shape) for alg, reward in zip(algs, rewards): alg.add_observation(action, reward) expected_rewards[t] = expected_reward return actions, expected_rewards scalarizations = { linear-uniform : (linear_scalarizer, uniform_weights), linear-boxed : (linear_scalarizer, boxed_linear_weights), chebyshev-uniform : (chebyshev_scalarizer, uniform_weights), chebyshev-boxed : (chebyshev_scalarizer, boxed_chebyshev_weights), hypervolume-uniform : (hypervolume_scalarizer, uniform_weights), hypervolume-boxed : (hypervolume_scalarizer, boxed_linear_weights), } from vizier.pyvizier import multimetric from vizier.pyvizier.multimetric import xla_pareto pareto_algo = xla_pareto.Jax Pareto Optimal Algorithm() import matplotlib.pyplot as plt num_rounds = 100 actions, linear_rewards = run_linear_bandit( thetas, scalarizer=hypervolume_scalarizer, weight_generator=uniform_weights ) all_values = np.inner(actions, thetas) all_frontier = all_values[pareto_algo.is_pareto_optimal(all_values)] frontier = linear_rewards[pareto_algo.is_pareto_optimal(linear_rewards)] plt.scatter(all_values[:, 0], all_values[:, 1], color= grey ) plt.scatter( all_frontier[:, 0], all_frontier[:, 1], color= red , label= Pareto-optimal points , s=30, ) plt.scatter( frontier[:, 0], frontier[:, 1], color= green , label= Discovered Frontier , s=30, ) plt.xlabel( y_1 , fontsize=13) plt.ylabel( y_2 , fontsize=13) plt.title( f Hypervolume scalarizer optimization at T={num_rounds} , fontsize=18 ) plt.legend() origin = -2 * np.ones(shape=num_metrics) num_repeats = 5 fig, ax = plt.subplots(1, 1, figsize=(12, 8)) for key, (scalarizer, weight_generator) in scalarizations.items(): hvs = [] for _ in range(num_repeats): _, linear_rewards = run_linear_bandit( thetas, scalarizer, weight_generator, num_rounds=200 ) front = multimetric.Pareto Frontier( points=linear_rewards, origin=origin, cum_hypervolume_base=xla_pareto.jax_cum_hypervolume_origin, ) hvs.append(front.hypervolume(is_cumulative=True)) y = np.array(hvs) analysis.plot_median_convergence( ax, y, xs=np.array(range(y.shape[1])), label=key, percentiles=((30, 70),) ) ax.legend() ax.set_title(f Cumulative HV from {origin} , fontsize=15) ax.set_ylabel( hypervolume (HV) ) ax.set_xlabel( # of Trials ) Neur IPS Paper Checklist The checklist is designed to encourage best practices for responsible machine learning research, addressing issues of reproducibility, transparency, research ethics, and societal impact. Do not remove the checklist: The papers not including the checklist will be desk rejected. The checklist should follow the references and follow the (optional) supplemental material. The checklist does NOT count towards the page limit. Please read the checklist guidelines carefully for information on how to answer these questions. For each question in the checklist: You should answer [Yes] , [No] , or [NA] . [NA] means either that the question is Not Applicable for that particular paper or the relevant information is Not Available. Please provide a short (1 2 sentence) justification right after your answer (even for NA). The checklist answers are an integral part of your paper submission. They are visible to the reviewers, area chairs, senior area chairs, and ethics reviewers. You will be asked to also include it (after eventual revisions) with the final version of your paper, and its final version will be published with the paper. The reviewers of your paper will be asked to use the checklist as one of the factors in their evaluation. While "[Yes] " is generally preferable to "[No] ", it is perfectly acceptable to answer "[No] " provided a proper justification is given (e.g., "error bars are not reported because it would be too computationally expensive" or "we were unable to find the license for the dataset we used"). In general, answering "[No] " or "[NA] " is not grounds for rejection. While the questions are phrased in a binary way, we acknowledge that the true answer is often more nuanced, so please just use your best judgment and write a justification to elaborate. All supporting evidence can appear either in the main paper or the supplemental material, provided in appendix. If you answer [Yes] to a question, in the justification please point to the section(s) where related material for the question can be found. IMPORTANT, please: Delete this instruction block, but keep the section heading Neur IPS paper checklist", Keep the checklist subsection headings, questions/answers and guidelines below. Do not modify the questions and only use the provided macros for your answers. Question: Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? Answer: [Yes] Justification: Abstract clearly state claims and introduction has a list of main contributions, both theoretical and experimental. Guidelines: The answer NA means that the abstract and introduction do not include the claims made in the paper. The abstract and/or introduction should clearly state the claims made, including the contributions made in the paper and important assumptions and limitations. A No or NA answer to this question will not be perceived well by the reviewers. The claims made should match theoretical and experimental results, and reflect how much the results can be expected to generalize to other settings. It is fine to include aspirational goals as motivation as long as it is clear that these goals are not attained by the paper. 2. Limitations Question: Does the paper discuss the limitations of the work performed by the authors? Answer: [Yes] Justification: In our introduction, we provide limitations of works given in the beyond linear scalarization and hypervolume regret subsections. Guidelines: The answer NA means that the paper has no limitation while the answer No means that the paper has limitations, but those are not discussed in the paper. The authors are encouraged to create a separate "Limitations" section in their paper. The paper should point out any strong assumptions and how robust the results are to violations of these assumptions (e.g., independence assumptions, noiseless settings, model well-specification, asymptotic approximations only holding locally). The authors should reflect on how these assumptions might be violated in practice and what the implications would be. The authors should reflect on the scope of the claims made, e.g., if the approach was only tested on a few datasets or with a few runs. In general, empirical results often depend on implicit assumptions, which should be articulated. The authors should reflect on the factors that influence the performance of the approach. For example, a facial recognition algorithm may perform poorly when image resolution is low or images are taken in low lighting. Or a speech-to-text system might not be used reliably to provide closed captions for online lectures because it fails to handle technical jargon. The authors should discuss the computational efficiency of the proposed algorithms and how they scale with dataset size. If applicable, the authors should discuss possible limitations of their approach to address problems of privacy and fairness. While the authors might fear that complete honesty about limitations might be used by reviewers as grounds for rejection, a worse outcome might be that reviewers discover limitations that aren t acknowledged in the paper. The authors should use their best judgment and recognize that individual actions in favor of transparency play an important role in developing norms that preserve the integrity of the community. Reviewers will be specifically instructed to not penalize honesty concerning limitations. 3. Theory Assumptions and Proofs Question: For each theoretical result, does the paper provide the full set of assumptions and a complete (and correct) proof? Answer: [Yes] Justification: All the theorems, formulas, and proofs are given in the paper or appendix and are numbered and cross-referenced. All informal proofs have a formal complement and all assumptions are clearly stated. Guidelines: The answer NA means that the paper does not include theoretical results. All the theorems, formulas, and proofs in the paper should be numbered and crossreferenced. All assumptions should be clearly stated or referenced in the statement of any theorems. The proofs can either appear in the main paper or the supplemental material, but if they appear in the supplemental material, the authors are encouraged to provide a short proof sketch to provide intuition. Inversely, any informal proof provided in the core of the paper should be complemented by formal proofs provided in appendix or supplemental material. Theorems and Lemmas that the proof relies upon should be properly referenced. 4. Experimental Result Reproducibility Question: Does the paper fully disclose all the information needed to reproduce the main experimental results of the paper to the extent that it affects the main claims and/or conclusions of the paper (regardless of whether the code and data are provided or not)? Answer: [Yes] Justification: The hypervolume scalarization is easy to code up and try on a wide variety of benchmarks. All the code is released and all data is generated synthetic or via open-sourced benchmarks. Guidelines: The answer NA means that the paper does not include experiments. If the paper includes experiments, a No answer to this question will not be perceived well by the reviewers: Making the paper reproducible is important, regardless of whether the code and data are provided or not. If the contribution is a dataset and/or model, the authors should describe the steps taken to make their results reproducible or verifiable. Depending on the contribution, reproducibility can be accomplished in various ways. For example, if the contribution is a novel architecture, describing the architecture fully might suffice, or if the contribution is a specific model and empirical evaluation, it may be necessary to either make it possible for others to replicate the model with the same dataset, or provide access to the model. In general. releasing code and data is often one good way to accomplish this, but reproducibility can also be provided via detailed instructions for how to replicate the results, access to a hosted model (e.g., in the case of a large language model), releasing of a model checkpoint, or other means that are appropriate to the research performed. While Neur IPS does not require releasing code, the conference does require all submissions to provide some reasonable avenue for reproducibility, which may depend on the nature of the contribution. For example (a) If the contribution is primarily a new algorithm, the paper should make it clear how to reproduce that algorithm. (b) If the contribution is primarily a new model architecture, the paper should describe the architecture clearly and fully. (c) If the contribution is a new model (e.g., a large language model), then there should either be a way to access this model for reproducing the results or a way to reproduce the model (e.g., with an open-source dataset or instructions for how to construct the dataset). (d) We recognize that reproducibility may be tricky in some cases, in which case authors are welcome to describe the particular way they provide for reproducibility. In the case of closed-source models, it may be that access to the model is limited in some way (e.g., to registered users), but it should be possible for other researchers to have some path to reproducing or verifying the results. 5. Open access to data and code Question: Does the paper provide open access to the data and code, with sufficient instructions to faithfully reproduce the main experimental results, as described in supplemental material? Answer: [Yes] Justification: All the code is released and all data is generated synthetic or via open-sourced benchmarks. Guidelines: The answer NA means that paper does not include experiments requiring code. Please see the Neur IPS code and data submission guidelines (https://nips.cc/ public/guides/Code Submission Policy) for more details. While we encourage the release of code and data, we understand that this might not be possible, so No is an acceptable answer. Papers cannot be rejected simply for not including code, unless this is central to the contribution (e.g., for a new open-source benchmark). The instructions should contain the exact command and environment needed to run to reproduce the results. See the Neur IPS code and data submission guidelines (https: //nips.cc/public/guides/Code Submission Policy) for more details. The authors should provide instructions on data access and preparation, including how to access the raw data, preprocessed data, intermediate data, and generated data, etc. The authors should provide scripts to reproduce all experimental results for the new proposed method and baselines. If only a subset of experiments are reproducible, they should state which ones are omitted from the script and why. At submission time, to preserve anonymity, the authors should release anonymized versions (if applicable). Providing as much information as possible in supplemental material (appended to the paper) is recommended, but including URLs to data and code is permitted. 6. Experimental Setting/Details Question: Does the paper specify all the training and test details (e.g., data splits, hyperparameters, how they were chosen, type of optimizer, etc.) necessary to understand the results? Answer: [Yes] Justification: Yes, all experimental details are provided in the core of the paper or in the appendix/code. Guidelines: The answer NA means that the paper does not include experiments. The experimental setting should be presented in the core of the paper to a level of detail that is necessary to appreciate the results and make sense of them. The full details can be provided either with the code, in appendix, or as supplemental material. 7. Experiment Statistical Significance Question: Does the paper report error bars suitably and correctly defined or other appropriate information about the statistical significance of the experiments? Answer: [Yes] Justification: Yes error bars are provided and we assume the standard reporting of 30-70 percentile for error bars over independent repeats, as stated. Guidelines: The answer NA means that the paper does not include experiments. The authors should answer "Yes" if the results are accompanied by error bars, confidence intervals, or statistical significance tests, at least for the experiments that support the main claims of the paper. The factors of variability that the error bars are capturing should be clearly stated (for example, train/test split, initialization, random drawing of some parameter, or overall run with given experimental conditions). The method for calculating the error bars should be explained (closed form formula, call to a library function, bootstrap, etc.) The assumptions made should be given (e.g., Normally distributed errors). It should be clear whether the error bar is the standard deviation or the standard error of the mean. It is OK to report 1-sigma error bars, but one should state it. The authors should preferably report a 2-sigma error bar than state that they have a 96% CI, if the hypothesis of Normality of errors is not verified. For asymmetric distributions, the authors should be careful not to show in tables or figures symmetric error bars that would yield results that are out of range (e.g. negative error rates). If error bars are reported in tables or plots, The authors should explain in the text how they were calculated and reference the corresponding figures or tables in the text. 8. Experiments Compute Resources Question: For each experiment, does the paper provide sufficient information on the computer resources (type of compute workers, memory, time of execution) needed to reproduce the experiments? Answer: [No] Justification: Our paper do not compare resource usage, rather it improves in Trial efficiency. Furthermore, the use of different scalarizations does not affect resource usage. Guidelines: The answer NA means that the paper does not include experiments. The paper should indicate the type of compute workers CPU or GPU, internal cluster, or cloud provider, including relevant memory and storage. The paper should provide the amount of compute required for each of the individual experimental runs as well as estimate the total compute. The paper should disclose whether the full research project required more compute than the experiments reported in the paper (e.g., preliminary or failed experiments that didn t make it into the paper). 9. Code Of Ethics Question: Does the research conducted in the paper conform, in every respect, with the Neur IPS Code of Ethics https://neurips.cc/public/Ethics Guidelines? Answer: [Yes] Justification: Research is mainly theoretical and conforms with code of ethics. Guidelines: The answer NA means that the authors have not reviewed the Neur IPS Code of Ethics. If the authors answer No, they should explain the special circumstances that require a deviation from the Code of Ethics. The authors should make sure to preserve anonymity (e.g., if there is a special consideration due to laws or regulations in their jurisdiction). 10. Broader Impacts Question: Does the paper discuss both potential positive societal impacts and negative societal impacts of the work performed? Answer: [NA] Justification: Our research is mainly theoretical and improves optimization pipelines on synthetic data. Guidelines: The answer NA means that there is no societal impact of the work performed. If the authors answer NA or No, they should explain why their work has no societal impact or why the paper does not address societal impact. Examples of negative societal impacts include potential malicious or unintended uses (e.g., disinformation, generating fake profiles, surveillance), fairness considerations (e.g., deployment of technologies that could make decisions that unfairly impact specific groups), privacy considerations, and security considerations. The conference expects that many papers will be foundational research and not tied to particular applications, let alone deployments. However, if there is a direct path to any negative applications, the authors should point it out. For example, it is legitimate to point out that an improvement in the quality of generative models could be used to generate deepfakes for disinformation. On the other hand, it is not needed to point out that a generic algorithm for optimizing neural networks could enable people to train models that generate Deepfakes faster. The authors should consider possible harms that could arise when the technology is being used as intended and functioning correctly, harms that could arise when the technology is being used as intended but gives incorrect results, and harms following from (intentional or unintentional) misuse of the technology. If there are negative societal impacts, the authors could also discuss possible mitigation strategies (e.g., gated release of models, providing defenses in addition to attacks, mechanisms for monitoring misuse, mechanisms to monitor how a system learns from feedback over time, improving the efficiency and accessibility of ML). 11. Safeguards Question: Does the paper describe safeguards that have been put in place for responsible release of data or models that have a high risk for misuse (e.g., pretrained language models, image generators, or scraped datasets)? Answer: [NA] Justification: Our research is mainly theoretical and improves optimization pipelines on synthetic data. Guidelines: The answer NA means that the paper poses no such risks. Released models that have a high risk for misuse or dual-use should be released with necessary safeguards to allow for controlled use of the model, for example by requiring that users adhere to usage guidelines or restrictions to access the model or implementing safety filters. Datasets that have been scraped from the Internet could pose safety risks. The authors should describe how they avoided releasing unsafe images. We recognize that providing effective safeguards is challenging, and many papers do not require this, but we encourage authors to take this into account and make a best faith effort. 12. Licenses for existing assets Question: Are the creators or original owners of assets (e.g., code, data, models), used in the paper, properly credited and are the license and terms of use explicitly mentioned and properly respected? Answer: [Yes] Justification: Our research is mainly theoretical and all previous usages are cited. Guidelines: The answer NA means that the paper does not use existing assets. The authors should cite the original paper that produced the code package or dataset. The authors should state which version of the asset is used and, if possible, include a URL. The name of the license (e.g., CC-BY 4.0) should be included for each asset. For scraped data from a particular source (e.g., website), the copyright and terms of service of that source should be provided. If assets are released, the license, copyright information, and terms of use in the package should be provided. For popular datasets, paperswithcode.com/datasets has curated licenses for some datasets. Their licensing guide can help determine the license of a dataset. For existing datasets that are re-packaged, both the original license and the license of the derived asset (if it has changed) should be provided. If this information is not available online, the authors are encouraged to reach out to the asset s creators. 13. New Assets Question: Are new assets introduced in the paper well documented and is the documentation provided alongside the assets? Answer: [NA] Justification: No new assets are released. Guidelines: The answer NA means that the paper does not release new assets. Researchers should communicate the details of the dataset/code/model as part of their submissions via structured templates. This includes details about training, license, limitations, etc. The paper should discuss whether and how consent was obtained from people whose asset is used. At submission time, remember to anonymize your assets (if applicable). You can either create an anonymized URL or include an anonymized zip file. 14. Crowdsourcing and Research with Human Subjects Question: For crowdsourcing experiments and research with human subjects, does the paper include the full text of instructions given to participants and screenshots, if applicable, as well as details about compensation (if any)? Answer: [NA] Justification: The paper does not involve crowdsourcing nor research with human subjects. Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Including this information in the supplemental material is fine, but if the main contribution of the paper involves human subjects, then as much detail as possible should be included in the main paper. According to the Neur IPS Code of Ethics, workers involved in data collection, curation, or other labor should be paid at least the minimum wage in the country of the data collector. 15. Institutional Review Board (IRB) Approvals or Equivalent for Research with Human Subjects Question: Does the paper describe potential risks incurred by study participants, whether such risks were disclosed to the subjects, and whether Institutional Review Board (IRB) approvals (or an equivalent approval/review based on the requirements of your country or institution) were obtained? Answer: [NA] Justification: The paper does not involve crowdsourcing nor research with human subjects. Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Depending on the country in which research is conducted, IRB approval (or equivalent) may be required for any human subjects research. If you obtained IRB approval, you should clearly state this in the paper. We recognize that the procedures for this may vary significantly between institutions and locations, and we expect authors to adhere to the Neur IPS Code of Ethics and the guidelines for their institution. For initial submissions, do not include any information that would break anonymity (if applicable), such as the institution conducting the review.