# inversionbased_latent_bayesian_optimization__ef6a7ff7.pdf Inversion-based Latent Bayesian Optimization Jaewon Chu , Jinyoung Park , Seunghun Lee, Hyunwoo J. Kim Computer Science & Engineering Korea University {allonsy07, lpmn678, llsshh319, hyunwoojkim}@korea.ac.kr Latent Bayesian optimization (LBO) approaches have successfully adopted Bayesian optimization over a continuous latent space by employing an encoderdecoder architecture to address the challenge of optimization in a high dimensional or discrete input space. LBO learns a surrogate model to approximate the black-box objective function in the latent space. However, we observed that most LBO methods suffer from the misalignment problem , which is induced by the reconstruction error of the encoder-decoder architecture. It hinders learning an accurate surrogate model and generating high-quality solutions. In addition, several trust region-based LBO methods select the anchor, the center of the trust region, based solely on the objective function value without considering the trust region s potential to enhance the optimization process. To address these issues, we propose Inversion-based Latent Bayesian Optimization (Inv BO), a plug-and-play module for LBO. Inv BO consists of two components: an inversion method and a potential-aware trust region anchor selection. The inversion method searches the latent code that completely reconstructs the given target data. The potential-aware trust region anchor selection considers the potential capability of the trust region for better local optimization. Experimental results demonstrate the effectiveness of Inv BO on nine real-world benchmarks, such as molecule design and arithmetic expression fitting tasks. Code is available at https://github.com/mlvlab/Inv BO. 1 Introduction Bayesian optimization (BO) has been used in a wide range of applications such as material science [1], chemical design [2, 3], ,and hyperparameter optimization [4, 5]. The main idea of BO is probabilistically estimating the expensive black-box objective function using a surrogate model to find the optimal solution with minimum objective function evaluation. While BO has shown its success on continuous domains, applying BO over discrete input space is challenging [6, 7]. To address it, Latent Bayesian Optimization (LBO) has been proposed [8 14]. LBO performs BO over a latent space by mapping the discrete input space into the continual latent space with generative models such as Variational Auto Encoders (VAE) [15], consisting of an encoder qϕ and a decoder pθ. Unlike the standard BO, the surrogate model in LBO associates a latent vector z with an objective function value by emulating the composition of the objective function and the decoder of VAE. In LBO, however, the reconstruction error of the VAE often leads to one latent vector z being associated with two different objective function values as explained in Figure 1. We observe that the discrepancy between y and y (or x and x ) hinders learning an accurate surrogate model g and generating high-quality solutions. We name this the misalignment problem . Most prior works [9, 11, 12] use the surrogate model genc, which is trained with the encoder triplet (x, z, y), and equal contributions Corresponding author 38th Conference on Neural Information Processing Systems (Neur IPS 2024). Figure 1: Misalignment problem. In LBO, a latent vector z can be associated with two function values y and y due to the reconstruction error of the VAE, i.e., x = x . (a) In Encoder triplet (x, z, y), latent vector z is associated with f(x), where x is the original input to the encoder, i.e., z = qϕ(x). (b) In Decoder triplet (x , z, y ), z is associated with y = f(x ), which is the objective function value of reconstructed input value x using the decoder, i.e., x = pθ(z). The discrepancy between y and y hinders learning the accurate surrogate model g. We name this the misalignment problem . generate solutions x via decoder pθ. Since genc fails to estimate the composite function of pθ and f, this approach often results in suboptimal outcomes. Some works [13, 14] handle the misalignment problem by employing the surrogate model gdec trained with the decoder triplet (x , z, y ). However, they request a huge amount of additional oracle calls to obtain the decoder triplet, which leads to inefficient optimization. In addition, several existing LBO methods [13, 14, 16] adopt the trust region method to restrict the search space and have shown performance gain. Most prior works select the anchor, the center of the trust region, as the current optimal point. This objective function value-based anchor selection overlooks the potential to benefit the optimization performance of the latent vectors within the trust region. In this work, we propose an Inversion-based Latent Bayesian Optimization (Inv BO), a plug-and-play module for VAE-based LBO methods. Inv BO consists of two components: the inversion method and a potential-aware trust region anchor selection. The inversion method addresses the misalignment problem by inverting decoder pθ to find the latent code that yields x without any additional oracle call. We theoretically analyze that our inversion method decreases the upper bound of the error between the surrogate model and the objective function within the trust region. The potential-aware trust region anchor selection method selects the anchor considering not only the observed objective function value but also the potential to enhance the optimization process of the latent vectors that the trust region contains. We provide the experimental evaluation on nine different tasks, Guacamol, DRD3, and arithmetic expression fitting task to show the general effectiveness of Inv BO. Specifically, plug-and-play results of Inv BO over diverse prior LBO works show a large performance gain and achieved state-of-the-art performance. The contributions of our paper are as follows: We propose the inversion method to address the misalignment problem in LBO by generating the decoder triplet without using any additional oracle calls. We propose the potential-aware trust region anchor selection, aiming to select the centers of trust regions considering the latent vectors expected to benefit the optimization process within the trust regions. By combining the inversion method and potential-aware trust region anchor selection, we propose Inversion-based Latent Bayesian Optimization (Inv BO), a novel plug-and-play module for LBO, and achieve state-of-the-art performance on the nine different tasks. 2 Related Works 2.1 Latent Bayesian Optimization The goal of Latent Bayesian Optimization (LBO) [8, 9, 11 14, 16 19] is to learn a latent space to enable optimization over a continuous space from discrete or structured input (e.g., graph or image). LBO consists of a Variational Auto Encoder (VAE) to generate data from the latent representation and a surrogate model (e.g., Gaussian process) to map the latent representation into the objective score. Some works on the LBO have designed new decoder architectures [8, 20 23] to perform the reconstruction better, while other works have proposed learning mechanisms [9 14, 17] to alleviate the discrepancy between the latent space and input space. LOL-BO [13] adapts the concept of trust-region to the latent space and jointly learns a VAE and a surrogate model to search the data point in the local region. Co BO [14] designs new loss to encourage the correlation between the distances in the latent space and objective function. 2.2 Inversion in Generative Models Inversion has widely been applied to a variety of generative models such as Generative Adversarial Networks (GANs) [24, 25] and Diffusion models [26 29]. Inversion is the process of finding the latent code zinv of a given image to manipulate images with generative models. Formally, given an image x and the well-trained generator G, the inversion can be written as: zinv = arg min z Z d X (G(z), x), (1) where d X ( , ) denotes the distance metric in the image space X, and Z is the latent space. To solve Eq. (1), most inversion-based works can be generally classified as two approaches: optimizationbased and learning-based methods. The optimization-based inversion [30 33] iteratively finds a latent vector to reconstruct the target image x through the fixed generator. The learning-based inversion [25, 34, 35] trains the encoder for mapping the image x to the latent code z while fixing the decoder. In this work, we introduce the concept of inversion to find the latent vector that can generate a desired sample for constructing an aligned triplet and we use the optimization-based inversion. 3 Preliminaries Bayesian optimization. Bayesian optimization (BO) is a powerful and sample-efficient optimization algorithm that aims at searching the input x with a maximum objective value f(x), which is formulated as: x = arg max x X f(x), (2) where the black-box objective function f : X 7 Y is assumed expensive to evaluate, and X is a feasible set. Since the objective function f is unknown or cost-expensive, BO methods probabilistically emulate the objective function by a surrogate model g with observed dataset D = {(xi, yi)|yi = f(xi)}n i=1. With the surrogate model g, the acquisition function α selects the most promising point xn+1 as the next evaluation point while balancing exploration and exploitation. BO repeats this process until the oracle budget is exhausted. Trust region-based local Bayesian optimization. Classical Bayesian optimization methods often suffer from the difficulty of the optimization in a high dimensional space [36]. To address this problem, Tu RBO [36] adopts trust regions to limit the search space to small regions. The anchor (center) of trust region T is selected as a current optimal point, and the size of the trust region is scheduled during the optimization process. At the beginning of the optimization, the side length of all trust regions is set to Linit. When the trust region T updates the best score τsucc times in a row, the side length becomes twice until it reaches Lmax. Similarly, when it fails to update the best score τfail times in a row, the side length becomes half. When L falls below a Lmin, the side length of the trust region is set to Linit and restart the scheduling. Recently, LOL-BO [13] adapted trust region-based local optimization to LBO, and has shown performance gain. In this section, we present an Inversion-based Latent Bayesian Optimization (Inv BO) consisting of an inversion and a novel trust region anchor selection method for effective and efficient optimization. We first describe latent Bayesian optimization and the misalignment problem of it (Section 4.1). Then, we introduce the inversion method to address the misalignment problem without using any additional oracle budgets (Section 4.2). Lastly, we present a potential-aware trust region anchor selection for better local search space (Section 4.3). Figure 2: Comparison of solutions to the misalignment problem. (a) Some works [13, 14] solve the misalignment problem by the recentering technique that generates the aligned triplet (x , z, y ). However, it requests additional oracle calls as y = f(x ) is unevaluated, and does not fully use the evaluated function value y = f(x). (b) The inversion method (ours) aims to find zinv that generates the evaluated data x to get the aligned triplet (x, zinv, y) without any additional oracle calls. 4.1 Misalignment in Latent Bayesian Optimization BO has proven its effectiveness in various areas where input space X is continuous, however, BO over the discrete domain, such as chemical design, is a challenging problem. To handle this problem, VAE-based latent Bayesian optimization (LBO) has been proposed [8, 11, 13, 14] that leverages BO over a continuous space by mapping the discrete input space X to a continuous latent space Z. Variational autoencoder (VAE) is composed of encoder qϕ : X 7 Z to compute the latent representation z of the input data x and decoder pθ : Z 7 X to generate the data x from the latent z. Given the objective function f, latent Bayesian optimization can be formulated as: z = arg max z Z f(pθ(z)), (3) where pθ(z) is a generated data with the decoder pθ and Z is a latent space. Unlike the standard BO, the surrogate model g aims to emulate the function f pθ : Z 7 Y. To the end, the surrogate model is trained with aligned dataset D = {(xi, zi, yi)}n i=1, where xi = pθ(zi) is generated by the decoder pθ : Z 7 X and yi = f(xi) is the objective value of xi evaluated via the black box objective function f : X 7 Y. In the rest of our paper, we define that the dataset is aligned when all triplets satisfy the above conditions (i.e., all triplets are the decoder triplets explained in Figure 1), and the dataset is misaligned otherwise. We define the misalignment problem as the misaligned dataset hinders the accurate learning of the surrogate model g. Figure 3: (Left) The number of oracle calls to evaluate the queries selected by the acquisition function (blue) and during the recentering (Red). (Right) The number of objective function evaluation that updates the best score. Most existing LBO works [11 14] overlook the misalignment problem 1, which originates from two processes: (i) construction of initial dataset D0 and (ii) update of VAE. Construction of initial dataset D0. Since initial dataset D0 is composed of pairs of input data and its corresponding objective value {(xi, yi)|yi = f(xi)}n i=1, LBO requires latent vectors {zi}n i=1 to train the surrogate model. Most works compute a latent vector zi as zi = qϕ(xi) under the assumption that VAE completely reconstructs every data points (i.e., pθ(qϕ(xi)) = xi), which is difficult to be satisfied in every case. This results in the data misalignment (xi = pθ(zi)) during the construction of initial dataset D0. Update of VAE. In LBO works, updating VAE during the optimization plays a crucial role in adapting well to newly generated samples. However, due to the update of VAE, the previously generated triplet (x, z, y) cannot ensure the alignment since z was computed by the VAE before the update. Previous LBO works [13, 14] solve the misalignment problem originating from the VAE update with a recentering technique that requests additional oracle calls to generate the aligned dataset D = {(x i, qϕ(xi), f(x i))}n i=1 where x i = pθ(qϕ(xi)) as shown in Figure 2 (Left). But, it has a limitation to consuming a huge amount of additional oracle calls while they do not update the best score (Figure 3). Note that prior works do not explicitly mention the additional oracle calls during the recentering technique, but it can be verified by the official Git Hub code. Further details about oracle consumption of the recentering are provided in the supplement Section H. 4.2 Inversion-based Latent Bayesian Optimization Our primary goal is training the surrogate model g to correctly emulate the composite function f pθ : Z 7 Y via constructing an aligned dataset without consuming additional oracle calls. To the end, we propose an inversion method that inverts the target discrete data x into the latent vector z that satisfies x = pθ(z) for dataset alignment as shown in Figure 2 (Right). With a pre-trained frozen decoder pθ, the latent vector z can be optimized by: zinv = arg min z Z d X (x, pθ(z)), (4) where x is a target data and d X is a distance function in the input space X. We use the normalized Levenshtein distance [37] as our distance function, d X , which can be applied to any string-form data. Our inversion method, however, is flexible and can utilize any task-specific distance functions, such as Tanimoto similarity [38] for molecule design tasks. To solve the Eq. (4), we iteratively update a latent vector z to find zinv that reconstructs the target data x. We provide the overall pseudocode of the inversion method in Algorithm 1. Algorithm 1 Inversion Input: Encoder qϕ, decoder pθ, target data x, max iteration T, distance function d X , learning rate η, reconstruction loss L 1: Initialize z(0) qϕ(x) 2: for t = 0, 1, ..., T 1 do 3: z(t+1) z(t) η z(t)L pθ(z(t)), x 4: if d X (x, pθ(z(t+1))) < ϵ then Eq. (4) 5: return z(t+1) 6: end if 7: end for 8: return z(T ) The initialization strategy of latent vector z plays a key role in the optimization-based inversion process. We set the initialization point of latent vector z as an output of a pre-trained encoder qϕ(x) given target discrete data x as in line 1. We iteratively update the latent vector z with the cross-entropy loss used in VAE training in line 3 until it reaches the maximum number of iterations T. Before the iteration budget is exhausted, we finish the inversion process when the distance between the generated data pθ(z) and target data x is less than ϵ as our goal is finding the latent vector zinv that satisfies Eq. (4), which is denoted in line 4. The inversion method generates the aligned dataset during the construction of the initial dataset D0 and the update of VAE to handle the misalignment problem. We theoretically show that optimizing the latent vector z to satisfy d X (x, pθ(z)) 0 with inversion plays a crucial role in minimizing the upper bound of the error between the posterior mean of the surrogate model and the objective function value within the trust region centered at z. Proposition 1. Let f be a black-box objective function and m be a posterior mean of Gaussian process, pθ be a decoder of the variational autoencoder, c be an arbitrarily small constant, d X and d Z be the distance function on input X and latent Z spaces, respectively. The distance function d X is bounded between 0 and 1, inclusive. We assume that f, m and the composite function of f and pθ are L1, L2, and L3-Lipschitz continuous functions, respectively. Suppose the following assumptions are satisfied: |f(x) m(z)| c, d X (x, pθ(z)) γ. (5) Then the difference between the posterior mean of the arbitrary point z in the trust region centered at z with trust radius δ and the black box objective value is upper bounded as: |f(pθ(z )) m(z )| c + γ L1 + δ (L2 + L3), (6) where d Z(z, z ) δ. The proof is available in Section A. We assume that the black box function f, the posterior mean of Gaussian process m, and the objective function f pθ are Lipschitz continuous functions, which is a common assumption in Bayesian optimization [39, 40] or global optimization [41]. Proposition 1 implies that the upper bound of the Gaussian process prediction error within the trust region can be minimized by reducing the distance between x and pθ(z), denoted as γ. Our inversion method reduces γ by generating an aligned dataset without additional oracle calls. In Eq. (6), the constant c represents the accuracy of the surrogate model, which can be improved during training. Some LBO works [12, 13] introduce regularization terms to learn a smooth latent space, implicitly reducing L3, the Lipschitz constant of the composite function f pθ. Co BO [14] further proposes regularization losses that explicitly reduce L3 with theoretical grounding. Since the surrogate model m emulates the composite function f pθ, its Lipschitz constant L2 can also be reduced along with L3. 4.3 Potential-aware Trust Region Anchor Selection Here, we propose a potential-aware trust region anchor selection to consider both the objective function value y and the potential ability of the trust region to benefit the optimization. Previous trust region-based BO works [13, 14, 36, 42] select the anchor based on the corresponding objective function value only. However, this objective score-based anchor selection does not consider that the trust region contains latent points expected to enhance the optimization performance. We design a potential score to measure the potential ability to enhance the optimization process of the trust region. To compute it, we use the acquisition function value, which is generally employed to measure the potential ability to improve the optimization process of a given data point. Specifically, we employ Thompson Sampling, a well-established acquisition function used in previous trust regionbased methods [13, 14, 36]. Formally, the potential score of each trust region Ti is computed as follows: αi pot = max z Zi cand ˆf(z) where ˆf GP (µ(z), k(z, z )) , (7) where i is an index of the candidate anchor point, Zi cand is the candidate set sampled from the trust region Ti and ˆf is a sampled function from the surrogate model (e.g., GP) posterior. As the scale of objective function values Y = {y1, y2, ..., yn} and that of the potential ability of each trust region A = α1 pot, α2 pot, ..., αn pot is changed dynamically during the optimization process, we calculate a scaled potential score αi pot by adjusting the scale of αi max according to the Y : αi scaled = αi pot Amin Amax Amin (Ymax Ymin), (8) where Amax = maxi αi pot and Amin = mini αi pot denote the maximum and minimum value of A, respectively. Ymax = maxi yi and Ymin = mini yi indicate the maximum and minimum value of Y , respectively. Based on the scaled potential score αi scaled, the final score si of each anchor is calculated as: si = yi + αi scaled. (9) Our final score takes into account the objective function value of the anchor (observed value) and the potential score (model s prediction) for the better local search space. Finally, we select the anchors with the highest final score si. We summarize our potential-aware trust region anchor selection schema in Algorithm 2 of Section N. 5 Experiments We measure the performance of the proposed method named Inv BO on nine different tasks with three Bayesian optimization benchmarks: Guacamol [43], DRD3, and arithmetic expression fitting tasks [6, 8, 11 13, 44]. Guacamol and the DRD3 benchmark tasks aim to find molecules with the most necessary properties. For Guacamol benchmarks, we use seven challenging tasks, Median molecules 2 (med2), Zaleplon MPO (zale), Perindopril MPO (pdop), Amlodipine MPO (adip), Osimertinib MPO (osmb), Ranolazine MPO (rano), and Valsartan SMARTS (valt). To show the effectiveness of our Inv BO in various settings, we also conducted experiments in a large budget setting, which is used in previous works [13, 14]. The goal of the arithmetic expression fitting task is to generate single-variable expressions that minimize the distance from a target expression (e.g., 1/3 + x + sin (x x)). More details of each benchmark are provided in Section K. Table 1: Optimization results of applying Inv BO to several trust region-based LBOs on Guacamol benchmark tasks. A higher score is a better one. Task Median molecules 2 (med2) Valsartan SMARTS (valt) Num Oracle 100 300 500 100 300 500 Tu RBO-L 0.186 0.000 0.186 0.000 0.186 0.000 0.000 0.000 0.000 0.000 0.000 0.000 Tu RBO-L + Inv BO 0.186 0.000 0.194 0.002 0.202 0.001 0.000 0.000 0.024 0.017 0.212 0.092 LOL-BO 0.186 0.000 0.186 0.000 0.190 0.001 0.000 0.000 0.000 0.000 0.000 0.000 LOL-BO + Inv BO 0.189 0.002 0.204 0.005 0.227 0.010 0.000 0.000 0.007 0.005 0.171 0.039 Co BO 0.186 0.000 0.188 0.002 0.191 0.003 0.000 0.000 0.000 0.000 0.000 0.000 Co BO + Inv BO 0.187 0.001 0.203 0.004 0.214 0.006 0.000 0.000 0.042 0.013 0.348 0.107 Figure 4: Optimization results on Guacamol benchmark tasks. The lines and ranges indicate the average and standard error of ten runs under the same settings. A higher score is a better score. 5.2 Baselines We compare to six latent Bayesian Optimization methods: LS-BO, Tu RBO [36], W-LBO [11], LOLBO [13], Co BO [14], and PG-LBO [17]. We also compare with GB-GA [45], a widely used genetic algorithm for graph structure data. LOL-BO and Co BO employ the decoder triplet generated by the recentering technique, and the other baselines employ the encoder triplet during the optimization process. In the case of LOL-BO and Co BO, we substitute the recentering technique with the inversion method. We provide the details of each baseline in Section L. 5.3 Implementation Details For the arithmetic expression fitting task, we follow other works [8, 11, 13, 14] to employ Grammar VAE model [8]. For the de novo molecule design tasks such as Guacamol benchmark and DRD3 tasks, we use SELFIES VAE [13] following recent works [13, 14]. In all tasks, we adopt sparse variational GP [46] with deep kernel [47] as our surrogate model. In the inversion method, the learning rate is set to 0.1 in all experiments. The further implementation details are provided in Section M. 5.4 Experimental Results We apply our Inv BO to several trust region-based LBOs, Tu RBO-L, LOL-BO [13] and Co BO [14], and provide optimization results on two Guacamol benchmark tasks, med2 and valt. All results are average scores of ten runs under the identical settings. Table 1 demonstrates that our Inv BO consistently improves all LBO models in two tasks by a large margin. In particular, in the valt task, all baseline models with Inv BO demonstrate significant performance improvements and Co BO with Inv BO achieves a 0.348 score gain while the baseline models without Inv BO fail in optimization. More results of other baselines with Inv BO on other tasks are provided in Section E. Figure 4 provides the optimization results on four Guacamol benchmark tasks including med2, zale, osmb, and valt. Each subfigure shows the number of evaluations of the black box objective function (Number of Oracle) and the corresponding average and standard error of the objective score (Best Score). Our Inv BO built on Co BO achieves the best performance in all four tasks. Further results of other tasks are provided in Section F. We also conduct experiments to demonstrate the effectiveness of our Inv BO on DRD3 and arithmetic expression fitting tasks and large-budget settings. The experimental results are illustrated in Fig- (a) DRD3 and Arithmetic expression task (b) Guacamol task under the large budget setting Figure 5: Optimization results on various tasks and settings. Note that: (a) A lower score is a better score. (b) A higher score is a better score. Figure 6: Optimization results of component ablation on zale, med2, osmb, and valt. ure 5(a) and Figure 5(b), respectively. Please note that both DRD3 and arithmetic expression tasks aim to minimize the score while the goal of Guacamol tasks is increasing the score. Figure 5 shows that our Inv BO applied to Co BO achieves the best performance. We provide further optimization results on other tasks under the large budget settings in Section G. These results demonstrate that Inv BO is effective in diverse tasks and settings. 6.1 Ablation Study We conduct additional experiments to verify the contribution of each component in our Inv BO: the inversion method (INV), and the potential-aware trust region anchor selection method (PAS). Figure 6 shows the optimization results of the ablation study on med2, zale, osmb, and valt tasks. From the figure, models with the inversion method (i.e., Co BO with INV and Inv BO) replace the recentering technique as the inversion method, while models without the inversion method (i.e., vanilla Co BO and Co BO with PAS) employ the recentering technique. Notably, both components of our method contribute to the optimization process, and the combination work, Inv BO, consistently obtains better objective scores compared to other models. Specifically, in osmb task, the average best score achieved by the methods with the PAS, INV and both (i.e., Inv BO) shows 0.784, 0.792, and 0.804 score gains compared to vanilla Co BO, respectively. 6.2 Analysis on Misalignment Problem and Inversion To further prove that the misaligned dataset hinders the accurate learning of the surrogate model (i.e., misalignment problem), we compare the performance of the surrogate model trained with aligned and misaligned datasets on the med2 task. Figure 7 shows the fitting results of the surrogate model trained with encoder triplets (genc, left) and decoder triplets (gdec, right), respectively. Figure 7 demonstrates that the surrogate model gdec approximates the objective function accurately, while genc fails to emulate the objective function. Further details of an experiment are provided in Section I. In Figure 8, we compare the optimization results of Co BO using decoder triplets and encoder triplets on the valt and med2 tasks. Both models use the potential-aware anchor selection, and the decoder triplets are made by our inversion method. Co BO using the decoder triplets shows superior optimization performance over the Co BO using the encoder triplets on both tasks. This implies that Figure 7: Gaussian process fitting results trained with encoder triplets and decoder triplets. Figure 8: Optimization results using encoder triplet and decoder triplet on valt and med2 tasks. Figure 9: (Left) Gaussian process prediction error within the trust region. (Right) Optimization results on med2 tasks. Figure 10: Optimization results with diverse trust region anchor selections on valt and med2 tasks. the misaligned dataset hinders the accurate learning of the surrogate model, which leads to suboptimal optimization results, and the inversion method handles the problem by generating the decoder triplet. 6.3 Effects of Inversion on Proposition 1 In Section 4.2, we provide the upper bound of the error between the predicted and ground-truth objective value. In Eq. (6), the Lipschitz constant of the objective function L1 and the trust region radius δ is fixed or the hyper-parameter and the constant c can be improved by learning the surrogate model. In the end, we can reduce the upper bound of the objective value prediction error by reducing the three components: γ, L2, and L3, which implies the distance between x and pθ(z), the Lipschitz constant of function m and f pθ, respectively. Our inversion method reduce γ by searching the latent vector zinv that satisfies x = pθ(zinv). Previously, Co BO [14] proposed regularization losses that reduce L3. Since the surrogate model emulates the composite function f pθ, these regularization losses can reduce L2 along with L3. Figure 9 (left) shows the regularization losses of Co BO and our inversion method reduces the objective value prediction error. Co BO-based models (i.e., Co BO+PAS and Co BO+Inv BO) employ the Co BO regularization losses, and models with Inv BO (i.e., Tu RBO-L+Inv BO and Co BO+Inv BO) employ our inversion method. Tu RBO-L does not use the Co BO regularization losses nor our inversion method, and models with PAS (i.e., Tu RBO-L+PAS and Co BO+PAS) employ encoder triplets. Applying the regularization losses and our inversion method reduces the objective value prediction error, respectively, but our inversion method shows a larger error reduction. The combination of regularization losses and our inversion shows the smallest prediction error, which implies our inversion method robustly complements existing methods. We provide the optimization results of each model on the med2 task in Figure 9 (right). These results demonstrate that reducing the objective function prediction error plays a crucial role in optimization performance. 6.4 Comparing Diverse Anchor Selection Methods To further prove the importance of our potential-aware anchor selection method, we perform BO with the diverse anchor selection methods: random, acquisition α, and objective score y. Random indicates randomly selected anchors, and acquisition and objective indicate anchors are selected based on the max acquisition function value and objective score, respectively. All models use the inversion method, and the optimization results on valt and med2 tasks are in Figure 10. Ours and objective score-based anchor selection rapidly find high-quality data compared to the random and acquisitionbased selections. However, objective score-based anchor selection shows inferior performance than our potential-aware anchor selection. This indicates that both the uncertainty of the surrogate model and objective function value need to be considered for exploration and exploitation. 7 Conclusion We propose Inversion-based Latent Bayesian Optimization (Inv BO), a plug-and-play module for LBO. We introduce the inversion method that inverts the decoder to find the latent vector for generating the aligned dataset. Additionally, we present the potential-aware trust region anchor selection that considers not only the corresponding objective function value of the anchor but also the potential ability of the trust region. From our experimental results, Inv BO achieves state-of-the-art performance on nine LBO benchmark tasks. We also theoretically demonstrate the effectiveness of our inversion method and provide a comprehensive analysis to show the effectiveness of our Inv BO. Broader Impacts One of the contributions of this paper is molecular design optimization, which requires careful consideration due to its unintentional applications such as the generation of toxic. We believe that our work has a lot of positive aspects to accelerate the development of chemical and drug discovery with an inversion-based latent Bayesian optimization method. Limitations The performance of the methods proposed in the paper depends on the quality of the generative model. For example, if the objective function is related to the molecule property, the generated model such as the VAE should have the ability to generate the proper molecule to have good performance of optimization. Acknowledgement This work was partly supported by ICT Creative Consilience Program through the Institute of Information & Communications Technology Planning & Evaluation (IITP) (IITP-2024-RS-2020II201819, 10%) and the National Research Foundation of Korea (NRF) (NRF-2023R1A2C2005373, 45%) grant funded by the Korea government (MSIT), and the Virtual Engineering Platform Project (Grant No. P0022336, 45%), funded by the Ministry of Trade, Industry & Energy (Mo TIE, South Korea). [1] Sebastian Ament, Maximilian Amsler, Duncan R Sutherland, Ming-Chiang Chang, Dan Guevarra, Aine B Connolly, John M Gregoire, Michael O Thompson, Carla P Gomes, and R Bruce van Dover. Autonomous materials synthesis via hierarchical active learning of nonequilibrium phase diagrams. Sci. Adv., 7(51):eabg4930, 2021. [2] Ke Wang and Alexander W Dowling. Bayesian optimization for chemical products and functional materials. Curr. Opin. Chem. Eng., 36:100728, 2022. [3] Ryan-Rhys Griffiths and José Miguel Hernández-Lobato. Constrained bayesian optimization for automatic chemical design using variational autoencoders. Chem. Sci., 11(2):577 586, 2020. [4] Jiazhuo Wang, Jason Xu, and Xuejun Wang. Combination of hyperband and bayesian optimization for hyperparameter optimization in deep learning, 2018. [5] Jia Wu, Xiu-Yun Chen, Hao Zhang, Li-Dong Xiong, Hang Lei, and Si-Hao Deng. Hyperparameter optimization for machine learning models based on bayesian optimization. J. Electron. Sci. Technol., 17(1):26 40, 2019. [6] Aryan Deshwal and Jana Doppa. Combining latent space and structured kernels for bayesian optimization over combinatorial spaces. In Neur IPS, 2021. [7] Changyong Oh, Jakub Tomczak, Efstratios Gavves, and Max Welling. Combinatorial bayesian optimization using the graph cartesian product. Neur IPS, 2019. [8] Matt J Kusner, Brooks Paige, and José Miguel Hernández-Lobato. Grammar variational autoencoder. In ICML, 2017. [9] Rafael Gómez-Bombarelli, Jennifer N Wei, David Duvenaud, José Miguel Hernández-Lobato, Benjamín Sánchez-Lengeling, Dennis Sheberla, Jorge Aguilera-Iparraguirre, Timothy D Hirzel, Ryan P Adams, and Alán Aspuru-Guzik. Automatic chemical design using a data-driven continuous representation of molecules. ACS Cent. Sci., 4(2):268 276, 2018. [10] Stephan Eissman, Daniel Levy, Rui Shu, Stefan Bartzsch, and Stefano Ermon. Bayesian optimization and attribute adjustment. In UAI, 2018. [11] Austin Tripp, Erik Daxberger, and José Miguel Hernández-Lobato. Sample-efficient optimization in the latent space of deep generative models via weighted retraining. In Neur IPS, 2020. [12] Antoine Grosnit, Rasul Tutunov, Alexandre Max Maraval, Ryan-Rhys Griffiths, Alexander I Cowen-Rivers, Lin Yang, Lin Zhu, Wenlong Lyu, Zhitang Chen, Jun Wang, et al. Highdimensional bayesian optimisation with variational autoencoders and deep metric learning, 2021. [13] Natalie Maus, Haydn Jones, Juston Moore, Matt J Kusner, John Bradshaw, and Jacob Gardner. Local latent space bayesian optimization over structured inputs. In Neur IPS, 2022. [14] Seunghun Lee, Jaewon Chu, Sihyeon Kim, Juyeon Ko, and Hyunwoo J Kim. Advancing bayesian optimization via learning correlated latent space. In Neur IPS, 2023. [15] Diederik P Kingma. Auto-encoding variational bayes. ar Xiv preprint ar Xiv:1312.6114, 2013. [16] Natalie Maus, Kaiwen Wu, David Eriksson, and Jacob Gardner. Discovering many diverse solutions with bayesian optimization. In AISTATS, 2023. [17] Taicai Chen, Yue Duan, Dong Li, Lei Qi, Yinghuan Shi, and Yang Gao. Pg-lbo: Enhancing high-dimensional bayesian optimization with pseudo-label and gaussian process guidance. In AAAI, 2024. [18] Pascal Notin, José Miguel Hernández-Lobato, and Yarin Gal. Improving black-box optimization in vae latent space using decoder uncertainty. Neur IPS, 2021. [19] Xiaoyu Lu, Javier Gonzalez, Zhenwen Dai, and Neil D Lawrence. Structured variationally auto-encoded optimization. ICML, 2018. [20] Wengong Jin, Regina Barzilay, and Tommi Jaakkola. Junction tree variational autoencoder for molecular graph generation. In ICML, 2018. [21] Hanjun Dai, Yingtao Tian, Bo Dai, Steven Skiena, and Le Song. Syntax-directed variational autoencoder for structured data. In ICLR, 2018. [22] Hiroshi Kajino. Molecular hypergraph grammar with its application to molecular optimization. In ICML, 2019. [23] Bidisha Samanta, Abir De, Gourhari Jana, Vicenç Gómez, Pratim Chattaraj, Niloy Ganguly, and Manuel Gomez-Rodriguez. Nevae: A deep generative model for molecular graphs. JMLR, 21(114):1 33, 2020. [24] Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial nets. In Neur IPS, 2014. [25] Jun-Yan Zhu, Philipp Krähenbühl, Eli Shechtman, and Alexei A Efros. Generative visual manipulation on the natural image manifold. In ECCV, 2016. [26] Jascha Sohl-Dickstein, Eric Weiss, Niru Maheswaranathan, and Surya Ganguli. Deep unsupervised learning using nonequilibrium thermodynamics. In ICML, 2015. [27] Jonathan Ho, Ajay Jain, and Pieter Abbeel. Denoising diffusion probabilistic models. In Neur IPS, 2020. [28] Jooyoung Choi, Sungwon Kim, Yonghyun Jeong, Youngjune Gwon, and Sungroh Yoon. Ilvr: Conditioning method for denoising diffusion probabilistic models. In ICCV, 2021. [29] Rinon Gal, Yuval Alaluf, Yuval Atzmon, Or Patashnik, Amit H. Bermano, Gal Chechik, and Daniel Cohen-Or. An image is worth one word: Personalizing text-to-image generation using textual inversion. In ICLR, 2023. [30] Antonia Creswell and Anil Anthony Bharath. Inverting the generator of a generative adversarial network. TNNLS, 30(7):1967 1974, 2018. [31] Rameen Abdal, Yipeng Qin, and Peter Wonka. Image2stylegan: How to embed images into the stylegan latent space? In ICCV, 2019. [32] Rameen Abdal, Yipeng Qin, and Peter Wonka. Image2stylegan++: How to edit the embedded images? In CVPR, 2020. [33] Minyoung Huh, Richard Zhang, Jun-Yan Zhu, Sylvain Paris, and Aaron Hertzmann. Transforming and projecting images into class-conditional generative networks. In ECCV, 2020. [34] Guim Perarnau, Joost Van De Weijer, Bogdan Raducanu, and Jose M Álvarez. Invertible conditional gans for image editing. In Neur IPSW, 2016. [35] David Bau, Jun-Yan Zhu, Jonas Wulff, William Peebles, Hendrik Strobelt, Bolei Zhou, and Antonio Torralba. Inverting layers of a large generator. In ICLRW, 2019. [36] David Eriksson, Michael Pearce, Jacob Gardner, Ryan D Turner, and Matthias Poloczek. Scalable global optimization via local bayesian optimization. In Neur IPS, 2019. [37] Vladimir I Levenshtein et al. Binary codes capable of correcting deletions, insertions, and reversals. Dokl. Phys., 10(8):707 710, 1966. [38] Dávid Bajusz, Anita Rácz, and Károly Héberger. Why is tanimoto index an appropriate choice for fingerprint-based similarity calculations? Journal of cheminformatics, 2015. [39] Javier González, Zhenwen Dai, Philipp Hennig, and Neil Lawrence. Batch bayesian optimization via local penalization. In Artificial intelligence and statistics, 2016. [40] Trong Nghia Hoang, Quang Minh Hoang, Ruofei Ouyang, and Kian Hsiang Low. Decentralized high-dimensional bayesian optimization with factor graphs. In Proceedings of the AAAI Conference on Artificial Intelligence, 2018. [41] Christodoulos A Floudas and Panos M Pardalos. Encyclopedia of optimization. Springer Science & Business Media, 2008. [42] Samuel Daulton, David Eriksson, Maximilian Balandat, and Eytan Bakshy. Multi-objective bayesian optimization over high-dimensional search spaces. In UAI, 2022. [43] Nathan Brown, Marco Fiscato, Marwin HS Segler, and Alain C Vaucher. Guacamol: benchmarking models for de novo molecular design. J. Chem. Inf. Model., 59(3):1096 1108, 2019. [44] Sungsoo Ahn, Junsu Kim, Hankook Lee, and Jinwoo Shin. Guiding deep molecular optimization with genetic exploration. In Neur IPS, 2020. [45] Jan H Jensen. A graph-based genetic algorithm and generative model/monte carlo tree search for the exploration of chemical space. Chemical science, 2019. [46] Edward Snelson and Zoubin Ghahramani. Sparse gaussian processes using pseudo-inputs. Neur IPS, 2005. [47] Andrew Gordon Wilson, Zhiting Hu, Ruslan Salakhutdinov, and Eric P Xing. Deep kernel learning. In AISTATS, 2016. [48] Maximilian Balandat, Brian Karrer, Daniel Jiang, Samuel Daulton, Ben Letham, Andrew G Wilson, and Eytan Bakshy. Botorch: A framework for efficient monte-carlo bayesian optimization. Neur IPS, 2020. [49] Christopher Williams and Carl Rasmussen. Gaussian processes for regression. NIPS, 1995. [50] Miguel García-Ortegón, Gregor NC Simm, Austin J Tripp, José Miguel Hernández-Lobato, Andreas Bender, and Sergio Bacallado. Dockstring: easy molecular docking yields better benchmarks for ligand design. J. Chem. Inf. Model., 62(15):3486 3502, 2022. [51] Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, et al. Pytorch: An imperative style, high-performance deep learning library. Neur IPS, 2019. [52] Jacob Gardner, Geoff Pleiss, Kilian Q Weinberger, David Bindel, and Andrew G Wilson. Gpytorch: Blackbox matrix-matrix gaussian process inference with gpu acceleration. Neur IPS, 2018. A A Proof of Proposition 1 In this section, we provide the proof of Proposition 1. Proposition 1. Let f be a black-box objective function and m be a posterior mean of Gaussian process, pθ be a decoder of the variational autoencoder, c be an arbitrarily small constant, d X and d Z be the distance function on input X and latent Z spaces, respectively. The distance function d X is bounded between 0 and 1, inclusive. We assume that f, m and the composite function of f and pθ are L1, L2, and L3-Lipschitz continuous functions, respectively. Suppose the following assumptions are satisfied: |f(x) m(z)| c, d X (x, pθ(z)) γ. (10) Then the difference between the posterior mean of the arbitrary point z in the trust region centered at z with trust radius δ and the black box objective value is upper bounded as: |f(pθ(z )) m(z )| c + γ L1 + δ (L2 + L3), (11) where d Z(z, z ) δ. Proof. With L-Lipschitz continuity, we have: |f(x) f(pθ(z))| L1 d X (x, pθ(z)) γ L1, |m(z) m(z )| L2 d Z(z, z ) δ L2, |f(pθ(z)) f(pθ(z ))| L3 d Z(z, z ) δ L3. (12) Thus, the difference between the posterior mean of point within the trust region of length δ centered at z and the black-box objective value is upper bounded as follows: |f(pθ(z )) m(z )| = |(f(x) m(z)) + (f(pθ(z)) f(x)) + (f(pθ(z )) f(pθ(z))) + (m(z) m(z ))| |f(x) m(z)| + |f(pθ(z)) f(x)| + |f(pθ(z )) f(pθ(z))| + |m(z) m(z )| c + γ L1 + δ (L2 + L3) B Exploration of Potential-aware Trust Region Anchor Selection Figure 11: Exploration effectiveness ablation study of the potential aware anchor selection (PAS) on Co BO+Inv BO (INV, PAS). We measure the number of searched unique data samples for each iteration to validate the exploration ability of our potential-aware trust region anchor selection in 4.3. We compare it with the objective score-based anchor selection, which selects the best anchors based on their objective score, used in most prior works [13, 14, 36]. From the figure, our potential-aware anchor selection searches more diverse data compared to the objective score-based anchor selection on both tasks. These results demonstrate that our anchor selection method improves the exploration ability without the loss of the exploitation ability. C Dissimilarity in Latent Bayesian Optimization Figure 12: Dissimilarity between xi and pθ(zi) with and without inversion on med2 and valt tasks. The measurement of dissimilarity is the normalized Levenshtein distance between the SELFIES token. (x-axis: number of iterations, y-axis: normalized Levenshtein distance.) We here provide the comparison of with and without inversion to empirically prove that the inversion generates aligned data. Given a pair of data sample xi and its corresponding latent vector zi, we measure the dissimilarity between the input xi and the data pθ(zi) generated from the decoder. The latent vector generated with the inversion method zi is defined as zi = arg minz Z d X (xi, pθ(z)) (Section 4.2) The latent vector without inversion method zi = qϕ(xi) is constructed by feeding the input xi into the encoder qϕ similar to prior LBO works [11, 17]. Figure 12 shows the dissimilarity comparison results without and with the inversion on med2 and valt tasks. We measure the dissimilarity as the normalized Levenshtein distance between two SELFIES tokens. The x-axis indicates the iteration, and the y-axis indicates the dissimilarity between xi and pθ(zi). From the figure, the inversion achieves zero dissimilarity for every iteration on both tasks, which indicates that the inversion always generates aligned data. On the other hand, BO without the inversion mostly generates misaligned data. These results demonstrate the necessity of the inversion in LBO. D Applying PAS to Tu RBO on standard BO benchmark Figure 13: Optimization results of Tu RBO and applying PAS to Tu RBO on the synthetic Ackley function with 40 dimensions. The lines and ranges indicate the mean and a standard deviation of ten runs with different seeds. We analyze the effectiveness of PAS along with standard BO approaches. We provide the optimization performance of Tu RBO and Tu RBO with PAS in the Ackley benchmark function with 40 dimension, with input ranges from [-32.768, 32.768]. Our implementation is based on the codebase of Tu RBO [36] provided in the Bo Torch [48] tutorial, and we use the RBF kernel with Automatic Relevance Determination (ARD) lengthscale [49]. The number of initial data is 80 and the number of total oracle calls is 1,000. Figure 13 shows that applying PAS on Tu RBO consistently improves the optimization performance during the iteration. This result demonstrates that PAS is also effective in standard BO benchmark tasks. E Experiments of Plug-and-Play We provide further results of applying our Inv BO to previous LBO works, LS-BO, Tu RBO-L, WLBO [11], LOL-BO [13], PG-LBO [17] and Co BO [14] on Guacamol and arithmetic fitting tasks. All results are average scores of ten runs under identical settings. Since LS-BO, W-LBO, and PG-LBO are not trust region-based LBO, we report the optimization results applying inversion (INV) only on these baselines. The experimental results are illustrated in Table 2. The table shows that applying our Inv BO to all previous LBO works consistently improve the optimization performance in all Guacamol and arithmetic expression fitting tasks. F Experiments on Guacamol Benchmark with Small Budget Figure 14: Optimization results on Guacamol benchmark tasks, excluding the tasks in Figure 4. The lines and ranges indicate the average and standard error of ten runs under the same settings. A higher score is a better score. Figure 14 provides the optimization results on three Guacamol benchmark tasks including pdop, adip, and rano. The y-axis of each subfigure denotes the best-found score, and the x-axis denotes the number of the objective function evaluation. Our Inv BO built on the Co BO shows the state-of-the-art performance on all three tasks. In particular, Co BO with our Inv BO achieves 0.56 best score in the pdop task while the best scores of all other baselines are under 0.48. G Experiments on Guacamol Benchmark with Large Budget We provide further optimization results on five Guacamol benchmark tasks, adip, osmb, pdop, rano, and zale under the large budget setting to demonstrate the effectiveness of our Inv BO in diverse budget settings, which are used in previous works [13, 14]. The experimental results are illustrated in Figure 15. In the figure, our Inv BO to Co BO achieves superior performance on large-budget settings. Specifically, Inv BO built on the Co BO showed a large margin from the baseline and achieved a more than 0.9 best score in adip tasks. These results imply that our Inv BO is effective in diverse settings and tasks. H Analysis on Recentering Technique We here provide the details about additional oracle calls in the recentering technique [13, 14]. Firstly, Co BO explicitly mentioned the oracle call for recentering in L8-9 of Algorithm 1 in the paper. Secondly, LOL-BO s oracle calls for recentering can be verified by its official Git Hub code. See, L188228 of lolbo/lolbo/lolbo.py and L54 of lolbo/lolbo/latent_space_objective.py. Table 2: Optimization results of applying Inv BO or inversion method (INV) to several LBOs on Guacamol benchmark tasks and arithmetic expression task, including the task in Table 1. A higher score is better for all tasks except the arithmetic expression task. Task Aripiprazole similarity (adip) Median molecules 2 (med2) Num Oracle 100 300 500 100 300 500 LBO 0.493 0.003 0.497 0.005 0.499 0.006 0.186 0.000 0.186 0.000 0.186 0.000 LBO + INV 0.495 0.002 0.501 0.006 0.509 0.008 0.186 0.000 0.188 0.002 0.190 0.002 W-LBO 0.497 0.003 0.499 0.006 0.502 0.006 0.186 0.000 0.186 0.000 0.186 0.000 W-LBO + INV 0.507 0.001 0.523 0.002 0.526 0.002 0.186 0.000 0.206 0.001 0.207 0.001 PG-LBO 0.497 0.003 0.518 0.008 0.528 0.008 0.186 0.000 0.186 0.000 0.186 0.000 PG-LBO + INV 0.498 0.003 0.523 0.006 0.558 0.007 0.186 0.000 0.186 0.000 0.191 0.002 Tu RBO-L 0.491 0.000 0.491 0.000 0.491 0.000 0.186 0.000 0.186 0.000 0.186 0.000 Tu RBO-L + Inv BO 0.499 0.006 0.518 0.008 0.541 0.008 0.186 0.000 0.194 0.002 0.202 0.001 LOL-BO 0.491 0.000 0.501 0.004 0.512 0.005 0.186 0.000 0.186 0.000 0.190 0.001 LOL-BO + Inv BO 0.500 0.002 0.545 0.009 0.578 0.011 0.189 0.002 0.204 0.005 0.227 0.010 Co BO 0.491 0.000 0.501 0.004 0.509 0.005 0.186 0.000 0.188 0.002 0.191 0.003 Co BO + Inv BO 0.502 0.004 0.542 0.006 0.581 0.008 0.187 0.001 0.203 0.004 0.214 0.006 Task Osimertinib MPO (osmb) Valsartan SMARTS (valt) Num Oracle 100 300 500 100 300 500 LBO 0.762 0.000 0.763 0.001 0.769 0.005 0.000 0.000 0.000 0.000 0.000 0.000 LBO + INV 0.767 0.002 0.773 0.004 0.782 0.004 0.000 0.000 0.000 0.000 0.000 0.000 W-LBO 0.762 0.001 0.766 0.003 0.771 0.004 0.000 0.000 0.000 0.000 0.001 0.001 W-LBO + INV 0.770 0.001 0.786 0.001 0.789 0.001 0.000 0.000 0.002 0.001 0.116 0.010 PG-LBO 0.763 0.002 0.764 0.002 0.771 0.003 0.000 0.000 0.000 0.000 0.000 0.000 PG-LBO + INV 0.765 0.003 0.784 0.005 0.804 0.004 0.000 0.000 0.000 0.000 0.159 0.079 Tu RBO-L 0.762 0.000 0.762 0.000 0.762 0.000 0.000 0.000 0.000 0.000 0.000 0.000 Tu RBO-L + Inv BO 0.765 0.002 0.785 0.003 0.799 0.004 0.000 0.000 0.024 0.017 0.212 0.092 LOL-BO 0.762 0.000 0.769 0.002 0.777 0.003 0.000 0.000 0.000 0.000 0.000 0.000 LOL-BO + Inv BO 0.775 0.002 0.797 0.006 0.807 0.005 0.000 0.000 0.007 0.005 0.171 0.039 Co BO 0.762 0.000 0.763 0.001 0.772 0.003 0.000 0.000 0.000 0.000 0.000 0.000 Co BO + Inv BO 0.769 0.002 0.795 0.004 0.804 0.004 0.000 0.000 0.042 0.013 0.348 0.107 Task Perindopril MPO (pdop) Ranolazine MPO (rano) Num Oracle 100 300 500 100 300 500 LBO 0.458 0.002 0.458 0.002 0.458 0.002 0.650 0.006 0.655 0.007 0.664 0.012 LBO + INV 0.466 0.007 0.466 0.007 0.469 0.007 0.658 0.007 0.669 0.008 0.682 0.009 W-LBO 0.460 0.004 0.462 0.004 0.464 0.005 0.649 0.006 0.674 0.007 0.681 0.005 W-LBO + INV 0.468 0.001 0.478 0.002 0.483 0.001 0.683 0.002 0.700 0.003 0.707 0.003 PG-LBO 0.458 0.002 0.458 0.002 0.465 0.004 0.650 0.003 0.681 0.009 0.720 0.008 PG-LBO + INV 0.458 0.002 0.466 0.005 0.490 0.009 0.674 0.010 0.713 0.013 0.729 0.016 Tu RBO-L 0.456 0.000 0.456 0.000 0.456 0.000 0.642 0.000 0.642 0.000 0.650 0.007 Tu RBO-L + Inv BO 0.470 0.005 0.506 0.008 0.534 0.008 0.688 0.004 0.743 0.011 0.791 0.013 LOL-BO 0.456 0.000 0.462 0.004 0.470 0.004 0.642 0.000 0.671 0.007 0.688 0.010 LOL-BO + Inv BO 0.512 0.017 0.546 0.014 0.565 0.014 0.699 0.004 0.762 0.012 0.787 0.014 Co BO 0.456 0.000 0.460 0.004 0.461 0.004 0.642 0.000 0.654 0.006 0.705 0.007 Co BO + Inv BO 0.506 0.011 0.538 0.014 0.561 0.015 0.686 0.004 0.738 0.012 0.756 0.015 Task Zaleplon MPO (zale) Arithmetic expression Num Oracle 100 300 500 100 300 500 LBO 0.344 0.009 0.357 0.010 0.368 0.009 1.274 0.147 0.512 0.046 0.425 0.020 LBO + INV 0.370 0.015 0.387 0.010 0.394 0.011 1.306 0.128 0.506 0.114 0.317 0.048 W-LBO 0.369 0.014 0.390 0.013 0.407 0.014 1.280 0.169 0.686 0.103 0.538 0.068 W-LBO + INV 0.383 0.003 0.416 0.002 0.423 0.002 1.524 0.008 0.466 0.006 0.445 0.006 PG-LBO 0.369 0.009 0.410 0.012 0.415 0.009 1.290 0.146 0.913 0.140 0.832 0.146 PG-LBO + INV 0.372 0.016 0.420 0.007 0.452 0.006 1.515 0.060 0.872 0.107 0.595 0.106 Tu RBO-L 0.327 0.000 0.332 0.005 0.332 0.005 1.424 0.126 0.885 0.083 0.605 0.094 Tu RBO-L + Inv BO 0.411 0.017 0.461 0.015 0.475 0.016 0.898 0.161 0.692 0.139 0.374 0.060 LOL-BO 0.327 0.000 0.398 0.012 0.431 0.006 0.953 0.109 0.578 0.053 0.412 0.039 LOL-BO + Inv BO 0.400 0.011 0.456 0.011 0.477 0.011 0.739 0.142 0.510 0.031 0.392 0.023 Co BO 0.327 0.000 0.393 0.014 0.421 0.010 0.752 0.122 0.526 0.057 0.391 0.006 Co BO + Inv BO 0.435 0.014 0.495 0.023 0.513 0.012 0.840 0.173 0.513 0.075 0.252 0.067 Figure 15: Optimization results on Guacamol task under the large budget setting, excluding the tasks in Figure 5(b). The lines and ranges indicate the average and standard error of five runs under the same settings. A higher score is a better score. We present the experiments on additional oracle calls in recentering in Figure 16. The figure shows the optimization results of Co BO that use recentering with additional oracle calls and recentering without additional oracle calls on med2 and valt tasks. In both tasks, Co BO that uses the recentering technique without additional oracle calls fails in optimization while consuming additional oracle calls shows progress. These results show that the additional oracle calls are essential in the recentering technique, while the previous works do not explicitly mention it. I Analysis on Misalignment Problem We here provide the experimental details about the Gaussian process fitting with the encoder triplets and the decoder triplets. We train the surrogate model on 300 train points and predict on 100 test points, all points are randomly sampled. The encoder triplets {(xi, qϕ(xi), f(xi))}n i=1 and the decoder triplets {(xi, zi inv, f(xi))}n i=1 share the same discrete data x, where xi = pθ(zi inv). The dots in the figure denote the mean predictions of the surrogate model and bars denote 95% confidence intervals. Figure 17 provides the Gaussian process fitting results trained with the encoder triplets (Left) and decoder triplets (Right) that predict the training points on the med2 task. The figure implies that the Gaussian process trained with encoder triplet fails to fit the composite function f pθ, while trained with decoder triplet emulates the function accurately. J Efficiency Analysis Although Bayesian optimization assumes the objective function is cost-expensive, it is still important to consider the efficiency of the algorithm. We present an efficiency analysis comparing with baseline methods [11, 13, 14, 17], and Tu RBO-L, LS-BO. We conducted experiments under the same condition for fair comparison: a single NVIDIA RTX 2080 TI with the CPU of AMD EPYC 7742. Wall-clock time comparison with the baseline model is in Table 3. The table shows that applying our Inv BO to Co BO achieves state-of-the-art performance not only under the same oracle calls but also under the same wall-clock time. Figure 16: Optimizaiton results on Guacamol tasks, med2 and valt. We compare the Co BO applying Inv BO that uses the recentering technique with and without additional oracle calls. The lines and ranges indicate the average and standard error of ten runs under the same settings. A higher score is a better score. Figure 17: Gaussian process fitting results trained with encoder triplets and decoder triplets. Note that the datapoints are in the training set. K Details of Benchmark Tasks We demonstrate the effectiveness of Inv BO through three optimization benchmarks: Guacamol [43], DRD3, and arithmetic expression fitting task [8, 11, 13, 14]. Guacamol benchmark tasks aim to design molecules with the desired properties, which are scored with a range between 0 to 1. The number of initial points is 100 and the number of oracle calls is 500 in a small-budget setting, and the number of initial points is 10,000 and the number of oracle calls is 70,000 in a large-budget setting. In the DRD3 benchmark task, we aim to design ligands (molecules that specifically bind to a protein) that bind to dopamine receptor D3 (DRD3). We evaluate the docking score of ligands with the Dockstring library [50]. The number of the observed data points is 100 and the number of max oracle calls is 3,000. Lastly, the arithmetic expression fitting task is to generate single-variable expressions that are close to a target expression (e.g., 1/3 + x + sin (x x)). The number of initial access to oracle is 100 and the number of max oracle calls is 500. All tasks are assumed to be noiseless settings, and we select the initial points randomly from the dataset provided in [14] and the same initial data was used in all tasks. L Details of Baselines We compare our Inv BO with six latent Bayesian Optimization methods: LS-BO, Tu RBO [36], WLBO [11], LOL-BO [13], Co BO [14], and PG-LBO [17], and one graph-based genetic algorithm: Table 3: Wall-clock time and corresponding best score for each model. Task Model Co BO+Inv BO Co BO PG-LBO LOL-BO Tu RBO LS-BO W-LBO Found Best Score 0.2143 0.1907 0.1856 0.1904 0.1856 0.1856 0.1856 Oracle call 500 500 500 500 500 500 500 Wall-clock time (s) 420.12 285.02 1582.23 308.45 842.25 152.31 573.21 Found Best Score 0.1938 0.1883 0.1856 0.1856 0.1856 0.1856 0.1856 Oracle calls 284 314 55 286 182 400 236 Wall-clock time (s) 152.31 152.31 152.31 152.31 152.31 152.31 152.31 Found Best Score 0.7325 0.0000 0.0000 0.0000 0.0000 0.0000 0.0000 Oracle calls 500 500 500 500 500 500 500 Wall-clock time (s) 246.12 148.23 443.65 163.67 101.58 152.96 133.48 Found Best Score 0.1173 0.0000 0.0000 0.0000 0.0000 0.0000 0.0000 Oracle calls 262 382 184 403 426 500 439 Wall-clock time (s) 101.58 101.58 101.58 101.58 101.58 101.58 101.58 GB-GA [45]. We adopt the standard LBO for LS-BO approach following other work [14] with VAE update during the optimization process. Tu RBO leverages trust regions to alleviate the overexploration. To adapt Tu RBO to the latent space, we employ Tu RBO-L, LS-BO with trust region, following other latent Bayesian optimization baselines [13, 14]. W-LBO weights the data based on importance in order to focus on samples with high objective values. LOL-BO adapts the concept of the trust region to the latent space and proposes VAE learning methods to inject the prior of the sparse GP models for better optimization. Co BO designs novel regularization losses based on the Lipschitz condition to boost the correlation between the distance in the latent space and the distance within the objective function. PG-LBO utilizes unobserved data with a pseudo-labeling technique and integrates Gaussian Process guidance into VAE training to learn a latent space for better optimization. In PG-LBO, we sample the pseudo data by adding the Gaussian noise, and dynamic thresholding. M Implementation Details Our implementation is based on the codebase of [13]. We use Py Torch3, Bo Torch4 [48], GPy Torch5 [52], and Guacamol6 software packages. In the Guacamol tasks and DRD3 task [50], we use the SELFIES VAE [13] which is pretrained in an unsupervised manner with 1.27M molecules from Guacamol benchmark. The Grammar VAE [8] is pre-trained with 40K expression data in an unsupervised manner. The dimension of latent space is 256 in the SELFIES VAE and 25 in the Grammar VAE. The size of the data obtained from the acquisition function and k are hyperparameters, which are presented in Table 4. M.1 Hyperparameters Setting The learning rate used in the inversion method is 0.1 in all tasks, as we empirically observe that it always finds the latent vector that generates the target discrete data. The maximum iteration number of the inversion method is 1,000 in all tasks. The rest of the hyperparameters follow the setting used in [14]. Since arithmetic fitting tasks and Guacamol with the small budget tasks use different initial data numbers used in [14], we set the number of top-k data and the number of query points Nq same as DRD3 task, as they use the same number of initial data points. N Pseudocode of Potential-aware Trust Region Anchor Selection Here, we provide the pseudocode of the potential-aware trust region anchor selection method. We get the max acquisition function value, which is a Thompson Sampling, of each trust region T i given a 3Copyright (c) 2016-Facebook, Inc (Adam Paszke) [51], Licensed under BSD-style license 4Copyright (c) Meta Platforms, Inc. and affiliates. Licensed under MIT License 5Copyright (c) 2017 Jake Gardner. Licensed under MIT License 6Copyright (c) 2020 The Apache Software Foundation, Licensed under the Apache License, Version 2.0. Table 4: Other hyperparameters for various benchmarks. Hyperparameters Guacamol-small Guacamol-large DRD3 Arithmetic Number of initial data points |D0| 100 10000 100 100 Number of query points Nq 5 10 5 5 Number of top-k data 50 1000 10 10 VAE update interval Nfail 10 10 10 10 Max oracle calls 500 70000 3000 500 Algorithm 2 Potential-aware Anchor Selection Input: Data history {xi, zi, yi}n i=1, surrogate model GP 1: for i = 1, 2, ..., n do 2: Get a candidate set Zi cand with random points in the trust region T i around zi 3: αi pot = max z Zi cand ˆf(z) where ˆf GP (µ(z), k(z, z )) Eq. (7) 4: end for 5: Y = {yi}n i=1 6: A = {αi pot}n i=1 7: for i = 1, 2, ..., n do 8: αi scaled Calculate(αi pot, Y, A) Eq. (8) 9: si yi + αi scaled Eq. (9) 10: end for 11: S = {si}n i=1 12: return zi where i is index of max (S) candidate set Zi cand in lines 2-3. After scaling the max acquisition function values, we calculate the final score of each trust region si in lines 8-9. Finally, we sort the anchors in the descending order of the final score and select the anchor. O Pseudocode of Inv BO Applied to Co BO We provide the pseudocode of Co BO applying our method, Inv BO, which consists of the inversion method and the potential-aware anchor selection. The inversion method is used in lines 1 and 8, and the potential-aware anchor selection method is used in line 11. As in Co BO, we use the subset of data history, which is composed of the top k scored data and the most recently evaluated data during the optimization process in line 4. When we fail Nfail times to update the best score, we train the VAE with Co BO loss LCo BO and update the dataset to be aligned using the inversion method. After that, we train the surrogate model with Dt and negative log-likelihood loss Lsurr except for the first iteration and select the trust region anchor with the proposed potential-aware trust region anchor selection in line 11. We select Nq next query points from the trust region in line 12. Then, we evaluate the next query points with the objective function and update the data history in lines 17, and 18. Algorithm 3 Inv BO built on Co BO Input: Pretrained encoder qϕ, decoder pθ, black-box function f, surrogate model g, acquisition function α, oracle budget T, latent update interval Nfail, number of query point Nq, initial data D0 = (xi, yi) n i=1, learning rate for inversion η, distance function d X 1: D0 xi, zi inv, yi | xi, yi D0, zi inv Inversion xi, qϕ, pθ, η, d X n i=1 Algorithm 1 2: nfail 0 3: for t = 1, 2, ..., T do 4: Dt CONCAT Dt 1[ Nq :], topk(Dt 1) 5: if nfail Nfail then 6: nfail 0 7: Train qϕ and pθ with LCo BO, Dt 8: Dt xi, zi inv, yi | xi, zi, yi Dt, zi inv Inversion xi, qϕ, pθ, η, d X |Dt| i=1 Algorithm 1 9: end if 10: Train g with Lsurr, Dt if t = 1 else D0 11: Zanchor Trust Region Anchor Selection(Dt, g) Algorithm 2 12: Znext Select Nq query points in trust region centered on Zanchor 13: y max(x,z,y) Dt y 14: if f(pθ(zi)) y , zi Znext then 15: nfail nfail + 1 16: end if 17: Dnew pθ(zi), zi, f(pθ(zi)) |zi Znext |Znext| i=1 18: Dt CONCAT (Dt, Dnew) 19: end for 20: (x , z , y ) arg max(x,z,y) DT y 21: return x Neur IPS Paper Checklist Question: Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? Answer: [Yes] Justification: The main claims reflect the paper s contributions and scope sufficiently. Guidelines: The answer NA means that the abstract and introduction do not include the claims made in the paper. The abstract and/or introduction should clearly state the claims made, including the contributions made in the paper and important assumptions and limitations. A No or NA answer to this question will not be perceived well by the reviewers. The claims made should match theoretical and experimental results, and reflect how much the results can be expected to generalize to other settings. It is fine to include aspirational goals as motivation as long as it is clear that these goals are not attained by the paper. 2. Limitations Question: Does the paper discuss the limitations of the work performed by the authors? Answer: [Yes] Justification: The paper provide the limitation after the Section 7. Guidelines: The answer NA means that the paper has no limitation while the answer No means that the paper has limitations, but those are not discussed in the paper. The authors are encouraged to create a separate "Limitations" section in their paper. The paper should point out any strong assumptions and how robust the results are to violations of these assumptions (e.g., independence assumptions, noiseless settings, model well-specification, asymptotic approximations only holding locally). The authors should reflect on how these assumptions might be violated in practice and what the implications would be. The authors should reflect on the scope of the claims made, e.g., if the approach was only tested on a few datasets or with a few runs. In general, empirical results often depend on implicit assumptions, which should be articulated. The authors should reflect on the factors that influence the performance of the approach. For example, a facial recognition algorithm may perform poorly when image resolution is low or images are taken in low lighting. Or a speech-to-text system might not be used reliably to provide closed captions for online lectures because it fails to handle technical jargon. The authors should discuss the computational efficiency of the proposed algorithms and how they scale with dataset size. If applicable, the authors should discuss possible limitations of their approach to address problems of privacy and fairness. While the authors might fear that complete honesty about limitations might be used by reviewers as grounds for rejection, a worse outcome might be that reviewers discover limitations that aren t acknowledged in the paper. The authors should use their best judgment and recognize that individual actions in favor of transparency play an important role in developing norms that preserve the integrity of the community. Reviewers will be specifically instructed to not penalize honesty concerning limitations. 3. Theory Assumptions and Proofs Question: For each theoretical result, does the paper provide the full set of assumptions and a complete (and correct) proof? Answer: [Yes] Justification: The paper provide the theory and The full set of assumption on Proposition 1. A complete proof of the proposition in Section A. Guidelines: The answer NA means that the paper does not include theoretical results. All the theorems, formulas, and proofs in the paper should be numbered and crossreferenced. All assumptions should be clearly stated or referenced in the statement of any theorems. The proofs can either appear in the main paper or the supplemental material, but if they appear in the supplemental material, the authors are encouraged to provide a short proof sketch to provide intuition. Inversely, any informal proof provided in the core of the paper should be complemented by formal proofs provided in appendix or supplemental material. Theorems and Lemmas that the proof relies upon should be properly referenced. 4. Experimental Result Reproducibility Question: Does the paper fully disclose all the information needed to reproduce the main experimental results of the paper to the extent that it affects the main claims and/or conclusions of the paper (regardless of whether the code and data are provided or not)? Answer: [Yes] Justification: The paper provide the implementation details including the hyperparameter settings in Section 5.3, M and the details of baselines in Section L, and the details of benchmark tasks in Section K. Guidelines: The answer NA means that the paper does not include experiments. If the paper includes experiments, a No answer to this question will not be perceived well by the reviewers: Making the paper reproducible is important, regardless of whether the code and data are provided or not. If the contribution is a dataset and/or model, the authors should describe the steps taken to make their results reproducible or verifiable. Depending on the contribution, reproducibility can be accomplished in various ways. For example, if the contribution is a novel architecture, describing the architecture fully might suffice, or if the contribution is a specific model and empirical evaluation, it may be necessary to either make it possible for others to replicate the model with the same dataset, or provide access to the model. In general. releasing code and data is often one good way to accomplish this, but reproducibility can also be provided via detailed instructions for how to replicate the results, access to a hosted model (e.g., in the case of a large language model), releasing of a model checkpoint, or other means that are appropriate to the research performed. While Neur IPS does not require releasing code, the conference does require all submissions to provide some reasonable avenue for reproducibility, which may depend on the nature of the contribution. For example (a) If the contribution is primarily a new algorithm, the paper should make it clear how to reproduce that algorithm. (b) If the contribution is primarily a new model architecture, the paper should describe the architecture clearly and fully. (c) If the contribution is a new model (e.g., a large language model), then there should either be a way to access this model for reproducing the results or a way to reproduce the model (e.g., with an open-source dataset or instructions for how to construct the dataset). (d) We recognize that reproducibility may be tricky in some cases, in which case authors are welcome to describe the particular way they provide for reproducibility. In the case of closed-source models, it may be that access to the model is limited in some way (e.g., to registered users), but it should be possible for other researchers to have some path to reproducing or verifying the results. 5. Open access to data and code Question: Does the paper provide open access to the data and code, with sufficient instructions to faithfully reproduce the main experimental results, as described in supplemental material? Answer: [Yes] Justification: The paper provide anonymized code that satisfied the submission guidelines. Guidelines: The answer NA means that paper does not include experiments requiring code. Please see the Neur IPS code and data submission guidelines (https://nips.cc/ public/guides/Code Submission Policy) for more details. While we encourage the release of code and data, we understand that this might not be possible, so No is an acceptable answer. Papers cannot be rejected simply for not including code, unless this is central to the contribution (e.g., for a new open-source benchmark). The instructions should contain the exact command and environment needed to run to reproduce the results. See the Neur IPS code and data submission guidelines (https: //nips.cc/public/guides/Code Submission Policy) for more details. The authors should provide instructions on data access and preparation, including how to access the raw data, preprocessed data, intermediate data, and generated data, etc. The authors should provide scripts to reproduce all experimental results for the new proposed method and baselines. If only a subset of experiments are reproducible, they should state which ones are omitted from the script and why. At submission time, to preserve anonymity, the authors should release anonymized versions (if applicable). Providing as much information as possible in supplemental material (appended to the paper) is recommended, but including URLs to data and code is permitted. 6. Experimental Setting/Details Question: Does the paper specify all the training and test details (e.g., data splits, hyperparameters, how they were chosen, type of optimizer, etc.) necessary to understand the results? Answer: [Yes] Justification: The paper provide the implementation details including the hyperparameter settings in Section 5.3, M and the details of baselines in Section L, and the details of benchmark tasks in Section K. Guidelines: The answer NA means that the paper does not include experiments. The experimental setting should be presented in the core of the paper to a level of detail that is necessary to appreciate the results and make sense of them. The full details can be provided either with the code, in appendix, or as supplemental material. 7. Experiment Statistical Significance Question: Does the paper report error bars suitably and correctly defined or other appropriate information about the statistical significance of the experiments? Answer: [Yes] Justification: The paper depict the errorbar and the error statistic on every performance results. Guidelines: The answer NA means that the paper does not include experiments. The authors should answer "Yes" if the results are accompanied by error bars, confidence intervals, or statistical significance tests, at least for the experiments that support the main claims of the paper. The factors of variability that the error bars are capturing should be clearly stated (for example, train/test split, initialization, random drawing of some parameter, or overall run with given experimental conditions). The method for calculating the error bars should be explained (closed form formula, call to a library function, bootstrap, etc.) The assumptions made should be given (e.g., Normally distributed errors). It should be clear whether the error bar is the standard deviation or the standard error of the mean. It is OK to report 1-sigma error bars, but one should state it. The authors should preferably report a 2-sigma error bar than state that they have a 96% CI, if the hypothesis of Normality of errors is not verified. For asymmetric distributions, the authors should be careful not to show in tables or figures symmetric error bars that would yield results that are out of range (e.g. negative error rates). If error bars are reported in tables or plots, The authors should explain in the text how they were calculated and reference the corresponding figures or tables in the text. 8. Experiments Compute Resources Question: For each experiment, does the paper provide sufficient information on the computer resources (type of compute workers, memory, time of execution) needed to reproduce the experiments? Answer: [Yes] Justification: The paper provide the computer resources including the time of excution in Section J. Guidelines: The answer NA means that the paper does not include experiments. The paper should indicate the type of compute workers CPU or GPU, internal cluster, or cloud provider, including relevant memory and storage. The paper should provide the amount of compute required for each of the individual experimental runs as well as estimate the total compute. The paper should disclose whether the full research project required more compute than the experiments reported in the paper (e.g., preliminary or failed experiments that didn t make it into the paper). 9. Code Of Ethics Question: Does the research conducted in the paper conform, in every respect, with the Neur IPS Code of Ethics https://neurips.cc/public/Ethics Guidelines? Answer: [Yes] Justification: The paper conform the Neur IPS Code of Ethics. Guidelines: The answer NA means that the authors have not reviewed the Neur IPS Code of Ethics. If the authors answer No, they should explain the special circumstances that require a deviation from the Code of Ethics. The authors should make sure to preserve anonymity (e.g., if there is a special consideration due to laws or regulations in their jurisdiction). 10. Broader Impacts Question: Does the paper discuss both potential positive societal impacts and negative societal impacts of the work performed? Answer: [Yes] Justification: The paper provide the broader impacts after the Section 7. Guidelines: The answer NA means that there is no societal impact of the work performed. If the authors answer NA or No, they should explain why their work has no societal impact or why the paper does not address societal impact. Examples of negative societal impacts include potential malicious or unintended uses (e.g., disinformation, generating fake profiles, surveillance), fairness considerations (e.g., deployment of technologies that could make decisions that unfairly impact specific groups), privacy considerations, and security considerations. The conference expects that many papers will be foundational research and not tied to particular applications, let alone deployments. However, if there is a direct path to any negative applications, the authors should point it out. For example, it is legitimate to point out that an improvement in the quality of generative models could be used to generate deepfakes for disinformation. On the other hand, it is not needed to point out that a generic algorithm for optimizing neural networks could enable people to train models that generate Deepfakes faster. The authors should consider possible harms that could arise when the technology is being used as intended and functioning correctly, harms that could arise when the technology is being used as intended but gives incorrect results, and harms following from (intentional or unintentional) misuse of the technology. If there are negative societal impacts, the authors could also discuss possible mitigation strategies (e.g., gated release of models, providing defenses in addition to attacks, mechanisms for monitoring misuse, mechanisms to monitor how a system learns from feedback over time, improving the efficiency and accessibility of ML). 11. Safeguards Question: Does the paper describe safeguards that have been put in place for responsible release of data or models that have a high risk for misuse (e.g., pretrained language models, image generators, or scraped datasets)? Answer: [No] Justification: The paper does not include the safeguards for misusing. Guidelines: The answer NA means that the paper poses no such risks. Released models that have a high risk for misuse or dual-use should be released with necessary safeguards to allow for controlled use of the model, for example by requiring that users adhere to usage guidelines or restrictions to access the model or implementing safety filters. Datasets that have been scraped from the Internet could pose safety risks. The authors should describe how they avoided releasing unsafe images. We recognize that providing effective safeguards is challenging, and many papers do not require this, but we encourage authors to take this into account and make a best faith effort. 12. Licenses for existing assets Question: Are the creators or original owners of assets (e.g., code, data, models), used in the paper, properly credited and are the license and terms of use explicitly mentioned and properly respected? Answer: [Yes] Justification: The paper provide the license in Section M and used the assets properly. Guidelines: The answer NA means that the paper does not use existing assets. The authors should cite the original paper that produced the code package or dataset. The authors should state which version of the asset is used and, if possible, include a URL. The name of the license (e.g., CC-BY 4.0) should be included for each asset. For scraped data from a particular source (e.g., website), the copyright and terms of service of that source should be provided. If assets are released, the license, copyright information, and terms of use in the package should be provided. For popular datasets, paperswithcode.com/datasets has curated licenses for some datasets. Their licensing guide can help determine the license of a dataset. For existing datasets that are re-packaged, both the original license and the license of the derived asset (if it has changed) should be provided. If this information is not available online, the authors are encouraged to reach out to the asset s creators. 13. New Assets Question: Are new assets introduced in the paper well documented and is the documentation provided alongside the assets? Answer: [Yes] Justification: The paper provide the code and documentation of it. Guidelines: The answer NA means that the paper does not release new assets. Researchers should communicate the details of the dataset/code/model as part of their submissions via structured templates. This includes details about training, license, limitations, etc. The paper should discuss whether and how consent was obtained from people whose asset is used. At submission time, remember to anonymize your assets (if applicable). You can either create an anonymized URL or include an anonymized zip file. 14. Crowdsourcing and Research with Human Subjects Question: For crowdsourcing experiments and research with human subjects, does the paper include the full text of instructions given to participants and screenshots, if applicable, as well as details about compensation (if any)? Answer: [NA] Justification: The paper does not involve crowdsourcing nor research with human subjects. Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Including this information in the supplemental material is fine, but if the main contribution of the paper involves human subjects, then as much detail as possible should be included in the main paper. According to the Neur IPS Code of Ethics, workers involved in data collection, curation, or other labor should be paid at least the minimum wage in the country of the data collector. 15. Institutional Review Board (IRB) Approvals or Equivalent for Research with Human Subjects Question: Does the paper describe potential risks incurred by study participants, whether such risks were disclosed to the subjects, and whether Institutional Review Board (IRB) approvals (or an equivalent approval/review based on the requirements of your country or institution) were obtained? Answer: [NA] Justification: The paper does not involve crowdsourcing nor research with human subjects. Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Depending on the country in which research is conducted, IRB approval (or equivalent) may be required for any human subjects research. If you obtained IRB approval, you should clearly state this in the paper. We recognize that the procedures for this may vary significantly between institutions and locations, and we expect authors to adhere to the Neur IPS Code of Ethics and the guidelines for their institution. For initial submissions, do not include any information that would break anonymity (if applicable), such as the institution conducting the review.