# randomized_strategic_facility_location_with_predictions__c00e30be.pdf Randomized Strategic Facility Location with Predictions Eric Balkanski Columbia University, IEOR eb3224@columbia.edu Vasilis Gkatzelis Drexel University, Computer Science gkatz@drexel.edu Golnoosh Shahkarami Max Planck Institut für Informatik, Universität des Saarlandes gshahkar@mpi-inf.mpg.de In the strategic facility location problem, a set of agents report their locations in a metric space and the goal is to use these reports to open a new facility, minimizing an aggregate distance measure from the agents to the facility. However, agents are strategic and may misreport their locations to influence the facility s placement in their favor. The aim is to design truthful mechanisms, ensuring agents cannot gain by misreporting. This problem was recently revisited through the learningaugmented framework, aiming to move beyond worst-case analysis and design truthful mechanisms that are augmented with (machine-learned) predictions. The focus of this prior work was on mechanisms that are deterministic and augmented with a prediction regarding the optimal facility location. In this paper, we provide a deeper understanding of this problem by exploring the power of randomization as well as the impact of different types of predictions on the performance of truthful learning-augmented mechanisms. We study both the single-dimensional and the Euclidean case and provide upper and lower bounds regarding the achievable approximation of the optimal egalitarian social cost. 1 Introduction In the classic facility location problem the goal is to determine the ideal location for a new facility, taking as input the preferences of a group of n agents. Due to the wide variety of applications that this problem captures, it has received a lot of attention from different perspectives. A notable example is the strategic version of this problem which is motivated by the fact that the participating agents may be able to strategically misreport their preferences, leading to a facility location choice that they prefer over the one that would be chosen if they were truthful. In the strategic facility location problem, the goal is to design truthful mechanisms, i.e., mechanisms that elicit the agents true preferences by carefully removing any incentive for the agents to lie [Moulin, 1980, Procaccia and Tennenholtz, 2013]. This problem has played a central role in the broader literature on mechanism design without money and a lot of prior work has focused on optimizing the quality of the returned location subject to the truthfulness constraint (see Section 1.2 for a brief overview). However, this work has mostly focused on worst-case analysis, often leading to unnecessarily pessimistic impossibility results. Aiming to overcome the limitations of worst-case analysis, a surge of work during the last few years has focused on the design of algorithms in the learning-augmented framework [Mitzenmacher and Vassilvitskii, 2022]. According to this framework, the designer is provided with some machinelearned prediction regarding the instance at hand and the goal is to leverage this additional information in the design process. However, crucially, this information is not guaranteed to be accurate and can, 38th Conference on Neural Information Processing Systems (Neur IPS 2024). in fact, be arbitrarily inaccurate. The goal is to achieve stronger performance guarantees whenever the prediction happens to be accurate (known as the consistency guarantee), while at the same time maintaining some worst-case guarantee even if the prediction is arbitrarily inaccurate (known as the robustness guarantee). Agrawal et al. [2022] and Xu and Lu [2022] recently introduced the learning-augmented framework to mechanism design problems involving self-interested strategic agents with private information, and both of these works studied the strategic facility location problem. Agrawal et al. [2022] considered both the egalitarian and the utilitarian social cost objectives and provided tight bounds on the bestfeasible robustness and consistency trade-off for truthful mechanisms augmented with a prediction regarding what the optimal facility location would be. However, the achievable trade-offs between robustness and consistency may heavily depend on the specific type of prediction that the mechanism is augmented with: note that the consistency guarantee only binds if the prediction is accurate, so more refined predictions bind on fewer instances (the subset of instances where even the refined prediction is accurate) and could, therefore, enable the design of learning-augmented mechanisms with improved guarantees. This leaves open the question of whether mechanisms equipped with more refined predictions can, indeed, achieve better trade-offs between robustness and consistency. Also, Agrawal et al. [2022] restricted their attention to deterministic mechanisms, leaving open the possibility for truthful randomized mechanisms to achieve even stronger guarantees. In this work we significantly expand our understanding of this problem by studying both the power of randomization and the power of alternative predictions. 1.1 Our Results We revisit the problem of designing truthful learning-augmented mechanisms for the strategic facility location problem. These mechanisms ask each agent to report their preferred location in some metric space (which is private information) and they are also augmented with some (unreliable) prediction related to this private information. Using the agents reported locations, along with the prediction, the mechanism chooses a facility location in this metric space, aiming to (approximately) minimize some social cost measure. Our focus in this paper is on the egalitarian social cost, i.e., the maximum distance between an agent s preferred location and the location chosen for the facility. Our goal is to evaluate the performance of mechanisms that can leverage randomization, as well as the power of different types of predictions. We first consider the well-studied single-dimensional version of the problem, where all agents lie on a line and the mechanism needs to choose a facility location on that line. Prior work on the strategic facility location problem without predictions, showed that for this class of instances no deterministic truthful mechanism can achieve an approximation better than 2 and no randomized truthful1 mechanism can achieve an approximation better than 1.5; both of these results are shown to be tight [Procaccia and Tennenholtz, 2013]. In the learning-augmented framework, Agrawal et al. [2022] showed there exists a truthful deterministic mechanism provided with a prediction regarding the optimal facility location that achieves the best of both worlds: a perfect consistency of 1 and an optimal robustness of 2. Our main result for the single-dimensional version shows that even if the mechanism is provided with the strongest possible prediction (i.e., a prediction regarding every agent s preferred location), any randomized truthful mechanism that is (1 + δ)-consistent for some δ [0, 0.5] can be no better than (2 δ)-robust. This implies that the previously proposed deterministic learning-augmented mechanism that is 1-consistent and 2-robust and the randomized non-learning-augmented mechanism that is 1.5-robust (and hence also 1.5-consistent) are both Pareto optimal among all randomized mechanisms, even if they are augmented with the strongest possible predictions. Beyond these two extreme points, this result proves a lower bound for the achievable trade-off between robustness and consistency for any randomized mechanism, even with the strongest predictions. We complement this bound by observing that this trade-off can, in fact, be achieved by a truthful randomized mechanism that chooses between the optimal learning-augmented deterministic mechanism and the optimal non-augmented randomized mechanism, thus verifying that our bound fully characterizes this Pareto frontier. 1A truthful randomized mechanism guarantees truthfulness in expectation. See Section 2 for more details. We then consider the two-dimensional case, for which much less is known about randomized mechanisms. The best-known truthful randomized mechanism is the Centroid, introduced by Tang et al. [2020], which guarantees a 2 1/n approximation. We provide a truthful randomized mechanism that is equipped with a prediction regarding the identities of the most extreme agents (those who would incur the maximum cost in the optimal solution) and, using this prediction, our mechanism achieves 1.67 consistency and 2 robustness. The idea is to use these predictions to select at most three extreme agents, apply the Centroid mechanism to them, and guarantee a 1.67 approximation if the predictions are accurate not only for these three agents but for all other agents as well, utilizing the properties of the Euler line. An interesting observation is that the lower bound of 1.5 from the single-dimensional case does not extend to two dimensions, so prior work does not imply any lower bound for two dimensions! We prove a lower bound of 1.118 for all truthful randomized mechanisms and then that no deterministic mechanism can simultaneously guarantee better than 2 consistency and better than 1+ 2 robustness, even if it is augmented with the strongest predictions. Similarly, no truthful randomized mechanism can simultaneously guarantee 1 consistency and better than 2 robustness. The former result proves the optimality of a previously introduced 1-consistent and 1 + 2-robust truthful deterministic mechanism provided only with a prediction regarding the optimal facility location [Agrawal et al., 2022]. We also show that the latter is tight. 1.2 Related Work Strategic facility location. Moulin [1980] provided a characterization of deterministic truthful facility location mechanisms on the line and introduced the median mechanism, which returns the median of the agent location profile x = x1, , xn . The median mechanism is known to be truthful, providing an optimal solution for the Utilitarian Social Cost and achieving a 2-approximation for the Egalitarian Social Cost, as demonstrated by Procaccia and Tennenholtz [2013]. Notably, Procaccia and Tennenholtz [2013] showed that this 2-approximation factor represents the best performance achievable by any deterministic truthful mechanism. Building on this, Border and Jordan [1983] extended these results to the Euclidean space by employing median schemes independently in each dimension. Subsequently, Barberà et al. [1993] further generalized this outcome to any L1norms. Other settings explored include general metric spaces [Alon et al., 2010] and d-dimensional Euclidean spaces [Meir, 2019, Walsh, 2020, El-Mhamdi et al., 2023, Goel and Hann-Caruthers, 2020], as well as circles [Alon et al., 2010, Meir, 2019] and trees [Alon et al., 2010, Feldman and Wilf, 2013]. Fundamental results in truthful facility location often focus on characterizing the space of truthful mechanisms. For the one-dimensional case, Moulin s characterization [Moulin, 1980] reveals that all deterministic truthful mechanisms belong to the general median mechanisms (GCM) family. For the two-dimensional case, a similar characterization was provided by Peters et al. [1993]. For a comprehensive review of previous work on this problem, see the survey by Chan et al. [2021]. Learning-augmented algorithms. Worst-case analysis on its own is often not informative enough, and several alternative measures have been proposed to overcome its limitations [Roughgarden, 2021]. Learning-augmented algorithms, or algorithms with predictions [Mitzenmacher and Vassilvitskii, 2022], aim to address these limitations by incorporating predictions into the algorithm s design. Lykouris and Vassilvitskii [2021] studied the online caching problem and introduced two main metrics, consistency and robustness, to evaluate the performance of these algorithms. The online version of the facility location problem, introduced by Meyerson [2001], involves points arriving online, requiring us to assign them irrevocably to either an existing facility or open a new facility. The objective is to minimize the distance of each agent to the assigned facility along with the cost of opening the facilities. Almanza et al. [2021], Jiang et al. [2022], and Fotakis et al. [2021] studied this problem augmented with predictions regarding the location of the optimal facility for each incoming point. We direct interested readers to a curated and frequently updated list of papers in this area [Lindermayr and Megow]. Learning-augmented mechanism design. The framework of learning-augmented mechanism design was first introduced by Agrawal et al. [2022] and Xu and Lu [2022]. Agrawal et al. [2022] focused on the strategic facility location problem, proposing mechanisms that leverage predictions to improve performance while maintaining truthfulness. Their work achieves the best possible consistency and robustness trade-off given a prediction of the optimal facility location. They also evaluated the performance of their mechanisms as a function of the prediction error. More recently, Christodoulou et al. [2024] provided performance bounds based on an alternative measure of prediction error. Barak et al. [2024] studied this problem assuming that the predictions are regarding each agent s location, but a small fraction of these predictions may be arbitrarily inaccurate. Meanwhile, Chen et al. [2024] evaluated non-truthful mechanisms with respect to their price of anarchy. Istrate and Bonchis [2022] and Fang et al. [2024] studied the obnoxious facility location version of this problem, aiming to design truthful mechanisms. Apart from the facility location problem, other works in learning-augmented mechanism design have been conducted in various contexts, including strategic scheduling [Xu and Lu, 2022, Balkanski et al., 2023a], auctions [Xu and Lu, 2022, Balkanski et al., 2023b,a, Lu et al., 2023, Caragiannis and Kalantzis, 2024, Gkatzelis et al., 2024], bicriteria mechanism design [Balcan et al., 2024], graph problems with private input [Colini-Baldeschi et al., 2024], and equilibrium analysis [Gkatzelis et al., 2022, Istrate et al., 2024]. Balkanski et al. [2023b] brought together the line of work on learning-augmented mechanism design with the literature on online algorithms with predictions by studying online mechanism design with predictions. 2 Preliminaries In the single facility location problem, there are n strategic agents with a location profile denoted by x = x1, , xn , where xi corresponds to the location of agent i. A mechanism f(x) outputs a, potentially randomized, location for the facility. The cost incurred by agent i is the expected Euclidean distance E[d(xi, f(x))] between the facility location f(x) and their location xi. Two renowned cost functions considered in this scenario are Egalitarian Social Cost and Utilitarian Social Cost. In this work, the goal is to optimize the Egalitarian Social Cost C(f, x) = E[maxxi x d(xi, f(x))], representing the expected maximum cost experienced by a single agent for each possible outcome of f(x). The optimal location for the facility in the two-dimensional Euclidean space corresponds to the center of the smallest circle that encloses all the points, denoted by o(x). To minimize the social cost, a mechanism needs to ask the agents to report their preferred locations, x R2n, and then use this information to determine the facility location f(x) R2. However, the preferred location xi of each agent i is private information, and they can choose to misreport their preferred location if that can reduce their own cost. A mechanism f : R2n R2 is considered truthful when no individual agent can benefit by misreporting their location, i.e., for all instances x R2n, every agent i [n], and every deviation x i R2, we have that d(xi, f(x)) d(xi, f(x i, x i)), where x i = x1, , xi 1, xi+1, , xn is the vector of the locations of all agents except agent i. In the context of randomized mechanisms, we can distinguish between two forms of truthfulness: universally truthful and truthful in expectation. A universally truthful mechanism involves a randomization over deterministic truthful mechanisms, with weights that may depend on the input. On the other hand, a mechanism is considered truthful in expectation if truth-telling yields the agent the maximum expected value. Specifically, a mechanism f : R2n R2 is truthful in expectation if for all instances x R2n, every agent i [n], and every deviation x i R2, we have that E[d(xi, f(x))] E[d(xi, f(x i, x i))]. We focus on mechanisms that are both unanimous and anonymous. A mechanism is unanimous if, when all points x are located at the same position (xi = xj for all i, j [n]), it places the facility at that specific location, i.e., f(x) = xi. To ensure bounded robustness, a mechanism must be unanimous, as the optimal cost is zero when the facility is at the same location as all points, whereas placing it elsewhere incurs a positive cost. A mechanism is anonymous if its outcome is independent of the agents identities, meaning it remains unchanged under any permutation of the agents. Finally, we also assume the mechanism is scale-independent, i.e., if we multiply every coordinate of every agent by the same factor, the coordinate of the chosen facility location are scaled in the same way; this captures the fact that it is only the relative distances that really matter in this problem. Learning-augmented algorithms encompass a class of algorithms that enhance their decision-making process by integrating pre-computed predictions or forecasts. These predictions, obtained from various sources, like statistical models, serve as inputs without requiring real-time learning from new data. For instance, in the single facility location problem, one might consider predictions ˆx regarding all of the agent s preferred locations, a prediction F regarding the optimal facility location, or a prediction regarding the identities of the most extreme agents ˆe (the ones that suffer the maximum cost in the optimal solution), which, as demonstrated in this paper, proves to be quite useful. We denote mechanisms enhanced with predictions in general as f(x, ). Specifically, for each prediction setting like ˆx or F , we denote a mechanism enhanced with ˆx and F by f(x, ˆx) and f(x, F ), respectively. We define mechanisms enhanced with other kinds of predictions in a similar way. The effectiveness of learning-augmented mechanisms is evaluated using their consistency and robustness. If x denotes all instances x for which prediction is accurate, a mechanism is α-consistent if it achieves an α-approximation when the prediction is correct, i.e., C(f(x, ), x) β-robust if it maintains a β-approximation regardless of the quality of the prediction, i.e., C(f(x, ), x) 3 Results for the Line In this section, we provide a lower bound regarding the performance of any randomized mechanism for the line, even if it is equipped with the strongest type of prediction, i.e., a prediction ˆx regarding the preferred location of every agent. Theorem 1. No mechanism for the line that is truthful in expectation and guarantees 1+δ consistency for some δ [0, 0.5] can also guarantee robustness better than 2 δ, even if it is provided with full predictions ˆx containing each of the agents locations. The proof consists of two main parts. We first show that, for any instance involving two agents, we can without loss of generality restrict our attention to a class of mechanisms that we call ONLYM mechanisms. We then show the desired lower bound for ONLYM mechanisms. We introduce the following notations specific to this section. Let x L be the leftmost reported location and x R be the rightmost reported location on the line. Therefore, we have x L x R. Let M denote the midpoint of these two extreme points, i.e., M = (x L + x R)/2, which would also correspond to the optimal facility location. For simplicity, we sometimes write f(x1, . . . , xn), dropping the angle brackets . ONLYM is the class of mechanisms that, whenever they choose a location within the interval (x L, x R), then this location is always M. Definition 1 (ONLYM mechanisms). A mechanism f for the line is an ONLYM mechanism if P[f(x) (x L, x R) \ {M}] = 0. The main lemma for the proof of Theorem 1 is the following reduction that allows to restrict our attention to ONLYM mechanisms. The complete proof of Lemma 1 can be found in Appendix A. Lemma 1. For any problem instance involving two agents with reported locations x = x L, x R on the line, and any randomized truthful in expectation mechanism achieving α-consistency and β-robustness over this class of instances, there exists a randomized ONLYM mechanism that is truthful in expectation and achieves the same consistency and robustness guarantees. Proof Sketch. Consider a randomized mechanism f(x L, x R) with probabilities pℓ= P[f(x L, x R) (x L, M)] and pr = P[f(x L, x R) (M, x R)], representing the chances of selecting a location in (x L, M) and (M, x R), respectively. If pℓ= pr = 0, the mechanism already satisfies the desired property. Otherwise, for pℓ> 0 and pr > 0, define the expected locations πℓ= E[f(x L, x R) | f(x L, x R) (x L, M)] and πr = E[f(x L, x R) | f(x L, x R) (M, x R)], which can be expressed as convex combinations πℓ= qℓx L +(1 qℓ)M and πr = qrx R +(1 qr)M for some qℓ, qr (0, 1). We then construct a new mechanism f that modifies f by reassigning probability masses to x L, M, and x R as follows: P[f (x L, x R) = x] = P[f(x L, x R) = x] if x < x L or x > x R, 0 if x (x L, M) (M, x R), P[f(x L, x R) = x] + qℓpℓ if x = x L, P[f(x L, x R) = x] + (1 qℓ)pℓ+ (1 qr)pr if x = M, P[f(x L, x R) = x] + qrpr if x = x R. To show f retains the consistency and robustness guarantees of f, we show that the expected costs of f and f are identical for instances with two agents (Lemma 2). Finally, we show that f maintains the truthfulness in expectation property by verifying that agent costs are unchanged between f and f (Lemma 3). Equipped with Lemma 1, we can now prove Theorem 1. Proof of Theorem 1. We start by proving the result for two-agent instances on the line using any ONLYM mechanism. Then, using Lemma 1, the result follows for any randomized mechanism. First, note that if the chosen facility location y is at distance d(M, y) from the point M = (x L + x R)/2, then its egalitarian social cost is equal to C(o(x), x) + d(M, y). To verify this fact, assume without loss of generality that this location is on the left of M and note that its distance from the agent located at x R is d(x R, M) + d(M, y). As a result, the expected social cost of a mechanism f with agent locations x is C(f, x) = C(o(x), x) + E[d(M, f(x))]. (1) Now, assume that there exists a mechanism that is (1 + δ)-consistent and better than (2 δ)-robust. For robustness, this would imply that for every instance x, irrespective of the prediction, the mechanism must guarantee that C(f, x) C(o(x), x) < 2 δ E[d(M, f(x))] C(o(x), x) < 1 δ. (2) For this mechanism to be truthful in expectation, it must ensure that no agent has an incentive to misreport their location. Consider the instance x = x L, x R , and note that the agent at x L has the option to misreport their location as x L = x L d(x L, x R), which would shift the new midpoint M to x L in the new instance x = x L, x R . This deviation would double the optimal cost, i.e., C(o(x ), x ) = 2 C(o(x), x). Inequality (2), then guarantees that the expected cost for this agent after the deviation would be at most 2(1 δ) C(o(x), x) since we have E[d(M , f(x ))] C(o(x ), x ) = E[d(x L, f(x ))] 2 C(o(x), x) < 1 δ. (3) As a result, to ensure that this agent will not misreport, the mechanism needs to ensure that the expected cost of the agent if they report the truth is strictly less than 2(1 δ) C(o(x), x). If we let P( L) = P[f(x) x L] denote the probability that the chosen location is weakly on the left of x L, and P(M) = P[f(x) = M] denote the probability that the chosen location is M, then the expected cost of the agent located at x L is at least C(o(x), x)P(M) + 2C(o(x), x)(1 P(M) P( L)) because the mechanism is an ONLYM mechanism. This is even if we assume that the only chosen facility locations weakly on the left of x L (and weakly on the right of x R, respectively) are exactly on x L (and exactly on x R, respectively). Therefore, the mechanism needs to always satisfy C(o(x), x)(P(M) + 2(1 P(M) P( L))) < 2(1 δ)C(o(x), x) P(M) + 2(1 P(M) P( L)) < 2(1 δ) P( L) > δ P(M) If we consider the same instance x = x L, x R , and assume that the mechanism is also provided with accurate predictions ˆx L = x L and ˆx R = x R regarding the agent locations, to guarantee 1 + δ consistency, Inequality (1) implies that C(f(x, ˆx = x), x) C(o(x), x) 1 + δ E[d(M, f(x, ˆx))] C(o(x), x) δ. (5) Finally, if we once again consider the same instance with inaccurate predictions ˆx L = x L and ˆx R = x R + d(x L, x R), we observe that in this case the agent located at x R would have the option to instead report x R = x R + d(x L, x R) = ˆx R, and the predictions would then appear to be accurate, forcing the mechanism to satisfy Inequality (5) in order to maintain the required consistency bound. Since the true location x R would now coincide with the middle point M of the misreported instance x = x L, x R , this would yield the agent located at x R who misreported an expected cost of 2δ C(o(x), x) since we have E[d(M , f(x , ˆx))] C(o(x ), x ) = E[d(x R, f(x, ˆx))] 2 C(o(x), x) δ. (6) As a result, to ensure that this agent will not misreport, the mechanism needs to ensure that the expected cost of the agent if they report the truth is at most 2δ C(o(x), x). If we once again let P( L) denote the probability that the chosen location is weakly on the left of x L, and P(M) denote the probability that the chosen location is M, then the expected cost of the agent located at x R is at least C(o(x), x)P(M) + 2C(o(x), x)P( L). This is even if we once again assume that the only chosen facility locations weakly on the left of x L (and weakly on the right of x R, respectively) are exactly on x L (and exactly on x R, respectively). Therefore, the mechanism needs to always satisfy P(M) + 2P( L) 2δ P( L) δ P(M) However, this contradicts Inequality (4), so we conclude that no mechanism can simultaneously guarantee (1 + δ) consistency and a robustness better than (2 δ). 3.1 The hardness result for the line is tight We observe that the lower bound regarding the robustness and consistency trade-off shown in Theorem 1 is actually tight. Specifically, it can be achieved by an appropriate randomization between the optimal deterministic learning-augmented mechanism and the optimal non-learning-augmented randomized mechanism. Proposition 1. For any δ [0, 0.5], there exists a randomized mechanism on the line which, given prediction F , is truthful in expectation, (1 + δ)-consistent, and (2 δ)-robust. As a result, the bound of Theorem 1 precisely captures the optimal robustness consistency trade-off over all truthful in expectation mechanisms for instances on the line. 3.2 Other prediction settings To gain a more complete picture of different prediction settings, we study an alternative strong prediction that is not strictly stronger than the predicted optimal facility location F . Specifically, we consider a setting where predictions are available for all pieces of information except for one of the extreme locations. We show that these predictions are not helpful, even without any robustness constraints, to improve the consistency guarantee alone. Theorem 2. Given a prediction set that provides the identities of all n agents, there exist n 1 agents such that, even if we have their exact locations, there is no deterministic truthful mechanism on the line that is better than 2-consistent, and there is no randomized mechanism on the line that is truthful in expectation and better than 1.5-consistent. 4 Results for the Plane We now consider instances in the Euclidean plane, with a location profile x = x1, . . . , xn , where xi R2 for each agent i. Missing proofs of this section can be found in Appendix B. 4.1 Impossibility Results Before considering the robustness and consistency guarantees achievable by randomized mechanisms augmented with different types of predictions, we start off by reducing the gap on what is known for mechanisms without predictions. Since the problem of designing good facility location mechanisms in two dimensions is harder than the one-dimensional case, it may seem counterintuitive at first, but the lower bound that prior work proved for all one-dimensional instances does not extend to two dimensions. The reason is that the design space for two-dimensional mechanisms is much richer, and the known lower bounds apply only to the more restricted class of one-dimensional mechanisms. For example, consider a simple instance with just two agents in two dimensions. For this two-dimensional instance, the mechanism can return a location for the facility that is not on the line containing the two agents locations, which it, of course, cannot do in one dimension. Note that, in terms of social cost alone, returning such a location is always Pareto-dominated by returning an appropriate location on the line (e.g., its projection onto the line). However, returning a location that is not on this line also allows the designer to affect the incentives of the participating agents in new and non-trivial ways that are impossible in the one-dimensional case. Specifically, returning points that are not on the line allows us to induce previously impossible cost vectors. For a simple example, in a one-dimensional instance involving two agents at distance 1 from each other, the only facility location that yields the same cost to both agents is the midpoint between them, leading to a cost vector of (0.5, 0.5). However, in two dimensions, we can induce a cost vector of (c, c) for any c 0.5 by simply returning the appropriate point on the interval s perpendicular bisector. This additional flexibility could potentially provide the designer with novel ways to ensure that the agents do not misreport their locations; for example, the classic technique of money burning, used to achieve incentive compatibility without monetary payments, heavily depends on the designer s ability to penalize and reward agents based on their reports. Although we conjecture that truthful mechanisms are always better off returning a facility on the line containing the agents locations in this instance, proving such a result appears non-trivial. Due to this additional flexibility in two dimensions, proving inapproximability results for this broader class of mechanisms is a more challenging problem and seems to require new techniques and insights. Theorem 3. Any randomized mechanism that is truthful in expectation in the Euclidean metric space has an approximation ratio of at least 1.118. Our following two results provide impossibility results for both deterministic and randomized mechanisms augmented with perfect predictions. For deterministic mechanisms, Agrawal et al. [2022] showed that there is a mechanism that, given a prediction about the optimal facility location, is 1-consistent and 1 + 2 -robust. Our impossibility result for deterministic mechanisms shows that stronger predictions do not help: even with predictions about the location of each agent, there is no deterministic mechanism that achieves a robustness better than 1 + 2 and a consistency that improves over the best approximation achievable without predictions. Theorem 4. For any δ > 0, there is no deterministic truthful mechanism that is (2 δ)-consistent and 1 + 2 δ -robust, even if it is provided with full predictions ˆx containing each of the agents locations. The next result shows that there is no hope of achieving the best of both worlds in the randomized setting; to obtain 1 consistency, we would need to sacrifice robustness. Theorem 5. For any δ > 0, there is no randomized mechanism that is truthful in expectation, 1-consistent, and (2 δ)-robust, even for two-agent instances and even if it is provided with full predictions ˆx (the location of each agent). This is tight, i.e., there exists a randomized mechanism that is truthful in expectation, 1-consistent, and 2-robust for two-agent instances. 4.2 Positive Results Using Extreme ID Prediction We now turn to positive results, and provide a learning-augmented randomized mechanism that is provided with a new type of prediction: the prediction does not provide us with any actual location, but instead only provides us with the identities ˆe = e1, , ek of the k agents who would suffer the maximum cost in the optimal solution (we refer to them as extreme agents ), i.e., {e1, . . . , ek} = arg maxei [n] d(xei, o(x)). Note that the smallest circle that encloses all the points is the circumcircle of the locations of the extreme agents. Therefore, the center of this circle is o(x), the optimum solution. We propose a mechanism leveraging predictions derived from IDs of extreme agents. The main idea is to return the centroid of the extreme agents, denoted by G with a probability of half, and each of the extreme points with a probability of 1/2k to prevent misreporting incentives. Tang et al. [2020] run this mechanism over all agents (not only the extreme ones) and achieve a 2 1/n approximation. We improve the approximation factor (in case of having good predictions) to Mechanism 1: Centroid Mechanism on Extreme Agents Input :Location profile x = x1, , xn , Predictions ˆe = e1, , ek Output :Probability distribution P on location of the facility With probability 1/2: return the centroid G = xe1+ +xek k With probability 1/2k: return each point xe1, , xek 3 1.67 by running this mechanism only on extreme agents. We use their ideas to maintain truthfulness and we use new techniques to show the approximation guarantees for an arbitrarily number of agents. Moreover, as long as any returned location falls within the minimum enclosing circle of the agents predicted to be extreme (which is the case for Mechanism 1), this ensures an approximation factor of 2 in case of having bad predictions. First, we argue that k 2, meaning that there are at least two extreme agents on the minimum enclosing circle. Otherwise, one can find a smaller circle containing all the points. As a warm-up, we first consider the k = 2 case and propose a randomized mechanism that is truthful in expectation and achieves 1.5 consistency and 2 robustness for any number of agents. Theorem 6. Assume there are only two extreme agents, i.e., k = 2. Then, given predictions ˆe = e1, e2 that provides the IDs of the only two extreme agents, there exists a randomized mechanism that is truthful in expectation and achieves 1.5 consistency and 2 robustness for any number of agents. We then show that it is sufficient to only consider the case where we have exactly three extreme agents on the minimum enclosing circle, meaning k = 3. If we have more than three agents located on the minimum enclosing circle, a continuous perturbation of the points makes the probability of having at least four points lying on a circle infinitesimally small [Berg et al., 2008]. For k = 3, we use the properties of the Euler line to prove the 1.67 consistency. Theorem 7. Given a prediction set that provides the IDs of the extreme agents, there exists a randomized mechanism that is truthful in expectation and achieves 1.67 consistency and 2 robustness for any number of agents. Proof. Let us first consider instances with only three extreme agents. Mechanism 1 is truthful since other agents apart from xe1, xe2, xek cannot influence the result and Mechanism 1 is equivalent to the mechanism of Tang et al. [2020] over agents e1, . . . , ek and since this mechanism is truthful in expectation, agents e1, . . . , ek cannot benefit from misreporting their locations. Since the mechanism returns the reported locations, rather than their predictions, and the centroid is guaranteed to be inside the circumcircle, we can conclude that the mechanism has a robustness factor of 2. The technical aspect of this proof involves establishing the consistency guarantee by considering the Euler line. Euler line. Given three arbitrary points x1, . . . , x3, their circumcircle is the smallest circle that encloses the three points, their centroid is (x1 + x2 + x3)/3, and their orthocenter is the point where the three altitudes (the perpendicular line segments from a vertex to the line that contains the opposite side) intersect. In any triangle, the center of the circumcircle (O), the centroid (G), and the orthocenter (H) are collinear, forming the Euler line. One property of the Euler line is that G is positioned midway between O and H. Additionally, d(O, G) = d(G, H)/2, implying d(O, G) = d(O, H)/3. Let R denote the radius of the circumcircle of xe1, xe2, xe3 and assume without loss of generality that the optimum cost is R = 1. For any j [n], we have d(xj, G) d(xj, O) + d(O, G) R + d(O, G) where the first inequality is by triangle inequality and the second is because all the points are within the circumcircle of the locations of the extreme agents when the predictions are correct. Since H lies within the circumcircle, we infer d(O, H) R, conclusively showing that d(O, G) R/3. Since R is the optimum cost and any agent can reach O by paying that cost, the cost of returning G for any agent is upper bounded by R + d(O, G), which is at most 1.34R. Consequently, the approximation of Mechanism 1 in case of having accurate predictions is less or equal than Altitudes Medians Perpendicular Bisectors Figure 1: G = (xe1 + xe2 + xe3)/3 is the centroid of xe1, . . . , xe3, which is the intersection of the three medians. H is the orthocenter, which is the intersection of the three altitudes, and O is the center of the circumcircle, which is the intersection of the three perpendicular bisectors. In any triangle, the circumcenter (O), the centroid (G), and the orthocenter (H) are collinear, forming the Euler line. Moreover, d(O, G) = 1 i=1 max j [n] d(xei, xj) + 1 2 1 R max j [n] d(xj, G) 1 + 1 2 R + d(O, G) where the first inequality is since d(xei, xj) d(xei, O) + d(O, xj) 2R when the predictions are correct. Next, we show how we can modify the mechanism to achieve the same approximation guarantees as having three extreme agents while maintaining truthfulness. As mentioned previously, the key concept involves perturbing the instance before requesting predictions to have at most three extreme agents. Given any instance x = x1, , xn , define a perturbed instance x = x1, , xn , which is the result of a continuous perturbation of x. Consider the set of predictions for the IDs of extreme agents in x. Although the extreme agents of the perturbed instance x might differ from those of the original instance x, their costs remain very close to the maximum cost in case of having good predictions. Therefore, the circumcircle of them is a good representation of the minimum enclosing circle and results in the same approximation guarantees. Since we run the mechanism on the main instance x, truthfulness holds as before. Note that since we ask for the ID of extreme agents of the perturbed instance x, we have consistency guarantees if predictions are accurate based on the perturbed instance x, and in any case we can ensure the robustness factor of 2. 5 Future Directions The problem of designing randomized facility location mechanisms in the Euclidean space is very natural and several open questions remain, even without predictions. While numerous studies have addressed this problem in restricted spaces, there remains a gap regarding between the best possible approximation guarantee in (1.118, 2 1/n). The possibility of a truthful in expectation mechanism achieving better than a 2 1/n approximation is intriguing. Obtaining a stronger lower bound than 1.118 would also be very interesting. Augmenting these mechanisms with predictions introduces a new dimension to the problem. While we have explored various prediction settings and provided both positive and negative results, many of the current findings in Euclidean space lack tightness. Finding tight results for different types of predictions would enable meaningful comparisons and help identify which prediction strategy is most effective in different scenarios. Acknowledgments Eric Balkanski and Vasilis Gkatzelis were partially supported by NSF grant CCF-2210502. Eric Balkanski was also supported by NSF grant IIS-2147361. Vasilis Gkatzelis was also supported by NSF CAREER award CCF-2047907. Priyank Agrawal, Eric Balkanski, Vasilis Gkatzelis, Tingting Ou, and Xizhi Tan. Learning-augmented mechanism design: Leveraging predictions for facility location. In EC 22: The 23rd ACM Conference on Economics and Computation. ACM, 2022. doi: 10.1145/3490486.3538306. Matteo Almanza, Flavio Chierichetti, Silvio Lattanzi, Alessandro Panconesi, and Giuseppe Re. Online facility location with multiple advice. In Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, Neur IPS 2021, pages 4661 4673, 2021. Noga Alon, Michal Feldman, Ariel D Procaccia, and Moshe Tennenholtz. Strategyproof approximation of the minimax on networks. Mathematics of Operations Research, 35(3):513 526, 2010. Maria-Florina Balcan, Siddharth Prasad, and Tuomas Sandholm. Bicriteria multidimensional mechanism design with side information. In Proceedings of the 37th International Conference on Neural Information Processing Systems, NIPS 23, Red Hook, NY, USA, 2024. Curran Associates Inc. Eric Balkanski, Vasilis Gkatzelis, and Xizhi Tan. Strategyproof Scheduling with Predictions. In 14th Innovations in Theoretical Computer Science Conference (ITCS 2023), 2023a. doi: 10.4230/ LIPIcs.ITCS.2023.11. Eric Balkanski, Vasilis Gkatzelis, Xizhi Tan, and Cherlin Zhu. Online mechanism design with predictions. ar Xiv preprint:2310.02879 (to appear in the 25th ACM Conference on Economics and Computation (EC 24)), 2023b. Zohar Barak, Anupam Gupta, and Inbal Talgam-Cohen. Mac advice for facility location mechanism design. ar Xiv preprint:2403.12181 (to appear in 38th Conference on Neural Information Processing Systems (Neur IPS)), 2024. Salvador Barberà, Faruk Gul, and Stacchetti Ennio. Generalized median voter schemes and committees. Journal of Economic Theory, 61(2):262 289, 1993. Mark Berg, Otfried Cheong, Marc van Kreveld, and Mark Overmars. Computational Geometry: Algorithms and Applications. Springer, Berlin, Germany, 3rd edition, 2008. ISBN 978-3-54077973-5. Kim C. Border and James Jordan. Straightforward elections, unanimity and phantom voters. Review of Economic Studies, 50(1):153 170, 1983. Ioannis Caragiannis and Georgios Kalantzis. Randomized learning-augmented auctions with revenue guarantees. In Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence, IJCAI-24, 2024. doi: 10.24963/ijcai.2024/297. Hau Chan, Aris Filos-Ratsikas, Bo Li, Minming Li, and Chenhao Wang. Mechanism design for facility location problems: A survey. In Zhi-Hua Zhou, editor, Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, IJCAI-21, pages 4356 4365. International Joint Conferences on Artificial Intelligence Organization, 2021. Qingyun Chen, Nick Gravin, and Sungjin Im. Strategic facility location via predictions. ar Xiv preprint:2410.07497 (to appear in the 20th International Conference of Web and Internet Economics, WINE), 2024. George Christodoulou, Alkmini Sgouritsa, and Ioannis Vlachos. Mechanism design augmented with output advice. ar Xiv preprint:2406.14165 (to appear in 38th Conference on Neural Information Processing Systems (Neur IPS)), 2024. Riccardo Colini-Baldeschi, Sophie Klumper, Guido Schäfer, and Artem Tsikiridis. To trust or not to trust: Assignment mechanisms with predictions in the private graph model. ar Xiv preprint:2403.03725 (to appear in the 25th ACM Conference on Economics and Computation(EC 24)), 2024. El-Mahdi El-Mhamdi, Sadegh Farhadkhani, Rachid Guerraoui, and Lê-Nguyên Hoang. On the strategyproofness of the geometric median. In Proceedings of The 26th International Conference on Artificial Intelligence and Statistics, pages 2603 2640, 2023. Jiazhu Fang, Qizhi Fang, Wenjing Liu, and Qingqin Nong. Mechanism design with predictions for facility location games with candidate locations. In Annual Conference on Theory and Applications of Models of Computation, pages 38 49. Springer, 2024. Michal Feldman and Yoav Wilf. Strategyproof facility location and the least squares objective. In Proceedings of the fourteenth ACM conference on Electronic commerce, pages 873 890, 2013. Dimitris Fotakis, Evangelia Gergatsouli, Themis Gouleakis, and Nikolas Patris. Learning augmented online facility location. Co RR, abs/2107.08277, 2021. Vasilis Gkatzelis, Kostas Kollias, Alkmini Sgouritsa, and Xizhi Tan. Improved price of anarchy via predictions. In Proceedings of the 23rd ACM Conference on Economics and Computation, EC 22, page 529 557, New York, NY, USA, 2022. Association for Computing Machinery. ISBN 9781450391504. doi: 10.1145/3490486.3538296. Vasilis Gkatzelis, Daniel Schoepflin, and Xizhi Tan. Clock auctions augmented with unreliable advice. ar Xiv preprint:2408.06483 (to appear in the 2025 ACM-SIAM Symposium on Discrete Algorithms, SODA), 2024. Sumit Goel and Wade Hann-Caruthers. Optimality of the coordinate-wise median mechanism for strategyproof facility location in two dimensions. Social Choice and Welfare, 2020. Gabriel Istrate and Cosmin Bonchis. Mechanism design with predictions for obnoxious facility location. ar Xiv preprint ar Xiv:2212.09521, 2022. Gabriel Istrate, Cosmin Bonchi s, and Victor Bogdan. Equilibria in multiagent online problems with predictions, 2024. Shaofeng H.-C. Jiang, Erzhi Liu, You Lyu, Zhihao Gavin Tang, and Yubo Zhang. Online facility location with predictions. In The Tenth International Conference on Learning Representations, ICLR 2022. Open Review.net, 2022. Alexander Lindermayr and Nicole Megow. Alps. https://algorithms-with-predictions. github.io/. Pinyan Lu, Zongqi Wan, and Jialin Zhang. Competitive auctions with imperfect predictions. ar Xiv preprint:2309.15414 (to appear in the 25th ACM Conference on Economics and Computation(EC 24)), 2023. Thodoris Lykouris and Sergei Vassilvitskii. Competitive caching with machine learned advice. J. ACM, 68(4):24:1 24:25, 2021. Reshef Meir. Strategyproof facility location for three agents on a circle. In International Symposium on Algorithmic Game Theory, pages 18 33. Springer, 2019. Adam Meyerson. Online facility location. In 42nd Annual Symposium on Foundations of Computer Science, FOCS 2001, pages 426 431. IEEE Computer Society, 2001. doi: 10.1109/SFCS.2001. 959917. Michael Mitzenmacher and Sergei Vassilvitskii. Algorithms with predictions. Commun. ACM, 65(7): 33 35, jun 2022. ISSN 0001-0782. doi: 10.1145/3528087. Hervé Moulin. On strategy-proofness and single peakedness. Public Choice, 35(4):437 455, 1980. Hans Peters, Hans van der Stel, and Ton Storcken. Range convexity, continuity, and strategy-proofness of voting schemes. Zeitschrift für Operations Research, 38:213 229, 1993. Ariel D. Procaccia and Moshe Tennenholtz. Approximate mechanism design without money. ACM Trans. Economics and Comput., 1(4):18:1 18:26, 2013. doi: 10.1145/2542174.2542175. Tim Roughgarden. Beyond the Worst-Case Analysis of Algorithms. Cambridge University Press, 2021. Pingzhong Tang, Dingli Yu, and Shengyu Zhao. Characterization of group-strategyproof mechanisms for facility location in strictly convex space. In EC 20: The 21st ACM Conference on Economics and Computation, pages 133 157. ACM, 2020. doi: 10.1145/3391403.3399471. Toby Walsh. Strategy proof mechanisms for facility location in Euclidean and Manhattan space. ar Xiv preprint ar Xiv:2009.07983, 2020. Chenyang Xu and Pinyan Lu. Mechanism design with predictions. In Lud De Raedt, editor, IJCAI 22: The 31st International Joint Conference on Artificial Intelligence,, pages 571 577, 2022. Agents reported locations Non-zero probability outcomes Arbitrary Mechanism Figure 2: The original mechanism f(.) and the new ONLYM mechanism f (.) in the proof of Lemma 1, where pℓ= P[f(x L, x R) (x L, M)] and pr = P[f(x L, x R) (M, x R)], πℓ= E[f(x L, x R) | f(x L, x R) (x L, M)] and πr = E[f(x L, x R) | f(x L, x R) (M, x R)], and qℓand qr are such that πℓ= qℓx L + (1 qℓ)M and πr = qrx R + (1 qr)M. A Missing Proof of Section 3 Lemma 1. For any problem instance involving two agents with reported locations x = x L, x R on the line, and any randomized truthful in expectation mechanism achieving α-consistency and β-robustness over this class of instances, there exists a randomized ONLYM mechanism that is truthful in expectation and achieves the same consistency and robustness guarantees. Proof. Consider any randomized mechanism f(x L, x R) and let pℓ= P[f(x L, x R) (x L, M)] and pr = P[f(x L, x R) (M, x R)] represent the probabilities that this mechanism chooses a facility location in (x L, M) and (M, x R), respectively. If pℓ= pr = 0, the mechanism already satisfies the desired property, and the proof is complete. Otherwise, if pℓ> 0, define πℓ= E[f(x L, x R) | f(x L, x R) (x L, M)] and, if pr > 0, define πr = E[f(x L, x R) | f(x L, x R) (M, x R)] as the expected locations returned by the mechanism when restricted to (x L, M) and (M, x R), respectively. Since πℓlies in (x L, M) and πr lies in (M, x R), we can express these two points as convex combinations of x L, M, and x R: πℓ= qℓx L + (1 qℓ)M and πr = qrx R + (1 qr)M for some qℓ, qr (0, 1). We then modify the original mechanism f to a new ONLYM mechanism f defined as P[f (x L, x R) = x] = P[f(x L, x R) = x] if x < x L or x > x R 0 if x (x L, M) (M, x R) P[f(x L, x R) = x] + qℓpℓ if x = x L P[f(x L, x R) = x] + (1 qℓ)pℓ+ (1 qr)pr if x = M P[f(x L, x R) = x] + qrpr if x = x R The construction of f is illustrated in Figure 2. To show that mechanism f achieves the same consistency and robustness guarantees as mechanism f, we show that their expected costs are equal on all instances with two agents. Lemma 2. For all mechanisms f for the line and instances x = x L, x R with two agents, the expected costs of f and f over x are equal, i.e., C(f, x) = C(f , x). Proof. First, note that E max xi {x L,x R} d(xi, f(x)) | f(x) (x L, M) P[f(x) (x L, M)] = E [d(x R, f(x)) | f(x) (x L, M)] pℓ = d(x R, πℓ) pℓ. Similarly, we have E max xi {x} d(xi, f(x)) | f(x) (M, x R) P[f(x) (M, x R)] = d(πr, x L) pr, which implies that C(f , x) C(f, x) = E max xi {x L,x R} d(xi, f (x)) E max xi {x L,x R} d(xi, f(x)) = d(x R, x L) (P[f (x) {x L, x R}] P[f(x) {x L, x R}]) + d(x R, M) (P[f (x) = M] P[f(x) = M]) d(x R, πℓ) pℓ d(x L, πr) pr = d(x R, x L) (qℓpℓ+ qrpr) + d(x R, M) ((1 qℓ)pℓ+ (1 qr)pr) d(x R, πℓ) pℓ d(x L, πr) pr = d(x R, x L) (qℓpℓ+ qrpr) + d(x R, M) ((1 qℓ)pℓ+ (1 qr)pr) d(x R, x L) qℓpℓ d(x R, M) (1 qℓ)pℓ d(x L, x R) qrpr d(x L, M) (1 qr)pr = 0. Next, to show that mechanism f is also truthful in expectation, we first show that the costs of the agents do not change between f and f . Lemma 3. For all mechanisms f for the line and instances x = x L, x R with two agents, the cost of the agent at location x L is identical under both f and f , i.e., E [d(x L, f(x))] = E [d(x L, f (x))] . Similarly, we have E [d(x R, f(x))] = E [d(x R, f (x))] . Proof. By definition of f , we have E [d(x L, f(x))] E [d(x L, f (x))] = d (x L, E [f(x) | f(x) (x L, M)] = πℓ) P [f(x) (x L, M)] + d (x L, E [f(x) | f(x) (M, x R)] = πr) P [f(x) (M, x R)] + d (x L, M) (P [f(x) = M] P [f (x) = M]) + d (x L, x R) (P [f(x) = x R] P [f (x) = x R]) = d(x L, qℓx L + (1 qℓ)M) pℓ+ d(x L, qrx R + (1 qr)M) pr d(x L, M)((1 qℓ)pℓ+ (1 qr)pr) d(x L, x R)(qrpr) = d(x L, M)(1 qℓ)pℓ+ d(x L, x R)qrpr + d(x L, M)(1 qr)pr d(x L, M)(1 qℓ)pℓ d(x L, M)(1 qr)pr d(x L, x R)qrpr = 0. Next, we use the previous lemma to show that mechanism f preserves the truthful in expectation guarantee of f. Lemma 4. If a mechanism f for the line is truthful in expectation over instances with two agents, then f is also truthful in expectation over instances with two agents. Proof. Assume that f is truthful in expectation over instances with two agents. Assume that one of the agents deviates and reports a false location in mechanism f . Due to symmetry, and without loss of generality, assume that the agent located at x L deviates and reports a false location x L. We need to show E [d(x L, f (x L, x R))] E [d(x L, f (x L, x R))] . Since the original mechanism f(x L, x R) is truthful in expectation, we have E [d(x L, f(x L, x R))] E [d(x L, f(x L, x R))] . (7) Combining these two inequalities with Lemma 3, it suffices to show E [d(x L, f (x L, x R))] E [d(x L, f(x L, x R))] . To prove the above inequality, we consider two main cases. First, we focus on the scenario where the agent located at x L deviates to the right, i.e., x L < x L. Let M = x L+x R 2 . Assume x L x R, define p ℓ= P[f(x L, x R) (x L, M )] and p r = P[f(x L, x R) (M , x R)]. If p ℓ= p r = 0, then f (x L, x R) = f(x L, x R), and hence E [d(x L, f (x L, x R))] = E [d(x L, f(x L, x R))]. Therefore, the proof is complete in this case. Next, consider the distribution where the mechanism f(x L, x R) returns a random facility location within the intervals (x L, M ) if p ℓ> 0, or within (M , x R) if p r > 0. The expected locations within these intervals are π ℓ (x L, M ) and π r (M , x R), which can be expressed as π ℓ= q ℓx L + (1 q ℓ)M and π r = q rx R + (1 q r)M , where q ℓand q r are the respective convex coefficients. In the first case, where x L < x L, we show that the cost to the left agent is the same across the two mechanisms. The key idea is that the difference between the two mechanisms, in terms of their returned locations, lies within the intervals (x L, M ) and (M , x R), both on the right side of x L. Thus, the agent s cost is fully determined by the expected locations in these intervals. Specifically we have E [d(x L, f(x L, x R))] E [d(x L, f (x L, x R))] = d (x L, E [f(x L, x R) | f(x L, x R) (x L, M )] = π ℓ) P [f(x L, x R) (x L, M )] + d (x L, E [f(x L, x R) | f(x L, x R) (M , x R)] = π r) P [f(x L, x R) (M , x R)] + d (x L, x L) (P [f(x L, x R) = x L] P [f (x L, x R) = x L]) + d (x L, M ) (P [f(x L, x R) = M ] P [f (x L, x R) = M ]) + d (x L, x R) (P [f(x L, x R) = x R] P [f (x L, x R) = x R]) = d(x L, q ℓx L + (1 q ℓ)M ) p ℓ+ d(x L, q rx R + (1 q r)M ) p r d(x L, x L)q ℓp ℓ d(x L, M )((1 q ℓ)p ℓ+ (1 q r)p r) d(x L, x R)q rp r = d(x L, x L)q ℓp ℓ+ d(x L, M )(1 q ℓ)p ℓ+ d(x L, x R)q rp r + d(x L, M )(1 q r)p r d(x L, x L)q ℓp ℓ d(x L, M )((1 q ℓ)p ℓ+ (1 q r)p r) d(x L, x R)q rp r = 0. Note that in the case of x R < x L, the same arguments hold with a minor change in notation. Define p ℓ= P[f(x L, x R) (M , x L)] and p r = P[f(x L, x R) (x R, M )]. Consider the expected locations π ℓand π r within the intervals (M , x L) and (x R, M ), respectively. Focusing our attention on the more interesting case where x L < x L, we aim to show that E [d(x L, f (x L, x R))] E [d(x L, f(x L, x R))]. If we focus on the interval (x L, M ) and assume x L M , the expected location of the facility returned by the two mechanisms, f(x L, x R) and f (x L, x R), conditioned on the facility being in the interval (x L, M ), is the same; this is true by construction (our reduction maintains the expected location π ℓbetween the reported location and the midpoint). The crucial difference between the two mechanisms is that although f(x L, x R) may return any facility location in the (x L, M ) interval, the mechanism f (x L, x R) returns the same expected location using only x L and M in its support. We show that the cost to an agent located at any point x L in the (x L, M ) interval is weakly higher in f (x L, x R) than it is in f(x L, x R), thus maintaining the truthfulness guarantee. Intuitively, this is true because the two mechanisms return the same expected location π ℓin that interval, but f (x L, x R) only returns the extreme points of the interval, which hurt the agent at x L the most. More formally, we replace any probability assigned to a point in (x L, M ) with a convex combination between x L and M , and each time we do this, we weakly increase the cost to an agent who is located at any point in (x L, M ). Specifically, for any outcome y (x L, M ) of the mechanism f(x L, x R) and any x L (x L, M ), if we have y = w x L + (1 w) M then d(x L, y) < w d(x L, x L) + (1 w) d(x L, M ). Since π ℓ= q ℓx L + (1 q ℓ)M , at the end of this process, we will end up with q ℓp ℓprobability increase on x L, and (1 q ℓ)p ℓprobability increase on M . Therefore, we have E [d(x L, f(x L, x R))] < d(x L, x L)q ℓp ℓ+ d(x L, M ) ((1 q ℓ)p ℓ+ (1 q r)p r) + d(x L, x R)q rp r = E [d(x L, f (x L, x R))] . In the case where x L > M , similar arguments apply in the interval (M , x R). The proof of Lemma 1 then immediately follows from Lemma 2 and Lemma 4. Proposition 1. For any δ [0, 0.5], there exists a randomized mechanism on the line which, given prediction F , is truthful in expectation, (1 + δ)-consistent, and (2 δ)-robust. Proof. We show that the bound of the robustness consistency trade-off in Theorem 1 is tight. To verify the tightness for δ = 0.5, note that it is possible to achieve a robustness of 1.5 (and therefore also a consistency of 1.5) using the following randomized truthful in expectation mechanism of Procaccia and Tennenholtz [2013]: Mechanism 1 (LRM). Given an input x, the mechanism returns x L with probability 1/4, x R with probability 1/4, and M with probability 1/2. On the other extreme, to verify the tightness of the theorem for δ = 0, note that it is possible to achieve a robustness of 2 with a consistency of 1 using the following truthful deterministic learning-augmented mechanism of Agrawal et al. [2022]: Mechanism 2 (Min Max P). Given an input x and a prediction F regarding the optimal facility location, the mechanism returns F if F [x L, x R], x L if F < x L, and x R if F > x R. In fact, for any δ [0, 0.5], we can achieve a robustness of 2 δ and a consistency of 1 + δ by randomly choosing which one of these two mechanisms to run. Specifically, we can achieve these bounds by running the LRM mechanism with probability 2δ and the Min Max P mechanism with probability 1 2δ. Note that since the decision regarding which one of the two truthful mechanisms to run is independent of the agents reports, the resulting randomized mechanism is truthful as well. Theorem 2. Given a prediction set that provides the identities of all n agents, there exist n 1 agents such that, even if we have their exact locations, there is no deterministic truthful mechanism on the line that is better than 2-consistent, and there is no randomized mechanism on the line that is truthful in expectation and better than 1.5-consistent. Proof. We first address the deterministic case with two agents whose locations are given by x = x L, x R . Assume we have a prediction ˆx L, which predicts the location of the leftmost agent. We will show that even if ˆx L is accurate, there is no truthful mechanism with a consistency better than 2. We use the same argument as in Theorem 3.2 of Procaccia and Tennenholtz [2013]. Assume, for contradiction, that f : R2 R is a truthful mechanism with a consistency less than 2. Consider the location profile x = x L = 0, x R = 1 and the prediction ˆx L = 0. To achieve consistency better than 2, the mechanism needs to choose a facility location y = f(x) such that y (0, 1). Now consider an alternative location profile x = x L = 0, x R = y with the prediction ˆx L = 0. The optimal facility location for this profile is at y/2, yielding a maximum cost of y/2. To maintain consistency better than 2, the mechanism should place the facility within the interval (0, y). However, if this were to occur, the rightmost agent would have an incentive to report a location of 1 instead of y, as this lie would place the facility exactly at y, rather than within (0, y). This contradicts the truthfulness of the mechanism. This argument extends to cases with arbitrary n by situating all other agents at 0 in both profiles x and x , with accurate predictions for them at 0. Similar reasoning holds true. For the randomized setting, we use the idea from the proof of Theorem 3.4 in Procaccia and Tennenholtz [2013]. We first focus on the case with two agents and then extend the result to any Figure 3: Instances x = x1 = (0, 0), x2 = = xn 1 = (1/2, 0), xn = (1, 0) , and x = x 1 = (0, 0), x 2 = = x n 1 = (1/2, 0), x n = (2, 0) in the proof of Theorem 2. It is assumed that d(xn, f(x)) 1/2 and x arg miny:d(xn,y) 1/2 C(y, x) number of agents. Let f be a randomized truthful in expectation mechanism. Consider the location profile x = x L = 0, x R = 1 . We have f(x) = P, where P is a probability distribution over R. There exists xi x such that E[d(xi, P)] 1/2. If this agent s location prediction is unavailable, we can prove that consistency better than 1.5 cannot be achieved. Assume E[d(x R, P)] 1/2 and that we have an accurate prediction ˆx L = 0 for the leftmost agent x L. Consider the profile x = x L = 0, x R = 2 , with an accurate prediction ˆx L = 0. For truthfulness, the expected distance from the location 1 should still be 1/2, otherwise the rightmost agent would lie in profile x. We know that if E[d(M, P)] = , then the expected maximum cost is + 1. With 1/2, the expected cost is at least 1.5, whereas the optimal cost is 1, resulting in a consistency of at least 1.5. This argument extends to arbitrary n by situating all other agents at 1/2 in both profiles x and x , with accurate predictions for them at 1/2. Similar reasoning holds true. B Missing Proofs of Section 4 Theorem 3. Any randomized mechanism that is truthful in expectation in the Euclidean metric space has an approximation ratio of at least 1.118. Proof. Let f(x) = P be a randomized truthful in expectation mechanism, where P is a probability distribution over R2. Consider the location profile x = x1 = (0, 0), x2 = = xn 1 = (1/2, 0), xn = (1, 0) . There exists an xi (either x1 or xn) such that d(xi, f(x)) = Ey Pd(xi, y) 1/2. Without loss of generality, assume that d(xn, f(x)) 1/2. Now consider the location profile x = x 1 = (0, 0), x 2 = = x n 1 = (1/2, 0), x n = (2, 0) . To maintain truthfulness, we must have d(xn = (1, 0), f(x )) 1/2, preventing the agent at location xn = (1, 0) in the profile x from having an incentive to lie about being at location x n = (2, 0). Extending the result of Theorem 3.4 in Procaccia and Tennenholtz [2013] from the line to the Euclidean metric, if we have d(o(x), f(x)) , then the expected maximum cost is at least q 2 + C(o(x), x)2. Therefore, the expected cost of x is at least q 2 2, resulting in a 1.25 1.118 approximation, as the optimum cost is d(x 1,x n) 2 = 1. Theorem 4. For any δ > 0, there is no deterministic truthful mechanism that is (2 δ)-consistent and 1 + 2 δ -robust, even if it is provided with full predictions ˆx containing each of the agents locations. Proof. Using the characterization provided by Peters et al. [1993], we know that any deterministic, truthful, anonymous, and unanimous mechanism can be expressed as a Generalized Coordinatewise Median (GCM) mechanism with n 1 constant points in P. The GCM mechanism takes as input the reported locations x of the n agents and a multiset P of fixed points (known as phantom points), which are constant and independent of the agents reported locations. It then outputs the coordinatewise median of the combined set x P. Consider a five-agent instance with x containing four agents at (0, 20) and one agent at (0, 10). Our proof first shows that to achieve a consistency better than 2, this mechanism needs to introduce four constant points all with y-coordinates greater than 10. Otherwise, even if the predictions are correct, the y-coordinate of GCM(x, P) will be at most 10, leading to consistency greater than or equal to 2, since the optimal point would be (0, 15). Then, we show that if all these four constant points have y-coordinates greater than 10, the robustness of the GCM mechanism is at least 1 + 2. Let x be the median of the x-coordinates of the four constant points, and now consider the following instance, where the optimal cost is 1. Consider a five-agent instance with x containing three agents at ( x, 0), one agent at ( x 1, 1), and one agent at x 1 1 Since x is the median of the x-coordinates of both true agent locations x and constant points P, x will be the x-coordinate of the GCM(x, P). Since all the constant points have y-coordinates greater than 10, the y-coordinate of the GCM(x, P) will be 1. Since ( x, 1) has a distance of 1 + 2 from x 1 1 while ( x 1, 0), the optimal point, has a distance of 1 from all the points, we can conclude that the robustness factor is at least 1 + Theorem 5. For any δ > 0, there is no randomized mechanism that is truthful in expectation, 1-consistent, and (2 δ)-robust, even for two-agent instances and even if it is provided with full predictions ˆx (the location of each agent). This is tight, i.e., there exists a randomized mechanism that is truthful in expectation, 1-consistent, and 2-robust for two-agent instances. Proof. Consider the instance x = x L = 0, x R = 2 , and assume we have predictions ˆx = ˆx L = 0, ˆx R = 2 . To achieve 1-consistency, a mechanism should return M = 1 with probability 1. Now consider another instance x = x L = 0, x R = 1 with incorrect predictions ˆx = ˆx L = 0, ˆx R = 2 . To ensure that the rightmost agent will not misreport to x R = 2 and have a cost of zero, the mechanism needs to ensure that the expected cost of the agent if they report the truth is at most zero, which means returning x R with probability 1. This results in a robustness factor of 2. Agrawal et al. [2022] proposed a deterministic Minimum Bounding Box mechanism that is 1consistent and 1 + 2-robust. We argue that this mechanism is 1-consistent and 2-robust for two-agent instances. Mechanism 3 (Minimum Bounding Box). Given n points x = (x1, y1), , (xn, yn) , and a prediction F = (x F , y F ), return (Min Max P( x1, , xn , x F ), Min Max P( y1, , yn , y F )). Observation 1. Mechanism 3 is 1-consistent and 2-robust for instances with only two agents. Proof. One of the two diagonals of the minimum bounding box has two agents on its vertices. Therefore, any point inside or on this box has the approximation factor of 2. Therefore, the upper and lower bounds match for instances with two agents, but it remains unclear where the exact value is even for three agents. Theorem 6. Assume there are only two extreme agents, i.e., k = 2. Then, given predictions ˆe = e1, e2 that provides the IDs of the only two extreme agents, there exists a randomized mechanism that is truthful in expectation and achieves 1.5 consistency and 2 robustness for any number of agents. Proof. If we only have two agents xe1 and xe2 located on the minimum enclosing circle, xe1xe2 represent a diameter of this circle; otherwise, we can find a smaller circle containing all the points. Therefore, the middle point of these two locations, (xe1 + xe2)/2, is the optimum location of the facility. Mechanism 1 runs LRM mechanism on the diameter of the minimum enclosing circle, meaning that it returns the optimum location with a probability of 1/2 and each of the reported locations of the extreme agents with a probability of 1/4, resulting in a 1.5 consistency. In terms of truthfulness, since no other agent besides xe1 and xe2 impacts the mechanism, it suffices to show that they do not have an incentive to lie. Agents xe1 and xe2 lack such incentive for reasons similar to those in LRM Mechanism. Additionally, as any returned location falls within the minimum enclosing circle, it ensures a robustness factor of 2. Neur IPS Paper Checklist Question: Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? Answer: [Yes] Justification: We describe all our results in the introduction section, specifically in the Our Results subsection. Guidelines: The answer NA means that the abstract and introduction do not include the claims made in the paper. The abstract and/or introduction should clearly state the claims made, including the contributions made in the paper and important assumptions and limitations. A No or NA answer to this question will not be perceived well by the reviewers. The claims made should match theoretical and experimental results, and reflect how much the results can be expected to generalize to other settings. It is fine to include aspirational goals as motivation as long as it is clear that these goals are not attained by the paper. 2. Limitations Question: Does the paper discuss the limitations of the work performed by the authors? Answer: [Yes] Justification: We discuss the limitations of our results in the last section, the Future Directions. Guidelines: The answer NA means that the paper has no limitation while the answer No means that the paper has limitations, but those are not discussed in the paper. The authors are encouraged to create a separate "Limitations" section in their paper. The paper should point out any strong assumptions and how robust the results are to violations of these assumptions (e.g., independence assumptions, noiseless settings, model well-specification, asymptotic approximations only holding locally). The authors should reflect on how these assumptions might be violated in practice and what the implications would be. The authors should reflect on the scope of the claims made, e.g., if the approach was only tested on a few datasets or with a few runs. In general, empirical results often depend on implicit assumptions, which should be articulated. The authors should reflect on the factors that influence the performance of the approach. For example, a facial recognition algorithm may perform poorly when image resolution is low or images are taken in low lighting. Or a speech-to-text system might not be used reliably to provide closed captions for online lectures because it fails to handle technical jargon. The authors should discuss the computational efficiency of the proposed algorithms and how they scale with dataset size. If applicable, the authors should discuss possible limitations of their approach to address problems of privacy and fairness. While the authors might fear that complete honesty about limitations might be used by reviewers as grounds for rejection, a worse outcome might be that reviewers discover limitations that aren t acknowledged in the paper. The authors should use their best judgment and recognize that individual actions in favor of transparency play an important role in developing norms that preserve the integrity of the community. Reviewers will be specifically instructed to not penalize honesty concerning limitations. 3. Theory Assumptions and Proofs Question: For each theoretical result, does the paper provide the full set of assumptions and a complete (and correct) proof? Answer: [Yes] Justification: Due to the page limit, we have moved some of our proofs to the appendix. Guidelines: The answer NA means that the paper does not include theoretical results. All the theorems, formulas, and proofs in the paper should be numbered and crossreferenced. All assumptions should be clearly stated or referenced in the statement of any theorems. The proofs can either appear in the main paper or the supplemental material, but if they appear in the supplemental material, the authors are encouraged to provide a short proof sketch to provide intuition. Inversely, any informal proof provided in the core of the paper should be complemented by formal proofs provided in appendix or supplemental material. Theorems and Lemmas that the proof relies upon should be properly referenced. 4. Experimental Result Reproducibility Question: Does the paper fully disclose all the information needed to reproduce the main experimental results of the paper to the extent that it affects the main claims and/or conclusions of the paper (regardless of whether the code and data are provided or not)? Answer: [NA] Justification: The contributions of our paper are mainly theoretical. Guidelines: The answer NA means that the paper does not include experiments. If the paper includes experiments, a No answer to this question will not be perceived well by the reviewers: Making the paper reproducible is important, regardless of whether the code and data are provided or not. If the contribution is a dataset and/or model, the authors should describe the steps taken to make their results reproducible or verifiable. Depending on the contribution, reproducibility can be accomplished in various ways. For example, if the contribution is a novel architecture, describing the architecture fully might suffice, or if the contribution is a specific model and empirical evaluation, it may be necessary to either make it possible for others to replicate the model with the same dataset, or provide access to the model. In general. releasing code and data is often one good way to accomplish this, but reproducibility can also be provided via detailed instructions for how to replicate the results, access to a hosted model (e.g., in the case of a large language model), releasing of a model checkpoint, or other means that are appropriate to the research performed. While Neur IPS does not require releasing code, the conference does require all submissions to provide some reasonable avenue for reproducibility, which may depend on the nature of the contribution. For example (a) If the contribution is primarily a new algorithm, the paper should make it clear how to reproduce that algorithm. (b) If the contribution is primarily a new model architecture, the paper should describe the architecture clearly and fully. (c) If the contribution is a new model (e.g., a large language model), then there should either be a way to access this model for reproducing the results or a way to reproduce the model (e.g., with an open-source dataset or instructions for how to construct the dataset). (d) We recognize that reproducibility may be tricky in some cases, in which case authors are welcome to describe the particular way they provide for reproducibility. In the case of closed-source models, it may be that access to the model is limited in some way (e.g., to registered users), but it should be possible for other researchers to have some path to reproducing or verifying the results. 5. Open access to data and code Question: Does the paper provide open access to the data and code, with sufficient instructions to faithfully reproduce the main experimental results, as described in supplemental material? Answer: [NA] Justification: The contributions of our paper are mainly theoretical. Guidelines: The answer NA means that paper does not include experiments requiring code. Please see the Neur IPS code and data submission guidelines (https://nips.cc/ public/guides/Code Submission Policy) for more details. While we encourage the release of code and data, we understand that this might not be possible, so No is an acceptable answer. Papers cannot be rejected simply for not including code, unless this is central to the contribution (e.g., for a new open-source benchmark). The instructions should contain the exact command and environment needed to run to reproduce the results. See the Neur IPS code and data submission guidelines (https: //nips.cc/public/guides/Code Submission Policy) for more details. The authors should provide instructions on data access and preparation, including how to access the raw data, preprocessed data, intermediate data, and generated data, etc. The authors should provide scripts to reproduce all experimental results for the new proposed method and baselines. If only a subset of experiments are reproducible, they should state which ones are omitted from the script and why. At submission time, to preserve anonymity, the authors should release anonymized versions (if applicable). Providing as much information as possible in supplemental material (appended to the paper) is recommended, but including URLs to data and code is permitted. 6. Experimental Setting/Details Question: Does the paper specify all the training and test details (e.g., data splits, hyperparameters, how they were chosen, type of optimizer, etc.) necessary to understand the results? Answer: [NA] Justification: The contributions of our paper are mainly theoretical. Guidelines: The answer NA means that the paper does not include experiments. The experimental setting should be presented in the core of the paper to a level of detail that is necessary to appreciate the results and make sense of them. The full details can be provided either with the code, in appendix, or as supplemental material. 7. Experiment Statistical Significance Question: Does the paper report error bars suitably and correctly defined or other appropriate information about the statistical significance of the experiments? Answer: [NA] Justification: The contributions of our paper are mainly theoretical. Guidelines: The answer NA means that the paper does not include experiments. The authors should answer "Yes" if the results are accompanied by error bars, confidence intervals, or statistical significance tests, at least for the experiments that support the main claims of the paper. The factors of variability that the error bars are capturing should be clearly stated (for example, train/test split, initialization, random drawing of some parameter, or overall run with given experimental conditions). The method for calculating the error bars should be explained (closed form formula, call to a library function, bootstrap, etc.) The assumptions made should be given (e.g., Normally distributed errors). It should be clear whether the error bar is the standard deviation or the standard error of the mean. It is OK to report 1-sigma error bars, but one should state it. The authors should preferably report a 2-sigma error bar than state that they have a 96% CI, if the hypothesis of Normality of errors is not verified. For asymmetric distributions, the authors should be careful not to show in tables or figures symmetric error bars that would yield results that are out of range (e.g. negative error rates). If error bars are reported in tables or plots, The authors should explain in the text how they were calculated and reference the corresponding figures or tables in the text. 8. Experiments Compute Resources Question: For each experiment, does the paper provide sufficient information on the computer resources (type of compute workers, memory, time of execution) needed to reproduce the experiments? Answer: [NA] Justification: The contributions of our paper are mainly theoretical. Guidelines: The answer NA means that the paper does not include experiments. The paper should indicate the type of compute workers CPU or GPU, internal cluster, or cloud provider, including relevant memory and storage. The paper should provide the amount of compute required for each of the individual experimental runs as well as estimate the total compute. The paper should disclose whether the full research project required more compute than the experiments reported in the paper (e.g., preliminary or failed experiments that didn t make it into the paper). 9. Code Of Ethics Question: Does the research conducted in the paper conform, in every respect, with the Neur IPS Code of Ethics https://neurips.cc/public/Ethics Guidelines? Answer: [Yes] Justification: We have reviewed the Neur IPS Code of Ethics, and our research was conducted in accordance with it. Guidelines: The answer NA means that the authors have not reviewed the Neur IPS Code of Ethics. If the authors answer No, they should explain the special circumstances that require a deviation from the Code of Ethics. The authors should make sure to preserve anonymity (e.g., if there is a special consideration due to laws or regulations in their jurisdiction). 10. Broader Impacts Question: Does the paper discuss both potential positive societal impacts and negative societal impacts of the work performed? Answer: [Yes] Justification: Although the contributions of our paper are mainly theoretical, we explain in the introduction section how they can positively impact society. Guidelines: The answer NA means that there is no societal impact of the work performed. If the authors answer NA or No, they should explain why their work has no societal impact or why the paper does not address societal impact. Examples of negative societal impacts include potential malicious or unintended uses (e.g., disinformation, generating fake profiles, surveillance), fairness considerations (e.g., deployment of technologies that could make decisions that unfairly impact specific groups), privacy considerations, and security considerations. The conference expects that many papers will be foundational research and not tied to particular applications, let alone deployments. However, if there is a direct path to any negative applications, the authors should point it out. For example, it is legitimate to point out that an improvement in the quality of generative models could be used to generate deepfakes for disinformation. On the other hand, it is not needed to point out that a generic algorithm for optimizing neural networks could enable people to train models that generate Deepfakes faster. The authors should consider possible harms that could arise when the technology is being used as intended and functioning correctly, harms that could arise when the technology is being used as intended but gives incorrect results, and harms following from (intentional or unintentional) misuse of the technology. If there are negative societal impacts, the authors could also discuss possible mitigation strategies (e.g., gated release of models, providing defenses in addition to attacks, mechanisms for monitoring misuse, mechanisms to monitor how a system learns from feedback over time, improving the efficiency and accessibility of ML). 11. Safeguards Question: Does the paper describe safeguards that have been put in place for responsible release of data or models that have a high risk for misuse (e.g., pretrained language models, image generators, or scraped datasets)? Answer: [NA] Justification: The contributions of our paper are mainly theoretical and pose no such risks. Guidelines: The answer NA means that the paper poses no such risks. Released models that have a high risk for misuse or dual-use should be released with necessary safeguards to allow for controlled use of the model, for example by requiring that users adhere to usage guidelines or restrictions to access the model or implementing safety filters. Datasets that have been scraped from the Internet could pose safety risks. The authors should describe how they avoided releasing unsafe images. We recognize that providing effective safeguards is challenging, and many papers do not require this, but we encourage authors to take this into account and make a best faith effort. 12. Licenses for existing assets Question: Are the creators or original owners of assets (e.g., code, data, models), used in the paper, properly credited and are the license and terms of use explicitly mentioned and properly respected? Answer: [Yes] Justification: We utilize models of learning-augmented algorithms and learning-augmented mechanisms, and our related work discussion appropriately cites all prior work and its contributions. Guidelines: The answer NA means that the paper does not use existing assets. The authors should cite the original paper that produced the code package or dataset. The authors should state which version of the asset is used and, if possible, include a URL. The name of the license (e.g., CC-BY 4.0) should be included for each asset. For scraped data from a particular source (e.g., website), the copyright and terms of service of that source should be provided. If assets are released, the license, copyright information, and terms of use in the package should be provided. For popular datasets, paperswithcode.com/datasets has curated licenses for some datasets. Their licensing guide can help determine the license of a dataset. For existing datasets that are re-packaged, both the original license and the license of the derived asset (if it has changed) should be provided. If this information is not available online, the authors are encouraged to reach out to the asset s creators. 13. New Assets Question: Are new assets introduced in the paper well documented and is the documentation provided alongside the assets? Answer: [NA] Justification: The contributions of our paper are mainly theoretical and we did not introduce a new assets. Guidelines: The answer NA means that the paper does not release new assets. Researchers should communicate the details of the dataset/code/model as part of their submissions via structured templates. This includes details about training, license, limitations, etc. The paper should discuss whether and how consent was obtained from people whose asset is used. At submission time, remember to anonymize your assets (if applicable). You can either create an anonymized URL or include an anonymized zip file. 14. Crowdsourcing and Research with Human Subjects Question: For crowdsourcing experiments and research with human subjects, does the paper include the full text of instructions given to participants and screenshots, if applicable, as well as details about compensation (if any)? Answer: [NA] Justification: Our paper does not involve crowdsourcing nor research with human subjects. Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Including this information in the supplemental material is fine, but if the main contribution of the paper involves human subjects, then as much detail as possible should be included in the main paper. According to the Neur IPS Code of Ethics, workers involved in data collection, curation, or other labor should be paid at least the minimum wage in the country of the data collector. 15. Institutional Review Board (IRB) Approvals or Equivalent for Research with Human Subjects Question: Does the paper describe potential risks incurred by study participants, whether such risks were disclosed to the subjects, and whether Institutional Review Board (IRB) approvals (or an equivalent approval/review based on the requirements of your country or institution) were obtained? Answer: [NA] Justification: Our paper does not involve crowdsourcing nor research with human subjects. Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Depending on the country in which research is conducted, IRB approval (or equivalent) may be required for any human subjects research. If you obtained IRB approval, you should clearly state this in the paper. We recognize that the procedures for this may vary significantly between institutions and locations, and we expect authors to adhere to the Neur IPS Code of Ethics and the guidelines for their institution. For initial submissions, do not include any information that would break anonymity (if applicable), such as the institution conducting the review.