# on_speeding_up_language_model_evaluation__1254dd53.pdf Published as a conference paper at ICLR 2025 ON SPEEDING UP LANGUAGE MODEL EVALUATION Jin Peng Zhou Cornell University Christian K. Belardi Cornell University Ruihan Wu University of California, San Diego Travis Zhang Cornell University Carla P. Gomes Cornell University Wen Sun Cornell University Kilian Q. Weinberger Cornell University Developing prompt-based methods with Large Language Models (LLMs) requires making numerous decisions, which give rise to a combinatorial search problem over hyper-parameters. This exhaustive evaluation can be time-consuming and costly. In this paper, we propose an adaptive approach to explore this space. We are exploiting the fact that often only few samples are needed to identify clearly superior or inferior settings, and that many evaluation tests are highly correlated. We lean on multiarmed bandits to sequentially identify the next (method, validation sample)-pair to evaluate and utilize low-rank matrix factorization to fill in missing evaluations. We carefully assess the efficacy of our approach on several competitive benchmark problems and show that it can identify the top-performing method using only 5-15% of the typical resources resulting in 85-95% LLM cost savings. Our code is available at https://github.com/kilian-group/banditeval. 1 INTRODUCTION Figure 1: Cost comparison for finding the best model or prompt between our proposed algorithms (UCB-E, UCB-E-LRF) and Full Evaluation on two datasets. The overhead computation time for UCB-E and UCB-E-LRF is 2.4 and 142.6 seconds, respectively. See text for details. Large language models (LLMs) have demonstrated remarkable proficiency in diverse tasks such as question answering, machine translation, and mathematical reasoning (Brown et al., 2020; Chowdhery et al., 2023; Lewkowycz et al., 2022). They are employed across numerous applications, ranging from automated customer support to content generation (Liang et al., 2024). As the development of new LLMs continues, practitioners find themselves with a plethora of options, selecting the optimal model, prompt (Kojima et al., 2022), and hyperparameters for their specific needs. Querying an LLM is resource-intensive, and extensive evaluation requires significant investment in capital, time, and compute. Figure 1 (left) shows the estimated cost of a Full Evaluation of the 153 models officially included in Alpaca Eval (Li et al., 2023) as of May 20, 2024 to be almost 800 USD. GPT4-turbo is the LLM judge (Zheng et al., 2024) and the monetary cost is calculated based on the Equal contribution. Correspondence to {jz563, ckb73}@cornell.edu and ruw076@ucsd.edu. Published as a conference paper at ICLR 2025 Table 1: Selected benchmarks where LLMs are state-of-the-art methods. We group them based on the task type and whether additional LLM-based / human-based scoring is needed for evaluation. Task Type LLM Inference + Rule-based Scoring LLM Inference + LLM-based Scoring Natural Language Understanding BOLD Dhamala et al. (2021) GLUE Wang et al. (2019) Hella Swag Zellers et al. (2019) SQu AD Rajpurkar et al. (2016) Trivia QA Joshi et al. (2017) Wino Grande ai2 (2019) CNN/Dailymail Nallapati et al. (2016) Newsroom Grusky et al. (2018) XSUM Narayan et al. (2018) Open-ended QA Google NQ Kwiatkowski et al. (2019) Hotpot QA Yang et al. (2018) QASPER Dasigi et al. (2021) Alpaca Eval Li et al. (2023) Chatbot Arena Chiang et al. (2024) MT-Bench Zheng et al. (2024) STEM and Reasoning APPS Hendrycks et al. (2021a) Arc Clark et al. (2018) GPQA Rein et al. (2023) GSM8K Cobbe et al. (2021) MATH Hendrycks et al. (2021c) MMLU Hendrycks et al. (2021b) PIQA Bisk et al. (2020) inference cost of only proprietary models and GPT4-turbo judging cost. Similarly, 78 Nvidia A6000 GPU hours are needed for evaluating 205 zero-shot prompts on 784 GSM8K (Cobbe et al., 2021) questions using Mistral-7B (Jiang et al., 2023), Figure 1 (right). In this paper we investigate the familiar setting where given a set of methods we aim to identify the best performing across a set of validation examples given a fixed evaluation budget. For example, the different methods could be LLM prompts or hyperparameter settings and the validation examples could be translation tasks or K12 math problems. The budget could be a fixed monetary sum or GPU time. We argue that this setting is extremely common among LLM applications, where the practitioner is primarily interested in identifying the single best method to deploy. A simple baseline is to evenly split the budget for each method. However, this can be very inefficient as it would spend a lot of the budget on obvious non-performers and too little effort on identifying the very best method. Intuitively, methods that consistently underperform other methods are unlikely to end up as the best. Therefore, a more efficient budget allocation strategy is to spend less on low-performing methods and more on promising ones. Multi-armed bandit algorithms enable this dynamic allocation by actively selecting the next method-example pair to evaluate based on the results of previous evaluations. We propose two active selection algorithms UCB-E and UCB-E-LRF. Our first algorithm is an extension of the classical UCB-E (Audibert and Bubeck, 2010) to solve the multi-armed bandit problem. It estimates the upper confidence bound (UCB) to guide the selection of the method which is paired with a randomly chosen example for the next evaluation in order to efficiently estimate the best method. This algorithm has a strong theoretical guarantee that the probability of selecting the best arm approaches 100% with an exponential convergence in the number of evaluations. Our second algorithm, UCB-E-LRF (for Low-Rank Factorization), leverages the fact that in practice many validation samples are similar, and their results are highly correlated. In other words, if a method performs well on one sample, it will predictably also perform well on similar samples. We can therefore speculatively fill in those results and evaluate the method on a sample for which the performance is more uncertain. In our method, we capture this insight through a low-rank approximation of the evaluation matrix. By deploying this approach in addition to UCB-E, UCE-ELRF actively selects both the method and example to evaluate, which leads to even bigger budget savings in settings with correlated validation samples. We evaluate the efficacy of these two algorithms, in addition to common non-active baselines, in a number of practical settings. Our empirical analysis shows that our two active selection algorithms are much better than all baselines. Not surprisingly, the benefits of active selection are larger if the top methods are separated by a bigger performance gap. Notably, we observe that UCB-E works best for such easier settings; and UCB-E-LRF shines in harder settings, when the performance differences between methods are more subtle. 2 NOTATION AND PROBLEM FORMULATION A typical evaluation workflow in large language model (LLM) applications involves three steps: inference, scoring, and performance aggregation. Inference. Given a dataset of examples X = {x1, . . . , xn}, outputs are generated by specific LLM-based methods F = {f1, . . . , fm}. Here, methods can differ based on their LLM architecture Published as a conference paper at ICLR 2025 (architecture search), a prompt (prompt engineering (Kojima et al., 2022)), or specific decoding parameters (e.g. temperature, decoding strategies), or other hyper-parameters. Scoring. The outputs from different methods are scored with a scoring function s : Y [0, 1]. Without loss of generality, we assume s(f(x)) > s(f (x)) indicates f has better performance than f on an example x. The scoring function can either be computational (exact string match, BLEU (Papineni et al., 2002), ROUGE (Lin, 2004)), LLM-based (BERTScore (Zhang et al., 2019), LLM judge (Zheng et al., 2024)), or human annotators. Depending on the task and dataset format, researchers have employed different types of scoring functions. In Table 1, we classify a selected group of commonly used datasets according to the task type and scoring function applied to them. We group LLM-based and human-based scoring together since they are usually considered alternatives to each other and have been shown to have relatively high correlation (Li et al., 2023). Performance aggregation. Once the underlying scoring matrix S [0, 1]m n is computed as Sij := s(fi(xj)), the aggregate performance of method fi is its average score across all validation samples, µi = 1 n Pn j=1 Sij. Alternative aggregations are possible but not considered in this work. Both inference and scoring often rely heavily on LLM-based scoring through black-box APIs (Lambert et al., 2024) inducing large costs and resource consumption. However, in many practical scenarios, we are only interested in identifying the best method among all methods. For example, for prompt engineering and hyperparameter tuning, knowing which method (prompt / configuration) is the best is usually sufficient for the next round of iteration. Identifying exactly which method is ranked 84 out of 100 is irrelevant and a waste of resources. We therefore claim that evaluating all methods on all examples of a dataset is excessive for this purpose, which motivates the research question of how to identify the best method i = arg maxi µi with limited evaluations. Formally, given a finite evaluation budget T, we want an algorithm A(T, F, X) that maximizes PA(A(T, F, X) = i ), the probability of returning the best method i within budget T. We introduce the following additional notation to allow for a clear description of the algorithms. Let Sobs ([0, 1] {?})m n denote the partially observed scoring matrix, and O {0, 1}m n denote the observation matrix where Oij indicates whether the method-example pair (fi, xj) has been observed. Finally, let us denote our estimate of S by ˆS. For ease of reference, We also list the notations used throughout the paper in Appendix E Table 4. 3 ALGORITHMS One simple baseline for designing the evaluation algorithm A is: uniformly sample and evaluate T examples from X, for each method fi F, estimate the performance ˆµi = 1 Pn j=1 Oij Pn j=1 Oij Sobs ij , and pick the method fˆi with the highest estimated mean ˆµˆi as the prediction for the best method. This baseline ignores any information gathered in the process of sequentially evaluating examples, and, as expected, can be very inefficient. We will see in the experiments that for some datasets, this baseline needs to evaluate at least 90% of all method-example pairs to predict the best method correctly with high probability. In the following section, we will introduce two adaptive algorithms that actively select the next method-example pair to evaluate based on previous results; Figure 2 illustrates this idea on a high level for prompt tuning. The methods in the figure are different prompts, and the examples are simple math problems. 3.1 ALGORITHM 1 UCB-E The main limitation of the simple baseline is that in expectation it samples T/n examples per method; evenly distributing its total budget across all different methods fi. In order to distinguish the best method fi from other strong methods fi (i.e. µi µi is small), we may need more than T/n examples, while T/n examples are sufficient to distinguish fi from obviously bad methods (i.e. µi µi is large). We address this limitation by adaptively selecting example t based on our previous t 1 observations until our budget T is expended. Our first algorithm UCB-E (Aue; Algorithm 1) is a simple extension of the classic multi-arm bandit algorithm UCB-E (Audibert and Bubeck, 2010). For every method fi, Published as a conference paper at ICLR 2025 Evaluate 2 3 Update Let's think step by step... I'll tip $10 if you solve... Explain how to solve... Natalia sold clips to 48 of... Weng earns $12 an hour... Gail has two fish tanks... LLM Evaluation Repeat Until Budget T Expended Observed Scoring Matrix Figure 2: Active method-example pair selection: After LLM evaluated t method-example pairs, we then call Algorithm A to select the next method-example pair. Then we query LLM for evaluating this pair and fill the scoring received from LLM into the scoring matrix. Algorithm A then updates its internal status prepared for the next method-example pair selection. This process is repeated T times and, in the end, the algorithm A predicts the best method fˆi . Algorithm 1 UCB-E (Aue(T, F, X; a)) Input: The evaluation budget T, a set of methods F, a set of examples X, exploration parameter a. Output: The prediction ˆi for best method i . 1: The upper confidence bounds B := {+ }m, the observation matrix O := {0}m n, the observed scoring matrix Sobs := {?}m n. 2: for t = 1, , T do 3: Select: Draw i arg maxk | (Pn j=1 Okj) =n Bk; Draw uniformly at random j {k [n]|Oik = 0}. 4: Evaluate: Run inference for the method-example pair (fi, xj), score the result, and receive s(fi(xj)); Sobs ij s(fi(xj)). 5: Update: Oij 1; Bi Pn j=1 Oij Sobs ij Pn j=1 Oij + q a Pn j=1 Oij . Return: ˆi = arg maxi Pn j=1 Oij Sobs ij Pn j=1 Oij at every step t, we estimate the upper confidence bound, Pn j=1 Oij Sobs ij Pn j=1 Oij + a Pn j=1 Oi,j . Here, a is a hyperparameter that controls the tradeoff between exploration and exploitation, where a larger a value encourages more exploration. In the subsequent iteration, we pick the method fi with the largest upper confidence bound (ties are resolved by uniform random sampling), and then sample an example xj uniformly at random from the set of examples not yet observed, that is the set {j [n]|Oij = 0}. We run the inference procedure for fi(xj) and score the result, obtaining Sobs ij = s(fi(xj)). Comparison with the simple baseline. The UCB-E algorithm addresses the limitations of the simple baseline by minimizing evaluations on suboptimal methods. For a bad method (i.e. µi µi is large), only a few evaluations (typically proportional to 1 (µi µi)2 ) are needed to ensure that its upper confidence bound will never exceed µi . Consequently, the UCB-E algorithm will not select this method again, thus saving the evaluation budget and focusing on more promising methods. Theoretical results. We state the theoretical guarantee of UCB-E in our context, which is a corollary of the theorem for UCB-E in the original multi-arm bandit setting (Audibert and Bubeck, 2010). Published as a conference paper at ICLR 2025 Corollary 1 (Lower bound on success probability of UCB-E) Define H1 = Pm i=1,i =i 1 (µi µ i )2 and suppose a = 25 H1 , PAue(Aue(T, F, X; a) = i ) 1 2Tm exp T m Comment. H1 in the corollary indicates the hardness of the problem specified by (F, X, s). Intuitively, the more methods that have a smaller performance gap with the best method, the harder it is to identify which method is the best. We observe that datasets with smaller H1 tend to have a higher chance of predicting i correctly, given the same evaluation budget T. 3.2 ALGORITHM 2 UCB-E WITH LOW-RANK FACTORIZATION (UCB-E-LRF) Algorithm 1, the UCB-E algorithm, treats each method independently and samples examples for a method uniformly at random. In practice, however, methods and examples are correlated with each other and we can predict the outcome of a particular method-example pair from previous observations. Being able to accurately predict the outcome gives a better estimate of the score of a method. Additionally, instead of sampling examples uniformly at random, prioritizing examples with large prediction uncertainties reduces the overall uncertainty in the scores of each method, which provides more accurate score estimates in the subsequent steps. In this section, we propose a low-rank factorization based algorithm (UCB-E-LRF) to capture the above intuition. Low-rank factorization. A natural approach to modeling these method-example interactions is through low-rank matrix factorization. Intuitively, if the method-examples are very correlated, there should exist U Rm r and V Rn r and rank r min(m, n) such that S UV (Candes and Recht, 2012; Chen et al., 2020; Cai et al., 2019). Obtaining U and V from the observed scoring matrix Sobs allows us to estimate the scores of all method-example pairs from just a few scores. For instance, if the matrix S were an exact rank-1 matrix, only m + n 1 scores would be needed to recover the full scoring matrix S, which is far more efficient than exhaustively evaluating all m n combinations. To find U and V , we use alternating least squares (Hastie et al., 2015) to optimize the following problem: O UV T Sobs 2 Score estimation. We can then combine the observed matrix Sobs and estimated matrix ˆS = UV T , using O as a gating function, to arrive at a prediction of the score µi as follows: j=1 Oij Sobs ij + (1 Oij) ˆSij. (2) Uncertainty estimation. The score estimation gives us a better way to estimate the score of every method by exploiting the correlation between methods and examples at every step. We now discuss how we estimate prediction uncertainties with ensembles. By fitting many factorization models, we obtain an empirical distribution of predictions for each method-example pair. Consequently, the standard deviation of the predictions can be used to quantify uncertainty. If the predictions are very similar to each other, this suggests the ensembles tend to agree with each other, and hence there is little uncertainty. In order to achieve sufficient model variance to approximate uncertainty, we uniformly at random drop some values from Sobs when solving the factorization problem in Equation 1 for each model (Efron and Tibshirani, 1994; Zoubir and Boashash, 1998). For an ensemble of size C the estimate for ˆS becomes the following: c=1 U (c)(V (c))T . (3) Then the uncertainty matrix R Rm n is given by: " 1 O U (c)(V (c))T ˆS #2 Published as a conference paper at ICLR 2025 Algorithm 2 UCB-E-LRF (Auel(T, F, X; M, r, C, T0, η)) Input: The evaluation budget T, a set of methods F, a set of examples X, the low-rank factorization model M, the rank r, the ensemble size C, the warm-up budget T0, the uncertainty scaling η. Output: The prediction ˆi for best method i . 1: Uniformly draw T0 method-example pairs from [m] [n] and get the observation matrix O {0, 1}m n and observed scoring matrix Sobs ([0, 1] {?})m n w.r.t. these T0 evaluations. 2: ˆS, R M(Sobs, O; r, C); fi F, Bi 1 n Pn j=1(Oij Sobs ij + (1 Oij) ˆSij + ηRij) 3: for t = T0, , T do 4: Select: Draw i arg maxk | (Pn j=1 Okj) =n Bk; Draw j arg maxk | Oik=0 Rik. 5: Evaluate: Run inference for the method-example pair (fi, xj), score the result, and receive s(fi(xj)); Sobs ij s(fi(xj)). 6: Update: Oij 1; ˆS, R M(Sobs, O; r, C); fi F, Bi 1 n Pn j=1(Oij Sobs ij + (1 Oij) ˆSij + ηRij). 7: end for Return: ˆi = arg maxi 1 n Pn j=1(Oij Sobs ij + (1 Oij) ˆSij) Adaptive selection. Bringing score estimation and uncertainty estimation together, we propose our second algorithm UCB-E-LRF (Auel; Algorithm 2), an extension of Algorithm 1 leveraging the low-rank factorization model M. The high level structure of the algorithms are very similar to each other. Specifically, we estimate the upper confidence bound (UCB) for each method based on a linear combination of score and uncertainty estimation, select the method with the greatest UCB, and evaluate the method on an example that has the highest uncertainty. Lastly, we introduce a warm-up budget T0. This is because M requires a minimum number of initial observations in order to estimate the low-rank representation accurately. Therefore, before the active selection phase (line 3-9 in Algorithm 2), we have a warm-up phase (line 1-2 in Algorithm 2), where T0 method-example pairs are sample uniformly at random to evaluate before moving on to the active phase. 4 EXPERIMENTS 4.1 DATASETS Table 2: Statistics of each dataset used in our experiments. The examples X in each dataset correspond to the questions from their respective evaluation benchmarks. H1 as defined in Section 3.1, reflects the difficulty of identifying the best method, with larger values indicating more challenging settings. More dataset details can be found in Appendix C. DA stands for Drop Annotator. Method Set F Dataset Name Size m n Scoring Function s H1 Various LLMs Alpaca Eval 154 805 GPT4-turbo annotator 966 Various LLMs excluding GPT4-turbo Alpaca Eval (DA) 153 805 GPT4-turbo annotator 4462 Mistral-7B with different prompts GSM8K Prompts 205 784 regex match with correct answer 107445 Various LLMs and sampling configurations GSM8K Models 122 1000 regex match with correct answer 20562 Tulu-7B with different prompts PIQA Prompts 177 1546 regex match with correct choice 66284 Various LLMs and sampling configurations PIQA Models 103 1000 regex match with correct choice 10273 To assess the performance of our algorithms under a variety of use cases, we test with three datasets Alpaca Eval (Li et al., 2023), Grade School Math 8K (GSM8K) (Cobbe et al., 2021) and Physical Interaction: Question Answering (PIQA) (Bisk et al., 2020), together with different settings of the method set F and the scoring function s; Table 2 summarizes each dataset. For Alpaca Eval, we design two sets of F. The first set F contains all LLMs reported by Li et al. (2023). The second F contains the same LLMs except for the annotator model GPT-4 Turbo. The latter F that does not contain GPT-4 Turbo makes the learning more challenging and interesting since the annotator, GPT-4 Turbo, is a clear winner among all LLMs. For GSM8K and PIQA, we set F as a set of prompts to simulate prompt engineering, or a set of LLMs with various sampling configurations, to mimic model selection and hyperparameter tuning. We also observe that the datasets indeed exhibit low-rank properties as seen from ratios of singular values (Table 3) and explained variance of each principal component (Figure 6) for the data matrices. More information about these datasets is in Appendix C. Published as a conference paper at ICLR 2025 Performance GSM8K Prompts H1 = 107445 10% 100% 0.00 PIQA Prompts GSM8K Models PIQA Models H1 = 10273 Alpaca Eval (Drop Annotator) Alpaca Eval 10% 100% 0.00 Performance 10% 100% 0.00 Statistical Significance 0.01 10% 100% 0.30 10% 100% 0.30 10% 100% 0.40 Statistical Significance 0.1 10% 100% 0.30 10% 100% 0.40 10% 100% Percentage of Matrix Evaluated 10% 100% Percentage of Matrix Evaluated 10% 100% Percentage of Matrix Evaluated 10% 100% Percentage of Matrix Evaluated 10% 100% Percentage of Matrix Evaluated 10% 100% Percentage of Matrix Evaluated UCB-E UCB-E-LRF LRF Row Mean Imputation Filled Subset H1 := Dataset Hardness Quantity Figure 3: Comparison of algorithms on six datasets evaluated with various metrics. The vertical axes represents performance of a metric and horizontal axes represents the percentage of method-example pairs evaluated. All results are aggregated based on 50 trials with different random seeds. The datasets are ordered by decreasing H1 (see Section 3.1). Larger H1 values indicate settings where finding the best method is more difficult. Our proposed algorithms: UCB-E and UCB-E-LRF consistently require much less evaluation to achieve the same performance as baselines. 4.2 METRICS Top 1 Precision. To determine if an algorithm finds the best method, we can directly check if the algorithm s prediction matches with the best method from our empirical data. However, because every method is empirically evaluated on a limited number of examples, it is possible that one method has a slightly lower average performance than another method whereas if we were to evaluate on more examples, the former would achieve a higher performance. To this end, we calculate top 1 precision by first determining a set of methods we consider equally good and checking if the predicted best method from an algorithm is a member of that set. We propose two ways to determine if a set of methods are equally good: performance gap and statistical significance. For performance gap ϵ top 1 precision, we consider all methods whose performance is within ϵ of the best empirical method to be equally good. For the statistical significance p top 1 precision, we perform Mc Nemar s statistical test (Mc Nemar, 1947) for each method against the best empirical method. If we cannot reject the null hypothesis that the performance of one method is the same as the best empirical method up to a significance level p, then that method is considered equally good as the best method. We evaluate all baselines and our algorithms on two ϵ values {0.001, 0.01}, and two p values {0.01, 0.1} simulating a diverse need of precision. NDCG@K. Although our focus is to identify the best method quickly, it is sometimes also desirable to generally rank the top K promising methods high. We therefore evaluate with normalized discounted Published as a conference paper at ICLR 2025 10% 100% Percentage of Matrix Evaluated Performance GSM8K Prompts 10% 100% Percentage of Matrix Evaluated 1.0 Alpaca Eval (Drop Annotator) Score Only a UCB-E UCB-E-LRF UCB-E-LRF (Score Only) 10% 100% Percentage of Matrix Evaluated Performance GSM8K Prompts 10% 100% Percentage of Matrix Evaluated 1.0 Alpaca Eval (Drop Annotator) Uncertainty Scaling Factor d 0.1 0.5 1 5 10 10% 100% Percentage of Matrix Evaluated Performance GSM8K Prompts 10% 100% Percentage of Matrix Evaluated 1.0 Alpaca Eval (Drop Annotator) Batch Size b b 1 32 128 1 32 128 10% 100% Percentage of Matrix Evaluated Performance GSM8K Prompts 10% 100% Percentage of Matrix Evaluated 1.0 Alpaca Eval (Drop Annotator) Warm-Up Budget T0 e 0.5% 1.0% 5.0% 10.0% 10% 100% Percentage of Matrix Evaluated Performance GSM8K Prompts 10% 100% Percentage of Matrix Evaluated 1.0 Alpaca Eval (Drop Annotator) Ensemble Size C c 4 16 64 256 10% 100% Percentage of Matrix Evaluated Performance GSM8K Prompts 10% 100% Percentage of Matrix Evaluated 1.0 Alpaca Eval (Drop Annotator) Rank r f Figure 4: Ablations of the proposed algorithms and hyperparameters. UCB-E and UCB-E-LRF are denoted by blue and red lines, respectively, consistent with Figure 3. cumulative gain (NDCG) at K. NDCG@K takes as input the top K prediction by an algorithm and the higher their true ranks are, the higher the NDCG is. We choose K = 10 to evaluate all algorithms. 4.3 SET-UP OF OUR ALGORITHMS AND BASELINES UCB-E: We use a = 1 since this consistently yields the best performance across all datasets. This parameter is ablated in Figure 7 of Appendix D. UCB-E-LRF: For all datasets, we use rank r = 1 for low rank factorization with an ensemble size of C = 64. We use 5% of data for warm up i.e. T0 = 0.05 m n and η = 5. Row Mean Imputation: We select method-example pairs uniformly at random from all pairs. The score for each method is calculated as the average of all the scores evaluated so far for that method. Filled Subset: Instead of randomly selecting method-example pairs, filled subset first selects an example index j uniformly at random. Then all methods are evaluated on example j. If all methods have been evaluated, the algorithm selects a new example index from remaining ones. The score for each method is calculated as the average of all the scores evaluated so far for that method. Although row mean imputation and filled subset do not have learning component, they both produce an unbiased estimate of the real score for each method at all times. LRF: Similar to row mean imputation, LRF (low rank factorization) randomly selects method-example pairs to evaluate. For estimating the score of each method, LRF calculates the average of all scores for that method. That is, for evaluated examples, the score is the actual observed score, and for examples yet to be evaluated, the score is the estimated score from LRF. The details for three baselines Row Mean Imputation, Filled Subset, and LRF are found in Appendix D Algorithm 3, 4, and 5, respectively. Finally, to take advantaged of the parallelism available in modern computing hardware, all algorithms and baselines are implemented with a batch size b. For algorithms that randomly draw an example, they instead randomly draw b examples without replacement. For algorithms that draw an example with the arg max operator, they instead draw with the Top b operator. We use b = 32 for all experiments unless explicitly stated otherwise. 4.4 MAIN RESULTS In Figure 3, we plot the performance of baselines and our algorithms on all six datasets (columns) as the budget T increases from 5% to 100% of the total number of method-example pairs. In each row, we evaluate the algorithms on a different metric (either top 1 precision with different ϵ, p or NDCG@10). Each line in the figure is the average result over 50 independent trials with different seeds. For example, an average top 1 precision of 0.9 indicates that 45 out of 50 trials predict the best method correctly at the budget level that its x-coordinate represents. In addition, we calculate H1 as defined in Corollary 1 that quantifies the difficulty of finding the best method on a dataset. Intuitively, a higher H1 value suggests that the method set F is large or there are many methods that have similar performance with the best one and distinguishing them can be challenging. In Appendix C Figure 5, we also plot the histogram to show the distribution of performance among F on these datasets. Published as a conference paper at ICLR 2025 How do our algorithms compare with the baselines? As seen from Figure 3, both UCB-E and UCB-E-LRF consistently achieve high precision and NDCG with much less budget compared to the baselines. For example, on Alpaca Eval (Drop Annotator), our proposed algorithms can reach a precision of 1 with just 8% budget whereas the baselines require 80-90%, an order of magnitude more budget needed. These results suggest that it is entirely possible to identify the best method without exhaustively evaluating all method-example pairs. They further demonstrate that the active selection algorithms (our two algorithms) are more efficient than the non-active algorithms (three baselines). Additionally, the better NDCG performance from our proposed algorithms shows that our methods can more correctly rank top-performance methods. How is the comparison between our two algorithms UCB-E and UCB-E-LRF? The datasets from left to right are ranked by the hardness indicated by H1. Interestingly, we find that on easier datasets such as Alpaca Eval, UCB-E performs better and saves 2-3% more on budget compared to UCB-ELRF, while on harder datasets, such as GSM8K Prompts and PIQA Prompts, UCB-E-LRF achieves higher precision faster than UCB-E, saving about 10% in absolute budget. These observations give us a hint on what algorithm to apply in practice. 4.5 MORE EMPIRICAL ANALYSIS We provide more empirical analysis and the ablation results can be found in Figure 4. Does H1 correctly reflect the hardness of a dataset in the empirical experiments? Yes, going through Figure 3 from left to right, as H1 decreases, the percentage of matrix evaluation needed to reach a precision of 1 also decreases from more than 20% to just under 5%. Moreover, the H1 values seem to be related to the tasks. Prompt engineering datasets typically have higher H1 possibly due to the homogeneity of prompt performance with the same LLM. In contrast, datasets that benchmark different LLMs such as Alpaca Eval make it much easier to find the best performing model. Score Only Ablation. To study the effect of actively selecting both the next method and next example to evaluate using uncertainty matrix R, we consider an ablation of UCB-E-LRF. In this variant, called UCB-E-LRF (Score Only), we modify the UCB-E algorithm by replacing its mean calculation with the mean calculated via low-rank factorization (as described in Equation 3). The upper confidence bound calculation and next example selection are the same as the original UCB-E algorithm. The detailed algorithmic description can be found in Algorithm 6. We see that on the hard dataset GSM8K Prompts where UCB-E-LRF has a certain benefit over UCB-E, there is a significant gap between UCB-E-LRF and UCB-E-LRF (Score Only). This means that the component of uncertainty matrix R in UCB-E-LRF is crucial to contributing the benefit over UBC-E. Batch Size b Ablation. Intuitively, a smaller batch size is more flexible and can have more fine grained selection. We experiment with b {1, 32, 128}, and as shown in the plot, the smallest batch size gives the best performance, especially on harder datasets whereas a batch size of 128 significantly degrades the performance. However, it incurs more overall computational cost due to fitting low rank factorization more often than a smaller batch size; for example when b = 2, it takes almost 16 times as much time as b = 32 to achieve 1.0 precision, which is ineffective. Ensemble Size C Ablation. We vary the ensemble size C {4, 16, 64, 256} and find that very small ensemble size gives a suboptimal performance on hard datasets. There is almost no performance difference on easy datasets like Alpaca Eval (Drop Annotator). The performance is robust to different C as long as it is larger than 64. Uncertainty Scaling Factor η Ablation. We experiment with different uncertainty scaling values {0.1, 0.5, 1, 5, 10}. In Figure 4d, it can be seen that a certain range of η, from 0.1 to 5, gives similar performance on both two datasets. This demonstrates the robustness of selecting η. Warm-up Budget T0 Ablation. Our algorithm UCB-E-LRF by default randomly evaluates 5% of the method-example pairs in the data matrix before selecting actively. We analyze the effect of varying the budget T0 among {0.5, 1, 5, 10}% on the algorithm performance on the two datasets. A very small warm-up budget with 0.5% of data can achieve decent precision initially, but fall behind compared to larger warm-up budget as more data are evaluated. In contrast, a very large warm-up budget of 10% delays the active selection algorithm too much and also achieves suboptimal performance. We therefore use 5% as a generally strong starting point. On Alpaca Eval, it is possible to achieve the same performance with an even smaller warm-up budget, suggesting more savings is possible. Published as a conference paper at ICLR 2025 Rank r Ablation. Hyperparameter r adjusts the bias-variance trade-off. Empirically we experiment with r {1, 2, 5} and find that r = 1 is consistently better than larger value counterparts. The results can be explained by the fact that a larger r requires more evaluated data in order to prevent overfitting, which might not have an advantage in the limited budget setting. Therefore, we recommend using r = 1 for all datasets. 5 RELATED WORK Best-arm identification in multi-arm bandits. The goal of best-arm identification (Bubeck et al., 2009; Audibert and Bubeck, 2010) is to find the arm with the highest reward by pulling these arms and getting feedback. By making an analogy, in our problem method fi is the arm and the score µi of fi is the reward. There are two ways to define the best-arm identification problem: fixed budget and fixed confidence. In the fixed budget setting, the budget for the arm pulls is fixed and the algorithm is designed for better chances to identify the correct best arm our problem defined in this paper has a similar evaluation budget. UCB-E (Audibert and Bubeck, 2010) and Successive Elimination (Audibert and Bubeck, 2010) are two pioneering algorithms proposed for this setting, followed by a line of improvement (Honda and Takemura, 2011; Kaufmann et al., 2012; Capp e et al., 2013; Kaufmann et al., 2016). Qin (2022) states the optimal fixed budget best-arm identification as an open problem. In the setting of fixed confidence, the algorithms (Kalyanakrishnan et al., 2012; Kaufmann and Kalyanakrishnan, 2013; Jamieson et al., 2014) work towards fewer number of arms to guarantee the given confidence of getting the best arm. Garivier and Kaufmann (2016) gives an optimal algorithm in terms of the minimum of arms to pull. Another extension beyond the setting of fixed budget or confidence is the PAC learning framework, where the target is to maximize the chance of getting an mostly best arm, with a tolerance of ε gap to the highest reward (Karnin et al., 2013; Jamieson et al., 2014; Chaudhuri and Kalyanakrishnan, 2019). LLM evaluation with multi-arm bandits. Concurrently, Shi et al. (2024) explore using multi-arm bandit for prompt optimization in a limited budget setting. Although there are similarities in the approach, it is fair to say that their work differs substantially in focus. In contrast, our main goal is to speed up model evaluation rather than prompt optimization. Further, we also propose an algorithm that leverages low-rank factorization for better best-arm identification performance. We discuss additional related work on LLM performance evaluation functions and benchmarks in Appendix A, which are relevant but orthogonal to our problem. 6 DISCUSSION AND CONCLUSION In this work we focus on identifying the top-performer with a limited budget as it is a very common problem; covering a wide range of applications particularly for practitioners who simply want to find the best prompt or model for deployment purposes. One interesting direction for future work is recovering a full or top K ranking of performers, as it provides more fine-grained information that is necessary for specific use cases such as iterating on LLM design, and tracking LLM ranking over time. Of course the opportunity for resource savings is less than the top-performer setting. In Figure 3, we evaluate our algorithms on NDCG@K as preliminary evidence supporting the possibility of resource saving recovering a top K ranking. We leave a more thorough study as future work. A limitation of our proposed algorithms is that we assume there is a two-dimensional fixed-size scoring matrix for active method-example selection. In some real-world applications, new rows or columns can be gradually incorporated, making the matrix size dynamic. In other scenarios such as Chatbot Arena, instead of being able to decide what examples to select, we can only select a pair of models to compare for a user-specified example. We leave studying how to efficiently select the best method in these settings as another direction for future work. In conclusion, we tackle the challenge of resource-constrained evaluation, where assessing methodexample pairs is costly in terms of money, compute, and time. Our goal is to identify the best method with high probability while staying within budget. We propose two sequential decision-making algorithms: UCB-E, inspired by the classic multi-arm bandit approach, which provides a theoretical guarantee on success probability, and UCB-E-LRF, which extends UCB-E by leveraging low-rank factorization to approximate the scoring matrix. Experiments show both outperform random sampling, and we identify conditions where UCB-E or UCB-E-LRF excels. Published as a conference paper at ICLR 2025 7 REPRODUCIBILITY STATEMENT Our code is available at https://github.com/kilian-group/banditeval. ACKNOWLEDGEMENT We thank Justin Lovelace, Varsha Kishore, and Katie Luo for their helpful discussion and feedback on the paper draft. JPZ is supported by a grant from the Natural Sciences and Engineering Research Council of Canada (NSERC) (567916). CKB and CPG are supported by Schmidt Sciences programs, an AI2050 Senior Fellowship and two Eric and Wendy Schmidt AI in Science Postdoctoral Fellowships; the National Science Foundation (NSF) including the NSF Research Traineeship program; the National Institute of Food and Agriculture; and the Air Force Office of Scientific Research. RW is supported by grants from NSF (CIF-2402817, CNS-1804829), Sa TC-2241100, CCF-2217058, ARO-MURI (W911NF2110317), and ONR under N00014-24-1-2304. WS is supported by NSF IIS-2154711, NSF CAREER 2339395 and DARPA LANCER: Le Arning Network Cyb ERagents. This research is also supported by grants from the NSF (IIS-2107161, and IIS-1724282, HDR2118310), the Cornell Center for Materials Research with funding from the NSF MRSEC program (DMR-1719875), DARPA, ar Xiv, Linked In, and the New York Presbyterian Hospital. Winogrande: An adversarial winograd schema challenge at scale. 2019. J.-Y. Audibert and S. Bubeck. Best arm identification in multi-armed bandits. In COLT-23th Conference on learning theory-2010, pages 13 p, 2010. Y. Bisk, R. Zellers, R. L. Bras, J. Gao, and Y. Choi. Piqa: Reasoning about physical commonsense in natural language. In Thirty-Fourth AAAI Conference on Artificial Intelligence, 2020. O. r. Bojar, R. Chatterjee, C. Federmann, Y. Graham, B. Haddow, M. Huck, A. Jimeno Yepes, P. Koehn, V. Logacheva, C. Monz, M. Negri, A. Neveol, M. Neves, M. Popel, M. Post, R. Rubino, C. Scarton, L. Specia, M. Turchi, K. Verspoor, and M. Zampieri. Findings of the 2016 conference on machine translation. In Proceedings of the First Conference on Machine Translation, pages 131 198, Berlin, Germany, August 2016. Association for Computational Linguistics. URL http://www.aclweb.org/anthology/W/W16/W16-2301. T. Brown, B. Mann, N. Ryder, M. Subbiah, J. D. Kaplan, P. Dhariwal, A. Neelakantan, P. Shyam, G. Sastry, A. Askell, et al. Language models are few-shot learners. Advances in neural information processing systems, 33:1877 1901, 2020. S. Bubeck, R. Munos, and G. Stoltz. Pure exploration in multi-armed bandits problems. In Algorithmic Learning Theory: 20th International Conference, ALT 2009, Porto, Portugal, October 3-5, 2009. Proceedings 20, pages 23 37. Springer, 2009. C. Cai, G. Li, H. V. Poor, and Y. Chen. Nonconvex low-rank tensor completion from noisy data. Advances in neural information processing systems, 32, 2019. E. Candes and B. Recht. Exact matrix completion via convex optimization. Communications of the ACM, 55(6):111 119, 2012. E. J. Cand es and T. Tao. The power of convex relaxation: Near-optimal matrix completion. IEEE transactions on information theory, 56(5):2053 2080, 2010. O. Capp e, A. Garivier, O.-A. Maillard, R. Munos, and G. Stoltz. Kullback-leibler upper confidence bounds for optimal sequential allocation. The Annals of Statistics, pages 1516 1541, 2013. Y. Chang, X. Wang, J. Wang, Y. Wu, L. Yang, K. Zhu, H. Chen, X. Yi, C. Wang, Y. Wang, et al. A survey on evaluation of large language models. ACM Transactions on Intelligent Systems and Technology, 15(3):1 45, 2024. A. R. Chaudhuri and S. Kalyanakrishnan. Pac identification of many good arms in stochastic multi-armed bandits. In International Conference on Machine Learning, pages 991 1000. PMLR, 2019. Published as a conference paper at ICLR 2025 Y. Chen, Y. Chi, J. Fan, C. Ma, and Y. Yan. Noisy matrix completion: Understanding statistical guarantees for convex relaxation via nonconvex optimization. SIAM journal on optimization, 30 (4):3098 3121, 2020. Y. Chi, Y. M. Lu, and Y. Chen. Nonconvex optimization meets low-rank matrix factorization: An overview. IEEE Transactions on Signal Processing, 67(20):5239 5269, 2019. W.-L. Chiang, L. Zheng, Y. Sheng, A. N. Angelopoulos, T. Li, D. Li, H. Zhang, B. Zhu, M. Jordan, J. E. Gonzalez, et al. Chatbot arena: An open platform for evaluating llms by human preference. ar Xiv preprint ar Xiv:2403.04132, 2024. A. Chowdhery, S. Narang, J. Devlin, M. Bosma, G. Mishra, A. Roberts, P. Barham, H. W. Chung, C. Sutton, S. Gehrmann, et al. Palm: Scaling language modeling with pathways. Journal of Machine Learning Research, 24(240):1 113, 2023. P. Clark, I. Cowhey, O. Etzioni, T. Khot, A. Sabharwal, C. Schoenick, and O. Tafjord. Think you have solved question answering? try arc, the ai2 reasoning challenge. ar Xiv preprint ar Xiv:1803.05457, 2018. K. Cobbe, V. Kosaraju, M. Bavarian, M. Chen, H. Jun, L. Kaiser, M. Plappert, J. Tworek, J. Hilton, R. Nakano, C. Hesse, and J. Schulman. Training verifiers to solve math word problems. ar Xiv preprint ar Xiv:2110.14168, 2021. P. Dasigi, K. Lo, I. Beltagy, A. Cohan, N. A. Smith, and M. Gardner. A dataset of information-seeking questions and answers anchored in research papers. 2021. M. Dell, J. Carlson, T. Bryan, E. Silcock, A. Arora, Z. Shen, L. D Amico-Wong, Q. Le, P. Querubin, and L. Heldring. American stories: A large-scale structured text dataset of historical u.s. newspapers, 2023. J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova. Bert: Pre-training of deep bidirectional transformers for language understanding. ar Xiv preprint ar Xiv:1810.04805, 2018. J. Dhamala, T. Sun, V. Kumar, S. Krishna, Y. Pruksachatkun, K.-W. Chang, and R. Gupta. Bold: Dataset and metrics for measuring biases in open-ended language generation. In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, FAcc T 21, page 862 872, New York, NY, USA, 2021. Association for Computing Machinery. ISBN 9781450383097. doi: 10.1145/3442188.3445924. URL https://doi.org/10.1145/3442188.3445924. B. Efron and R. J. Tibshirani. An introduction to the bootstrap. Chapman and Hall/CRC, 1994. A. Garivier and E. Kaufmann. Optimal best arm identification with fixed confidence. In Conference on Learning Theory, pages 998 1027. PMLR, 2016. M. Grusky, M. Naaman, and Y. Artzi. Newsroom: A dataset of 1.3 million summaries with diverse extractive strategies. ar Xiv preprint ar Xi v:1804.11283, 2018. T. Hastie, R. Mazumder, J. D. Lee, and R. Zadeh. Matrix completion and low-rank svd via fast alternating least squares. The Journal of Machine Learning Research, 16(1):3367 3402, 2015. D. Hendrycks, S. Basart, S. Kadavath, M. Mazeika, A. Arora, E. Guo, C. Burns, S. Puranik, H. He, D. Song, and J. Steinhardt. Measuring coding challenge competence with apps, 2021a. D. Hendrycks, C. Burns, S. Basart, A. Zou, M. Mazeika, D. Song, and J. Steinhardt. Measuring massive multitask language understanding. Proceedings of the International Conference on Learning Representations (ICLR), 2021b. D. Hendrycks, C. Burns, S. Kadavath, A. Arora, S. Basart, E. Tang, D. Song, and J. Steinhardt. Measuring mathematical problem solving with the math dataset. ar Xiv preprint ar Xiv:2103.03874, 2021c. W. Hoeffding. Probability inequalities for sums of bounded random variables. The collected works of Wassily Hoeffding, pages 409 426, 1994. Published as a conference paper at ICLR 2025 J. Honda and A. Takemura. An asymptotically optimal policy for finite support models in the multiarmed bandit problem. Machine Learning, 85:361 391, 2011. K. Jamieson, M. Malloy, R. Nowak, and S. Bubeck. lil ucb: An optimal exploration algorithm for multi-armed bandits. In Conference on Learning Theory, pages 423 439. PMLR, 2014. A. Q. Jiang, A. Sablayrolles, A. Mensch, C. Bamford, D. S. Chaplot, D. de las Casas, F. Bressand, G. Lengyel, G. Lample, L. Saulnier, L. R. Lavaud, M.-A. Lachaux, P. Stock, T. L. Scao, T. Lavril, T. Wang, T. Lacroix, and W. E. Sayed. Mistral 7b, 2023. Q. Jin, B. Dhingra, Z. Liu, W. W. Cohen, and X. Lu. Pubmedqa: A dataset for biomedical research question answering. ar Xiv preprint ar Xiv:1909.06146, 2019. M. Joshi, E. Choi, D. Weld, and L. Zettlemoyer. triviaqa: A Large Scale Distantly Supervised Challenge Dataset for Reading Comprehension. ar Xiv e-prints, art. ar Xiv:1705.03551, 2017. S. Kalyanakrishnan, A. Tewari, P. Auer, and P. Stone. Pac subset selection in stochastic multi-armed bandits. In ICML, volume 12, pages 655 662, 2012. Z. Karnin, T. Koren, and O. Somekh. Almost optimal exploration in multi-armed bandits. In International conference on machine learning, pages 1238 1246. PMLR, 2013. E. Kaufmann and S. Kalyanakrishnan. Information complexity in bandit subset selection. In Conference on Learning Theory, pages 228 251. PMLR, 2013. E. Kaufmann, O. Capp e, and A. Garivier. On bayesian upper confidence bounds for bandit problems. In Artificial intelligence and statistics, pages 592 600. PMLR, 2012. E. Kaufmann, O. Capp e, and A. Garivier. On the complexity of best-arm identification in multi-armed bandit models. The Journal of Machine Learning Research, 17(1):1 42, 2016. T. Kojima, S. S. Gu, M. Reid, Y. Matsuo, and Y. Iwasawa. Large language models are zero-shot reasoners. Advances in neural information processing systems, 35:22199 22213, 2022. T. Kwiatkowski, J. Palomaki, O. Redfield, M. Collins, A. Parikh, C. Alberti, D. Epstein, I. Polosukhin, M. Kelcey, J. Devlin, K. Lee, K. N. Toutanova, L. Jones, M.-W. Chang, A. Dai, J. Uszkoreit, Q. Le, and S. Petrov. Natural questions: a benchmark for question answering research. Transactions of the Association of Computational Linguistics, 2019. N. Lambert, V. Pyatkin, J. Morrison, L. Miranda, B. Y. Lin, K. Chandu, N. Dziri, S. Kumar, T. Zick, Y. Choi, N. A. Smith, and H. Hajishirzi. Rewardbench: Evaluating reward models for language modeling, 2024. A. Lewkowycz, A. Andreassen, D. Dohan, E. Dyer, H. Michalewski, V. Ramasesh, A. Slone, C. Anil, I. Schlag, T. Gutman-Solo, et al. Solving quantitative reasoning problems with language models. Advances in Neural Information Processing Systems, 35:3843 3857, 2022. X. Li, T. Zhang, Y. Dubois, R. Taori, I. Gulrajani, C. Guestrin, P. Liang, and T. B. Hashimoto. Alpacaeval: An automatic evaluator of instruction-following models. https://github.com/ tatsu-lab/alpaca_eval, 2023. W. Liang, Y. Zhang, Z. Wu, H. Lepp, W. Ji, X. Zhao, H. Cao, S. Liu, S. He, Z. Huang, et al. Mapping the increasing use of llms in scientific papers. ar Xiv preprint ar Xiv:2404.01268, 2024. C.-Y. Lin. Rouge: A package for automatic evaluation of summaries. In Text summarization branches out, pages 74 81, 2004. Q. Mc Nemar. Note on the sampling error of the difference between correlated proportions or percentages. Psychometrika, 12(2):153 157, 1947. R. Nallapati, B. Zhou, C. Gulcehre, B. Xiang, et al. Abstractive text summarization using sequenceto-sequence rnns and beyond. ar Xiv preprint ar Xiv:1602.06023, 2016. S. Narayan, S. B. Cohen, and M. Lapata. Don t give me the details, just the summary! topic-aware convolutional neural networks for extreme summarization. ar Xiv preprint ar Xiv:1808.08745, 2018. Published as a conference paper at ICLR 2025 K. Papineni, S. Roukos, T. Ward, and W.-J. Zhu. Bleu: a method for automatic evaluation of machine translation. In Proceedings of the 40th annual meeting of the Association for Computational Linguistics, pages 311 318, 2002. C. Qin. Open problem: Optimal best arm identification with fixed-budget. In Conference on Learning Theory, pages 5650 5654. PMLR, 2022. P. Rajpurkar, J. Zhang, K. Lopyrev, and P. Liang. SQu AD: 100,000+ questions for machine comprehension of text. In J. Su, K. Duh, and X. Carreras, editors, Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, pages 2383 2392, Austin, Texas, Nov. 2016. Association for Computational Linguistics. doi: 10.18653/v1/D16-1264. URL https://aclanthology.org/D16-1264. D. Rein, B. L. Hou, A. C. Stickland, J. Petty, R. Y. Pang, J. Dirani, J. Michael, and S. R. Bowman. Gpqa: A graduate-level google-proof q&a benchmark, 2023. V. Sahayak, V. Shete, and A. Pathan. Sentiment analysis on twitter data. International Journal of Innovative Research in Advanced Engineering (IJIRAE), 2(1):178 183, 2015. T. Sellam, D. Das, and A. P. Parikh. Bleurt: Learning robust metrics for text generation. ar Xiv preprint ar Xiv:2004.04696, 2020. C. Shi, K. Yang, J. Yang, and C. Shen. Best arm identification for prompt learning under a limited budget. ar Xiv preprint ar Xiv:2402.09723, 2024. S. Singh, F. Vargus, D. Dsouza, B. F. Karlsson, A. Mahendiran, W.-Y. Ko, H. Shandilya, J. Patel, D. Mataciunas, L. OMahony, M. Zhang, R. Hettiarachchi, J. Wilson, M. Machado, L. S. Moura, D. Krzemi nski, H. Fadaei, I. Erg un, I. Okoh, A. Alaagib, O. Mudannayake, Z. Alyafeai, V. M. Chien, S. Ruder, S. Guthikonda, E. A. Alghamdi, S. Gehrmann, N. Muennighoff, M. Bartolo, J. Kreutzer, A. Ust un, M. Fadaee, and S. Hooker. Aya dataset: An open-access collection for multilingual instruction tuning, 2024. A. Wang, A. Singh, J. Michael, F. Hill, O. Levy, and S. R. Bowman. GLUE: A multi-task benchmark and analysis platform for natural language understanding. 2019. In the Proceedings of ICLR. Y. Wang, H. Ivison, P. Dasigi, J. Hessel, T. Khot, K. R. Chandu, D. Wadden, K. Mac Millan, N. A. Smith, I. Beltagy, and H. Hajishirzi. How far can camels go? exploring the state of instruction tuning on open resources, 2023. C. Yang, X. Wang, Y. Lu, H. Liu, Q. V. Le, D. Zhou, and X. Chen. Large language models as optimizers. ar Xiv preprint ar Xiv:2309.03409, 2023. Y. Yang, W.-t. Yih, and C. Meek. Wikiqa: A challenge dataset for open-domain question answering. In Proceedings of the 2015 conference on empirical methods in natural language processing, pages 2013 2018, 2015. Z. Yang, P. Qi, S. Zhang, Y. Bengio, W. W. Cohen, R. Salakhutdinov, and C. D. Manning. Hotpot QA: A dataset for diverse, explainable multi-hop question answering. In Conference on Empirical Methods in Natural Language Processing (EMNLP), 2018. W. Yuan, G. Neubig, and P. Liu. Bartscore: Evaluating generated text as text generation. Advances in Neural Information Processing Systems, 34:27263 27277, 2021. R. Zellers, A. Holtzman, Y. Bisk, A. Farhadi, and Y. Choi. Hellaswag: Can a machine really finish your sentence? ar Xiv preprint ar Xiv:1905.07830, 2019. T. Zhang, V. Kishore, F. Wu, K. Q. Weinberger, and Y. Artzi. Bertscore: Evaluating text generation with bert. ar Xiv preprint ar Xiv:1904.09675, 2019. L. Zheng, W.-L. Chiang, Y. Sheng, S. Zhuang, Z. Wu, Y. Zhuang, Z. Lin, Z. Li, D. Li, E. Xing, et al. Judging llm-as-a-judge with mt-bench and chatbot arena. Advances in Neural Information Processing Systems, 36, 2024. A. M. Zoubir and B. Boashash. The bootstrap and its application in signal processing. IEEE signal processing magazine, 15(1):56 76, 1998. Published as a conference paper at ICLR 2025 A FURTHER DISCUSSION ON RELATED WORK Low-rank factorization for (noisy) matrix completion. As shown in the objective function of Equation 1, low-rank factorization is a non-convex optimization problem. Therefore a line of work focus on how to solve this non-convex optimization (Cand es and Tao, 2010; Candes and Recht, 2012; Chi et al., 2019), while another line of work (Chen et al., 2020; Cai et al., 2019) study the approximation error between the estimated low-rank matrix and the target matrix in terms of p when assuming the observations are i.i.d. sampled with a pre-assumed chance p and the additive noise to each observation is i.i.d. Gaussian noise. LLM performance evaluation functions. Tremendous effort has been devoted to developing effective evaluation functions to assess the quality of open-ended generations from language models. Early work in this direction such as BLEU (Papineni et al., 2002) and ROUGE (Lin, 2004) are rule-based that use lexical overlaps to quantify similarity between a generated response and reference. However, lexical overlaps may not align well with the underlying semantics of the text. The shortcomings of these rule-based evaluation functions motivated a line of work studying using language models (Zhang et al., 2019; Sellam et al., 2020; Yuan et al., 2021) to evaluate generations. A seminal work in this area is BERTScore which uses embeddings from a BERT model (Devlin et al., 2018) to compute similarity. More recently, LLM-as-a-Judge (Zheng et al., 2024) proposes using instruction-tuned large language models to evaluate generations. Zheng et al. (2024) finds that costly proprietary models such as GPT4 have high agreement rate with human. LLM performance evaluation benchmarks. Diverse benchmarks have been developed to evaluate LLM performance across various domains including natural language understanding on translations and sentiment analysis (Bojar et al., 2016; Sahayak et al., 2015; Singh et al., 2024), mathematical and common sense reasoning (Cobbe et al., 2021; Hendrycks et al., 2021c; Bisk et al., 2020), text retrieval and question answering (Yang et al., 2015; Jin et al., 2019; Dell et al., 2023). We refer readers to Chang et al. (2024) for a more comprehensive discussion on existing benchmarks. The commonality of these benchmarks is that a ground truth answer is typically given and once the responses from LLMs are generated, they can be evaluated fairly efficiently without incurring large amount of overhead such as money and compute. Recently, a new type of benchmark or more precisely leaderboard has been created attempting to compare various LLMs to each other. The typical setup involves providing the same prompt to two LLMs and then having their responses compared side by side with another LLM, which determines which response is considered superior and thus wins. Two notable leaderboards are Alpaca Eval (Li et al., 2023) and Chatbot Arena (Chiang et al., 2024). Alpaca Eval has a static prompt dataset and for each prompt and LLM, the response is compared with that of a baseline model (GPT3 or GPT4). For Chatbot Arena, the prompts are submitted by users and two random LLMs are selected to respond to the user-written prompt. For both benchmarks, to aggregate the overall performance of LLMs, average win rate or ELO score can be calculated from the pairwise comparison statistics. Since the evaluation requires using another LLM to decide the better response, the evaluation incurs additional compute and / or monetary cost than the benchmarks that evaluate models based on ground truth answers. B DERIVATION OF COROLLARY 1 By mostly following the proof of the original theorem of UCB-E (Section 6.2 in Audibert and Bubeck (2010)), we only need to show that Oij Sobs ij Pn j=1 Oij a Pn j=1 Oij This is equivalent to prove if there are p random variables {X1, , Xp} sampling without replacement from the finite set {Si1, , Sin}, This can be implied by the analogous Hoeffding inequality (Hoeffding, 1994) that works for the sum of random variables sampling without replacement. Published as a conference paper at ICLR 2025 C ADDITIONAL DETAILS - DATASETS GSM8K Prompts [H1=107445] PIQA Prompts [H1=66284] GSM8K Models [H1=20562] PIQA Models [H1=10273] Alpaca Eval (Drop Annotator) [H1=4462] 1 25 50 75 100 125 150 175 200 Method Rank Alpaca Eval [H1=966] Figure 5: Bar plot of model performance on all examples in all six datasets. Examples are ordered by their rank. Dataset H1 values are displayed next to dataset name. It can be seen that datasets, such as Alpaca Eval, that have a large gap between the best and second best method have small H1 values. The H1 value indicates the difficulty of identifying the best method with smaller values indicating easier identification. Alpaca Eval and Alpaca Eval (Drop Annotator). Alpaca Eval benchmarks various models on a fixed set of 805 questions. First, for each model, the responses to these questions are collected. Then, each response is compared against that of a baseline model (GPT4-turbo) by using an LLM judge (in this case, GPT4-turbo as well). We collected the Alpaca Eval dataset on May 20, 2024 from the official repository 1. At that time, there were in total 154 models benchmarked, and hence our dataset size is 154 805. To remove the bias of favoring its own responses, we also create a derived dataset called Alpaca Eval (Drop Annotator) where we drop the responses of the annotator model, which leads to a size of 153 805. GSM8K Prompts and PIQA Prompts. To simulate a prompt engineering use case, we create two datasets: GSM8K Prompts and PIQA Prompts. For these datasets, we ask GPT4 to generate 205 and 177 prompts following the prompt engineering work from Yang et al. (2023). The LLM used to perform inference on the datasets are Mistra-7B Jiang et al. (2023) and Tulu-7B Wang et al. (2023) respectively. We evaluate the prompts on 784 and 1546 questions on the training set (about 10% of the size of the two training sets). For scoring, we extract the final answers from the responses generated by the LLMs and compare them with the ground truth answers (for GSM8K) or ground truth choice (for PIQA). GSM8K Models and PIQA Models. Lastly, for the model selection and hyperparameter tuning use case, we similarly create two more datasets on GSM8K and PIQA. Instead of each method being a prompt, each method is now a combination of LLM with different sampling configurations. Specifically, we use 11 public available models: GPT2 2, GPT2-Large 3, Code LLa MA 4, Tulu-7B 1https://github.com/tatsu-lab/alpaca eval 2https://huggingface.co/openai-community/gpt2 3https://huggingface.co/openai-community/gpt2-large 4https://huggingface.co/codellama/Code Llama-7b-python-hf Published as a conference paper at ICLR 2025 5, Tulu-2-7B 6, Gemma-7B 7, Phi2 8, Llema-7B 9, LLa MA-2-7B 10, Mistral-7B 11 and Star Coder7B 12. We also have three temperature choices {0, 0.5, 1}, two maximum decoding length choices {128, 512} and two zero-shot prompt choices (directly asking for the answer and Let s think step by step). The Cartesian product of all these choices give in total 132 different combinations as our methods. Depending on the dataset, some of these configurations experienced out-of-memory error on a Nvidia 3090 when we collected our data which we drop to simulate real-world scenarios. For the examples, we randomly select 1000 questions from each dataset. Table 3: Ratio of ith singular value to the largest singular value of S in each setting. σi represents the ith largest singular value of S. The ratio σr+1 σ1 can be interpreted as the relative reconstruction error with the best rank-r approximation of S. For many datasets, the small numerical differences between the first row and subsequent rows suggest that using rank-1 approximation for UCB-E-LRF is appropriate. Singular Value Ratio GSM8K Prompts PIQA Prompts GSM8K Models PIQA Models Alpaca Eval (Drop Annotator) Alpaca Eval σ2/σ1 0.1647 0.0889 0.3328 0.1972 0.3500 0.3661 σ3/σ1 0.1588 0.0763 0.2667 0.1758 0.2153 0.2120 σ4/σ1 0.1538 0.0739 0.2611 0.1715 0.1746 0.1754 GSM8K Prompts PIQA Prompts GSM8K Models PIQA Models Alpaca Eval (Drop Annotator) 1 25 50 75 100 125 150 175 200 Principal Component Alpaca Eval Figure 6: Bar plot of the explained variance of each principal component for all six datasets. The first principal component is much more pronounced than other principal components, suggesting the dataset matrices are generally low-rank. D ADDITIONAL DETAILS - ALGORITHMS The algorithms for Row Mean Imputation, Filled Subset, LRF and UCB-E-LRF (Score Only) are shown in Algorithm 3, 4, 5 and 6 respectively. 5https://huggingface.co/The Bloke/tulu-7B-fp16 6https://huggingface.co/allenai/tulu-2-7b 7https://huggingface.co/google/gemma-7b 8https://huggingface.co/microsoft/phi-2 9https://huggingface.co/Eleuther AI/llemma 7b 10https://huggingface.co/meta-llama/Llama-2-7b-chat-hf 11https://huggingface.co/mistralai/Mistral-7B-v0.1 12https://huggingface.co/bigcode/starcoder2-7b Published as a conference paper at ICLR 2025 10% 100% Percentage of Matrix Evaluated Performance GSM8K Prompts 10% 100% Percentage of Matrix Evaluated Alpaca Eval (Drop Annotator) Exploration Parameter a 0.1 0.5 1 5 10 Figure 7: Ablation of the exploration parameter a of UCB-E, Algorithm 1. Algorithm 3 Row Mean Imputation (Armi(T, F, X)) Input: The evaluation budget T, a set of methods F, a set of examples X. Output: The prediction ˆi for best method i . 1: The observation matrix O := {0}m n, the observed scoring matrix Sobs := {?}m n. 2: for t = 1, , T do 3: Select: Draw uniformly at random (i, j) {[m] [n]|Oij = 0}. 4: Evaluate: Run inference for the method-example pair (fi, xj), score the result, and receive s(fi(xj)); Sobs ij s(fi(xj)). 5: Update: Oij 1; 6: end for Return: ˆi = arg maxi Pn j=1 Oij Sobs ij Pn j=1 Oij . Algorithm 4 Filled Subset (Afs(T, F, X)) Input: The evaluation budget T, a set of methods F, a set of examples X. Output: The prediction ˆi for best method i . 1: The observation matrix O := {0}m n, the observed scoring matrix Sobs := {?}m n, initial random example index j. 2: for t = 1, , T do 3: Select: Draw uniformly at random i {[m]|Oij = 0}. 4: Evaluate: Run inference for the method-example pair (fi, xj), score the result, and receive s(fi(xj)); Sobs ij s(fi(xj)). 5: Update: Oij 1; If Pm i=1 Oij = m, draw uniformly at random j {[n]| Pm i=1 Oij = 0}. 6: end for Return: ˆi = arg maxi Pn j=1 Oij Sobs ij Pn j=1 Oij . Published as a conference paper at ICLR 2025 Algorithm 5 LRF (Alrf(T, F, X; M, r, C, T0)) Input: The evaluation budget T, a set of methods F, a set of examples X, the low-rank factorization model M, the rank r, the ensemble size C, the warm-up budget T0. Output: The prediction ˆi for best method i . 1: Uniformly draw T0 method-example pairs from [m] [n] and get the observation matrix O {0, 1}m n and observed scoring matrix Sobs ([0, 1] {?})m n w.r.t. these T0 evaluations. 2: for t = T0, , T do 3: Select: Draw uniformly at random (i, j) {[m] [n]|Oij = 0}. 4: Evaluate: Run inference for the method-example pair (fi, xj), score the result, and receive s(fi(xj)); Sobs ij s(fi(xj)). 5: Update: Oij 1; ˆS M(Sobs, O; r, C). 6: end for Return: ˆi = arg maxi 1 n Pn j=1(Oij Sobs ij + (1 Oij) ˆSij). Algorithm 6 UCB-E-LRF (Score Only) (Auel so(T, F, X; M, r, C, T0, η)) Input: The evaluation budget T, a set of methods F, a set of examples X, the low-rank factorization model M, the rank r, the ensemble size C, the warm-up budget T0, the uncertainty scaling η. Output: The prediction ˆi for best method i . 1: Uniformly draw T0 method-example pairs from [m] [n] and get the observation matrix O {0, 1}m n and observed scoring matrix Sobs ([0, 1] {?})m n w.r.t. these T0 evaluations. 2: ˆS, R M(Sobs, O; r, C); fi F, Bi 1 n Pn j=1(Oij Sobs ij + (1 Oij) ˆSij + ηRij) 3: for t = T0, , T do 4: Select: Draw i arg maxk | (Pn j=1 Okj) =n Bk; Draw uniformly at random j {k [n]|Oik = 0}. 5: Evaluate: Run inference for the method-example pair (fi, xj), score the result, and receive s(fi(xj)); Sobs ij s(fi(xj)). 6: Update: Oij 1; ˆS, R M(Sobs, O; r, C); fi F, Bi 1 n Pn j=1(Oij Sobs ij + (1 Oij) ˆSij + ηRij). 7: end for Return: ˆi = arg maxi 1 n Pn j=1(Oij Sobs ij + (1 Oij) ˆSij). Published as a conference paper at ICLR 2025 E TABLE OF NOTATIONS The notations used in the paper are described below. Table 4: Notations used in the paper. Symbol Description F = {f1, . . . , fm} A list of LLM-based methods X = {x1, . . . , xn} A dataset of examples s A scoring function S ([0, 1] {?})m n A scoring matrix computed as Sij := s(fi(xj)) µi The average score of a method fi on all examples as 1 n Pn j=1 Sij µ i The average score of the best method T An evaluation budget with T method-example queries A(T, F, X) An algorithm that predicts the best method among F on X with T query budget Sobs ([0, 1] {?})m n A partially observed scoring matrix O {0, 1}m n An observation matrix where 1 means the corresponding entry in S has been observed Bi An upper confidence bound for a method fi a A hyperparameter for UCB-E algorithm that balances exploration and exploitation H1 = Pm i=1,i =i 1 (µi µ i )2 The dataset hardness quantity U Rm r, V Rn r The low-rank factorization of Sobs such that Sobs UV ˆS Rm n An estimated scoring matrix ˆS = UV T for UCB-E-LRF ˆµi An estimated average score of a method fi for UCB-E-LRF C The ensemble size for UCB-E-LRF R Rm n The uncertainty matrix for UCB-E-LRF T0 The warmup budget for UCB-E-LRF r The rank of low-rank factorization M(Sobs, O; r, C) The low-rank factorization model that outputs ˆS and R for UCB-E-LRF η The uncertainty scaling factor for UCB-E-LRF