# better_instructionfollowing_through_minimum_bayes_risk__c3b95ffe.pdf Published as a conference paper at ICLR 2025 BETTER INSTRUCTION-FOLLOWING THROUGH MINIMUM BAYES RISK Ian Wu1 Patrick Fernandes2,3,4 Amanda Bertsch2 Seungone Kim2 Sina Pakazad1 Graham Neubig2 1C3 AI 2Carnegie Mellon University 3Instituto de Telecomunicações 4Instituto Superior Técnico, University of Lisbon {ian.wu, sina.pakazad}@c3.ai {pfernand, gneubig}@cs.cmu.edu {abertsch, seungonk}@andrew.cmu.edu General-purpose LLM judges capable of human-level evaluation provide not only a scalable and accurate way of evaluating instruction-following LLMs but also new avenues for supervising and improving their performance. One promising way of leveraging LLM judges for supervision is through Minimum Bayes Risk (MBR) decoding, which uses a reference-based evaluator to select a high-quality output from amongst a set of candidate outputs. In the first part of this work, we explore using MBR decoding as a method for improving the test-time performance of instruction-following LLMs. We find that MBR decoding with reference-based LLM judges substantially improves over greedy decoding, best-of-N decoding with reference-free judges and MBR decoding with lexical and embedding-based metrics on Alpaca Eval and MT-Bench. These gains are consistent across LLMs with up to 70B parameters, demonstrating that smaller LLM judges can be used to supervise much larger LLMs. Then, seeking to retain the improvements from MBR decoding while mitigating additional test-time costs, we explore iterative self-training on MBR-decoded outputs. We find that self-training using Direct Preference Optimisation leads to significant performance gains, such that the selftrained models with greedy decoding generally match and sometimes exceed the performance of their base models with MBR decoding. 1 INTRODUCTION Instruction-following large language models (LLMs) (Chung et al., 2022; Wei et al., 2022) have shown remarkable potential as generalist problem-solvers, prompting extensive efforts to improve their performance. One task that has seen tremendous progress due to LLMs is the evaluation of text generation itself. Recent works find that LLM-as-a-Judge frameworks (Zheng et al., 2023; Li et al., 2023; Dubois et al., 2024a) demonstrate strong correlation with human evaluations and significantly outperform lexical (Lin, 2004b; Papineni et al., 2002) and embedding-based methods (Zhang et al.; Yuan et al., 2021; Qin et al., 2023) across a wide range of instruction-following tasks. While the use of LLM judges has largely focused on the evaluation of outputs, LLM judges can also provide a way to supervise the generations of other LLMs. This generally involves using the judge as a reference-free evaluator to score candidate outputs produced by the LLM and then selecting the highest-scoring candidate as the final output in what is known as best-of-N (Bo N) decoding (Song et al., 2024). However, prior works find that LLM judges, including powerful proprietary LLMs such as GPT-4, significantly underperform when no human-curated reference answer is available (Ye et al., 2024; Zheng et al., 2023). In contrast, reference-based evaluation, where a human-curated reference answer is available, shows significantly higher correlation with human evaluations of outputs. This poses a chicken-and-egg problem: how can we leverage reference-based LLM judges for test time generation if no human references are available? Minimum Bayes Risk (MBR) decoding (Bickel & Doksum, 1977) provides a way of overcoming this problem. In place of the inaccessible human references, MBR decoding leverages other candidate Published as a conference paper at ICLR 2025 Max five-word description of a black hole, for a 5-year-old. Region of ultra dense spacetime. Big hole in space. Dark hole in space. Pseudo-References Marginalise Decoded Output Dark hole in space. Big hole in space. Region of ultra dense spacetime. Dark hole in space. Section 4: MBR Distillation Section 3: MBR Inference Figure 1: Illustration of MBR decoding. Multiple candidates are first sampled from an LLM. Then, each candidate is evaluated against all other candidates (pseudo-references) using a reference-based evaluator. The pseudo-references are marginalised to produce final scores, and the candidate with the highest score is selected. outputs as pseudo-references, and uses the evaluator, also known as the utility metric, to conduct reference-based evaluation of all candidate outputs against all pseudo-references. The final output is then chosen as the candidate output with the highest average score: see Figure 1. In this work, we explore whether MBR decoding using LLM judges as utility metrics can be used to enhance instruction-following LLMs. We divide our work into two main parts. First, inspired by the effectiveness of scaling inference time compute (Welleck et al., 2024; Snell et al., 2024), we investigate whether MBR decoding with LLM judges can improve the performance of instructionfollowing LLMs during test time (denoted as MBR inference ) (Section 3). Second, following recent works demonstrating that iterative self-training can improve LLM performance (Xu et al., 2023; Chen et al., 2024; Yuan et al., 2024a; Wu et al., 2024), we examine whether MBR decoding with LLM judges can be used to select high-quality model outputs for use in subsequent iterations of self-training (denoted as MBR distillation ) (Section 4). This both provides a way of training models without access to external labels and also allows us to mitigate the inference-time costs associated with MBR inference. From our MBR inference experiments, we find that MBR decoding with LLM judge Prometheus2-7B (Kim et al., 2024) improves performance by +3.6% on Alpaca Eval and +0.28 on MT-Bench on average across five LLMs relative to greedy decoding. Notably, Llama-2-7b with Prometheus MBR decoding outperforms Llama-2-13b with greedy decoding on MT-Bench, while Prometheus MBR decoding with Llama-2-13b outperforms Llama-2-70b with greedy decoding on Alpaca Eval 2.0. Gains persist even for large 70B models, demonstrating that small LLM judges can supervise larger LLMs through MBR decoding. We also compare MBR decoding against other methods that use LLM judges for supervision. We show that Prometheus MBR decoding is far more effective than MBR decoding with word match-based metrics (e.g. ROUGE) or semantic similarity-based metrics (e.g. BERTScore). Comparing MBR to Bo N decoding, we find that MBR decoding consistently outperforms Bo N decoding across multiple LLMs and LLM judges, and that the gains from MBR decoding increase as the supervising LLM judge increases in size and ability. From our MBR distillation experiments, we find that self-training with Direct Preference Optimisation (DPO) (Rafailov et al., 2024) on preference pairs selected using MBR decoding (Yang et al., 2024) with Prometheus-2-7B substantially improves greedy decoding performance. For instance, MBR self-trained Llama-2-13b improves by +7.1% on Alpaca Eval 2.0 and +0.90 on MT-Bench relative to its baseline SFT counterpart when evaluated using only greedy decoding, far surpassing the corresponding gains from Bo N self-training. We also find that MBR self-trained models evaluated with greedy decoding generally match and sometimes exceed the performance of their base models evaluated with MBR decoding, thereby demonstrating that MBR distillation is an effective way of mitigating the inference-time costs of MBR decoding while retaining improved performance. 2 BACKGROUND Language models are autoregressive probabilistic models; i.e., the probability of a token yi depends on prompt x and all previous tokens in the sequence: i 1 ppyi|yi 1, . . . , y1, xq. (1) Published as a conference paper at ICLR 2025 During inference, outputs are typically obtained either using maximum a-posteriori-based decoding methods that attempt to maximise probability, such as greedy decoding (yi arg maxyi ppyi|yăi, xq) and beam search (Graves, 2012), or by tokenwise sampling from the distribution (yi ppyi|yăi, xq). Both rely on the model s distribution as indicative of output quality. Alternatively, we can first obtain a hypothesis set Hhyp comprising Ncand candidate outputs from the model (for example, by sampling multiple times), and then select the final output from Hhyp based on some external criteria. For example, given some reference-free evaluator u (e.g. an LLM judge), best-of-N (Bo N) decoding selects the output ˆy P Hhyp such that ˆy arg max y PHhyp upyq. (2) As reference-free estimation of output quality can be a difficult problem, MBR decoding replaces the reference-free evaluator with a reference-based evaluator upy, y q (e.g. a reference-based LLM judge) that evaluates candidate y relative to a reference y .1 In the MBR literature, this evaluator is known as a utility metric (Freitag et al., 2022; Fernandes et al., 2022; Finkelstein et al., 2024). MBR decoding selects the final output ˆy that maximises expected utility under the model distribution: ˆy arg max y PHhyp Ey ppy|xqrupy, y qs looooooooooomooooooooooon 1 Ncand řNcand j 1 upy, ypjqq where the expectation is approximated as a Monte Carlo sum using model samples yp1q, . . . , yp Ncandq ppy|xq. In practice, this amounts to computing the utility of each candidate in Hhyp using all other candidates as (pseudo-)references, and then selecting the candidate with the highest average utility as the final output2 - see Appendix I.1 for an algorithmic description of MBR decoding. The MBR-decoded output can therefore be interpreted as being the candidate with the highest consensus utility as measured by the utility metric, as it achieves the highest average utility when evaluated against all other candidate outputs. It is therefore crucial to choose a reference-based metric that is a good proxy for human preferences as our utility function, as this ensures that a high-consensus output corresponds to a high-quality output. 3 MBR INFERENCE In this experiment, we investigate using MBR decoding with LLM judge utility metrics to improve instruction-following LLMs at test time. 3.1 EXPERIMENTAL SETUP 3.1.1 MODELS AND GENERATION PARAMETERS We use the chat and instruct variants of the Llama-2 (Touvron et al., 2023b) and Llama-3 (Dubey et al., 2024) models in this experiment. All models have undergone prior SFT and demonstrate strong instruction-following and conversation abilities. We generate Ncand 30 candidates using temperature sampling with t 0.3 for all MBR decoding experiments unless otherwise specified. 3.1.2 MBR UTILITY METRICS LLM judge We choose Prometheus-2-7B (Kim et al., 2024) as our representative reference-based LLM judge. Prometheus is a specialist judge model finetuned from Mistral-7b (Jiang et al., 2023) that correlates strongly with human judges and GPT-4. It takes as inputs a task prompt, a scoring rubric (see Appendix C), a candidate and a reference, and outputs an explanation of its judgement followed by a score from 1 to 5, which we interpret as a utility score. Crucially, Prometheus can also act as a reference-free judge by simply omitting the reference from its input. This allows us to directly compare MBR with Bo N decoding using the same LLM judge utility metric. 1Certain evaluators (e.g. LLM judges) are task-aware , and take prompt x as an input when performing evaluation. Such utility metrics can then be written as upy; xq and upy, y ; xq. 2The expectation can also be computed over a separate set of model outputs known as the evidence set (Eikema & Aziz, 2020; Bertsch et al., 2023). We do not explore this setting in our work. Published as a conference paper at ICLR 2025 We compare using LLM judges for MBR decoding with three other classes of utility metrics: ROUGE ROUGE (Lin, 2004a) is a word-level F-measure designed for measuring summarisation and machine translation performance (Lin, 2004a). We use ROUGE-1 in our main experiments but include results for other ROUGE variants in Appendix A.3. BERTScore BERTScore is a neural evaluation metric that computes the token-level contextual similarity between a candidate and a reference. Like ROUGE, BERTScore is not task-aware. As our model outputs may be longer than 512 tokens, we use Longformer-large (Beltagy et al., 2020) as our BERTScore model, which supports inputs up to 4094 tokens long. Dense embedders Dense embedders generate contextual embeddings of text passages for use in downstream tasks. One such task is measuring the level of semantic similarity between two text passages (Agirre et al., 2012). This task is directly relevant to MBR decoding, as we can treat pairs of candidates as text passages and their similarity score as the utility. To the best of our knowledge, using dense embedders as a utility metric for MBR decoding has never been explored before. In our work, we use the instruction-following embedder SFR-Embedder-2_R (Meng et al., 2024) as our representative dense embedder. We include results for two other dense embedders in Appendix A.3. 3.1.3 BASELINES In addition to greedy decoding and beam search (BS) with k 10 beams, we also experiment with LONGEST decoding, where we select the longest candidate from the hypothesis set (as measured in characters) as the final output, and best-of-N (Bo N) decoding. We generate Ncand 30 candidates using temperature sampling with t 0.3 for both longest and Bo N decoding. See Appendices A.2 and A.3 for comparisons with additional baselines. 3.1.4 EVALUATION Alpaca Eval 2.0 Alpaca Eval (Li et al., 2023) is an LLM-based evaluation metric. It consists of an 805-sample, highly diverse single-turn instruction-following conversational dataset and an associated evaluation framework. In Alpaca Eval 2.0 (Dubois et al., 2024b), evaluation is conducted by performing head-to-head comparison of candidate answers against GPT-4-generated answers facilitated by a judge LLM. The judge model is prompted to output a single token representing its choice of winner, with the log-probabilities of the token used to compute a weighted win rate. In addition to standard win rates, Alpaca Eval 2.0 also provides length-controlled (LC) win rates, which are debiased versions of the standard win rates that control for the length of the outputs. Both the Alpaca Eval standard and LC evaluation demonstrate strong correlation with human judgements. MT-Bench MT-Bench (Zheng et al., 2023) is an 80-sample, two-turn instruction-following conversational dataset. It can be evaluated using either head-to-head comparison or direct assessment with an LLM judge. In the direct assessment setting, the judge LLM is prompted to generate an explanation followed by a score between 1 and 10, with no reference answer used. MT-Bench with GPT-4-judge matches crowdsourced human preferences well, achieving over 80% agreement, which is the same level of agreement between human evaluators (Zheng et al., 2023). We use GPT-4o (Open AI et al., 2024) as the LLM judge for both Alpaca Eval 2.0 and MT-Bench. For Alpaca Eval, we report LC win rates unless otherwise stated. For MT-Bench, we use direct assessment for all experiments. See Appendix B for further details on our evaluation strategy, and Appendix H for human study findings verifying the alignment of our automatic LLM evaluation results with human judgements. 3.2 EXPERIMENTAL RESULTS Our main experimental results are documented in Tables 1 and 2. Prometheus MBR decoding provides significant and consistent gains The gains associated with Prometheus MBR decoding are significantly larger than those associated with other utility metrics, yielding an average improvement of +3.6% on Alpaca Eval 2.0 and +0.28 on MT-Bench. For comparison, the performance gap between Llama-2-7b and Llama-2-13b with greedy decoding is +4.6% on Alpaca Eval 2.0 and +0.18 on MT-Bench, while the corresponding gap between Llama-2-13b and Llama-2-70b is +3.8% and +0.60. Notably, Llama-2-7b with Prometheus MBR Published as a conference paper at ICLR 2025 2-7B 2-13B 2-70B 3-8B 3-70B Avg. Greedy 14.4 19.0 22.8 34.4 42.7 0 BS 14.8 18.2 21.5 33.9 42.4 -0.50 Longest 10.5 15.2 19.8 29.8 40.4 -3.51 Prometheus Bo N 16.4 20.8 25.0 35.5 44.3 1.74 ROUGE MBR 16.2 20.0 24.7 35.4 43.7 1.33 BERTScore MBR 16.2 20.5 24.4 35.7 44.0 1.50 SFR-Embedder MBR 12.1 16.6 22.2 32.5 42.8 -1.42 Prometheus MBR 17.7 23.4 26.2 37.9 46.0 3.62 Table 1: Alpaca Eval 2.0 win rates (%) for various models and decoding strategies, along with the average win rate differences compared to greedy decoding across all models (denoted as Avg. ). MBR decoding with Prometheus consistently outperforms all baseline methods and other MBR decoding methods. 2-7B 2-13B 2-70B 3-8B 3-70B Avg. Greedy 5.72 5.90 6.50 7.54 8.29 0 BS 5.58 5.95 6.49 7.30 8.20 -0.09 Longest 5.67 6.03 6.59 7.22 8.22 -0.04 Prometheus Bo N 5.77 6.08 6.65 7.66 8.42 0.13 ROUGE MBR 5.78 6.11 6.68 7.63 8.31 0.11 BERTScore MBR 5.68 6.02 6.72 7.52 8.42 0.08 SFR-Embedder MBR 5.73 6.04 6.54 7.45 8.33 0.03 Prometheus MBR 6.10 6.26 6.79 7.69 8.50 0.28 Table 2: MT-Bench scores for various models and decoding strategies, along with the average score differences compared to greedy decoding across all models (denoted as Avg. ). MBR decoding with Prometheus consistently outperforms all baseline methods and other MBR decoding methods. decoding outperforms Llama-2-13b with greedy decoding on MT-Bench, while Prometheus MBR decoding with Llama-2-13b outperforms Llama-2-70b - a model over five times bigger - with greedy decoding on Alpaca Eval 2.0. We also find that Prometheus MBR decoding yields larger gains than Prometheus Bo N decoding; we explore this further in Section 3.3.1. We also highlight that the performance gains associated with Prometheus MBR decoding are significant across models of all sizes, even for much larger models such as Llama-3-70b. This scaling property suggests that small judge models can still be used to supervise much larger models. ROUGE and BERTScore MBR decoding provide small but consistent gains ROUGE and BERTScore MBR decoding improve average performance relative to greedy decoding by +1.3% and +1.5% on Alpaca Eval 2.0 and by +0.11 and +0.08 on MT-Bench respectively. This benefit is present for all models. This improvement suggests that selecting outputs without awareness of the task and using only wordor token-level measures of consistency can still yield meaningful improvements even in the instruction-following setting. SFR-Embedder MBR decoding fails to yield consistent gains SFR-Embedder MBR decoding reduces performance relative to greedy decoding by -1.4% on Alpaca Eval 2.0 while improving performance by +0.03 on MT-Bench on average. We hypothesise that embedder models, which are trained to distinguish at a high level between text passages, cannot to detect nuanced differences between semantically-similar outputs. We also note that embedder MBR decoding generally selects for longer outputs, which may explain the discrepancy between its performance on Alpaca Eval 2.0 (which is length-controlled) and MT-Bench. See Appendix A.4 for analysis on the generation lengths of various decoding strategies. Beam search and LONGEST decoding degrade performance Beam search and LONGEST decoding reduce performance relative to greedy decoding by -0.5% and -3.5% on Alpaca Eval 2.0 and -0.09 and -0.04 on MT-Bench respectively. The poor performance of beam search further underscores the idea that optimising for output probability alone is not enough to improve output quality. 3.3 ANALYSIS OF PROMETHEUS MBR DECODING Given the promise shown by Prometheus MBR decoding, we conduct additional experiments to better understand its properties. Published as a conference paper at ICLR 2025 0 5 10 15 20 25 30 Ncand Win Rate vs. GPT-4 0.0 0.2 0.4 0.6 0.8 1.0 1.2 1.4 t Win Rate vs. GPT-4 Figure 2: Alpaca Eval 2.0 win rates (%) for Llama-2-70b with varying hypothesis set size Ncand (left) and generation temperature t (right) values for Prometheus MBR and Bo N decoding. Performance for both methods initially increases with Ncand and plateaus at around Ncand 20. Performance also initially increases with t, but drops rapidly after t 1.0. Prometheus MBR decoding performance vs. t and Ncand We plot Alpaca Eval 2.0 win rates as a function of Ncand and t in Figure 2 for Llama-2-70b with Prometheus MBR and Bo N decoding. We find that performance initially increases with Ncand but plateaus at around Ncand 20, suggesting that expanding the size of the hypothesis set beyond this yields little benefit. Performance also initially increases with t, highlighting the benefits of increased candidate diversity, although it rapidly degrades at high temperatures as the individual candidate outputs decline in quality. Technical writing Recommendations Creative writing Mathematical reasoning Argumentation Puzzles & logical reasoning Win Rate vs. GPT-4 Figure 3: Difference in Alpaca Eval 2.0 win rates (%) between Prometheus MBR decoding and greedy decoding averaged over all five LLMs and broken down by question category. A positive value indicates that MBR decoding outperforms greedy decoding on the given category. Orange bars represent the standard error. We find that Prometheus MBR decoding improves performance across a wide range of question categories. Prometheus MBR decoding performance by question category We classified questions from the Alpaca Eval dataset into one of ten categories using GPT-4o (see Appendix D for details), and then computed the differences in win rates between Prometheus MBR decoding and greedy decoding by question category, averaged over all five LLMs. We find that MBR decoding improves output quality across most question categories. These include reasoning-based categories such as coding and mathematical reasoning, although the largest improvements are seen across writingbased categories such as technical, recommendations and creative writing. We hypothesise that this discrepancy arises due to (1) the higher bar for correctness associated with reasoning tasks which limits the number of good answers that can be found amongst candidate outputs; and (2) limitations of existing utility functions, which may struggle to handle difficult reasoning tasks. 3.3.1 FURTHER COMPARISONS WITH BEST-OF-N DECODING As Bo N decoding can also leverage LLM judges as a utility metric, we conduct additional experiments to compare its performance against MBR decoding. We compare Bo N and MBR decoding for five different LLM judges on MT-Bench and report the results in Table 3. In addition to Prometheus-2-7B, we also evaluate its larger sibling Prometheus-2-8x7B, as well as Judge LM-7b and Judge LM-33b (Zhu et al., 2023c), which are two judge models finetuned from LLa MA models (Touvron et al., 2023a). We also assess Llama-3-8b-Instruct and Llama-3-70b-Instruct as zero- Published as a conference paper at ICLR 2025 2-7B 2-13B 2-70B 3-8B 3-70B Avg. Greedy 5.72 5.90 6.50 7.54 8.29 0 Prometheus-2-7B-Bo N 5.77 6.08 6.65 7.66 8.42 0.13 Prometheus-2-7B-MBR 6.10 6.26 6.79 7.69 8.50 0.28 Prometheus-2-8x7B-Bo N 6.01 6.17 6.80 7.75 8.41 0.24 Prometheus-2-8x7B-MBR 6.26 6.32 6.87 7.79 8.64 0.39 Judge LM-7b-Bo N 5.63 5.95 6.69 7.37 8.26 -0.01 Judge LM-7b-MBR 6.00 6.11 6.79 7.69 8.44 0.22 Judge LM-33b-Bo N 5.68 6.03 6.58 7.37 8.35 0.01 Judge LM-33b-MBR 5.94 6.27 6.88 7.92 8.50 0.31 Llama-3-8b-Instruct-Bo N 5.83 6.05 6.61 7.60 8.38 0.10 Llama-3-8b-Instruct-MBR 5.96 6.28 6.84 7.80 8.47 0.28 Llama-3-70b-Instruct-Bo N 5.77 6.16 6.57 7.39 8.35 0.06 Llama-3-70b-Instruct-MBR 6.22 6.43 6.94 7.87 8.52 0.41 Table 3: MT-Bench scores for Bo N and MBR decoding with various judge LLMs as utility metrics, along with the average score differences compared to greedy decoding across all models (denoted Avg. ). MBR decoding consistently outperforms Bo N decoding across all comparable utility metrics. shot judges for MBR decoding (see Appendix E for our prompts). All chosen judges can act as both reference-free and reference-based judges, allowing us to compare MBR and Bo N decoding fairly.3 We find that MBR decoding consistently outperforms Bo N decoding across all selected judge models. This difference is especially large for the Judge LM models and for Llama-3-70b-Instruct, where Bo N fails to significantly improve on greedy decoding. One explanation for this discrepancy is that our LLM judges are insufficiently good at reference-free evaluation for Bo N decoding to be effective. This idea is supported by prior studies comparing reference-free and reference-based evaluation, which consistently show that reference-free methods tend to underperform, even when using strong judge models like GPT-4 (Ye et al., 2024; Kim et al., 2024; Zheng et al., 2023). Another explanation is that MBR decoding provides a smoothing effect that arises from our use of expected utility in place of utility point estimates for output selection, tying back to our hypothesis that selecting high-consensus outputs yields significant benefit. This averaging process reduces the impact of individual mistakes made by the imperfect LLM judge, thereby providing for a more stable and reliable measure of quality. We leave further exploration of these ideas to future work. Notably, in Table 3, MBR performance improves by scaling the size of the LLM judge, with Prometheus-2-8x7B outperforming Prometheus-2-7B, Judge LM-33b outperforming Judge LM-7b, and Llama-3-70b-Instruct outperforming Llama-3-8b-Instruct. This suggests that improving the LLM judge utility metric directly improves MBR decoding performance and that MBR decoding will benefit as newer and better LLM judges are developed. 4 MBR DISTILLATION Our results so far demonstrate the potential of MBR decoding to significantly improve the test-time performance of instruction-following models, but this comes at the cost of substantial inference-time compute costs due to the linear cost for generating Ncand candidate outputs and the quadratic cost for computing utility across these candidates. To mitigate this, we explore distilling MBR-decoded outputs back into the model itself and aim to obtain MBR decoding-level (or better) performance without needing to perform MBR decoding at test time. 4.1 EXPERIMENTAL SETUP 4.1.1 TRAINING AN SFT MODEL We start by performing SFT on the base Llama-2-7b and Llama-2-13b models. This is necessary to instil instruction-following behaviour in the models so that they can be used to generate instructionfollowing self-training data. We choose not to use the official chat variants of these models as we 3Because sequence-classifier reward models (Stiennon et al., 2022) do not support reference-based evaluation, it is not possible to fairly compare Bo N decoding with these methods to MBR. We therefore do not discuss this in the main text and report our findings in Appendix F instead. Published as a conference paper at ICLR 2025 wish to retain control over the training procedure and avoid inheriting any biases introduced through prior finetuning and alignment. We use 3000 random samples from Ultra Chat (Ding et al., 2023) for SFT. Ultra Chat is a diverse conversational instruction-following dataset created using GPT-3.5Turbo. Each sample consists of multi-turn prompts and responses, although we only take the first turn of each sample in order to simplify experimentation. We designate our SFT models as sft. 4.1.2 ITERATIVE DPO ON MBR-DECODED OUTPUTS Having obtained SFT models, we now conduct DPO to improve the models on their own MBRdecoded outputs, an approach first proposed by Yang et al. (2024) for improving machine translation. We start by randomly drawing a further 3000 prompts from Ultra Chat (excluding the samples that have already been selected for SFT). Next, we generate Ncand 12 candidate outputs from our sft models using these prompts. We use a smaller Ncand than for MBR inference to balance performance and compute cost, as we know from Figure 2 that using an Ncand value above 10 already yields significant gains. Following Yang et al. (2024), we then score the candidate outputs using a utility metric and form preference pairs from the highest-scoring and lowest-scoring outputs. This preference pair dataset is then used for DPO on the sft models, yielding dpo-MBR-1. We extend upon Yang et al. (2024) by iteratively repeating this process twice more, each time using the latest dpo models as the base model paired with a fresh set of 3000 prompts, yielding the models dpo MBR-2 and dpo-MBR-3. See Appendix K for a summary of our SFT and DPO hyperparameters, Appendix G.5 for experimental results from using another preference pair selection strategy, and Appendix I.2 for mathematical and algorithmic overviews of MBR distillation. 4.1.3 UTILITY METRICS AND EVALUATION We use Prometheus-2-7B as our utility metric, although we also try MBR self-training with ROUGE and Llama-3-8b-Instruct as the utility metrics (Appendix G.3 and G.4). We compare our dpo models with greedy decoding against the sft models with greedy decoding, beam search and MBR decoding. For MBR decoding, we use Ncand 30 and t 0.3 with Prometheus-2-7B as the utility metric. We also baseline against models trained with SFT on 12000 Ultra Chat samples (sft-full). Finally, we experiment with Bo N self-training, again using Prometheus-2-7B as the utility metric and following the same procedure as for MBR self-training, which yields the models dpo-Bo N-1, dpo-Bo N-2 and dpo-Bo N-3. We evaluate our trained models using greedy decoding on Alpaca Eval 2.0, once again reporting length-controlled win rates vs. GPT-4, and MT-Bench. As we only train our models to engage in single-turn conversations we evaluate only on the first turn of MT-Bench. We report additional evaluation results in the Appendix, including head-to-head results between various self-trained models and sft with greedy decoding (Appendix G.1), evaluation results on a selection of popular NLP benchmarks (Appendix G.2), and human study results (Appendix H). 4.2 RESULTS Alpaca Eval 2.0 MT-Bench 7B 13B 7B 13B sft w. Greedy 5.18 8.24 5.43 5.85 sft w. MBR 9.99 13.6 5.78 6.31 sft-full 6.35 9.40 5.55 6.26 dpo-1-Bo N 5.78 10.3 5.78 6.08 dpo-2-Bo N 6.22 11.2 5.91 6.41 dpo-3-Bo N 6.40 12.8 5.88 6.56 dpo-1-MBR 5.68 10.8 5.78 6.48 dpo-2-MBR 7.22 13.9 6.11 6.73 dpo-3-MBR 8.86 15.3 6.14 6.75 Alpaca Eval 2.0 MT-Bench sft-1-MBR 5.52 5.48 sft-2-MBR 6.75 5.43 sft-3-MBR 6.48 5.51 Table 4: (Left) Alpaca Eval 2.0 win rates (%) and MT-Bench scores for models self-trained using DPO. After three rounds of training, the self-trained models consistently outperform their Bo N counterparts and SFT baselines. (Top) Alpaca Eval 2.0 win rates (%) and MT-Bench scores for models self-trained using SFT. Self-training with SFT yields substantially worse results than self-training with DPO. DPO self-training significantly improves model performance We report the results of our selftraining experiment in the left subtable of Table 4. We find that three rounds of MBR self-training with DPO significantly improves model performance, with the 7B dpo-3-MBR model outperforming Published as a conference paper at ICLR 2025 Argumentation Recommendations Creative writing Technical writing Puzzles & logical reasoning Mathematical reasoning Win Rate vs. GPT-4 Figure 4: Differences in Alpaca Eval 2.0 win rates (%) between dpo-3-MBR models and their respective sft with greedy decoding baselines on different question categories. The largest improvements are seen in open-ended writing tasks, with less improvement on reasoning-focussed tasks (e.g. mathematical reasoning and coding). 20 40 60 80 100 120 tokens / s Win Rate vs. sft + Greedy sft + Greedy sft + BS sft + Bo N sft + MBR dpo-3-MBR + Greedy sft + Greedy sft + BS sft + Bo N sft + MBR dpo-3-MBR + Greedy 0.0 Avg. Generation Time Avg. Decoding Time Figure 5: (Left) Alpaca Eval 2.0 win rate (%) vs. sft with greedy decoding against generation throughput. dpo-3-MBR with greedy decoding matches the performance of sft with MBR decoding, but with significantly higher throughput. (Right) Average generation and decoding times on the Alpaca Eval dataset. The decoding step in MBR decoding takes disproportionately long. This problem can mitigated through MBR self-training. the 13B sft model with greedy decoding on both Alpaca Eval 2.0 and MT-Bench. The improvements saturate by the third round of self-training as measured on MT-Bench (6.11 vs. 6.14 (7B) and 6.73 vs. 6.75 (13B) in Table 4), although there appears to be room for further improvement on Alpaca Eval 2.0 (7.22 vs. 8.86 (7B) and 13.9 vs. 15.3 (13B) in Table 4). Both dpo-3-MBR models outperform their sft-full counterparts, which suggests that training on MBR-decoded outputs is more beneficial than SFT on a larger split of Ultra Chat. The dpo-3-MBR models also generally outperform sft with MBR decoding, and this is especially prominent for MT-Bench, which suggests that DPO on MBRdecoded outputs enables models to recover and then exceed their MBR-decoding performances. We find that DPO on Bo N-decoded outputs also improves model performance, although less so than DPO with MBR-decoded outputs. We attribute this to the relative strength of MBR decoding. SFT self-training yields smaller gains than DPO self-training We experiment with iterative SFT self-training, using the 7B sft model. We document our results in the right subtable of Table 4. We use the same sets of prompts as for DPO and select as our SFT labels the highest-scoring sample as determined by MBR, following Finkelstein et al. (2024). As before, we conduct three rounds of iterative training, yielding sft-1-MBR, sft-2-MBR and sft-3-MBR. We find that SFT training yields significantly less improvement than DPO. This indicates that MBR self-training benefits most from preference learning, where the model learns to contrast its highestand lowest-quality outputs. DPO self-trained model performance by question category We repeat the analysis on performance by question category for dpo-3-MBR in Figure 4. Self-training improves performance on almost all question categories, with generally larger improvement on writing-based categories and smaller improvement on reasoning-based categories. We attribute this difference to the writingskewed distribution of question categories in our Ultra Chat training data (see Appendix G.6). Analysis of compute costs We illustrate the savings on compute introduced by self-training in Figure 5. We perform inference with various 7B models and decoding strategies on Alpaca Eval 2.0, using 2x A100 GPUs and v LLM (Kwon et al., 2023) as our inference engine. We use a generation batch size of 1, a LLM judge utility calculation batch size of 32, and Ncand 12. We find that MBR decoding imposes significant overhead, largely due to the quadratic Op N 2 candq cost incurred during Published as a conference paper at ICLR 2025 the utility calculation step. This overhead is removed through MBR self-training, which nonetheless retains performance gains. Note that dpo-3-MBR generates longer outputs than sft, which explains why its average generation time as seen in the right-hand plot of Figure 5 is higher. 5 RELATED WORK MBR decoding MBR decoding has been explored in the context of machine translation using a variety of translation metrics such as COMET (Rei et al., 2020) and BLEURT (Sellam et al., 2020), with promising results (Freitag et al., 2022; 2023; Farinhas et al., 2023; Stanojevi c & Sima an, 2014). Prior works (Bertsch et al., 2023; Jinnai et al., 2023) also study MBR decoding for summarisation, using ROUGE and BERTScore as metrics. Suzgun et al. (2022) apply MBR decoding to several tasks, including summarisation, machine translation and three different BIG-Bench tasks (Srivastava et al., 2023). None of these works explore the use of MBR decoding in the more openended instruction-following domain, nor do they consider using LLM judges as utility metrics. LLM judges Based on the strong instruction-following capabilities of LLMs, recent works explore prompting LLMs to judge responses from other LLMs (Li et al., 2023; Zheng et al., 2023). Follow-up works suggest that training on the evaluation traces of strong models may equip smaller models with strong evaluation capabilities (Kim et al., 2023; 2024; Zhu et al., 2023b; Vu et al., 2024; Wang et al., 2024b). These works focus on training LLMs to produce scoring decisions matching those of humans. In our work, instead of viewing evaluation as an end goal, we explore utilising the evaluation capabilities of LLM judges as supervision to improve instruction-following LLMs. Inference-time algorithms Many inference-time algorithms generate candidate outputs and select a final output based on external criteria. In addition to MBR and Bo N decoding, examples include Self-Consistency (Wang et al., 2023), which selects the most self-consistent answers through marginalisation of chain-of-thought reasoning paths and Universal Self-Consistency (USC) (Chen et al., 2023), where the LLM is used to self-select consistent chain-of-thought reasoning paths from amongst many reasoning paths. Kuhn et al. (2023) propose an MBR-esque algorithm that uses dense embedders and clustering to measure semantic uncertainty. Other inference-time algorithms prompt the LLM to perform additional inference steps in a structured manner. Examples include Tree-of-Thoughts (Yao et al., 2023) and Graph-of-Thoughts (Besta et al., 2024), as well as recursive improvement strategies such as Self-Refine (Madaan et al., 2023) and Reflexion (Shinn et al., 2023). Self-training Self-training is a promising avenue for model improvement as it enables training without labelled data. Gulcehre et al. (2023) introduce an algorithm that generates samples from a policy and then updates the policy using offline RL. Yuan et al. (2024b) train models to score their own outputs, and then use these scores to create preference datasets which for distillation. Huang et al. (2022) train models on their own highest-confidence outputs as determined by majority voting. Selftraining on MBR-decoded outputs has also been explored for machine translation. Finkelstein et al. (2024) train models with SFT on their own MBR and quality estimation outputs for machine translation and demonstrate that this yields improvements over baseline models. Wang et al. (2024a) use MBR to generate targets for sequence-level distillation, again for machine translation. Yang et al. (2024) are the first to use DPO to upweight the model s own MBR generations, allowing them to recover much of their original MBR performances on translation using only greedy decoding. 6 CONCLUSION In this work, we investigate using LLM judges to supervise other LLMs on instruction-following tasks through MBR decoding, and find that this yields significant and consistent improvements to model performance relative to greedy decoding, beam search and Bo N decoding. These benefits persist across a wide range of question categories and are also consistent across models of various sizes, demonstrating that small LLM judges can be used to improve much larger LLMs at inference time. To mitigate the significant inference-time costs associated with MBR decoding, we also explore iterative self-training on MBR-decoded outputs. We find that MBR self-training using DPO, but not SFT, enables models to recover and even exceed their base MBR decoding performance using only greedy decoding. We hope our work further highlights the potential of using LLM judges for supervision and inspires future research into MBR decoding beyond its traditional domains and applications, particularly through the development of new utility metrics. Published as a conference paper at ICLR 2025 REPRODUCIBILITY STATEMENT In an effort to make our work reproducible, we document all prompts (Appendices E and B), as well as training and inference hyperparameters (Appendix K) used throughout our experiments. We also include version information for all API-based LLMs (Appendix B), and choose to use open-source models (the Llama-2, Llama-3, Prometheus-2 and Judge LM families) where possible. ACKNOWLEDGEMENTS AB was supported by a grant from the National Science Foundation Graduate Research Fellowship Program under Grant No. DGE2140739. PF was supported by the Portuguese Recovery and Resilience Plan through project C645008882-00000055 (Center for Responsible AI) and by FCT/MECI through national funds and when applicable co-funded EU funds under UID/50008: Instituto de Telecomunicações. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the sponsors. Eneko Agirre, Daniel Cer, Mona Diab, and Aitor Gonzalez-Agirre. Sem Eval-2012 task 6: A pilot on semantic textual similarity. In Eneko Agirre, Johan Bos, Mona Diab, Suresh Manandhar, Yuval Marton, and Deniz Yuret (eds.), *SEM 2012: The First Joint Conference on Lexical and Computational Semantics Volume 1: Proceedings of the main conference and the shared task, and Volume 2: Proceedings of the Sixth International Workshop on Semantic Evaluation (Sem Eval 2012), pp. 385 393, Montréal, Canada, 7-8 June 2012. Association for Computational Linguistics. URL https://aclanthology.org/S12-1051. Iz Beltagy, Matthew E. Peters, and Arman Cohan. Longformer: The long-document transformer, 2020. URL https://arxiv.org/abs/2004.05150. Amanda Bertsch, Alex Xie, Graham Neubig, and Matthew R. Gormley. It s mbr all the way down: Modern generation techniques through the lens of minimum bayes risk, 2023. Maciej Besta, Nils Blach, Ales Kubicek, Robert Gerstenberger, Michal Podstawski, Lukas Gianinazzi, Joanna Gajda, Tomasz Lehmann, Hubert Niewiadomski, Piotr Nyczyk, and Torsten Hoefler. Graph of thoughts: Solving elaborate problems with large language models. Proceedings of the AAAI Conference on Artificial Intelligence, 38(16):17682 17690, March 2024. ISSN 2159-5399. doi: 10.1609/aaai.v38i16.29720. URL http://dx.doi.org/10.1609/ aaai.v38i16.29720. P.J. Bickel and K.A. Doksum. Mathematical Statistics: Basic Ideas and Selected Topics. Prentice Hall, 1977. ISBN 9780135641477. URL https://books.google.com.sg/books?id= uc Mf AQAAIAAJ. Xinyun Chen, Renat Aksitov, Uri Alon, Jie Ren, Kefan Xiao, Pengcheng Yin, Sushant Prakash, Charles Sutton, Xuezhi Wang, and Denny Zhou. Universal self-consistency for large language model generation, 2023. Zixiang Chen, Yihe Deng, Huizhuo Yuan, Kaixuan Ji, and Quanquan Gu. Self-play fine-tuning converts weak language models to strong language models, 2024. URL https://arxiv. org/abs/2401.01335. Hyung Won Chung, Le Hou, Shayne Longpre, Barret Zoph, Yi Tay, William Fedus, Yunxuan Li, Xuezhi Wang, Mostafa Dehghani, Siddhartha Brahma, Albert Webson, Shixiang Shane Gu, Zhuyun Dai, Mirac Suzgun, Xinyun Chen, Aakanksha Chowdhery, Alex Castro-Ros, Marie Pellat, Kevin Robinson, Dasha Valter, Sharan Narang, Gaurav Mishra, Adams Yu, Vincent Zhao, Yanping Huang, Andrew Dai, Hongkun Yu, Slav Petrov, Ed H. Chi, Jeff Dean, Jacob Devlin, Adam Roberts, Denny Zhou, Quoc V. Le, and Jason Wei. Scaling instruction-finetuned language models, 2022. URL https://arxiv.org/abs/2210.11416. Published as a conference paper at ICLR 2025 Peter Clark, Isaac Cowhey, Oren Etzioni, Tushar Khot, Ashish Sabharwal, Carissa Schoenick, and Oyvind Tafjord. Think you have solved question answering? try arc, the ai2 reasoning challenge, 2018. URL https://arxiv.org/abs/1803.05457. Ning Ding, Yulin Chen, Bokai Xu, Yujia Qin, Zhi Zheng, Shengding Hu, Zhiyuan Liu, Maosong Sun, and Bowen Zhou. Enhancing chat language models by scaling high-quality instructional conversations, 2023. URL https://arxiv.org/abs/2305.14233. Abhimanyu Dubey, Abhinav Jauhri, Abhinav Pandey, Abhishek Kadian, Ahmad Al-Dahle, Aiesha Letman, Akhil Mathur, Alan Schelten, Amy Yang, Angela Fan, Anirudh Goyal, Anthony Hartshorn, Aobo Yang, Archi Mitra, Archie Sravankumar, Artem Korenev, Arthur Hinsvark, Arun Rao, Aston Zhang, Aurelien Rodriguez, Austen Gregerson, Ava Spataru, Baptiste Roziere, Bethany Biron, Binh Tang, Bobbie Chern, Charlotte Caucheteux, Chaya Nayak, Chloe Bi, Chris Marra, Chris Mc Connell, Christian Keller, Christophe Touret, Chunyang Wu, Corinne Wong, Cristian Canton Ferrer, Cyrus Nikolaidis, Damien Allonsius, Daniel Song, Danielle Pintz, Danny Livshits, David Esiobu, Dhruv Choudhary, Dhruv Mahajan, Diego Garcia-Olano, Diego Perino, Dieuwke Hupkes, Egor Lakomkin, Ehab Al Badawy, Elina Lobanova, Emily Dinan, Eric Michael Smith, Filip Radenovic, Frank Zhang, Gabriel Synnaeve, Gabrielle Lee, Georgia Lewis Anderson, Graeme Nail, Gregoire Mialon, Guan Pang, Guillem Cucurell, Hailey Nguyen, Hannah Korevaar, Hu Xu, Hugo Touvron, Iliyan Zarov, Imanol Arrieta Ibarra, Isabel Kloumann, Ishan Misra, Ivan Evtimov, Jade Copet, Jaewon Lee, Jan Geffert, Jana Vranes, Jason Park, Jay Mahadeokar, Jeet Shah, Jelmer van der Linde, Jennifer Billock, Jenny Hong, Jenya Lee, Jeremy Fu, Jianfeng Chi, Jianyu Huang, Jiawen Liu, Jie Wang, Jiecao Yu, Joanna Bitton, Joe Spisak, Jongsoo Park, Joseph Rocca, Joshua Johnstun, Joshua Saxe, Junteng Jia, Kalyan Vasuden Alwala, Kartikeya Upasani, Kate Plawiak, Ke Li, Kenneth Heafield, Kevin Stone, Khalid El-Arini, Krithika Iyer, Kshitiz Malik, Kuenley Chiu, Kunal Bhalla, Lauren Rantala-Yeary, Laurens van der Maaten, Lawrence Chen, Liang Tan, Liz Jenkins, Louis Martin, Lovish Madaan, Lubo Malo, Lukas Blecher, Lukas Landzaat, Luke de Oliveira, Madeline Muzzi, Mahesh Pasupuleti, Mannat Singh, Manohar Paluri, Marcin Kardas, Mathew Oldham, Mathieu Rita, Maya Pavlova, Melanie Kambadur, Mike Lewis, Min Si, Mitesh Kumar Singh, Mona Hassan, Naman Goyal, Narjes Torabi, Nikolay Bashlykov, Nikolay Bogoychev, Niladri Chatterji, Olivier Duchenne, Onur Çelebi, Patrick Alrassy, Pengchuan Zhang, Pengwei Li, Petar Vasic, Peter Weng, Prajjwal Bhargava, Pratik Dubal, Praveen Krishnan, Punit Singh Koura, Puxin Xu, Qing He, Qingxiao Dong, Ragavan Srinivasan, Raj Ganapathy, Ramon Calderer, Ricardo Silveira Cabral, Robert Stojnic, Roberta Raileanu, Rohit Girdhar, Rohit Patel, Romain Sauvestre, Ronnie Polidoro, Roshan Sumbaly, Ross Taylor, Ruan Silva, Rui Hou, Rui Wang, Saghar Hosseini, Sahana Chennabasappa, Sanjay Singh, Sean Bell, Seohyun Sonia Kim, Sergey Edunov, Shaoliang Nie, Sharan Narang, Sharath Raparthy, Sheng Shen, Shengye Wan, Shruti Bhosale, Shun Zhang, Simon Vandenhende, Soumya Batra, Spencer Whitman, Sten Sootla, Stephane Collot, Suchin Gururangan, Sydney Borodinsky, Tamar Herman, Tara Fowler, Tarek Sheasha, Thomas Georgiou, Thomas Scialom, Tobias Speckbacher, Todor Mihaylov, Tong Xiao, Ujjwal Karn, Vedanuj Goswami, Vibhor Gupta, Vignesh Ramanathan, Viktor Kerkez, Vincent Gonguet, Virginie Do, Vish Vogeti, Vladan Petrovic, Weiwei Chu, Wenhan Xiong, Wenyin Fu, Whitney Meers, Xavier Martinet, Xiaodong Wang, Xiaoqing Ellen Tan, Xinfeng Xie, Xuchao Jia, Xuewei Wang, Yaelle Goldschlag, Yashesh Gaur, Yasmine Babaei, Yi Wen, Yiwen Song, Yuchen Zhang, Yue Li, Yuning Mao, Zacharie Delpierre Coudert, Zheng Yan, Zhengxing Chen, Zoe Papakipos, Aaditya Singh, Aaron Grattafiori, Abha Jain, Adam Kelsey, Adam Shajnfeld, Adithya Gangidi, Adolfo Victoria, Ahuva Goldstand, Ajay Menon, Ajay Sharma, Alex Boesenberg, Alex Vaughan, Alexei Baevski, Allie Feinstein, Amanda Kallet, Amit Sangani, Anam Yunus, Andrei Lupu, Andres Alvarado, Andrew Caples, Andrew Gu, Andrew Ho, Andrew Poulton, Andrew Ryan, Ankit Ramchandani, Annie Franco, Aparajita Saraf, Arkabandhu Chowdhury, Ashley Gabriel, Ashwin Bharambe, Assaf Eisenman, Azadeh Yazdan, Beau James, Ben Maurer, Benjamin Leonhardi, Bernie Huang, Beth Loyd, Beto De Paola, Bhargavi Paranjape, Bing Liu, Bo Wu, Boyu Ni, Braden Hancock, Bram Wasti, Brandon Spence, Brani Stojkovic, Brian Gamido, Britt Montalvo, Carl Parker, Carly Burton, Catalina Mejia, Changhan Wang, Changkyu Kim, Chao Zhou, Chester Hu, Ching-Hsiang Chu, Chris Cai, Chris Tindal, Christoph Feichtenhofer, Damon Civin, Dana Beaty, Daniel Kreymer, Daniel Li, Danny Wyatt, David Adkins, David Xu, Davide Testuggine, Delia David, Devi Parikh, Diana Liskovich, Didem Foss, Dingkang Wang, Duc Le, Dustin Holland, Edward Dowling, Eissa Jamil, Elaine Montgomery, Eleonora Presani, Emily Hahn, Emily Wood, Erik Brinkman, Esteban Arcaute, Evan Dunbar, Evan Smothers, Fei Sun, Felix Kreuk, Feng Tian, Firat Ozgenel, Francesco Published as a conference paper at ICLR 2025 Caggioni, Francisco Guzmán, Frank Kanayet, Frank Seide, Gabriela Medina Florez, Gabriella Schwarz, Gada Badeer, Georgia Swee, Gil Halpern, Govind Thattai, Grant Herman, Grigory Sizov, Guangyi, Zhang, Guna Lakshminarayanan, Hamid Shojanazeri, Han Zou, Hannah Wang, Hanwen Zha, Haroun Habeeb, Harrison Rudolph, Helen Suk, Henry Aspegren, Hunter Goldman, Ibrahim Damlaj, Igor Molybog, Igor Tufanov, Irina-Elena Veliche, Itai Gat, Jake Weissman, James Geboski, James Kohli, Japhet Asher, Jean-Baptiste Gaya, Jeff Marcus, Jeff Tang, Jennifer Chan, Jenny Zhen, Jeremy Reizenstein, Jeremy Teboul, Jessica Zhong, Jian Jin, Jingyi Yang, Joe Cummings, Jon Carvill, Jon Shepard, Jonathan Mc Phie, Jonathan Torres, Josh Ginsburg, Junjie Wang, Kai Wu, Kam Hou U, Karan Saxena, Karthik Prasad, Kartikay Khandelwal, Katayoun Zand, Kathy Matosich, Kaushik Veeraraghavan, Kelly Michelena, Keqian Li, Kun Huang, Kunal Chawla, Kushal Lakhotia, Kyle Huang, Lailin Chen, Lakshya Garg, Lavender A, Leandro Silva, Lee Bell, Lei Zhang, Liangpeng Guo, Licheng Yu, Liron Moshkovich, Luca Wehrstedt, Madian Khabsa, Manav Avalani, Manish Bhatt, Maria Tsimpoukelli, Martynas Mankus, Matan Hasson, Matthew Lennie, Matthias Reso, Maxim Groshev, Maxim Naumov, Maya Lathi, Meghan Keneally, Michael L. Seltzer, Michal Valko, Michelle Restrepo, Mihir Patel, Mik Vyatskov, Mikayel Samvelyan, Mike Clark, Mike Macey, Mike Wang, Miquel Jubert Hermoso, Mo Metanat, Mohammad Rastegari, Munish Bansal, Nandhini Santhanam, Natascha Parks, Natasha White, Navyata Bawa, Nayan Singhal, Nick Egebo, Nicolas Usunier, Nikolay Pavlovich Laptev, Ning Dong, Ning Zhang, Norman Cheng, Oleg Chernoguz, Olivia Hart, Omkar Salpekar, Ozlem Kalinli, Parkin Kent, Parth Parekh, Paul Saab, Pavan Balaji, Pedro Rittner, Philip Bontrager, Pierre Roux, Piotr Dollar, Polina Zvyagina, Prashant Ratanchandani, Pritish Yuvraj, Qian Liang, Rachad Alao, Rachel Rodriguez, Rafi Ayub, Raghotham Murthy, Raghu Nayani, Rahul Mitra, Raymond Li, Rebekkah Hogan, Robin Battey, Rocky Wang, Rohan Maheswari, Russ Howes, Ruty Rinott, Sai Jayesh Bondu, Samyak Datta, Sara Chugh, Sara Hunt, Sargun Dhillon, Sasha Sidorov, Satadru Pan, Saurabh Verma, Seiji Yamamoto, Sharadh Ramaswamy, Shaun Lindsay, Shaun Lindsay, Sheng Feng, Shenghao Lin, Shengxin Cindy Zha, Shiva Shankar, Shuqiang Zhang, Shuqiang Zhang, Sinong Wang, Sneha Agarwal, Soji Sajuyigbe, Soumith Chintala, Stephanie Max, Stephen Chen, Steve Kehoe, Steve Satterfield, Sudarshan Govindaprasad, Sumit Gupta, Sungmin Cho, Sunny Virk, Suraj Subramanian, Sy Choudhury, Sydney Goldman, Tal Remez, Tamar Glaser, Tamara Best, Thilo Kohler, Thomas Robinson, Tianhe Li, Tianjun Zhang, Tim Matthews, Timothy Chou, Tzook Shaked, Varun Vontimitta, Victoria Ajayi, Victoria Montanez, Vijai Mohan, Vinay Satish Kumar, Vishal Mangla, Vítor Albiero, Vlad Ionescu, Vlad Poenaru, Vlad Tiberiu Mihailescu, Vladimir Ivanov, Wei Li, Wenchen Wang, Wenwen Jiang, Wes Bouaziz, Will Constable, Xiaocheng Tang, Xiaofang Wang, Xiaojian Wu, Xiaolan Wang, Xide Xia, Xilun Wu, Xinbo Gao, Yanjun Chen, Ye Hu, Ye Jia, Ye Qi, Yenda Li, Yilin Zhang, Ying Zhang, Yossi Adi, Youngjin Nam, Yu, Wang, Yuchen Hao, Yundi Qian, Yuzi He, Zach Rait, Zachary De Vito, Zef Rosnbrick, Zhaoduo Wen, Zhenyu Yang, and Zhiwei Zhao. The llama 3 herd of models, 2024. URL https://arxiv.org/abs/2407.21783. Yann Dubois, Balázs Galambosi, Percy Liang, and Tatsunori B Hashimoto. Length-controlled alpacaeval: A simple way to debias automatic evaluators. ar Xiv preprint ar Xiv:2404.04475, 2024a. Yann Dubois, Balázs Galambosi, Percy Liang, and Tatsunori B. Hashimoto. Length-controlled alpacaeval: A simple way to debias automatic evaluators, 2024b. URL https://arxiv. org/abs/2404.04475. Bryan Eikema and Wilker Aziz. Is map decoding all you need? the inadequacy of the mode in neural machine translation, 2020. Angela Fan, Mike Lewis, and Yann Dauphin. Hierarchical neural story generation, 2018. URL https://arxiv.org/abs/1805.04833. António Farinhas, José de Souza, and Andre Martins. An empirical study of translation hypothesis ensembling with large language models. In Houda Bouamor, Juan Pino, and Kalika Bali (eds.), Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pp. 11956 11970, Singapore, December 2023. Association for Computational Linguistics. doi: 10.18653/v1/2023.emnlp-main.733. URL https://aclanthology.org/2023. emnlp-main.733. Patrick Fernandes, António Farinhas, Ricardo Rei, José G. C. de Souza, Perez Ogayo, Graham Neubig, and Andre Martins. Quality-aware decoding for neural machine translation. In Ma- Published as a conference paper at ICLR 2025 rine Carpuat, Marie-Catherine de Marneffe, and Ivan Vladimir Meza Ruiz (eds.), Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 1396 1412, Seattle, United States, July 2022. Association for Computational Linguistics. doi: 10.18653/v1/2022.naacl-main.100. URL https://aclanthology.org/2022.naacl-main.100. Mara Finkelstein, Subhajit Naskar, Mehdi Mirzazadeh, Apurva Shah, and Markus Freitag. Mbr and qe finetuning: Training-time distillation of the best and most expensive decoding methods, 2024. URL https://arxiv.org/abs/2309.10966. Rudolf Flesch. A new readability yardstick. Journal of Applied Psychology, 32(3):221 233, 1948. doi: 10.1037/h0057532. Markus Freitag, David Grangier, Qijun Tan, and Bowen Liang. High quality rather than high model probability: Minimum Bayes risk decoding with neural metrics. Transactions of the Association for Computational Linguistics, 10:811 825, 2022. doi: 10.1162/tacl_a_00491. URL https: //aclanthology.org/2022.tacl-1.47. Markus Freitag, Behrooz Ghorbani, and Patrick Fernandes. Epsilon sampling rocks: Investigating sampling strategies for minimum bayes risk decoding for machine translation, 2023. Alex Graves. Sequence transduction with recurrent neural networks, 2012. URL https: //arxiv.org/abs/1211.3711. Caglar Gulcehre, Tom Le Paine, Srivatsan Srinivasan, Ksenia Konyushkova, Lotte Weerts, Abhishek Sharma, Aditya Siddhant, Alex Ahern, Miaosen Wang, Chenjie Gu, Wolfgang Macherey, Arnaud Doucet, Orhan Firat, and Nando de Freitas. Reinforced self-training (rest) for language modeling, 2023. URL https://arxiv.org/abs/2308.08998. Dan Hendrycks, Collin Burns, Steven Basart, Andy Zou, Mantas Mazeika, Dawn Song, and Jacob Steinhardt. Measuring massive multitask language understanding, 2021. URL https: //arxiv.org/abs/2009.03300. Ari Holtzman, Jan Buys, Li Du, Maxwell Forbes, and Yejin Choi. The curious case of neural text degeneration, 2020. URL https://arxiv.org/abs/1904.09751. Jiaxin Huang, Shixiang Shane Gu, Le Hou, Yuexin Wu, Xuezhi Wang, Hongkun Yu, and Jiawei Han. Large language models can self-improve, 2022. URL https://arxiv.org/abs/ 2210.11610. Albert Q. Jiang, Alexandre Sablayrolles, Arthur Mensch, Chris Bamford, Devendra Singh Chaplot, Diego de las Casas, Florian Bressand, Gianna Lengyel, Guillaume Lample, Lucile Saulnier, Lélio Renard Lavaud, Marie-Anne Lachaux, Pierre Stock, Teven Le Scao, Thibaut Lavril, Thomas Wang, Timothée Lacroix, and William El Sayed. Mistral 7b, 2023. URL https: //arxiv.org/abs/2310.06825. Yuu Jinnai, Tetsuro Morimura, Ukyo Honda, Kaito Ariu, and Kenshi Abe. Model-based minimum bayes risk decoding, 2023. Seungone Kim, Jamin Shin, Yejin Cho, Joel Jang, Shayne Longpre, Hwaran Lee, Sangdoo Yun, Seongjin Shin, Sungdong Kim, James Thorne, et al. Prometheus: Inducing fine-grained evaluation capability in language models. In The Twelfth International Conference on Learning Representations, 2023. Seungone Kim, Juyoung Suk, Shayne Longpre, Bill Yuchen Lin, Jamin Shin, Sean Welleck, Graham Neubig, Moontae Lee, Kyungjae Lee, and Minjoon Seo. Prometheus 2: An open source language model specialized in evaluating other language models, 2024. Lorenz Kuhn, Yarin Gal, and Sebastian Farquhar. Semantic uncertainty: Linguistic invariances for uncertainty estimation in natural language generation. ar Xiv preprint ar Xiv:2302.09664, 2023. Published as a conference paper at ICLR 2025 Woosuk Kwon, Zhuohan Li, Siyuan Zhuang, Ying Sheng, Lianmin Zheng, Cody Hao Yu, Joseph E. Gonzalez, Hao Zhang, and Ion Stoica. Efficient memory management for large language model serving with pagedattention. In Proceedings of the ACM SIGOPS 29th Symposium on Operating Systems Principles, 2023. Nathan Lambert, Valentina Pyatkin, Jacob Morrison, LJ Miranda, Bill Yuchen Lin, Khyathi Chandu, Nouha Dziri, Sachin Kumar, Tom Zick, Yejin Choi, Noah A. Smith, and Hannaneh Hajishirzi. Rewardbench: Evaluating reward models for language modeling, 2024. URL https://arxiv.org/abs/2403.13787. Chankyu Lee, Rajarshi Roy, Mengyao Xu, Jonathan Raiman, Mohammad Shoeybi, Bryan Catanzaro, and Wei Ping. Nv-embed: Improved techniques for training llms as generalist embedding models, 2024. URL https://arxiv.org/abs/2405.17428. Yaniv Leviathan, Matan Kalman, and Yossi Matias. Fast inference from transformers via speculative decoding, 2023. URL https://arxiv.org/abs/2211.17192. Xuechen Li, Tianyi Zhang, Yann Dubois, Rohan Taori, Ishaan Gulrajani, Carlos Guestrin, Percy Liang, and Tatsunori B. Hashimoto. Alpacaeval: An automatic evaluator of instruction-following models. https://github.com/tatsu-lab/alpaca_eval, 5 2023. Chin-Yew Lin. ROUGE: A package for automatic evaluation of summaries. In Text Summarization Branches Out, pp. 74 81, Barcelona, Spain, July 2004a. Association for Computational Linguistics. URL https://aclanthology.org/W04-1013. Chin-Yew Lin. Rouge: A package for automatic evaluation of summaries. In Text summarization branches out, pp. 74 81, 2004b. Stephanie Lin, Jacob Hilton, and Owain Evans. Truthfulqa: Measuring how models mimic human falsehoods, 2022. URL https://arxiv.org/abs/2109.07958. Aman Madaan, Niket Tandon, Prakhar Gupta, Skyler Hallinan, Luyu Gao, Sarah Wiegreffe, Uri Alon, Nouha Dziri, Shrimai Prabhumoye, Yiming Yang, Shashank Gupta, Bodhisattwa Prasad Majumder, Katherine Hermann, Sean Welleck, Amir Yazdanbakhsh, and Peter Clark. Selfrefine: Iterative refinement with self-feedback, 2023. URL https://arxiv.org/abs/ 2303.17651. Rui Meng, Ye Liu, Shafiq Rayhan Joty, Caiming Xiong, Yingbo Zhou, and Semih Yavuz. Sfr-embedding-mistral:enhance text retrieval with transfer learning. Salesforce AI Research Blog, 2024. URL https://blog.salesforceairesearch.com/ sfr-embedded-mistral/. Niklas Muennighoff, Thomas Wang, Lintang Sutawika, Adam Roberts, Stella Biderman, Teven Le Scao, M Saiful Bari, Sheng Shen, Zheng-Xin Yong, Hailey Schoelkopf, Xiangru Tang, Dragomir Radev, Alham Fikri Aji, Khalid Almubarak, Samuel Albanie, Zaid Alyafeai, Albert Webson, Edward Raff, and Colin Raffel. Crosslingual generalization through multitask finetuning, 2023. URL https://arxiv.org/abs/2211.01786. Zach Nussbaum, John X. Morris, Brandon Duderstadt, and Andriy Mulyar. Nomic embed: Training a reproducible long context text embedder, 2024. URL https://arxiv.org/abs/2402. 01613. Open AI, Josh Achiam, Steven Adler, Sandhini Agarwal, Lama Ahmad, Ilge Akkaya, Florencia Leoni Aleman, Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, Red Avila, Igor Babuschkin, Suchir Balaji, Valerie Balcom, Paul Baltescu, Haiming Bao, Mohammad Bavarian, Jeff Belgum, Irwan Bello, Jake Berdine, Gabriel Bernadett-Shapiro, Christopher Berner, Lenny Bogdonoff, Oleg Boiko, Madelaine Boyd, Anna-Luisa Brakman, Greg Brockman, Tim Brooks, Miles Brundage, Kevin Button, Trevor Cai, Rosie Campbell, Andrew Cann, Brittany Carey, Chelsea Carlson, Rory Carmichael, Brooke Chan, Che Chang, Fotis Chantzis, Derek Chen, Sully Chen, Ruby Chen, Jason Chen, Mark Chen, Ben Chess, Chester Cho, Casey Chu, Hyung Won Chung, Dave Cummings, Jeremiah Currier, Yunxing Dai, Cory Decareaux, Thomas Degry, Noah Deutsch, Damien Deville, Arka Dhar, David Dohan, Steve Dowling, Sheila Published as a conference paper at ICLR 2025 Dunning, Adrien Ecoffet, Atty Eleti, Tyna Eloundou, David Farhi, Liam Fedus, Niko Felix, Simón Posada Fishman, Juston Forte, Isabella Fulford, Leo Gao, Elie Georges, Christian Gibson, Vik Goel, Tarun Gogineni, Gabriel Goh, Rapha Gontijo-Lopes, Jonathan Gordon, Morgan Grafstein, Scott Gray, Ryan Greene, Joshua Gross, Shixiang Shane Gu, Yufei Guo, Chris Hallacy, Jesse Han, Jeff Harris, Yuchen He, Mike Heaton, Johannes Heidecke, Chris Hesse, Alan Hickey, Wade Hickey, Peter Hoeschele, Brandon Houghton, Kenny Hsu, Shengli Hu, Xin Hu, Joost Huizinga, Shantanu Jain, Shawn Jain, Joanne Jang, Angela Jiang, Roger Jiang, Haozhun Jin, Denny Jin, Shino Jomoto, Billie Jonn, Heewoo Jun, Tomer Kaftan, Łukasz Kaiser, Ali Kamali, Ingmar Kanitscheider, Nitish Shirish Keskar, Tabarak Khan, Logan Kilpatrick, Jong Wook Kim, Christina Kim, Yongjik Kim, Jan Hendrik Kirchner, Jamie Kiros, Matt Knight, Daniel Kokotajlo, Łukasz Kondraciuk, Andrew Kondrich, Aris Konstantinidis, Kyle Kosic, Gretchen Krueger, Vishal Kuo, Michael Lampe, Ikai Lan, Teddy Lee, Jan Leike, Jade Leung, Daniel Levy, Chak Ming Li, Rachel Lim, Molly Lin, Stephanie Lin, Mateusz Litwin, Theresa Lopez, Ryan Lowe, Patricia Lue, Anna Makanju, Kim Malfacini, Sam Manning, Todor Markov, Yaniv Markovski, Bianca Martin, Katie Mayer, Andrew Mayne, Bob Mc Grew, Scott Mayer Mc Kinney, Christine Mc Leavey, Paul Mc Millan, Jake Mc Neil, David Medina, Aalok Mehta, Jacob Menick, Luke Metz, Andrey Mishchenko, Pamela Mishkin, Vinnie Monaco, Evan Morikawa, Daniel Mossing, Tong Mu, Mira Murati, Oleg Murk, David Mély, Ashvin Nair, Reiichiro Nakano, Rajeev Nayak, Arvind Neelakantan, Richard Ngo, Hyeonwoo Noh, Long Ouyang, Cullen O Keefe, Jakub Pachocki, Alex Paino, Joe Palermo, Ashley Pantuliano, Giambattista Parascandolo, Joel Parish, Emy Parparita, Alex Passos, Mikhail Pavlov, Andrew Peng, Adam Perelman, Filipe de Avila Belbute Peres, Michael Petrov, Henrique Ponde de Oliveira Pinto, Michael, Pokorny, Michelle Pokrass, Vitchyr H. Pong, Tolly Powell, Alethea Power, Boris Power, Elizabeth Proehl, Raul Puri, Alec Radford, Jack Rae, Aditya Ramesh, Cameron Raymond, Francis Real, Kendra Rimbach, Carl Ross, Bob Rotsted, Henri Roussez, Nick Ryder, Mario Saltarelli, Ted Sanders, Shibani Santurkar, Girish Sastry, Heather Schmidt, David Schnurr, John Schulman, Daniel Selsam, Kyla Sheppard, Toki Sherbakov, Jessica Shieh, Sarah Shoker, Pranav Shyam, Szymon Sidor, Eric Sigler, Maddie Simens, Jordan Sitkin, Katarina Slama, Ian Sohl, Benjamin Sokolowsky, Yang Song, Natalie Staudacher, Felipe Petroski Such, Natalie Summers, Ilya Sutskever, Jie Tang, Nikolas Tezak, Madeleine B. Thompson, Phil Tillet, Amin Tootoonchian, Elizabeth Tseng, Preston Tuggle, Nick Turley, Jerry Tworek, Juan Felipe Cerón Uribe, Andrea Vallone, Arun Vijayvergiya, Chelsea Voss, Carroll Wainwright, Justin Jay Wang, Alvin Wang, Ben Wang, Jonathan Ward, Jason Wei, CJ Weinmann, Akila Welihinda, Peter Welinder, Jiayi Weng, Lilian Weng, Matt Wiethoff, Dave Willner, Clemens Winter, Samuel Wolrich, Hannah Wong, Lauren Workman, Sherwin Wu, Jeff Wu, Michael Wu, Kai Xiao, Tao Xu, Sarah Yoo, Kevin Yu, Qiming Yuan, Wojciech Zaremba, Rowan Zellers, Chong Zhang, Marvin Zhang, Shengjia Zhao, Tianhao Zheng, Juntang Zhuang, William Zhuk, and Barret Zoph. Gpt-4 technical report, 2024. URL https://arxiv.org/abs/2303.08774. Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu. Bleu: a method for automatic evaluation of machine translation. In Proceedings of the 40th annual meeting of the Association for Computational Linguistics, pp. 311 318, 2002. Yiwei Qin, Weizhe Yuan, Graham Neubig, and Pengfei Liu. T5score: Discriminative fine-tuning of generative evaluation metrics. In Findings of the Association for Computational Linguistics: EMNLP 2023, pp. 15185 15202, 2023. Rafael Rafailov, Archit Sharma, Eric Mitchell, Stefano Ermon, Christopher D. Manning, and Chelsea Finn. Direct preference optimization: Your language model is secretly a reward model, 2024. URL https://arxiv.org/abs/2305.18290. Ricardo Rei, Craig Stewart, Ana C Farinha, and Alon Lavie. Comet: A neural framework for mt evaluation, 2020. URL https://arxiv.org/abs/2009.09025. Brian Richards. Type/token ratios: What do they really tell us? Journal of Child Language, 14(2): 201 209, 1987. doi: 10.1017/S0305000900012885. Thibault Sellam, Dipanjan Das, and Ankur P. Parikh. Bleurt: Learning robust metrics for text generation, 2020. URL https://arxiv.org/abs/2004.04696. Published as a conference paper at ICLR 2025 Noah Shinn, Federico Cassano, Edward Berman, Ashwin Gopinath, Karthik Narasimhan, and Shunyu Yao. Reflexion: Language agents with verbal reinforcement learning, 2023. URL https://arxiv.org/abs/2303.11366. Charlie Snell, Jaehoon Lee, Kelvin Xu, and Aviral Kumar. Scaling llm test-time compute optimally can be more effective than scaling model parameters. ar Xiv preprint ar Xiv:2408.03314, 2024. Yifan Song, Guoyin Wang, Sujian Li, and Bill Yuchen Lin. The good, the bad, and the greedy: Evaluation of llms should not ignore non-determinism, 2024. URL https://arxiv.org/ abs/2407.10457. Aarohi Srivastava, Abhinav Rastogi, Abhishek Rao, Abu Awal Md Shoeb, Abubakar Abid, Adam Fisch, Adam R. Brown, Adam Santoro, Aditya Gupta, Adrià Garriga-Alonso, Agnieszka Kluska, Aitor Lewkowycz, Akshat Agarwal, Alethea Power, Alex Ray, Alex Warstadt, Alexander W. Kocurek, Ali Safaya, Ali Tazarv, Alice Xiang, Alicia Parrish, Allen Nie, Aman Hussain, Amanda Askell, Amanda Dsouza, Ambrose Slone, Ameet Rahane, Anantharaman S. Iyer, Anders Andreassen, Andrea Madotto, Andrea Santilli, Andreas Stuhlmüller, Andrew Dai, Andrew La, Andrew Lampinen, Andy Zou, Angela Jiang, Angelica Chen, Anh Vuong, Animesh Gupta, Anna Gottardi, Antonio Norelli, Anu Venkatesh, Arash Gholamidavoodi, Arfa Tabassum, Arul Menezes, Arun Kirubarajan, Asher Mullokandov, Ashish Sabharwal, Austin Herrick, Avia Efrat, Aykut Erdem, Ayla Karaka s, B. Ryan Roberts, Bao Sheng Loe, Barret Zoph, Bartłomiej Bojanowski, Batuhan Özyurt, Behnam Hedayatnia, Behnam Neyshabur, Benjamin Inden, Benno Stein, Berk Ekmekci, Bill Yuchen Lin, Blake Howald, Bryan Orinion, Cameron Diao, Cameron Dour, Catherine Stinson, Cedrick Argueta, César Ferri Ramírez, Chandan Singh, Charles Rathkopf, Chenlin Meng, Chitta Baral, Chiyu Wu, Chris Callison-Burch, Chris Waites, Christian Voigt, Christopher D. Manning, Christopher Potts, Cindy Ramirez, Clara E. Rivera, Clemencia Siro, Colin Raffel, Courtney Ashcraft, Cristina Garbacea, Damien Sileo, Dan Garrette, Dan Hendrycks, Dan Kilman, Dan Roth, Daniel Freeman, Daniel Khashabi, Daniel Levy, Daniel Moseguí González, Danielle Perszyk, Danny Hernandez, Danqi Chen, Daphne Ippolito, Dar Gilboa, David Dohan, David Drakard, David Jurgens, Debajyoti Datta, Deep Ganguli, Denis Emelin, Denis Kleyko, Deniz Yuret, Derek Chen, Derek Tam, Dieuwke Hupkes, Diganta Misra, Dilyar Buzan, Dimitri Coelho Mollo, Diyi Yang, Dong-Ho Lee, Dylan Schrader, Ekaterina Shutova, Ekin Dogus Cubuk, Elad Segal, Eleanor Hagerman, Elizabeth Barnes, Elizabeth Donoway, Ellie Pavlick, Emanuele Rodola, Emma Lam, Eric Chu, Eric Tang, Erkut Erdem, Ernie Chang, Ethan A. Chi, Ethan Dyer, Ethan Jerzak, Ethan Kim, Eunice Engefu Manyasi, Evgenii Zheltonozhskii, Fanyue Xia, Fatemeh Siar, Fernando Martínez-Plumed, Francesca Happé, Francois Chollet, Frieda Rong, Gaurav Mishra, Genta Indra Winata, Gerard de Melo, Germán Kruszewski, Giambattista Parascandolo, Giorgio Mariani, Gloria Wang, Gonzalo Jaimovitch López, Gregor Betz, Guy Gur-Ari, Hana Galijasevic, Hannah Kim, Hannah Rashkin, Hannaneh Hajishirzi, Harsh Mehta, Hayden Bogar, Henry Shevlin, Hinrich Schütze, Hiromu Yakura, Hongming Zhang, Hugh Mee Wong, Ian Ng, Isaac Noble, Jaap Jumelet, Jack Geissinger, Jackson Kernion, Jacob Hilton, Jaehoon Lee, Jaime Fernández Fisac, James B. Simon, James Koppel, James Zheng, James Zou, Jan Koco n, Jana Thompson, Janelle Wingfield, Jared Kaplan, Jarema Radom, Jascha Sohl-Dickstein, Jason Phang, Jason Wei, Jason Yosinski, Jekaterina Novikova, Jelle Bosscher, Jennifer Marsh, Jeremy Kim, Jeroen Taal, Jesse Engel, Jesujoba Alabi, Jiacheng Xu, Jiaming Song, Jillian Tang, Joan Waweru, John Burden, John Miller, John U. Balis, Jonathan Batchelder, Jonathan Berant, Jörg Frohberg, Jos Rozen, Jose Hernandez-Orallo, Joseph Boudeman, Joseph Guerr, Joseph Jones, Joshua B. Tenenbaum, Joshua S. Rule, Joyce Chua, Kamil Kanclerz, Karen Livescu, Karl Krauth, Karthik Gopalakrishnan, Katerina Ignatyeva, Katja Markert, Kaustubh D. Dhole, Kevin Gimpel, Kevin Omondi, Kory Mathewson, Kristen Chiafullo, Ksenia Shkaruta, Kumar Shridhar, Kyle Mc Donell, Kyle Richardson, Laria Reynolds, Leo Gao, Li Zhang, Liam Dugan, Lianhui Qin, Lidia Contreras-Ochando, Louis-Philippe Morency, Luca Moschella, Lucas Lam, Lucy Noble, Ludwig Schmidt, Luheng He, Luis Oliveros Colón, Luke Metz, Lütfi Kerem Senel, Maarten Bosma, Maarten Sap, Maartje ter Hoeve, Maheen Farooqi, Manaal Faruqui, Mantas Mazeika, Marco Baturan, Marco Marelli, Marco Maru, Maria Jose Ramírez Quintana, Marie Tolkiehn, Mario Giulianelli, Martha Lewis, Martin Potthast, Matthew L. Leavitt, Matthias Hagen, Mátyás Schubert, Medina Orduna Baitemirova, Melody Arnaud, Melvin Mc Elrath, Michael A. Yee, Michael Cohen, Michael Gu, Michael Ivanitskiy, Michael Starritt, Michael Strube, Michał Sw edrowski, Michele Bevilacqua, Michihiro Yasunaga, Mihir Kale, Mike Cain, Mimee Xu, Mirac Suzgun, Mitch Walker, Mo Tiwari, Mohit Bansal, Published as a conference paper at ICLR 2025 Moin Aminnaseri, Mor Geva, Mozhdeh Gheini, Mukund Varma T, Nanyun Peng, Nathan A. Chi, Nayeon Lee, Neta Gur-Ari Krakover, Nicholas Cameron, Nicholas Roberts, Nick Doiron, Nicole Martinez, Nikita Nangia, Niklas Deckers, Niklas Muennighoff, Nitish Shirish Keskar, Niveditha S. Iyer, Noah Constant, Noah Fiedel, Nuan Wen, Oliver Zhang, Omar Agha, Omar Elbaghdadi, Omer Levy, Owain Evans, Pablo Antonio Moreno Casares, Parth Doshi, Pascale Fung, Paul Pu Liang, Paul Vicol, Pegah Alipoormolabashi, Peiyuan Liao, Percy Liang, Peter Chang, Peter Eckersley, Phu Mon Htut, Pinyu Hwang, Piotr Miłkowski, Piyush Patil, Pouya Pezeshkpour, Priti Oli, Qiaozhu Mei, Qing Lyu, Qinlang Chen, Rabin Banjade, Rachel Etta Rudolph, Raefer Gabriel, Rahel Habacker, Ramon Risco, Raphaël Millière, Rhythm Garg, Richard Barnes, Rif A. Saurous, Riku Arakawa, Robbe Raymaekers, Robert Frank, Rohan Sikand, Roman Novak, Roman Sitelew, Ronan Le Bras, Rosanne Liu, Rowan Jacobs, Rui Zhang, Ruslan Salakhutdinov, Ryan Chi, Ryan Lee, Ryan Stovall, Ryan Teehan, Rylan Yang, Sahib Singh, Saif M. Mohammad, Sajant Anand, Sam Dillavou, Sam Shleifer, Sam Wiseman, Samuel Gruetter, Samuel R. Bowman, Samuel S. Schoenholz, Sanghyun Han, Sanjeev Kwatra, Sarah A. Rous, Sarik Ghazarian, Sayan Ghosh, Sean Casey, Sebastian Bischoff, Sebastian Gehrmann, Sebastian Schuster, Sepideh Sadeghi, Shadi Hamdan, Sharon Zhou, Shashank Srivastava, Sherry Shi, Shikhar Singh, Shima Asaadi, Shixiang Shane Gu, Shubh Pachchigar, Shubham Toshniwal, Shyam Upadhyay, Shyamolima, Debnath, Siamak Shakeri, Simon Thormeyer, Simone Melzi, Siva Reddy, Sneha Priscilla Makini, Soo-Hwan Lee, Spencer Torene, Sriharsha Hatwar, Stanislas Dehaene, Stefan Divic, Stefano Ermon, Stella Biderman, Stephanie Lin, Stephen Prasad, Steven T. Piantadosi, Stuart M. Shieber, Summer Misherghi, Svetlana Kiritchenko, Swaroop Mishra, Tal Linzen, Tal Schuster, Tao Li, Tao Yu, Tariq Ali, Tatsu Hashimoto, Te-Lin Wu, Théo Desbordes, Theodore Rothschild, Thomas Phan, Tianle Wang, Tiberius Nkinyili, Timo Schick, Timofei Kornev, Titus Tunduny, Tobias Gerstenberg, Trenton Chang, Trishala Neeraj, Tushar Khot, Tyler Shultz, Uri Shaham, Vedant Misra, Vera Demberg, Victoria Nyamai, Vikas Raunak, Vinay Ramasesh, Vinay Uday Prabhu, Vishakh Padmakumar, Vivek Srikumar, William Fedus, William Saunders, William Zhang, Wout Vossen, Xiang Ren, Xiaoyu Tong, Xinran Zhao, Xinyi Wu, Xudong Shen, Yadollah Yaghoobzadeh, Yair Lakretz, Yangqiu Song, Yasaman Bahri, Yejin Choi, Yichi Yang, Yiding Hao, Yifu Chen, Yonatan Belinkov, Yu Hou, Yufang Hou, Yuntao Bai, Zachary Seid, Zhuoye Zhao, Zijian Wang, Zijie J. Wang, Zirui Wang, and Ziyi Wu. Beyond the imitation game: Quantifying and extrapolating the capabilities of language models, 2023. URL https://arxiv.org/abs/2206.04615. Miloš Stanojevi c and Khalil Sima an. Fitting sentence level translation evaluation with many dense features. In Alessandro Moschitti, Bo Pang, and Walter Daelemans (eds.), Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 202 206, Doha, Qatar, October 2014. Association for Computational Linguistics. doi: 10.3115/v1/ D14-1025. URL https://aclanthology.org/D14-1025. Nisan Stiennon, Long Ouyang, Jeff Wu, Daniel M. Ziegler, Ryan Lowe, Chelsea Voss, Alec Radford, Dario Amodei, and Paul Christiano. Learning to summarize from human feedback, 2022. URL https://arxiv.org/abs/2009.01325. Mirac Suzgun, Luke Melas-Kyriazi, and Dan Jurafsky. Follow the wisdom of the crowd: Effective text generation via minimum bayes risk decoding, 2022. Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timothée Lacroix, Baptiste Rozière, Naman Goyal, Eric Hambro, Faisal Azhar, Aurelien Rodriguez, Armand Joulin, Edouard Grave, and Guillaume Lample. Llama: Open and efficient foundation language models, 2023a. URL https://arxiv.org/abs/2302.13971. Hugo Touvron, Louis Martin, Kevin Stone, Peter Albert, Amjad Almahairi, Yasmine Babaei, Nikolay Bashlykov, Soumya Batra, Prajjwal Bhargava, Shruti Bhosale, Dan Bikel, Lukas Blecher, Cristian Canton Ferrer, Moya Chen, Guillem Cucurull, David Esiobu, Jude Fernandes, Jeremy Fu, Wenyin Fu, Brian Fuller, Cynthia Gao, Vedanuj Goswami, Naman Goyal, Anthony Hartshorn, Saghar Hosseini, Rui Hou, Hakan Inan, Marcin Kardas, Viktor Kerkez, Madian Khabsa, Isabel Kloumann, Artem Korenev, Punit Singh Koura, Marie-Anne Lachaux, Thibaut Lavril, Jenya Lee, Diana Liskovich, Yinghai Lu, Yuning Mao, Xavier Martinet, Todor Mihaylov, Pushkar Mishra, Igor Molybog, Yixin Nie, Andrew Poulton, Jeremy Reizenstein, Rashi Rungta, Kalyan Saladi, Alan Schelten, Ruan Silva, Eric Michael Smith, Ranjan Subramanian, Xiaoqing Ellen Tan, Binh Published as a conference paper at ICLR 2025 Tang, Ross Taylor, Adina Williams, Jian Xiang Kuan, Puxin Xu, Zheng Yan, Iliyan Zarov, Yuchen Zhang, Angela Fan, Melanie Kambadur, Sharan Narang, Aurelien Rodriguez, Robert Stojnic, Sergey Edunov, and Thomas Scialom. Llama 2: Open foundation and fine-tuned chat models, 2023b. URL https://arxiv.org/abs/2307.09288. Tu Vu, Kalpesh Krishna, Salaheddin Alzubi, Chris Tar, Manaal Faruqui, and Yun-Hsuan Sung. Foundational autoraters: Taming large language models for better automatic evaluation. ar Xiv preprint ar Xiv:2407.10817, 2024. Jun Wang, Eleftheria Briakou, Hamid Dadkhahi, Rishabh Agarwal, Colin Cherry, and Trevor Cohn. Don t throw away data: Better sequence knowledge distillation, 2024a. URL https: //arxiv.org/abs/2407.10456. Peifeng Wang, Austin Xu, Yilun Zhou, Caiming Xiong, and Shafiq Joty. Direct judgement preference optimization. ar Xiv preprint ar Xiv:2409.14664, 2024b. Xuezhi Wang, Jason Wei, Dale Schuurmans, Quoc Le, Ed Chi, Sharan Narang, Aakanksha Chowdhery, and Denny Zhou. Self-consistency improves chain of thought reasoning in language models, 2023. Jason Wei, Maarten Bosma, Vincent Y. Zhao, Kelvin Guu, Adams Wei Yu, Brian Lester, Nan Du, Andrew M. Dai, and Quoc V. Le. Finetuned language models are zero-shot learners, 2022. URL https://arxiv.org/abs/2109.01652. Sean Welleck, Amanda Bertsch, Matthew Finlayson, Hailey Schoelkopf, Alex Xie, Graham Neubig, Ilia Kulikov, and Zaid Harchaoui. From decoding to meta-generation: Inference-time algorithms for large language models, 2024. URL https://arxiv.org/abs/2406.16838. Tianhao Wu, Weizhe Yuan, Olga Golovneva, Jing Xu, Yuandong Tian, Jiantao Jiao, Jason Weston, and Sainbayar Sukhbaatar. Meta-rewarding language models: Self-improving alignment with llm-as-a-meta-judge. ar Xiv preprint ar Xiv:2407.19594, 2024. Jing Xu, Andrew Lee, Sainbayar Sukhbaatar, and Jason Weston. Some things are more cringe than others: Preference optimization with the pairwise cringe loss. ar Xiv preprint ar Xiv:2312.16682, 2023. Guangyu Yang, Jinghong Chen, Weizhe Lin, and Bill Byrne. Direct preference optimization for neural machine translation with minimum bayes risk decoding, 2024. URL https://arxiv. org/abs/2311.08380. Shunyu Yao, Dian Yu, Jeffrey Zhao, Izhak Shafran, Thomas L. Griffiths, Yuan Cao, and Karthik Narasimhan. Tree of thoughts: Deliberate problem solving with large language models, 2023. URL https://arxiv.org/abs/2305.10601. Seonghyeon Ye, Doyoung Kim, Sungdong Kim, Hyeonbin Hwang, Seungone Kim, Yongrae Jo, James Thorne, Juho Kim, and Minjoon Seo. Flask: Fine-grained language model evaluation based on alignment skill sets, 2024. URL https://arxiv.org/abs/2307.10928. Weizhe Yuan, Graham Neubig, and Pengfei Liu. Bartscore: Evaluating generated text as text generation. Advances in Neural Information Processing Systems, 34:27263 27277, 2021. Weizhe Yuan, Richard Yuanzhe Pang, Kyunghyun Cho, Xian Li, Sainbayar Sukhbaatar, Jing Xu, and Jason Weston. Self-rewarding language models, 2024a. Weizhe Yuan, Richard Yuanzhe Pang, Kyunghyun Cho, Xian Li, Sainbayar Sukhbaatar, Jing Xu, and Jason Weston. Self-rewarding language models, 2024b. URL https://arxiv.org/ abs/2401.10020. Rowan Zellers, Ari Holtzman, Yonatan Bisk, Ali Farhadi, and Yejin Choi. Hellaswag: Can a machine really finish your sentence?, 2019. URL https://arxiv.org/abs/1905.07830. Tianyi Zhang, Varsha Kishore, Felix Wu, Kilian Q Weinberger, and Yoav Artzi. Bertscore: Evaluating text generation with bert. In International Conference on Learning Representations. Published as a conference paper at ICLR 2025 Lianmin Zheng, Wei-Lin Chiang, Ying Sheng, Siyuan Zhuang, Zhanghao Wu, Yonghao Zhuang, Zi Lin, Zhuohan Li, Dacheng Li, Eric P. Xing, Hao Zhang, Joseph E. Gonzalez, and Ion Stoica. Judging llm-as-a-judge with mt-bench and chatbot arena, 2023. Banghua Zhu, Evan Frick, Tianhao Wu, Hanlin Zhu, and Jiantao Jiao. Starling-7b: Improving llm helpfulness & harmlessness with rlaif, November 2023a. Lianghui Zhu, Xinggang Wang, and Xinlong Wang. Judgelm: Fine-tuned large language models are scalable judges. 2023b. Lianghui Zhu, Xinggang Wang, and Xinlong Wang. Judgelm: Fine-tuned large language models are scalable judges, 2023c. URL https://arxiv.org/abs/2310.17631. Published as a conference paper at ICLR 2025 A MBR DECODING: ADDITIONAL RESULTS This section presents additional results from our experiments exploring the use of MBR decoding to improve test-time performance. A.1 ADDITIONAL MBR UTILITY METRICS Llama-2-7B Llama-2-70B Avg. Greedy 14.4 22.8 0 ROUGE-1 MBR 16.2 24.7 1.85 SFR-Embedder MBR 12.1 22.2 -1.45 Prometheus MBR 17.7 26.2 3.35 ROUGE-2 MBR 16.6 24.6 2.00 ROUGE-L MBR 15.5 24.7 1.50 NVEmbed-Embedder MBR 14.1 22.1 -0.50 Nomic-Embedder MBR 16.3 24.1 1.60 Table 5: Alpaca Eval 2.0 win rates (%) for additional MBR decoding experiments, along with the average win rate differences compared to greedy decoding across all models (denoted Avg. ). We experiment with two additional ROUGE variants and two additional dense embedders as utility metrics. The two ROUGE variants are ROUGE-2 and ROUGE-L, which detect bigram overlap and longest co-occurring n-gram overlap respectively. The two dense embedders are NVEmbedv2 (Lee et al., 2024) and Nomic-Text-v1.5 (Nussbaum et al., 2024), which, like SFR Embedder, are strong, long-context dense embedders that rank highly on the MTEB leaderboard (Muennighoff et al., 2023). We find that MBR decoding with ROUGE-2 performs slightly better than MBR decoding with ROUGE-1, while MBR decoding with ROUGE-L performs slightly worse. MBR decoding with NVEmbed and Nomic both perform better than MBR decoding with SFR Embedder, with Nomic showing some improvement relative to greedy decoding. The latter result suggests that dense embedders could potentially be used for MBR decoding, although further work is required to understand what properties make for good MBR embedders. Overall, none of our additional utility metrics provide comparable improvements to MBR with Prometheus. A.2 COMPARISON WITH UNIVERSAL SELF-CONSISTENCY We compare MBR decoding to Universal Self-Consistency (USC) (Chen et al., 2023). In USC, Ncand outputs are sampled from the LLM and passed directly to the LLM for consistency detection. This entails prompting the LLM to choose the most consistent output. In Chen et al. (2023), the authors demonstrate that USC improves over greedy decoding for mathematical reasoning, code generation, summarisation and question-answering. The limited context lengths of LLMs poses a significant challenge when using USC, as it requires fitting all Ncand 30 samples into a single prompt. In contrast, MBR decoding only requires fitting two outputs into a single prompt as utility is computed pairwise. As our existing choice of models have limited context lengths (4096 tokens for Llama-2, 8192 tokens for Llama-3) and our outputs can be long (up to 1024 tokens), we are unable to assess USC on equal footing with MBR decoding using these models without significantly reducing Ncand. In order to facilitate a fair comparison, we therefore use the Llama-3.1 models (Dubey et al., 2024) in place of the Llama-2 and Llama-3 models for this experiment. The Llama-3.1 models possess context lengths of 128k, thereby allowing us to fit all Ncand samples into a single to prompt as required. Our USC prompt is as follows: Published as a conference paper at ICLR 2025 You are given a collection of 30 responses to a prompt. Select the most consistent response based on majority consensus. The most concistent response should be the most representative of all the responses provided. You should consider a variety of factors when evaluating consistency, including content, arguments and examples employed, style, structure and final answer, if relevant. Do not pass judgement on the quality or correctness of the response. Consider only consistency. Provide a short explanation of your choice, followed by your choice. Your choice should follow this format: "Most Consistent Response: [[Response ID]]", for example: "Most Consistent Response: [[15]]" if response 15 is the most consistent amongst all responses. Responses: {responses} Llama-3.1-8B Llama-3.1-70B Avg. Greedy 34.2 42.1 0 USC 37.3 43.7 2.35 MBR Prometheus 40.2 45.8 4.85 Table 6: Alpaca Eval 2.0 win rates (%) for Llama-3.1 models with greedy decoding, USC and MBR decoding with Prometheus, along with the average win rate differences compared to greedy decoding across all models (denoted Avg. ). We document our findings in Table 6. We find while that USC provides some improvements over greedy decoding, this improvement is smaller than the improvement provided by MBR decoding with Prometheus. A.3 SAMPLING BASELINES We conduct additional baseline experiments with top-p (Holtzman et al., 2020) and top-k (Fan et al., 2018) sampling. We use these sampling methods to directly obtain a final output, and do not explore replacing standard temperature sampling with these sampling strategies for generating the hypothesis set - we leave this to future work. Llama-2-7B Llama-2-70B Avg. Greedy 14.4 22.8 0 ROUGE-1 MBR 16.2 24.7 1.85 Prometheus MBR 17.7 26.2 3.35 top-p (p 0.9, t 0.3) 14.3 23.7 0.40 top-p (p 0.9, t 0.7) 14.7 23.9 0.70 top-p (p 0.5, t 0.3) 15.3 24.9 1.50 top-p (p 0.5, t 0.7) 15.0 24.9 1.35 top-k (k 50, t 0.3) 14.7 23.6 0.55 top-k (k 50, t 0.7) 15.0 23.5 0.65 top-k (k 20, t 0.3) 15.0 24.7 1.25 top-k (k 20, t 0.7) 14.7 24.0 0.75 Table 7: Alpaca Eval 2.0 win rates for top-p and top-k decoding, along with the average score differences compared to greedy decoding across all models (denoted Avg. ). Top-p and top-k sampling improve over greedy decoding, but do not match the performance improvements of Prometheus MBR decoding. We find that top-p and top-k sampling improves over greedy decoding, and can achieve performance close to that of MBR decoding with ROUGE. The improvements are nonetheless much smaller than MBR decoding with Prometheus. Nonetheless, these results demonstrate that top-p and topk sampling could be used to produce hypothesis sets containing higher quality candidates, which could in turn improve downstream MBR decoding performance. Published as a conference paper at ICLR 2025 A.4 GENERATION LENGTHS Bo N Prometheus MBR BERTScore MBR Embedder MBR Prometheus Avg. Generation Length / Words Figure 6: Average generation lengths (in words) for various decoding strategies, averaged over all five generator LLMs. MBR with Prometheus produces slightly longer outputs than most baseline methods, although these outputs are still far shorter than those produced by MBR with SFR-embedder and by the longest decoding baseline. Bo N Prometheus MBR BERTScore MBR Embedder MBR Prometheus Avg. Generation Length / Words Llama2-7b Llama2-13b Llama2-70b Llama3-8b Llama3-70b Figure 7: Average generation lengths (in words) for various models and decoding strategies on Alpaca Eval. MBR with Prometheus produces slightly longer outputs than most baseline methods, although these outputs are still far shorter than those produced by MBR with SFR-embedder and by the longest decoding baseline. We compute the average generation lengths (in words) of our five generator LLMs on Alpaca Eval 2.0 with different decoding strategies and plot the results in Figures 6 and 7. We find that MBR decoding with Prometheus yields outputs that are on average longer than those yielded by greedy decoding, beam search or Prometheus Bo N decoding. They are nonetheless much shorter than the outputs yielded by MBR decoding with SFR-Embedder and decoding by selecting the longest output. We hypothesise that MBR decoding with Prometheus encourages selection of more detailed responses which improves model performance, although we note that verbosity alone is not enough to improve performance on our benchmarks as the LONGEST baseline fails to yield any gains, and as we use length-controlled metrics for evaluation. Published as a conference paper at ICLR 2025 Bullet List Numbered List Both Neither Greedy 14.2 29.1 26.4 30.2 Prometheus Bo N 13.6 29.8 26.0 30.8 Prometheus MBR 13.9 30.1 28.1 27.8 Table 8: Percentage of responses to Alpaca Eval containing bullet lists, numbered lists, both bullet and numbered lists, and neither bullet nor numbered lists. These responses were generated by both Llama-3-8b-Instruct and Llama-3-70b-Instruct using various decoding strategies. We find that there is little difference in the formatting of MBR-generated, Bo N-generated and greedy-decoded outputs. TTR FK Greedy 0.514 12.13 Prometheus Bo N 0.520 12.40 Prometheus MBR 0.521 12.24 Table 9: Averaged type token ratios (TTR) and Flesch Kincaid readability scores (FK) across outputs generated by both Llama-3-8b-Instruct and Llama-3-70b-Instruct using various decoding strategies. We find that there is little difference in lexical diversity (as measured by TTR) and readability (as measured by FK scores) between MBR-generated, Bo N-generated and greedy-decoded outputs. A.5 LINGUISTIC FEATURES We attempt to identify linguistic differences (beyond generation length) between MBR-generated responses and responses produced using greedy decoding and Bo N decoding. We first consider looking for distinctive features in the formatting of MBR-generated responses by identifying responses to Alpaca Eval that make use of bullet and numbered lists and then computing the percentage of outputs that fall into each category. We use Llama-3-8b-Instruct and Llama-3-70b-Instruct as the generator models and Prometheus-2-7b as the MBR judge, and document the results in Table 8. We find that there is little difference in the formatting of MBR-generated, Bo N-generated and greedy-decoded outputs. Next, we analyse the lexical diversity and readability of MBR-generated outputs, using the same dataset, generator and judge model as for the formatting experiments. For lexical diversity, we measure the type token ratio (Richards, 1987) of outputs, and for readability we compute Flesch Kincaid readability scores (Flesch, 1948). We document our findings in Table 9, and leave further analysis of linguistic differences between various decoding strategies to future work. A.6 SPECULATIVE DECODING Avg time per generation (s) Avg tokens / s GPT-4o score Vanilla MBR 59.3 4.48 6.83 MBR + Spec. Decoding 52.9 5.21 6.90 Table 10: Results demonstrating that MBR decoding can be combined with speculative decoding to increase autoregressive decoding speed. We measure the average time per generation and the average tokens per second associated MBR decoding with and without speculative decoding, and show that speculative decoding improves decoding speed without sacrificing output quality. We can combine MBR decoding with speculative decoding (Leviathan et al., 2023) to speed up the MBR decoding step. We demonstrate this on 100 samples uniformly drawn from Alpaca Eval, using Llama-3-70b-Instruct as both the generator and the MBR judge (see huggingface.co/ibm-fms/Llama-3-70b-accelerator for our chosen draft model), and report our findings in Table 10. A.7 RANK CORRELATION WITH GPT-4O SCORES We used GPT-4o as a reference-free judge to score the outputs of Llama-3-8b-Instruct (Ncand 30, t 0.7) on the first turn of MT-Bench. Outputs were scored on a scale of 1 to 10, using the standard MT-bench evaluation approach. We used a GPT-4o judge temperature of 0.5 and generated three scores per generation, taking as our final scores the average of the three sampled scores. Then, for Published as a conference paper at ICLR 2025 Avg. over Greedy Avg. Rank Corr Prometheus-7b 0.28 0.119 Prometheus-8x7b 0.39 0.136 Judge LM-7b 0.22 0.053 Judge LM-33b 0.31 0.113 Llama-3-8b-Instruct 0.28 0.116 Llama-3-70b-Instruct 0.41 0.144 Table 11: Spearman s rank correlation between the GPT-4o scores and MBR scores generated using various MBR judge models, along with average over greedy decoding, computed on Llama-3-8b-Instruct outputs on turn 1 of MT-Bench. We find that stronger judges (as measured by the avg. over greedy) are associated with higher average rank correlations with GPT-4o scores. every sample, we computed the Spearman s rank correlation between the GPT-4o scores and the MBR scores. We find that stronger judges (higher avg. over greedy) are generally associated with better correlation with GPT-4o scores, although the absolute value of this correlation is not very high. This suggests that our judges are unlikely to be overly biased towards GPT-4o, although stronger judges will generally agree with GPT-4o more. B GPT-4O-JUDGE DETAILS We use GPT-4o as a judge for both Alpaca Eval 2.0 and MT-Bench. More specifically, we use gpt-4o-2024-05-13, accessed through an Azure endpoint. For Alpaca Eval 2.0, we use the official Alpaca Eval implementation4 to conduct evaluation. We use the weighted_alpaca_eval_gpt4_turbo evaluator and baseline results against the default gpt-4-1106-preview generations unless other specified. For MT-Bench, we use the singleand multi-turn judge prompts provided by LLMSYS Fast Chat5 as our judge prompts. Due to stochasticity in the outputs of the judge models, even with temperature set to zero, we generate three judgements per sample and take as the final score the median of the three judgement scores. C PROMETHEUS SCORING RUBRICS Prometheus takes as input a scoring rubric that defines scoring criteria to be used during evaluation. We use a single, generic scoring rubric for all our experiments: [Consider a wide range of factors such as the helpfulness, relevance, accuracy, depth, creativity, and level of detail of the response.] Score 1: The answer is completely unhelpful and incorrect. Nothing useful can be learned from it. Score 2: The answer contains some helpful and useful information, but major flaws, in terms of facuality, accuracy and relevance, are also present. Score 3: The answer is mostly helpful and relevant, although minor flaws exist. Score 4: The answer is accurate, relevant and helpful, although there are some clear improvements that can be made with respect to depth, creativity and detail. Score 5: The answer is excellent. It is completely accurate and relevant, and demonstrates a high degree of depth and creativity. 4github.com/tatsu-lab/alpaca_eval 5github.com/lm-sys/Fast Chat Published as a conference paper at ICLR 2025 We hypothesise that the performance of MBR and Bo N decoding with Prometheus could be improved through further optimisation of the scoring rubric, particularly with question-specific adjustments, where unique rubrics are tailored to each question. We leave this to future work. D ALPACAEVAL CATEGORIES We classify questions from Alpaca Eval and MT-Bench and then evaluate performance by category. For Alpaca Eval, we perform manual inspection on the dataset and identify ten common question categories. We then use GPT-4o to classify questions based on these categories, using the following prompt: Categorise an instruction based on the following list of categories. Only choose one category, and only return the category, nothing else. - Creative writing - Business, technical and scientific writing - Argumentation, debate and persuasion - Mathematical reasoning - Puzzles and logical reasoning - Coding - How-to and other guides - Recommendations and advice - Factual question-answering - Other Example Instruction: Bob has 5 sisters and 1 brother. How many siblings does one of Bob s sisters have? Category: Puzzles and logical reasoning Example Instruction: Write me a resignation email for my job as an accountant, explaining that I am leaving to pursue my dream of becoming a lion tamer. Category: Business, technical and scientific writing Example Instruction: What factors gave rise to the English Civil War. Category: Factual question-answering Example Instruction: I am visiting Kyoto next April. Recommend me 10 things to do! Category: Recommendations and advice Example Instruction: Pretend to be Donald Trump - write a speech announcing that you are becoming a Democrat. Category: Creative Writing Instruction: {instruction} Category: The percentage of questions assigned to them are listed in Table 12. The results of our performance by category analysis for MBR decoding with Prometheus on Alpaca Eval are illustrated in the main text, in Figure 3. E LLAMA-3 AS AN MBR AND BON UTILITY METRIC We experiment with using Llama-3-8b-Instruct and Llama-3-70b-Instruct as MBR and Bo N utility metrics in Section 3.3.1. This entails prompting the models to act as reference-based or referencefree evaluators. Our prompt for single-turn reference-based evaluation is as follows: Published as a conference paper at ICLR 2025 Category Percentage of Questions Factual question-answering 24.3 How-to and other guides 20.6 Recommendations and advice 12.1 Mathematical reasoning 3.2 Other 2.2 Creative writing 11.8 Business, technical and scientific writing 12.7 Puzzles and logical reasoning 3.5 Coding 6.7 Argumentation, debate and persuasion 2.7 Table 12: Alpaca Eval question categories identified by GPT-4o. [Instruction] Please act as an impartial judge and evaluate the quality of the response provided by an AI assistant to the user question displayed below. In addition to the user question, you are also given a reference answer. This is the best possible answer provided by a human expert. You should evaluate the assistant s response based on this. A good assistant s answer should share the content and style of the reference answer. Begin your evaluation by providing a short explanation. Be as objective as possible. After providing your explanation, you must rate the response on a scale of 1 to 10 by strictly following this format: "[[rating]]", for example: "Rating: [[5]]". [Question] {question} [Reference Answer] {reference} [The Start of Assistant s Answer] {answer} [The End of Assistant s Answer] Our prompt for single-turn reference-free evaluation is as follows: [Instruction] Please act as an impartial judge and evaluate the quality of the response provided by an AI assistant to the user question displayed below. Begin your evaluation by providing a short explanation. Be as objective as possible. After providing your explanation, you must rate the response on a scale of 1 to 10 by strictly following this format: "[[rating]]", for example: "Rating: [[5]]". [Question] {question} [The Start of Assistant s Answer] {answer} [The End of Assistant s Answer] For multi-turn evaluation, we use a system prompt to specify the rules. For reference-based evaluation: Published as a conference paper at ICLR 2025 Please act as an impartial judge and evaluate the quality of the response provided by an AI assistant to the user question displayed below. Your evaluation should focus on the assistant s answer to the second user question. In addition to the user question and conversation history, you are also given a reference answer. This is the best possible answer to the second user question provided by a human expert. You should evaluate the assistant s response based on this. A good assistant s answer should share the content and style of the reference answer. Begin your evaluation by providing a short explanation. Be as objective as possible. After providing your explanation, you must rate the response on a scale of 1 to 10 by strictly following this format: "[[rating]]", for example: "Rating: [[5]]". For multi-turn, reference-free evaluation: Please act as an impartial judge and evaluate the quality of the response provided by an AI assistant to the user question displayed below. Your evaluation should focus on the assistant s answer to the second user question. Begin your evaluation by providing a short explanation. Be as objective as possible. After providing your explanation, you must rate the response on a scale of 1 to 10 by strictly following this format: "[[rating]]", for example: "Rating: [[5]]". The prompt template for both reference-based and reference-free multi-turn evaluation is: <|The Start of Assistant A s Conversation with User|> ### User: {question_1} ### Assistant A: {answer_1} ### User: {question_2} ### Assistant A: {answer_2} <|The End of Assistant A s Conversation with User|> F REWARD MODEL AS A BON UTILITY METRIC 2-7B 2-13B 2-70B 3-8B 3-70B Avg. Greedy 5.72 5.90 6.50 7.54 8.29 0 BS 5.58 5.95 6.49 7.30 8.20 -0.09 Prometheus Bo N 5.77 6.08 6.65 7.66 8.42 0.13 Starling-RM-7b Bo N 5.99 6.49 6.85 7.88 8.46 0.34 Prometheus MBR 6.10 6.26 6.79 7.69 8.50 0.28 Table 13: MT-Bench scores for various models and decoding strategies, along with the average score differences compared to greedy decoding across all models (denoted Avg. ). Bo N decoding with reward model Starling RM-7b marginally outperforms MBR decoding with Prometheus. Published as a conference paper at ICLR 2025 We evaluate Bo N decoding using Starling-RM-7B-alpha (Zhu et al., 2023a) as the utility metric. Starling-RM is a strong 7B sequence classifier reward model trained to facilitate RLHF (Stiennon et al., 2022) that achieves a similar overall score to Prometheus-2-7B on Reward Bench (Lambert et al., 2024). It takes as inputs a prompt and a single candidate and outputs a scalar reward. As with reward models in general, Starling-RM does not support reference-based evaluation. We find that Bo N decoding with Starling-RM as a utility metric outperforms MBR decoding with Prometheus by a small margin. We have two possible explanations for this discrepancy. Firstly, Starling-RM achieves a much higher score than Prometheus-2-7B on Reward Bench s Chat task (likely due to the distribution of its training data), suggesting that it may simply be more suited to our particular benchmarks. Secondly, by providing continuous scalar rewards instead of discrete integer scores, Starling-RM enables more fine-grained evaluation of candidate outputs, allowing it to distinguish between outputs that Prometheus might have rated equally. We believe that using a reference-based reward model as the metric for MBR decoding could combine the advantages of fine-grained scoring with the increased reliability of consensus-based output selection. We leave the development of such models to future work. G SELF-TRAINING: ADDITIONAL RESULTS G.1 HEAD-TO-HEAD RESULTS sft + BS sft + MBR sft Full + Greedy dpo-MBR-1 dpo-MBR-2 dpo-MBR-3 dpo-Bo N-1 dpo-Bo N-2 dpo-Bo N-3 Win Rate vs. sft + Greedy Figure 8: Alpaca Eval 2.0 win rates (%) for self-trained models and various SFT baselines against sft with greedy decoding. Generation for all dpo models is done with greedy decoding. We find that MBR self-training with DPO allows models to match their MBR-decoding performance. We conduct head-to-head evaluation of our DPO self-trained models and various SFT baselines against sft with greedy decoding using the Alpaca Eval 2.0 and illustrate our findings in Figure 8. Our head-to-head results show that our MBR self-trained models outperform our Bo N self-trained models, the sft with beam search baseline, and the full sft with greedy decoding baseline. Our MBR self-trained models match the performance of the sft with MBR decoding baseline. G.2 RESULTS ON NLP BENCHMARKS Model MMLU pÒq ARC challenge pÒq Hella Swag pÒq Truthful QA pÒq sft 47.2 57.3 80.6 51.6 dpo-3-Bo N 47.2 57.5 80.6 53.5 dpo-3-MBR 47.3 57.1 80.7 52.6 sft 56.1 62.3 83.5 48.4 dpo-3-Bo N 56.3 62.5 83.5 48.2 dpo-3-MBR 56.1 62.6 83.5 47.4 Table 14: Evaluation results of Prometheus self-trained models on four different NLP benchmarks. We find that MBR and Bo N self-training maintains performance on across all four datasets compared with the sft models. We assess our self-trained models on four different NLP benchmarks: MMLU (Hendrycks et al., 2021), ARC challenge (Clark et al., 2018), Hella Swag (Zellers et al., 2019) and Truthful QA (Lin et al., 2022), and report our results in Table 14. We find that self-training maintains performance across all four benchmarks despite us using a training dataset that is irrelevant for these tasks. This Published as a conference paper at ICLR 2025 shows that MBR self-training can be used to improve the instruction-following abilities of models without jeopardising other skills. G.3 MBR SELF-TRAINING WITH ROUGE Alpaca Eval 2.0 MT-Bench 7B 13B 7B 13B sft 5.18 8.24 5.43 5.85 dpo-1-MBR-Prometheus 5.68 10.8 5.78 6.48 dpo-2-MBR-Prometheus 7.22 13.9 6.11 6.73 dpo-3-MBR-Prometheus 8.86 15.3 6.14 6.75 dpo-1-MBR-ROUGE 4.66 5.61 7.65 5.98 dpo-2-MBR-ROUGE 5.83 5.78 9.01 6.06 dpo-3-MBR-ROUGE 5.42 5.67 8.31 5.91 Table 15: Alpaca Eval 2.0 win rates (%) and MT-Bench scores for models self-trained using DPO with Prometheus and with ROUGE-1 as utility metrics. MBR self-training with ROUGE fails to yield substantial gains. We explore using ROUGE-1 as the utility metric for MBR self-training. We find that MBR selftraining with ROUGE fails to yield substantial gains. Improvements also saturate quickly, with model performance decreasing after the third training iteration. These results highlight the importance of choosing the correct MBR utility metric for self-training. G.4 MBR SELF-TRAINING WITH LLAMA-3-8B-INSTRUCT We also consider using Llama-3-8b-Instruct in place of Prometheus as the utility metric to self-train Llama-3-8b. We compare this approach to using Prometheus to train Llama-3-8b, and find that both approaches lead to significant gains. The fact that Llama-3-8b-Instruct as the judge can be used to improve Llama-3-8b suggests that self-improvement using MBR decoding is possible. We leave this to future work. Prometheus Llama-3-8b-Instruct sft 6.70 6.70 dpo-1-MBR 6.94 6.99 dpo-2-MBR 7.45 7.51 dpo-3-MBR 7.55 7.52 Table 16: MT-Bench scores for Llama-3-8b self-trained using DPO with either Prometheus or Llama-3-8b Instruct as the MBR judge. Generation is done using greedy decoding for all models. We find that both approaches lead to significant gains. G.5 ALTERNATIVE PAIR SELECTION STRATEGIES Alpaca Eval 2.0 MT-Bench sft 5.18 5.43 dpo-1-MBRBW 5.68 5.78 dpo-2-MBRBW 7.22 6.11 dpo-3-MBRBW 8.86 6.14 dpo-1-MBRBMW 4.95 5.78 dpo-2-MBRBMW 8.06 5.96 dpo-3-MBRBMW 8.88 5.97 Table 17: Alpaca Eval 2.0 win rates (%) and MT-Bench scores for models self-trained using DPO with BW and BMW preference pair selection strategies with sft-7b as the base model. Generation is done using greedy decoding for all models. We find that our BMW models perform similarly to the original BW models on Alpaca Eval 2.0 and slightly worse on MT-Bench. We investigate using an alternative MBR preference pair selection strategy. Following Yang et al. (2024), we create two preference pairs from each sample, one formed from the highest-scoring (best) and median-scoring (mid) outputs, and another formed from the median-scoring and lowest-scoring Published as a conference paper at ICLR 2025 (worst) outputs. We follow the exact same DPO training procedure as before, but replace our original BW training set with this new BMW training set. We document our results in Table 17. We find that our BMW models perform similarly to the original BW models on Alpaca Eval 2.0 and slightly worse on MT-Bench. We do not pursue this line of work further and leave additional investigations to future work. G.6 PERFORMANCE BY CATEGORY Category Percentage of Questions Factual question-answering 26.4 How-to and other guides 21.1 Recommendations and advice 4.4 Mathematical reasoning 0.1 Other 1.3 Creative writing 18.6 Business, technical and scientific writing 19.0 Puzzles and logical reasoning 0.1 Coding 6.3 Argumentation, debate and persuasion 2.7 Table 18: Ultra Chat-200k question categories identified by GPT-4o. We perform this analysis on a subsample of 1000 prompts sampled randomly from our 12000 training prompts. We replicate the question category analysis described in Section 3.3 and Appendix D for our Prometheus MBR self-trained models and report results in Figure 4. We find that performance relative to the sft models improves across almost all question categories, with performance on writingbased categories improving more than on reasoning-based categories. Performance on mathematical reasoning remains unchanged for the 7B model and decreases for the 13B model. To better understand this discrepancy, we sample 1000 prompts from our 12000 Ultra Chat training prompts and categorise them using GPT-4o, following the same procedure described in Section 3.3 and Appendix D. We document our results in Table 18. We find that reasoning-based questions (coding, puzzles and logical reasoning, mathematical reasoning) are underrepresented in this dataset, with mathematical reasoning and puzzles and logical reasoning especially underrepresented. We attribute our models inconsistent improvements in these areas to this lack of data. G.7 GENERATION LENGTHS sft Iter 1 Iter 2 Iter 3 Average Generation Length / Words Figure 9: Average generation lengths (in words) after iterative Prometheus MBR and Bo N self-training on Alpaca Eval. MBR self-training with DPO teaches the model to generate more detailed responses. We measure the generation lengths of our Prometheus MBR and Bo N self-trained models on Alpaca Eval. We find that MBR self-training with DPO teaches the model to generate longer and more detailed responses at every iteration. In contrast, Bo N self-training and MBR self-training with SFT only results in small increases in generation lengths. Published as a conference paper at ICLR 2025 H HUMAN STUDY Win Draw Loss Prometheus MBR vs. Greedy 30.0 53.3 16.7 Prometheus Bo N vs. Greedy 21.7 60.0 18.3 Table 19: Head-to-head evaluation of Prometheus MBR and Prometheus Bo N vs. greedy decoding for Llama2-70b conducted on the Alpaca Eval dataset by human evaluators. Win Draw Loss dpo-3-MBR vs. sft 46.7 43.3 10.0 dpo-3-Bo N vs. sft 33.3 55.0 11.7 Table 20: Head-to-head evaluation of 13B dpo-3-MBR and dpo-3-Bo N vs. sft conducted on the Alpaca Eval dataset by human evaluators. Greedy decoding used for all models. We conduct a human study for key MBR inference (Section 3) and MBR distillation (Section 4) experiments. The objective of this study is to verify that our LLM evaluation results (Alpaca Eval 2.0 and MT-Bench) align with real human judgements. We recruited 4 volunteers, each of whom were given 60 samples in total to evaluate. Each sample consisted of a randomly-sampled Alpaca Eval prompt and two corresponding generations displayed in a random order. Volunteers were asked to select their preferred generation and, if neither generation was preferred, to then rate the generations as equal. We conducted four head-to-head experiments. The first two experiments, corresponding to experiments in Section 3, were between Prometheus MBR vs. greedy decoding and Prometheus Bo N vs. greedy decoding with Llama-2-70b. The next two experiments, corresponding to experiments in Section 4, were between 13B dpo-3-MBR vs. sft and dpo-3-Bo N vs. sft. We used 60 samples for each experiment. Volunteers were given an even distribution of samples from each experiment. Our results show good alignment with our LLM evaluation results. We find that Prometheus MBR decoding performs well against greedy decoding, with its win rate higher than the corresponding win rate for Prometheus Bo N decoding vs. greedy decoding. We also find that dpo-3-MBR significantly outperforms sft, and its margin of victory is greater than the margin of victory of dpo-3-Bo N vs. sft. Furthermore, the margin of improvement associated with MBR distillation is larger than the corresponding margin for MBR inference. These findings align with our findings from our automatic LLM evaluation experiments. I ALGORITHMS I.1 MBR INFERENCE We provide the algorithm for MBR inference in Algorithm 1, complementing the mathematical overview provided in Section 2. I.2 MBR DISTILLATION In MBR distillation, we first gather a preference dataset using MBR decoding over a set of input prompts. Given input prompts Xk and a base model πθk 1, we first sample Ncand outputs per prompt ypiq πθk 1p |xq, where x P Xk, forming hypothesis set Hhyp typ1q, yp2q, . . . , yp Ncandqu. We then compute expected utility for elements in Hhyp upypiqq 1 Ncand j 1 upypiq, ypjqq (4) Published as a conference paper at ICLR 2025 Algorithm 1 MBR Inference Inputs: Prompt x, generator model p, reference-based utility metric u, number of candidates Ncand, sampling temperature t. Output: MBR output ˆy Initialise Hhyp Ð H for i P t1, 2, . . . , Ncandu do Sample ypiq pp |xq with temperature t Add ypiq to Hhyp // Form hypothesis set end for for ypiq P Hhyp do Compute upypiqq 1 Ncand řNcand j 1 upypiq, ypjqq // Compute expected utility end for Select ˆy arg maxy PHhyp upyq return ˆy and form preference pairs using the output that maximises expected utility ˆy and the output that minimises it ˆy . ˆy arg max y PHhyp upyq ˆy arg min y PHhyp upyq We collect these preference pairs and their prompt px, ˆy , ˆy q in a dataset we denote Yk. We then use these preference pairs for DPO (Rafailov et al., 2024) training. In DPO, we minimise the following policy objective: LDPO Epx,ˆy ,ˆy Ykq log σ ˆ β log πθpˆy |xq πrefpˆy |xq β log πθpˆy |xq πrefpˆy |xq where πθ is the policy and πref the reference model. We repeat this process iteratively with k 1, 2, . . . , K. The initial model πθ0 is the base sft model. We choose K 3 in our experiments. We also provide the algorithm for MBR distillation with DPO in Algorithm 2. J LIMITATIONS While our work demonstrates the significant potential of MBR decoding, there are limitations that should be addressed in future research. Firstly, although we demonstrate using existing judge LLMs utility metrics that MBR decoding consistently outperforms Bo N decoding, this does not preclude the existence of reference-free metrics that are powerful enough to match or surpass the performance of their direct reference-based counterparts. This relates to a possible broader limitation on the benefits of using consensus quality for output selection, as the consensus solution may not always be the optimal one. We encourage future work to train better utility metrics in order to better understand these limitations. A second limitation of our work is that we do not study the biases introduced by the utility metric. One particularly pernicious form of bias is reward-hacking behavior, where the utility metric (likely as a result of its own training) selects outputs that evaluate well on our benchmarks but that are actually worse in quality. While we preclude this from being the case in our experiments via our human study (Appendix H), this does not mean that such pernicious behavior cannot arise in other settings. Finally, we do not study limitations on scalability. Although we show that small judge LLMs (7B) can serve as utility metrics for much larger models (70B), it is likely that weaker utility metrics cease being useful for very strong LLMs and on very complex tasks. Further research is needed to determine when this breakdown occurs. Published as a conference paper at ICLR 2025 Algorithm 2 MBR Distillation with DPO Inputs: Prompt sets X1, X2, . . . , XK, sft model πθ0, reference-based utility metric u, number of candidates Ncand, sampling temperature t, number of self-training iterations K. Output: Self-trained model πθK for k P t1, 2, . . . , Ku do Initialise Yk Ð H for x P Xk do Initialise Hhyp Ð H for i P t1, 2, . . . , Ncandu do Sample ypiq πθk 1p |xq with temperature t Add ypiq to Hhyp // Form hypothesis set end for for ypiq P Hhyp do Compute upypiqq 1 Ncand řNcand j 1 upypiq, ypjqq // Compute expected utility end for Select ˆy arg maxy PHhyp upyq // Select highest scoring output Select ˆy arg miny PHhyp upyq // Select lowest scoring output Add pˆy , ˆy q to Yk // Form preference pairs end for Update πθk Ð DPOpπθk 1, Ykq // DPO training on preference pairs end for return πθK K TRAINING AND INFERENCE HYPERPARAMETERS Hyperparameter Value Learning Rate 5e-6 Num Epochs 3 Batch Size 32 Optimiser Adam W β1 0.9 β2 0.95 ϵ 1e-8 Weight Decay 0.1 Scheduler Cosine Table 21: Hyperparameters for SFT and MBR self-training with SFT. Hyperparameter Value Learning Rate 5e-7 Num Epochs 5 Batch Size 8 Optimiser RMSProp α 0.99 βDPO 0.1 Scheduler Constant with warmup Warmup Steps 150 Table 22: Hyperparameters for MBR self-training with DPO. Our SFT and DPO hyperparameters for our self-training experiments in Section 4.2 are provided in Tables 21 and 22. We use bf16 mixed precision training with 8x A100 GPUs for all experiments. For inference, we use 4x A100 GPUs with bf16 quantisation for all LLMs and judge LLMs, other than for the Analysis of compute costs experiments in Section 4.2, where we use 2x A100 GPUs. We use v LLM (Kwon et al., 2023) as the inference engine for all experiments.