# bootstrapping_language_models_with_dpo_implicit_rewards__ed68eef1.pdf Published as a conference paper at ICLR 2025 BOOTSTRAPPING LANGUAGE MODELS WITH DPO IMPLICIT REWARDS Changyu Chen 1, Zichen Liu 23, Chao Du 2, Tianyu Pang2, Qian Liu2, Arunesh Sinha 4, Pradeep Varakantham 1, Min Lin2 1Singapore Management University 2Sea AI Lab, Singapore 3National University of Singapore 4Rutgers University {chency,liuzc,duchao,tianyupang,liuqian,linmin}@sea.com; arunesh.sinha@rutgers.edu; pradeepv@smu.edu.sg Human alignment in large language models (LLMs) is an active area of research. A recent groundbreaking work, direct preference optimization (DPO), has greatly simplified the process from past work in reinforcement learning from human feedback (RLHF) by bypassing the reward learning stage in RLHF. DPO, after training, provides an implicit reward model. In this work, we make a novel observation that this implicit reward model can by itself be used in a bootstrapping fashion to further align the LLM. Our approach is to use the rewards from a current LLM to construct a preference dataset, which is then used in subsequent DPO rounds. We incorporate two refinements to further improve our approach: 1) length-regularized reward shaping to make the preference dataset length-unbiased; 2) experience replay to enhance the quality of the preference dataset. Our approach, named self-alignment with DPO Impli Cit r Ewards (DICE), shows great improvements in alignment. It achieves an increase of more than 8% in lengthcontrolled win rate on Alpaca Eval 2 for all the different base models that we tried, without relying on external feedback. Our code is available at https: //github.com/sail-sg/dice. 1 INTRODUCTION Direct preference optimization (DPO) (Rafailov et al., 2024b) presents a compelling alternative to reinforcement learning from human feedback (RLHF) (Stiennon et al., 2020) in large language models (LLMs). By circumventing the complexity of learning a reward model from given human preferences, DPO is simpler to implement and train compared to the RLHF approaches. Importantly, DPO, once trained, implicitly specifies a reward model. Mathematically, the reward for a response y to the prompt x can be expressed in terms of the optimal policy π and the reference policy πref: r (x, y) = β log π (y|x) πref(y|x) + β log Z(x), for parameter β and normalizing constant Z. Further, an implicit reward r(x, y) = β [log π (y|x) log πref(y|x)] is defined in DPO where the normalizing constant term can be ignored as it will be canceled out in the DPO objective, which only involves the difference of the rewards for the same prompt. In this work, we explore whether the above readily available implicit reward model after DPO training provides an opportunity to further improve the language model. This paper answers the research question in the affirmative, by using the above implicit rewards in a bootstrapping fashion to further improve the LLM alignment with human preferences. Specifically, our approach follows the iterative DPO framework (Tran et al., 2023), where implicit rewards serve as the preference signals, as illustrated in Figure 1. We start with a model that has been through one round of DPO using human preference data, referred to as a DPO-tuned model. We then use the implicit rewards induced by DPO itself to rank outputs from the current LLM, thereby, yielding a new Equal contribution. The project was done during Changyu Chen s internship at Sea AI Lab. Correspondence to Chao Du, Arunesh Sinha, and Pradeep Varakantham. Published as a conference paper at ICLR 2025 Sampled prompt , x Preference Signal DPO Training Bgyuowz0o Ak Enu AZXu HNm Tgvzrvzs Rgt OPn OMfy B8/k DCQOSg=r PE Prompting External LLM r PE = φ g(x, y) External Scalar RM r S = rφ(x, y) r S {yi}i2[K] (t 1)( |x) {x, yw, yl} LR Implicit RM r LR = β log (t 1)(y | x) ref (y | x) Prompting LLM Itself r PI = (t 1) g(x, y) Base Prompt Implicit Alpaca Eval 2 LC Win Rate (%) Zephyr Llama3 Figure 1: (Left) Iterative DPO (Tran et al., 2023) with various preference signals: under the iterative DPO framework, the policy model is iteratively trained on a newly generated preference dataset. This dataset can be constructed using various preference signals. A common source is a scalar reward model (RM) (Ouyang et al., 2022), denoted as rϕ. Alternatively, the dataset can be created by prompting a LLM to judge the responses. This LLM can either be an external model πϕ or the policy model itself from the previous iteration πθ(t 1). In this context, x and y are processed through a LLM-as-a-judge template g( , ) (Yuan et al., 2024). We propose utilizing the lengthregularized (LR) implicit rewards introduced in Section 3.1, where πθ(t 1) and π(t 1) ref represent the policy model and reference model from the previous DPO iteration, respectively. Our experiment excludes approaches requiring external models, such as rϕ and πϕ, as they are beyond this work s scope. (Right) Our method, which leverages implicit rewards, further improves DPO-tuned models by a large margin, resulting in superior performance compared to the prompting counterpart. preference dataset cheaply. We run DPO again with this newly generated preference dataset to obtain an updated LLM and then repeat the process. However, the above approach still needs further refinement to address practical issues. One is the known issue of length exploitation (Park et al., 2024) where LLMs generate long responses when the same content could be provided more succinctly. Another issue is that the implicit reward model is an approximate proxy for human preferences, hence relying on it strongly can result in corruption of the initial knowledge inbuilt into the LLM. Inspired by (Park et al., 2024), we address the length exploitation issue by length-regularized reward shaping, which discourages long responses from being preferred. Notably different from Park et al. (2024), who incorporate response length into the loss function, we directly search for a lengthunbiased dataset, avoiding the expensive hyper-parameter tuning. To fix the overreliance on implicit rewards, we leverage insights from continual learning (Rolnick et al., 2019) and replay high quality human preference data that was used in the first round of DPO (the round before bootstrapping began). Our method, named self-alignment with DPO Impli Cit r Ewards (DICE), significantly improves LLM alignment quality with different base models. On Alpaca Eval 2, we achieve 8.02% length-controlled (LC) win rate improvement with the Zephyr-based model and 9.35% improvement with the Llama3-based model. To summarize, our main contributions are as follows: We propose to utilize the implicit reward model readily available in a DPO-tuned LLM. The implicit reward model enables us to evaluate the responses generated by the current policy and construct a preference dataset for future rounds of DPO without any external feedback; We propose to apply two techniques together with our above proposed approach, lengthregularized reward shaping and experience replay; Empirical results show that our approach DICE enables significant (more than 8% LC win rate increase on Alpaca Eval 2) improvement in alignment with different base models; thus, we believe that the core idea of using DPO Implicit Reward in DICE is a general purpose approach that can improve alignment for any single DPO-tuned base model. Published as a conference paper at ICLR 2025 2 PRELIMINARIES We provide a brief review of the standard RLHF (Stiennon et al., 2020; Ouyang et al., 2022) and DPO algorithms (Rafailov et al., 2024b). Through the review, we demonstrate the implicit reward model induced by DPO, which will be used in our work. In preference tuning, the preference data typically takes the form of pairwise preferences. Each prompt x is paired with two possible responses, y1 and y2. The human annotator (Christiano et al., 2017) or AI annotator (Lee et al., 2023) provides the preference feedback o(y1 y2|x) {0, 1}, indicating whether y1 is preferred over y2. The preferred response is denoted as yw, while the other is denoted as yl. A common assumption is that the ground-truth human preferences follow the Bradley-Terry model (Bradley & Terry, 1952). Based on this assumption, we can train a parameterized reward model rϕ(x, y) using maximum likelihood: LR (rϕ, D) = E(x,yw,yl) D [log σ (rϕ (x, yw) rϕ (x, yl))] , (1) where σ is the logistic function. 2.1 REINFORCEMENT LEARNING FROM HUMAN FEEDBACK The standard RLHF algorithm treats the LLM as a policy and optimizes the policy using the reward model rϕ. The optimization objective is represented by the following equation: max πθ Ex D,y πθ(y|x)[rϕ(x, y)] β DKL [πθ(y|x) πref(y|x)] , (2) where πref(y|x) denotes a reference distribution, and β is a hyper-parameter. This objective balances the maximization of the reward rϕ(x, y) and divergence from the fixed reference distribution. The divergence term, given by the KL divergence (i.e., DKL [πθ(y|x) πref(y|x)]) acts as a regularizer to prevent the policy πθ from drifting too far away from the initial distribution πref(y|x). This objective is then optimized using a general-purpose RL algorithm, such as PPO (Schulman et al., 2017). 2.2 DIRECT PREFERENCE OPTIMIZATION DPO (Rafailov et al., 2024b) starts with the same objective as Eq. (2) but derives an analytical closed-form relation between the reward and the resulting optimal policy. This relation can be used to reparameterize the ground truth reward in terms of the corresponding optimal policy. This reparameterized formulation can be substituted back into the reward optimization objective in Eq. (1), enabling direct training of the optimal model on the feedback data using the following objective: LDPO (πθ; πref) = E(x,yw,yl) D log σ β log πθ (yw|x) πref (yw|x) β log πθ (yl|x) πref (yl|x) In this context, the parameter β is the same as in Eq. (2), balancing the expected reward and divergence from the initial model. The DPO objective is particularly advantageous as it facilitates the recovery of the optimal model through a standard classification loss, without the need for on-policy sampling or extensive hyper-parameter tuning. Observe that Eq. (3) resembles the reward modeling objective in Eq. (1) under the parameterization r(x, y) = β log πθ(y|x) πref(y|x). (4) This reward function is commonly referred to as an implicit reward (Rafailov et al., 2024a; Zhong et al., 2024). Theorem 1 in Rafailov et al. (2024b) demonstrates that this parameterization of a reward model is indeed valid without loss of generality. If we substitute this form of rθ(x, y) into the RL objective in Eq. (2), we can derive the optimal solution in a closed form, which is πθ. Consequently, once DPO optimization is completed, we obtain an implicit reward model as defined by Eq. (4). Published as a conference paper at ICLR 2025 3 BOOTSTRAPPING WITH DPO IMPLICIT REWARDS DPO is an attractive alternative to RLHF as it largely simplifies the implementation and training process of language model alignment. However, recent evidences (Guo et al., 2024; Tran et al., 2023) have shown that continuing DPO training on a fixed offline dataset results in inferior policy, while the policy can be further improved if one can collect new responses generated by the updated policy and provide preference labels to perform another round of DPO training. This can be understood as being closer to the on-policy sampling, which is generally preferred in preference fine-tuning (Tajwar et al., 2024; Tang et al., 2024). In this work, we employ such iterative DPO preference tuning framework, where we start from a base language model (a base policy) πθ(0) that is DPO-tuned from an initial reference model π(0) ref , commonly an SFT model. In each round t {1, 2, . . . }, we sample K on-policy responses from the latest policy πθ(t 1)( | x) given a prompt x. We then label the response with the highest and the lowest implicit rewards as winning and losing responses respectively, thus constructing a new preference dataset Dt. We further fine-tune the policy with DPO s objective (Eq. (3)) to obtain the updated policy πθ(t) with reference model π(t) ref = πθ(t 1). This process is repeated to iteratively improve the language model. Notably, by fine-tuning a DPO-tuned LM on the its own constructed dataset, we essentially align it without relying on any external preference feedback (e.g., RLHF or RLAIF), hence in a bootstrapping fashion. We thus refer to our method as iterative self-alignment (bootstrapping) with DPO Impli Cit r Ewards (DICE). We next introduce two ingredients that are proven critical in the self-alignment process. In Section 3.1, we present Length-Regularized Implicit Rewards, which augment the vanilla implicit rewards with a length-regularized reward shaping, to judge the on-policy sampled responses for constructing the preference dataset. Furthermore, to mitigate the potential catastrophic forgetting in the continual fine-tuning, we propose experience replay (Section 3.2) that mixes the generated data with the offline data for better performance. 3.1 LENGTH-REGULARIZED IMPLICIT REWARDS It is a known issue in the literature that preference tuning may introduce length bias (or length exploitation) (Park et al., 2024) , which is likely caused by the fact that the preference labels collected from human annotators favor more verbose responses. This problem is further compounded by the iterative self-alignment scheme such as the one in Yuan et al. (2024), because the generated responses that are long and preferred will be reinforced in the next round of DPO, leading the language model to generate increasingly longer responses. Inevitably, the vanilla DPO implicit rewards as in Eq. (4) would also exhibit length bias when generating preference dataset. In Figure 2, we show the distribution of the difference in string length of the winning and losing responses. We can see from the top figure that vanilla implicit rewards yield a skewed distribution (in green), with an average length difference 1031. In stark contrast, the length difference of a high quality preference dataset is almost normally distributed (as in the bottom figure). This observation motivates us to debias the distribution induced by vanilla implicit rewards so as to mitigate the length exploitation. We resort to reward shaping (Sutton & Barto, 2018) for this purpose. In particular, we introduce a length-regularized (LR) reward shaping term in the implicit Length Difference Distribution vanilla regularized 6000 4000 2000 0 2000 4000 6000 Figure 2: Distribution of the length difference between the winning and losing examples (|yw| |yl|). (Top) Distribution of the first round on-policy generated dataset. With LR reward shaping defined by Eq. (5), the length bias is mitigated and the length difference becomes more evenly distributed. The average length difference decreases from 1031 to 21 by setting α=0.023. (Bottom) Distribution of the high quality Ultra Feedback preference dataset (Cui et al., 2023) is almost unbiased. Published as a conference paper at ICLR 2025 Algorithm 1 Bootstrapping with DPO Implicit Rewards (DICE) 1: Input: prompt set X extracted from Doffline, initial DPO-tuned policy πθ(0), initial reference policy πref = πθ( 1), number of generated samples K, regularization weight β, experience replay weight γ (0, 1) 2: for t = 1, 2, . . . do 3: Generate responses by sampling x X and y1:K πθ(t 1); 4: Create preference dataset D(α ) by optimizing Eq. (6); // D(α) is constructed by taking the best and worst y based on r LR(x, yk; α), k [K]; // Evaluate r LR(x, yk; α) given πθ(t 1) and πθ(t 2) as the target policy and reference policy 5: Create the mixed dataset via experience replay Dt = {(xi, yi w, yi l)}i [N]; // Sample (x, yw, yl) p Dt, p Dt = (1 γ)p D(α ) + γp Doffline 6: Optimize πθ according to DPO loss, Eq. (3): θ(t) arg min θ E(x,yw,yl) Dt log σ β log πθ (yw|x) πθ(t 1) (yw|x) β log πθ (yl|x) πθ(t 1) (yl|x) reward that penalizes the length of the response to obtain the shaped reward: r LR(x, y; α) = β log πθ(y | x) πref(y | x) α |y| , (5) where α controls the penalty strength and |y| is the string length of the response. Based on the shaped rewards, we can construct many versions of the preference dataset D(α), following the principle that the response with the higest r LR(x, yi; α) is labeled as yw and the one with the lowest reward is labeled as yl. To find the most suitable α such that D(α) is (approximately) unbiased, we optimize α with the objective to minimize the average absolute difference in response length: α = arg min α E(yw,yl) D(α) (|yw| |yl|) We can employ any black-box optimizer to solve this non-differentiable objective function. In this work we find a simple random search suffices, and its solution effectively transforms the dataset into a more evenly distributed one (shown in the orange curve in the top of Figure 2). We will output D(α ) for the next round of DPO training. Details of the optimization can be found in Appendix C. Importantly, despite the resemblance of Eq. (5) to Park et al. (2024) where they incorporate the token length as a regularizer in the training objective, our reward shaping is conducted during the dataset construction stage, thereby avoiding the need for expensive hyper-parameter tuning. 3.2 EXPERIENCE REPLAY Though DICE enables us to learn from the response of the current policy, we know that the implicit reward model from the DPO training is not a perfect proxy for human preferences. Solely relying on the implicit reward model may result in forgetting the knowledge inbuilt in the initial policy at the first DPO stage. Inspired by the technique of experience replay (Rolnick et al., 2019) in continual learning for preventing catastrophic forgetting and making a good balance between old and new data, we propose to use a mixture of the generated data and the offline preference dataset. While offline preference data is considered to be high-quality, it is off-policy samples; the generated data is closer to the on-policy samples, but the imperfect implicit reward model may introduce noise in labeling the preference. Combining the two can make for a good balance. This idea is also related to deep Q-learning with demonstrations (Hester et al., 2018), which demonstrates that incorporating supervised learning from an offline dataset can accelerate online reinforcement learning. Our complete algorithm is summarized in Algorithm 1. During each iteration t, we generate K responses y1, y2, . . . , y K from the policy πθ(t 1) for each prompt x (Line 3). The preference dataset D(α) is constructed based on the LR rewards r LR, where the response with the highest LR reward is labeled yw and the one with the lowest reward is labeled yl. In this process, πθ(t 1) serves as the target policy, and πθ(t 2) serves as the reference policy. For t = 1, we denote the reference policy in the implicit reward model as πθ( 1), corresponding to the reference policy used in the initial DPO training. At Line 4, we construct the debiased dataset D(α ) by minimizing the average absolute Published as a conference paper at ICLR 2025 difference in response length, referring to Eq. (6). Subsequently, a mixed dataset Dt is created by sampling γ proportion of data from Doffline and (1 γ) proportion of data from D(α ) (Line 5). Finally, DPO training is conducted on Dt using πθ(t 1) as both the initial policy and the reference policy, resulting in the updated policy πθ(t) (Line 6). 4 EXPERIMENTS This section empirically investigates DICE. Our findings highlight several key points: (1) DICE significantly improves the model performance on the widely used leaderboard Alpaca Eval 2 (Li et al., 2023b), increasing length-controlled win rate by more than 8% for all the different base models; (2) our best model bootstrapped from a 8B base model (Llama-3-8B-DPO) achieves a better performance than Gemini Pro 1.0 (Team et al., 2023) without any extra human annotations (or external reward model) other than the preference dataset that is used in the initial DPO training of base model; (3) the two proposed techniques in Sections 3.1 and 3.2 are shown to be critical for DICE; (4) DICE demonstrates competitive performance relative to scalar reward models trained exclusively on the same seed data. 4.1 EXPERIMENT SETUP Base Models and Datasets. We adopt Llama-3-8B-DPO1 and zephyr-7B-beta2 as our base models. Both models are trained following the pipeline of Zephyr (Tunstall et al., 2023) on the Ultra Feedback (Cui et al., 2023) dataset. Llama-3-8B-DPO is trained based on metallama/Meta-Llama-3-8B , developed by Meng et al. (2024). zephyr-7B-beta is fine-tuned based on mistralai/Mistral-7B-v0.1 . We randomly sample a subset of around 10k preference pairs from Ultra Feedback as the offline dataset Doffline for our fine-tuning experiments. Our experiments aim to show how much the language model can improve from a DPO-tuned model and a subset of the preference dataset that was used to conduct the initial DPO training. Response Generation and Dataset Construction. At the start of each round, we sample responses from the current policy, with temperature T = 0.9, p = 1.0 for the Llama3 setting and T = 0.7, p = 0.9 for the Zephyr setting. We sample with different random seeds to get K = 16 diverse responses for each prompt. We then reward each prompt-response pair by the implicit reward model (Eq. (4)) and incorporate length-regularized reward shaping (Eq. (5)) to get the debiased dataset D(α ) with the optimal regularization strength α . The final dataset is a mixture of D(α ) and Doffline. Training Details. All experiments are conducted on 8 Nvidia A100 GPUs. For DICE, we trained two rounds in total. In each round, we train the model for 300 steps on a preference dataset with 9.6k preference pairs (either a solely generated dataset, or a mixture of the generated dataset and the offline preference dataset). The global training batch size is set to 32 and the learning rate is 5e-7 with a constant schedule and a warm-up of 50 steps. We hypertuned β {0.01, 0.1} based on the model performance on Alpaca Eval 2 for each method and model separately. For our approach, we additionally hypertuned the experience replay ratio γ using cross-validation to ensure fair assessment. Further details on hyperparameter tuning can be found in Appendix F. Baselines. We evaluate the following baseline methods applicable to the setting of this paper: Offline DPO: continue conducting DPO training with the offline preference dataset. Offline DPO w/ new ref: similar to offline DPO but we assign the trained policy after each round as the new reference model for the next round. This corresponds to γ = 1. DPO with prompted rewards: similar to self-rewarding LM (Yuan et al., 2024), where they prompt the LLM itself as a preference judge to construct new preference pairs and iteratively fine-tune the LLM with the DPO algorithm. In their case, the judge capability is learned by supervised fine-tuning on an evaluation fine-tuning dataset. We exploit our base model to perform LLM-as-a-Judge directly as it can follow the judge instructions well. In the experiment, we call it LLM-as-a-Judge. The LLM-as-a-Judge prompt template can be found in Appendix E. 1https://huggingface.co/princeton-nlp/Llama-3-Base-8B-SFT-DPO 2https://huggingface.co/Hugging Face H4/zephyr-7b-beta Published as a conference paper at ICLR 2025 Table 1: Results of Alpaca Eval 2 and Arena-Hard across two base models. LC and WR denote length-controlled and raw win rate in percentage (%) respectively. Method zephyr-7B-beta Llama-3-8B-DPO Alpaca Eval 2 Arena-Hard Alpaca Eval 2 Arena-Hard LC WR LC LC WR LC Base 12.69 10.71 10.17 18.20 15.50 22.32 Offline DPO Iter 1 13.40 11.10 14.13 20.22 18.33 24.78 Offline DPO Iter 2 4.96 5.47 1.89 21.04 19.21 24.04 Offline DPO (w/ new ref) Iter 1 13.40 11.10 13.10 22.29 19.96 22.2 Offline DPO (w/ new ref) Iter 2 4.58 5.27 5.15 22.50 20.18 23.55 LLM-as-a-Judge Iter 1 15.36 17.81 13.96 20.30 21.31 26.46 LLM-as-a-Judge Iter 2 14.14 17.89 13.47 21.80 22.42 27.99 DICE Iter 1 19.03 17.67 18.15 25.08 25.77 33.52 DICE Iter 2 20.71 20.16 18.45 27.55 30.99 39.13 * We note that the results of Llama-3-8B-DPO base are obtained from Meng et al. (2024). Evaluation. We evaluate our method by Alpaca Eval 2 (Li et al., 2023b) and Arena-Hard (Li et al., 2024). Both are LLM-based automatic evaluation benchmarks and have been widely adopted by the community. Alpaca Eval 2 employs Alpaca Farm (Dubois et al., 2023) as its prompts set composed of general human instructions. The model responses and the reference responses generated by GPT4-Turbo are fed into a GPT-4-Turbo-based annotator to be judged. We follow the standard approach and report both the win rate (WR) and the Length-Controlled win rate (LC) (Dubois et al., 2024) over the reference responses. Arena-Hard is a recently released benchmark, incorporating 500 welldefined technical problem-solving queries. Due to the expensive evaluation cost of Arena-Hard, we follow the official guidance and use a strong open-source model mistralai/Mistral-Large-Instruct2407 (123B)3 as the judge model. 4.2 MAIN RESULTS Table 2: Alpaca Eval 2 leaderboard results. Model LC WR GPT-4 0613 30.18 15.76 Mistral Medium 28.61 21.86 Claude 2 28.15 17.19 DICE-Llama3 8B Iter 2 27.55 30.99 Snorkel (Mistral-Pair RM-DPO) 26.39 30.22 Gemini Pro 24.38 18.18 Mixtral 8 7B v0.1 23.69 18.26 Llama 3 8B Instruct 22.92 22.57 GPT-3.5 Turbo 0613 22.35 14.10 Tulu 2+DPO 70B 21.24 15.98 DICE-Zephyr 7B Iter 2 20.71 20.16 GPT-3.5 Turbo 1106 19.30 9.18 Llama-3-8B-DPO 18.20 15.50 Vicuna 33B v1.3 17.58 12.71 zephyr-7B-beta 12.69 10.71 DICE Effectively Improves a DPO-tuned Model. With only a DPO-tuned model and a preference dataset that was used to train this model, one can choose to further improve the current policy via Offline DPO, or construct a new preference dataset via LLM-as-a-Judge. In Table 1, we compare the performance of the model fine-tuned by DICE in two rounds with the base model and other baselines. It shows all methods can improve the LC win rate on Alpaca Eval 2 over the base model while DICE leads to the most significant enhancement in both Zephyr and Llama3 settings, increasing by 8.02% and 9.35% respectively. We found that the LLM-as-a-Judge leads to good performance in the Zephyr setting while it has minor improvement in the Llama3 setting. We hypothesize this may be caused by the coarse rewards which are not able to provide effective preference signals when responses are of high quality (the prompt template requires LLM judge to provide a discrete score from 0 to 5, referring to Appendix E). We note that training on a fixed offline dataset for multiple rounds leads to even worse performance than the base model, possibly due to the increased data staleness and overfitting. 3https://huggingface.co/mistralai/Mistral-Large-Instruct-2407 Published as a conference paper at ICLR 2025 Table 3: Alpaca Eval 2 results showing DPO rewards are compatible with other DAP algorithms. Method Offline DICE LC WR LC WR DPO 13.40 11.10 19.03 17.67 IPO 14.83 16.16 18.51 19.49 KTO 13.92 10.65 14.88 12.16 Hinge 13.51 12.45 15.92 15.57 Table 4: Alpaca Eval 2 results showing effects of the Length Regularized Weight. γ α LC WR Avg. Len. 0.0 0.000 13.32 15.37 2570 0.023 (α ) 18.88 19.31 2109 0.046 (2α ) 14.40 9.08 876 0.5 0.000 15.92 19.08 2600 0.023 (α ) 19.03 17.67 1848 0.046 (2α ) 14.91 10.80 1185 Comparison with the models on the leaderboard. Compared with the models on the public leaderboard shown in Table 2, DICE-Llama3 8B performs better than the official instruct version of Llama3 by a non-trivial margin, 4.63%. Regarding the closed-source models, it achieves better performance than Gemini Pro with only 8B parameters and does not require any in-house data or external reward model. 4.3 DICE IS COMPATIBLE WITH OTHER DIRECT ALIGNMENT FROM PREFERENCE ALGORITHMS Though DICE works best with DPO as it makes the iterative training possible (because the implicit reward model for the next round can be naturally derived using the updated policy), we would like to check if the dataset generated by DICE can also improve the base model with other Direct Alignment from Preference (DAP) algorithms. In the Zephyr setting, we tune the policy model using DICE-generated dataset (at the first round) and the offline dataset with KTO (Ethayarajh et al., 2024), IPO (Azar et al., 2024), and Hinge loss proposed in Zhao et al. (2023). The training follows the protocol described in Section 4.1. The results in Table 3 show that all DAP algorithms benefit from the newly generated data by the current policy and DPO implicit reward model with LR reward shaping, demonstrating LC win rates higher than their offline counterparts. Notably, DICE shows the greatest improvement for DPO. 4.4 COMPARING DICE WITH INTERNAL REWARD MODEL Our experiment setup assumes access to an offline preference dataset and a DPO-tuned language model. To further improve the language model, one may choose to train a new scalar reward model and then conduct iterative DPO. As this scalar reward model is trained solely on the offline preference dataset without utilizing the external resources, we call it the internal reward model (Intl RM). In this section, we compare the Intl RM with implicit reward model to evaluate their performance in our self-alignment setting. Base Models and Datasets. We utilize mistral-7b-sft-beta4 as the base model. We adopt the full Ultra Feedback preference dataset (Cui et al., 2023), comprising 60k preference pairs. Training Details. For the implicit reward, we use zephyr-7B-beta along with its reference model mistral-7b-sft-beta. The scalar reward model is trained using Open RLHF (Hu et al., 2024) framework, adhering to the recommended training parameters5. This setup ensures that both the implicit reward model and Intl RM use the same base model (mistral-7b-sft-beta) and are trained on the same dataset (Ultra Feedback). Evaluation. We evaluate the reward models based on the alignment rate of the model with the preference labels provided by GPT-4o. The alignment rate is defined as m/n, where n is the total number of prompt-response pairs and m is the number of labels matching GPT-4o. As our goal is to provide preference labels for the model-generated responses to enable the subsequent DPO training, we sample 500 (x, y1, y2) tuples from the preference dataset during the first iteration of the DICE experiment under the Zephyr setting. Different reward models are then queried to provide the preference labels. 4https://huggingface.co/Hugging Face H4/mistral-7b-sft-beta 5https://github.com/Open RLHF/Open RLHF/blob/main/examples/scripts/train_ rm_llama.sh Published as a conference paper at ICLR 2025 Table 5: Alignment rate of the reward model with the preference labels provided by GPT-4o. Method Intl RM ERM-555k DPO Implicit Reward Alignment Rate 0.624 0.656 0.698 In the experiment, we include an external reward model, ERM-555k6, trained by Open RLHF maintainers on 555k preference data, as a stronger baseline. In Table 5, we observe that the DPO implicit reward achieves higher accuracy than the Intl RM trained on an equivalent amount of preference data. Furthermore, it surpasses the performance of the ERM-555k trained on substantially more data. This implies that the implicit reward model offers advantages when evaluating its own generated data, though the scalar reward models in general excel on a wider range of tasks when the preference data is abundant, as demonstrated by Armo RM-Llama3-8B-v0.1, which was trained on 1M preference data. Based on these findings, we argue that the implicit reward is a competitive option in self-alignment settings. 4.5 ABLATION STUDY In this section, we investigate the effects of LR reward shaping and experience replay. Effects of LR reward shaping. LR reward shaping (Eq. (5)) penalizes responses for being too verbose, and guides the construction of a debiased dataset with the optimal penalty strength α found by optimizing Eq. (6). To validate the effectiveness of the propose LR reward shaping as well as the α-searching procedure, we run experiments in the Zephyr setting with different mixture ratios (γ = 0 and γ = 0.5) and ablate three design choices: (1) no LR reward shaping (α = 0); (2) LR reward shaping with penalty strength found by Eq. (6), i.e., α = α = 0.023; (3) LR reward shaping with slightly larger penalty strength (α = 2α ). Results are presented in Table 4. For all values of the γ, we observe that α = 0.0 does lead to serious length exploitation. So, even if the policy can get a high win rate, it will suffer in the LC win rate due to the length exploitation issue (responses with longer average length get a lower LC win rate), e.g., the low LC win rate with γ = 0.5, α = 0.0. In contrast, a larger α seemingly mitigates the length exploitation issue even better, but it may adversely affect the response quality. Our proposed approach of finding α using the objective of minimizing the absolute difference of response length does provide the best performance. base iter 1 iter 2 5 20.71 LC Win Rate Figure 3: Alpaca Eval 2 LC Win Rate across different experience replay ratios (γ) in the Zephyr setting. The highest LC Win Rate is reported via text. Effects of experience replay. The experience replay results in a new mixed dataset in which γ fraction of the data is from the offline dataset and (1 γ) fraction of the data is from the generated dataset. E.g., we use data only from the generated dataset if γ = 0.0. We run experiments in the Zephyr setting with γ {0.0, 0.25, 0.5, 0.75, 1.0} and conduct DICE in total of two rounds. From the results shown in Figure 3, we find γ = 0.5 provides the best performance. The results satisfy our expectations. With only offline preference data, the DPO optimizes the current policy with offpolicy samples that are further away from its distribution. With only its own generated data, the model may keep reinforcing its current belief , potentially leading to catastrophic forgetting. An intermediate value γ = 0.5 finds a good balance. Additional results in the Llama3 setting can be found in Appendix D. 5 LIMITATION AND FUTURE WORK Limitations. One of the primary limitations of our work is the reliance on the DPO training prior to bootstrapping. If the implicit reward model is not well-trained, it can lead to a collapse of the training pipeline. This challenge is not unique to our approach; the classic RLHF pipeline is also 6https://huggingface.co/Open RLHF/Llama-3-8b-rm-mixture Published as a conference paper at ICLR 2025 struggling when its reward model is not well-trained. Another limitation is the lack of continued improvement over many iterations. Similar to other research, such as the work by Yuan et al. (2024), which enhances the policy model without an external reward model, we did not observe continuous improvement in our model beyond three iterations. This issue highlights an open question within this field regarding the iterative enhancement of policy models. Future Work. Future research could explore the rewarding capabilities of models trained using other DPO variants, such as KTO and IPO. Investigating whether these variants can offer general rewards similar to those provided by DPO-tuned policy would be valuable. Another promising direction involves developing methods that enable continuous improvement of the policy model over iterations without degradation. Additionally, investigating a theoretical understanding of the policy learned through self-bootstrapping could provide deeper insights into the mechanics of our approach and facilitate further advancements. 6 RELATED WORK Self-Improving Fine-Tuning. Many studies explore fine-tuning language models without a large amount of human annotation (Huang et al., 2022; Li et al., 2023a; Sun et al., 2023; 2024; Yuan et al., 2024; Chen et al., 2024b; Zhang et al., 2025). Starting from an SFT model, Sun et al. (2023) collect the preference labels by prompting the SFT model itself to choose the preferred response based on principles, training a principle-driven reward model, and optimizing via PPO. Yuan et al. (2024) construct a preference dataset by their own supervised fine-tuned model trained on instruction following data and evaluation fine-tuning data, followed by DPO training. Different from these approaches, which align models by leveraging pretrained knowledge with minimal seed data, our work focuses on further improving a DPO-tuned model through bootstrapping with its implicit rewards. On-Policy Sampling in Preference Tuning. DPO and its variants are widely used for their simplicity, but their offline nature often limits policy learning (Guo et al., 2024). Guo et al. (2024) show that offline DPO quickly overfits the preference dataset, while it performs much better and is more stable if online feedback can be provided to their on-policy samples7. Tajwar et al. (2024) further analyze the properties of different preference tuning methods, concluding that on-policy sampling enhances performance and efficiency. When pure on-policy sampling is infeasible, using data closer to on-policy samples, such as in iterative DPO with external rewards (Tran et al., 2023; Dong et al., 2024; Xiong et al., 2024; Ding et al., 2024; Pang et al., 2024; Muldrew et al., 2024) or internal rewards (Yuan et al., 2024; Furuta et al., 2025), can still be beneficial. Similarly, our approach enables us to train the policy model with the preference data which is closer to on-policy samples than the offline dataset without any external reward models. We hypothesize this is one of the main gain sources of our approach. DPO Implicit Rewards. Rafailov et al. (2024a) recently studied DPO from a token-level MDP perspective, revealing that DPO-tuned models implicitly parameterize token-wise dense reward functions. Building on similar theoretical fundations, Zhong et al. (2024) pretrains a DPO model to serve as a standalone dense reward generator for PPO training. Unlike these works, we leverage the implicit reward model as an outcome reward model to bootstrap itself for self-alignment. Implicit reward models have also been used for selecting the pairwise data (Yang et al., 2024; Chen et al., 2024a) to improve the annotation efficiency in preference tuning. 7 CONCLUSION In this paper, we introduce DICE, a novel approach that leverages the implicit reward model from DPO to further align LLMs with human preferences. Our method stands out in the current landscape of LLM alignment research, as it uses the implicit reward model to iteratively refine the policy model. Empirical results show that our approach DICE enables significant (more than 8% LC win rate increase on Alpaca Eval 2) improvement in alignment across different base models, without relying on any external feedback. 7On-policy samples: the data collected while following the current policy that is being optimized. Published as a conference paper at ICLR 2025 ETHICS STATEMENT This paper presents work whose goal is to advance the field of Machine Learning. There are many potential societal consequences of our work, none of which we feel must be specifically highlighted here. Mohammad Gheshlaghi Azar, Zhaohan Daniel Guo, Bilal Piot, Remi Munos, Mark Rowland, Michal Valko, and Daniele Calandriello. A general theoretical paradigm to understand learning from human preferences. In International Conference on Artificial Intelligence and Statistics, pp. 4447 4455. PMLR, 2024. Ralph Allan Bradley and Milton E Terry. Rank analysis of incomplete block designs: I. the method of paired comparisons. Biometrika, 39(3/4):324 345, 1952. Yifang Chen, Shuohang Wang, Ziyi Yang, Hiteshi Sharma, Nikos Karampatziakis, Donghan Yu, Kevin Jamieson, Simon Shaolei Du, and Yelong Shen. Cost-effective proxy reward model construction with on-policy and active learning. ar Xiv preprint ar Xiv:2407.02119, 2024a. Zixiang Chen, Yihe Deng, Huizhuo Yuan, Kaixuan Ji, and Quanquan Gu. Self-play fine-tuning converts weak language models to strong language models. In Forty-first International Conference on Machine Learning, 2024b. Paul F Christiano, Jan Leike, Tom Brown, Miljan Martic, Shane Legg, and Dario Amodei. Deep reinforcement learning from human preferences. Advances in neural information processing systems, 2017. Ganqu Cui, Lifan Yuan, Ning Ding, Guanming Yao, Wei Zhu, Yuan Ni, Guotong Xie, Zhiyuan Liu, and Maosong Sun. Ultrafeedback: Boosting language models with high-quality feedback, 2023. Mucong Ding, Souradip Chakraborty, Vibhu Agrawal, Zora Che, Alec Koppel, Mengdi Wang, Amrit Bedi, and Furong Huang. Sail: Self-improving efficient online alignment of large language models. ar Xiv preprint ar Xiv:2406.15567, 2024. Hanze Dong, Wei Xiong, Bo Pang, Haoxiang Wang, Han Zhao, Yingbo Zhou, Nan Jiang, Doyen Sahoo, Caiming Xiong, and Tong Zhang. Rlhf workflow: From reward modeling to online rlhf. ar Xiv preprint ar Xiv:2405.07863, 2024. Yann Dubois, Xuechen Li, Rohan Taori, Tianyi Zhang, Ishaan Gulrajani, Jimmy Ba, Carlos Guestrin, Percy Liang, and Tatsunori B. Hashimoto. Alpacafarm: A simulation framework for methods that learn from human feedback, 2023. Yann Dubois, Bal azs Galambosi, Percy Liang, and Tatsunori B Hashimoto. Length-controlled alpacaeval: A simple way to debias automatic evaluators. ar Xiv preprint ar Xiv:2404.04475, 2024. Kawin Ethayarajh, Winnie Xu, Niklas Muennighoff, Dan Jurafsky, and Douwe Kiela. Kto: Model alignment as prospect theoretic optimization. ar Xiv preprint ar Xiv:2402.01306, 2024. Hiroki Furuta, Kuang-Huei Lee, Shixiang Shane Gu, Yutaka Matsuo, Aleksandra Faust, Heiga Zen, and Izzeddin Gur. Geometric-averaged preference optimization for soft preference labels. Advances in Neural Information Processing Systems, 37:57076 57114, 2025. Shangmin Guo, Biao Zhang, Tianlin Liu, Tianqi Liu, Misha Khalman, Felipe Llinares, Alexandre Rame, Thomas Mesnard, Yao Zhao, Bilal Piot, et al. Direct language model alignment from online ai feedback. ar Xiv preprint ar Xiv:2402.04792, 2024. Todd Hester, Matej Vecerik, Olivier Pietquin, Marc Lanctot, Tom Schaul, Bilal Piot, Dan Horgan, John Quan, Andrew Sendonaris, Ian Osband, et al. Deep q-learning from demonstrations. In Proceedings of the AAAI conference on artificial intelligence, volume 32, 2018. Jian Hu, Xibin Wu, Weixun Wang, Dehao Zhang, Yu Cao, et al. Openrlhf: An easy-to-use, scalable and high-performance rlhf framework. ar Xiv preprint ar Xiv:2405.11143, 2024. Published as a conference paper at ICLR 2025 Jiaxin Huang, Shixiang Shane Gu, Le Hou, Yuexin Wu, Xuezhi Wang, Hongkun Yu, and Jiawei Han. Large language models can self-improve. ar Xiv preprint ar Xiv:2210.11610, 2022. Harrison Lee, Samrat Phatale, Hassan Mansoor, Kellie Lu, Thomas Mesnard, Colton Bishop, Victor Carbune, and Abhinav Rastogi. Rlaif: Scaling reinforcement learning from human feedback with ai feedback. ar Xiv preprint ar Xiv:2309.00267, 2023. Tianle Li, Wei-Lin Chiang, Evan Frick, Lisa Dunlap, Banghua Zhu, Joseph E Gonzalez, and Ion Stoica. From live data to high-quality benchmarks: The arena-hard pipeline, 2024. Xian Li, Ping Yu, Chunting Zhou, Timo Schick, Luke Zettlemoyer, Omer Levy, Jason Weston, and Mike Lewis. Self-alignment with instruction backtranslation. ar Xiv preprint ar Xiv:2308.06259, 2023a. Xuechen Li, Tianyi Zhang, Yann Dubois, Rohan Taori, Ishaan Gulrajani, Carlos Guestrin, Percy Liang, and Tatsunori B. Hashimoto. Alpacaeval: An automatic evaluator of instruction-following models, 2023b. Yu Meng, Mengzhou Xia, and Danqi Chen. Simpo: Simple preference optimization with a reference-free reward. ar Xiv preprint ar Xiv:2405.14734, 2024. William Muldrew, Peter Hayes, Mingtian Zhang, and David Barber. Active preference learning for large language models. In Proceedings of the 41st International Conference on Machine Learning, pp. 36577 36590, 2024. Long Ouyang, Jeffrey Wu, Xu Jiang, Diogo Almeida, Carroll Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, et al. Training language models to follow instructions with human feedback. Advances in neural information processing systems, 35: 27730 27744, 2022. Richard Yuanzhe Pang, Weizhe Yuan, Kyunghyun Cho, He He, Sainbayar Sukhbaatar, and Jason Weston. Iterative reasoning preference optimization. ar Xiv preprint ar Xiv:2404.19733, 2024. Ryan Park, Rafael Rafailov, Stefano Ermon, and Chelsea Finn. Disentangling length from quality in direct preference optimization. ar Xiv preprint ar Xiv:2403.19159, 2024. Rafael Rafailov, Joey Hejna, Ryan Park, and Chelsea Finn. From r to q : Your language model is secretly a q-function. ar Xiv preprint ar Xiv:2404.12358, 2024a. Rafael Rafailov, Archit Sharma, Eric Mitchell, Christopher D Manning, Stefano Ermon, and Chelsea Finn. Direct preference optimization: Your language model is secretly a reward model. Advances in Neural Information Processing Systems, 36, 2024b. David Rolnick, Arun Ahuja, Jonathan Schwarz, Timothy Lillicrap, and Gregory Wayne. Experience replay for continual learning. Advances in neural information processing systems, 2019. John Schulman, Filip Wolski, Prafulla Dhariwal, Alec Radford, and Oleg Klimov. Proximal policy optimization algorithms. ar Xiv preprint ar Xiv:1707.06347, 2017. Nisan Stiennon, Long Ouyang, Jeffrey Wu, Daniel Ziegler, Ryan Lowe, Chelsea Voss, Alec Radford, Dario Amodei, and Paul F Christiano. Learning to summarize with human feedback. Advances in Neural Information Processing Systems, 33:3008 3021, 2020. Zhiqing Sun, Yikang Shen, Hongxin Zhang, Qinhong Zhou, Zhenfang Chen, David Cox, Yiming Yang, and Chuang Gan. Salmon: Self-alignment with principle-following reward models. ar Xiv preprint ar Xiv:2310.05910, 2023. Zhiqing Sun, Yikang Shen, Qinhong Zhou, Hongxin Zhang, Zhenfang Chen, David Cox, Yiming Yang, and Chuang Gan. Principle-driven self-alignment of language models from scratch with minimal human supervision. Advances in Neural Information Processing Systems, 36, 2024. Richard S Sutton and Andrew G Barto. Reinforcement learning: An introduction. MIT press, 2018. Published as a conference paper at ICLR 2025 Fahim Tajwar, Anikait Singh, Archit Sharma, Rafael Rafailov, Jeff Schneider, Tengyang Xie, Stefano Ermon, Chelsea Finn, and Aviral Kumar. Preference fine-tuning of llms should leverage suboptimal, on-policy data. ar Xiv preprint ar Xiv:2404.14367, 2024. Yunhao Tang, Daniel Zhaohan Guo, Zeyu Zheng, Daniele Calandriello, Yuan Cao, Eugene Tarassov, R emi Munos, Bernardo Avila Pires, Michal Valko, Yong Cheng, and Will Dabney. Understanding the performance gap between online and offline alignment algorithms. ar Xiv preprint ar Xiv: 2405.08448, 2024. Gemini Team, Rohan Anil, Sebastian Borgeaud, Jean-Baptiste Alayrac, Jiahui Yu, Radu Soricut, Johan Schalkwyk, Andrew M Dai, Anja Hauth, Katie Millican, et al. Gemini: a family of highly capable multimodal models. ar Xiv preprint ar Xiv:2312.11805, 2023. Hoang Tran, Chris Glaze, and Braden Hancock. Iterative dpo alignment. Technical report, Snorkel AI, 2023. Lewis Tunstall, Edward Beeching, Nathan Lambert, Nazneen Rajani, Kashif Rasul, Younes Belkada, Shengyi Huang, Leandro von Werra, Cl ementine Fourrier, Nathan Habib, Nathan Sarrazin, Omar Sanseviero, Alexander M. Rush, and Thomas Wolf. Zephyr: Direct distillation of lm alignment, 2023. Yue Wu, Zhiqing Sun, Huizhuo Yuan, Kaixuan Ji, Yiming Yang, and Quanquan Gu. Self-play preference optimization for language model alignment. ar Xiv preprint ar Xiv:2405.00675, 2024. Yuxi Xie, Anirudh Goyal, Wenyue Zheng, Min-Yen Kan, Timothy P Lillicrap, Kenji Kawaguchi, and Michael Shieh. Monte carlo tree search boosts reasoning via iterative preference learning. ar Xiv preprint ar Xiv:2405.00451, 2024. Wei Xiong, Hanze Dong, Chenlu Ye, Ziqi Wang, Han Zhong, Heng Ji, Nan Jiang, and Tong Zhang. Iterative preference learning from human feedback: Bridging theory and practice for rlhf under kl-constraint. In Forty-first International Conference on Machine Learning, 2024. Sen Yang, Leyang Cui, Deng Cai, Xinting Huang, Shuming Shi, and Wai Lam. Not all preference pairs are created equal: A recipe for annotation-efficient iterative preference learning. ar Xiv preprint ar Xiv:2406.17312, 2024. Weizhe Yuan, Richard Yuanzhe Pang, Kyunghyun Cho, Sainbayar Sukhbaatar, Jing Xu, and Jason Weston. Self-rewarding language models. ar Xiv preprint ar Xiv:2401.10020, 2024. Xuan Zhang, Chao Du, Tianyu Pang, Qian Liu, Wei Gao, and Min Lin. Chain of preference optimization: Improving chain-of-thought reasoning in llms. Advances in Neural Information Processing Systems, 37:333 356, 2025. Yao Zhao, Rishabh Joshi, Tianqi Liu, Misha Khalman, Mohammad Saleh, and Peter J Liu. Slic-hf: Sequence likelihood calibration with human feedback. ar Xiv preprint ar Xiv:2305.10425, 2023. Han Zhong, Guhao Feng, Wei Xiong, Li Zhao, Di He, Jiang Bian, and Liwei Wang. Dpo meets ppo: Reinforced token optimization for rlhf. ar Xiv preprint ar Xiv: 2404.18922, 2024. Published as a conference paper at ICLR 2025 A BENEFITS OF ON-POLICY SAMPLING In this section, we provide a theoretical analysis similar to Xie et al. (2024) to show that learning with on-policy samples can be more effective than utilizing an offline dataset. We make a generalized notation for the sampling policy at the t-th round as π(t), and compare the on-policy sampling (π(t) = πθ(t 1)) with sampling from an offline dataset (π(t) =πµ8). For a prompt x, we denote its optimal response9 as y and suboptimal ones as S = {y i }. We also denote its preferences sampled from π(t) as (y(t) w , y(t) l ). By Eq. (3) and the definition of the logistic function, we can rewrite the DPO loss at round t for the sample (x, y(t) w , y(t) l ) as: L(t) DPO(πθ(t); π(t) ref) = log πθ(t) y(t) w | x β πθ(t) y(t) w | x β + πθ(t) y(t) l | x β Rβ , (7) where R = π(t) ref(y(t) w | x)/π(t) ref(y(t) l | x) is a constant. Observe that Eq. (7) can be minimized to zero by just minimizing πθ(t)(y(t) l |x) to be zero, without minimizing the likelihood of other suboptimal responses πθ(t)(y |x) for y = y(t) l . After t rounds of optimization, we are interested in the probability of outputting the optimal response: πθ(t)(y | x) = 1 X y i S πθ(t)(y i | x). (8) Eq. (8) allows us to reveal the deficiency of training on a fixed offline dataset. If there exists a suboptimal response y S that lies in the high likelihood region of πθ(t), say πθ(t)(y | x) p for some p [0, 1] that is close to 1, and y is never sampled from π(t) = πµ thus not optimized as y(t) l during all t rounds, we have: πθ(t)(y | x) 1 πθ(t)(y | x) 1 p. (9) Since πµ is zero except points that appear in Doffline, it is highly likely to find such a y not being sampled from πµ (hence never optimized) and therefore πθ(t)(y | x) can be very low with a large p. On the other hand, conducting on-policy sampling can alleviate the never-sampled issue and promote convergence to the optimal policy. This is because whenever πθ(t 1)(y i | x) is high, it is likely to sample y i from π(t) = πθ(t 1) and thus it can be optimized such that πθ(t) y(t) l = y i | x 0. In this way, the subtrahend of Eq. (8) is decreased per round, hence we can gradually improve the language model policy towards the optimal policy. B EXPERIMENT BEYOND TWO ITERATIONS We conduct DICE and LLM-as-a-judge baseline (the strongest baseline) for three iterations, evaluating them on Alpaca Eval 2 benchmark (see Table 6). From the table, we observe a drop in LC win rate at the third iteration for both our approach and the LLM-as-a-Judge baseline. This is a known challenge in the literature; both self-rewarding language model (Yuan et al., 2024) and SPPO (Wu et al., 2024) show successful improvement only up to three iterations, which corresponds to two iterations in our approach, as our method begins after one initial iteration of DPO training. Additionally, we use a fixed prompt set in every iteration, whereas other works employ different prompt sets across iterations. We hypothesize that using different prompt sets across iterations could help mitigate the performance drop observed in the third iteration. 8πµ is the behavior policy that collects the offline dataset, which can be a mixture of existing LLMs or human experts. 9Optimality is measured by the underlying BT reward model, instead of the implicit rewards. In general there can be multiple optimal responses, but here we assume a single optimal response for simplicity. Published as a conference paper at ICLR 2025 Table 6: Results of Alpaca Eval 2 across two base models over three iterations. LC and WR denote length-controlled and raw win rate in percentage (%) respectively. Method zephyr-7B-beta Llama-3-8B-DPO Alpaca Eval 2 Alpaca Eval 2 LC WR LC WR Base 12.69 10.71 18.20 15.50 LLM-as-a-Judge Iter 1 15.36 17.81 20.30 21.31 LLM-as-a-Judge Iter 2 14.14 17.89 21.80 22.42 LLM-as-a-Judge Iter 3 13.95 17.23 16.86 19.78 DICE Iter 1 19.03 17.67 25.08 25.77 DICE Iter 2 20.71 20.16 27.55 30.99 DICE Iter 3 14.02 15.30 20.61 26.18 * We note that the results of Llama-3-8B-DPO base are obtained from Meng et al. (2024). C OPTIMIZATION FOR LENGTH REGULARIZED REWARD SHAPING We solve the objective in Eq. (6) using a simple Bayesian optimization toolkit based on Gaussian process10. The objective landscape with respect to α is depicted in Figure 4, where we compare the proposed implicit reward model with the LLM-as-a-Judge reward model. With our lengthreguarlized implicit rewards, the optimizer is able to find the optimal solution quickly that debiases the length difference of the winning and losing responses. For LLM-as-a-Judge rewards, the optimal solution is obtained with α = 0, hence we do not explicitly debias the dataset for all the experiments. 0.00 0.05 0.10 α 0.00 0.05 0.10 α Average Absolute Length Difference Figure 4: The objective landscape for (Left) our implicit reward model and (Right) the LLM-as-a Judge reward model. D EXTENDED ABLATION STUDY Effects of experience replay with Llama3 backbone model. In the Llama3 setting, we also conduct a coarse sweeping for the experience ratio γ, and present the Alpaca Eval 2 LC win rate in Figure 5 for two self-alignment rounds. We observe similar trends to those in the Zephyr setting, which further justify the effectiveness of the proposed experience replay: it helps to keep a balance between the more on-policy generated data and the curated offline data. The best identified value of the mixture ratio is γ = 0.1. 10https://scikit-optimize.github.io/stable/modules/generated/skopt.gp_ minimize.html Published as a conference paper at ICLR 2025 base iter 1 iter 2 18 28 LC Win Rate Figure 5: Alpaca Eval 2 LC Win Rate across different experience replay ratio (γ) in the Llama3 setting. Comparison between Length-Regularized DPO and LR reward shaping. The lengthregularized DPO proposed by Park et al. (2024) incorporates a penalty term into the DPO objective to explicitly penalize length exploitation. In contrast, our approach mitigates this issue by applying reward shaping during preference dataset construction. Since both methods target the same problem, we compare their performance in the Zephyr setting with γ = 0. To clarify notation, we denote the length regularization coefficient in Park et al. (2024) as λ. Length-regularized DPO uses the generated dataset D(0) without reward shaping, and λ is hypertuned from {0.02, 0.05} as suggested in length-regularized DPO paper. In comparison, our method employs the dataset D(α ). All other settings remain consistent across both approaches. As shown in Table 7, we observe that our approach achieves better performance while effectively mitigating the length exploitation. In contrast, regularized DPO exhibits a quality-length trade-off, where larger regularization coefficients successfully control response length but lead to a decline in response quality. Table 7: Results of Alpaca Eval 2 comparing techniques for mitigating length exploitation. LC and WR denote length-controlled and raw win rate in percentage (%) respectively. Len denotes the average string length of the model response. Method λ Alpaca Eval 2 No mitigation - 13.32 15.37 2570 Regularized DPO 0.02 16.30 19.16 2629 0.05 16.03 16.14 2030 Ours - 18.88 19.31 2109 PPO with Implicit Reward. To assess whether implicit rewards can be effectively utilized in reward model-based methods such as PPO (Schulman et al., 2017), we conduct experiments using PPO with an implicit reward model. Specifically, the zephyr-7B-beta model and its SFT model are used to construct the implicit reward model. Additionally, zephyr-7B-beta serves as the base model to be optimized. We adopt the implementation in Open RLHF (Hu et al., 2024), adhering to the recommended training parameters. Evaluation results on Alpaca Eval 2 are presented in Table 8. The results indicate a slight improvement in length-controlled win rate for PPO, although the improvement is less pronounced compared to DPO. While PPO s performance can potentially be enhanced with additional hyperparameter tuning, such efforts would require substantial resources and time. We leave this for future work. E PROMPT USED BY LLM-AS-A-JUDGE We provide the prompt used by LLM-as-a-judge in Figure 6. Published as a conference paper at ICLR 2025 Table 8: Results of Alpaca Eval 2 for Zephyr backbone. LC and WR denote length-controlled and raw win rate in percentage (%) respectively. Method Alpaca Eval 2 Base 12.69 10.71 PPO 13.32 8.55 DPO 19.03 17.67 LLM-as-a-Judge Prompt Template Review the user s question and the corresponding response using the additive 5-point scoring system described below. Points are accumulated based on the satisfaction of each criterion: Add 1 point if the response is relevant and provides some information related to the user s inquiry, even if it is incomplete or contains some irrelevant content. Add another point if the response addresses a substantial portion of the user s question, but does not completely resolve the query or provide a direct answer. Award a third point if the response answers the basic elements of the user s question in a useful way, regardless of whether it seems to have been written by an AI Assistant or if it has elements typically found in blogs or search results. Grant a fourth point if the response is clearly written from an AI Assistant s perspective, addressing the user s question directly and comprehensively, and is wellorganized and helpful, even if there is slight room for improvement in clarity, conciseness or focus. Bestow a fifth point for a response that is impeccably tailored to the user s question by an AI Assistant, without extraneous information, reflecting expert knowledge, and demonstrating a high-quality, engaging, and insightful answer. User: {instruction} Response: {response} After examining the user s instruction and the response, provide brief step-by-step justifications and conclude with a score between 0 to 5. You must follow the format below: Evaluation: Score: Figure 6: We follow the prompt template used by Yuan et al. (2024) to use LLMs to judge model responses and construct paired dataset for further preference tuning. F HYPERPARAMETER TUNING In our experiments, the hyperparameters, including γ and β, were selected based on the model performance on Alpaca Eval 2 (AE2). These parameters were tuned for each method and model separately (with γ specific to our approach) but remained fixed across iterations. Specifically, β was tuned during the first iteration from the set {0.01, 0.1}, while γ was tuned using the ranges {0.0, 0.25, 0.50, 0.75, 1.0} for the Zephyr backbone and {0.0, 0.10, 0.25, 0.50, 0.75, 1.0} for the Llama-3 backbone. The best hyperparameter setting identified in Alpaca Eval 2 was directly applied in Arena-Hard (AH) for all methods and models without any further tuning. Compared to baseline approaches, our approach has the additional hyperparameter γ which may offer potential advantages during evaluation when hypertuning is performed on AE2. To address this concern, we treat the two benchmarks (AE2 and AH) as proxies for validation and test sets to ensure fair comparisons. Specifically, we validate on AE2 while testing on AH, and then reverse the roles of the two benchmarks. Below we show the results in Tables 9 and 10: Published as a conference paper at ICLR 2025 Table 9: Model performance on Alpaca Eval 2 (AE2) and Arena-Hard (AH) across different γ for the Zephyr Backbone. LC denotes the length-controlled win rate. CI denotes confidence interval. γ LC. AE2 LC.AH: Score (95% CI) 0.00 18.88 17.6 (-1.4, 1.7) 0.25 17.87 16.1 (-1.2, 1.4) 0.50 20.71 18.4 (-1.3, 1.6) 0.75 16.90 14.9 (-1.6, 1.4) 1.00 13.40 13.0 (-1.3, 1.3) Table 10: Model performance on Alpaca Eval 2 (AE2) and Arena-Hard (AH) across different γ for the Llama-3 Backbone. LC denotes the length-controlled win rate. CI denotes confidence interval. γ LC. AE2 LC.AH: Score (95% CI) 0.00 26.93 39.1 (-2.3, 2.5) 0.10 27.55 39.1 (-2.0, 2.4) 0.25 24.78 33.3 (-2.1, 1.8) 0.50 23.59 32.2 (-2.3, 2.2) 0.75 22.57 27.4 (-2.0, 1.8) 1.00 22.50 24.5 (-1.7, 1.8) The two benchmarks and two backbone models constitute four validation-test instances. We observe that the γ selected from the validation set/benchmark consistently achieves the best performance on the test set across all instances, demonstrating the optimal value of γ is robust to the choice of the validation set/benchmark. Meanwhile, we do observe that the optimal γ for different backbone models is varied, which is expected, as different backbones produce responses of varying quality. Higherquality generated data reduces the need for replayed experience, aligning with our observations for the Llama-3 backbone.