# scaling_laws_for_reward_model_overoptimization__214f66b8.pdf Scaling Laws for Reward Model Overoptimization Leo Gao 1 John Schulman 1 Jacob Hilton 1 In reinforcement learning from human feedback, it is common to optimize against a reward model trained to predict human preferences. Because the reward model is an imperfect proxy, optimizing its value too much can hinder ground truth performance, in accordance with Goodhart s law. This effect has been frequently observed, but not carefully measured due to the expense of collecting human preference data. In this work, we use a synthetic setup in which a fixed gold-standard reward model plays the role of humans, providing labels used to train a proxy reward model. We study how the gold reward model score changes as we optimize against the proxy reward model using either reinforcement learning or best-of-n sampling. We find that this relationship follows a different functional form depending on the method of optimization, and that in both cases its coefficients scale smoothly with the number of reward model parameters. We also study the effect on this relationship of the size of the reward model dataset, the number of reward model and policy parameters, and the coefficient of the KL penalty added to the reward in the reinforcement learning setup. We explore the implications of these empirical results for theoretical considerations in AI alignment. 1. Introduction Goodhart s law is an adage that states, When a measure becomes a target, it ceases to be a good measure. In machine learning, this effect arises with proxy objectives provided by static learned models, such as discriminators and reward models. Optimizing too much against such a model eventually hinders the true objective, a phenomenon we refer to as overoptimization. It is important to understand the size of this effect and how it scales, in order to predict how much a 1Open AI. Correspondence to: Leo Gao . Proceedings of the 40 th International Conference on Machine Learning, Honolulu, Hawaii, USA. PMLR 202, 2023. Copyright 2023 by the author(s). Figure 1: Diagram of the real and synthetic Reward Model (RM) training setups. Human labellers generate comparison data. In the real RLHF setting, this data is used to train a proxy RM that is optimized by RL/Bo N. In our synthetic setting, we instead use a Gold RM as our ground truth. In both settings, the proxy RM is a proxy for the ground truth process generating the labels (either the human or gold RM). learned model can be safely optimized against. Moreover, studying this effect empirically could aid in the development of theoretical models of Goodhart s law for neural networks, which could be critical for avoiding catastrophic misalignment of future AI systems. In this work, we study overoptimization in the context of large language models fine-tuned as reward models trained to predict which of two options a human will prefer. Such reward models have been used to train language models to perform a variety of complex tasks that are hard to judge automatically, including summarization (Stiennon et al., 2020), question-answering (Nakano et al., 2021; Menick et al., 2022), and general assistance (Ouyang et al., 2022; Bai et al., 2022; Glaese et al., 2022). Typically, the reward model score is optimized using either policy gradient-based reinforcement learning or best-of-n sampling, also known as rejection sampling or reranking. Overoptimization can occur with both methods, and we study both to better understand whether and how overoptimization behaves differently across both methods. A major challenge in studying overoptimization in this context is the expense of collecting human preference labels. A large number of labels are required to accurately estimate overall preference probabilities, and this is exacerbated by small effect sizes and the need to take many measurements in order to fit scaling laws. To overcome this, we use a synthetic setup that is described in Section 2, in which la- Scaling Laws for Reward Model Overoptimization bels are supplied by a gold-standard reward model (RM) instead of humans. Our main results are empirically validated functional forms for the gold reward model scores R as a function of the Kullback Leibler divergence from the initial policy to the optimized policy DKL (π πinit), which depends on the method of optimization used. This KL divergence between the initial and optimized policies increases monotonically during RL training (Figure 14), and can be computed analytically as a function of n for Bo N. Further, following Bai et al. (2022, Section 4.3), we will define d := p DKL (π πinit), and write our functional forms in terms of d. We find empirically that for best-of-n (Bo N) sampling, Rbon (d) = d (αbon βbond) , and for reinforcement learning,1 RRL (d) = d (αRL βRL log d) , Here, R(0) := 0 by convention because the reward is translation invariant, and αRL, βRL, αbon and βbon are parameters that may depend on the number of proxy reward model parameters, the size of the proxy reward model dataset, and so on. We see that these scaling laws make accurate predictions. We also find the following qualitative trends in addition to our quantitative results. RL versus best-of-n. As a function of the KL divergence, reinforcement learning tends to be slower than best-of-n sampling at both optimization and overoptimization. This suggests inadequacies with using KL to compare the amount of (over)optimization across methods. However, the relationship between the proxy reward model score and the gold reward model score is similar for both methods. Smooth coefficient scaling. The α and β coefficients in the Bo N and RL functional forms vary smoothly with the number of proxy reward model parameters, following approximate logarithmic trends.2 This allows prediction of attained gold RM score. Weak dependence on policy size. While larger policies perform better overall and benefit less from optimization against an RM as measured by increase 1We note that this form likely does not hold near the origin, as it has infinite slope there. We experimented with a number of different forms, but found worse fits and extrapolation. See Appendix B for more details. 2The coefficient αRL in particular being nearly independent of RM parameter count. in gold reward, they lead to very similar amounts of overoptimization, as measured through the gap between the proxy and gold scores (which indicates the shortfall between predicted and actual reward), and KL divergence at which the maximum gold RM score is attained. KL penalty ineffectiveness. In our reinforcement learning setup, using a KL penalty increases the proxy reward model score that can be achieved for a given KL divergence, but this does not correspond to a measurable improvement in the gold RM score KL frontier. However, we note this result could be particularly sensitive to hyperparameters. Finally, we discuss the implications of these findings for Reinforcement Learning From Human Feedback (RLHF), existing models of Goodhart s law, and AI Alignment more broadly. 2. Methodology The setting used throughout this paper is the same as for Instruct GPT (Ouyang et al., 2022). In our environment, the observations are text prompts and the policy is used to generate a response to the prompt. The prompts are drawn from a broad range of natural language instructions describing different language model tasks. Then, a learned RM is used to provide the reward signal for the response, which is used by either RL or Bo N for optimization. For all experiments, we use pretrained GPT-3 series language models as the initial checkpoint (Brown et al., 2020). All initial policies are trained with supervised fine-tuning (SFT) on human-generated Instruct GPT demonstrations (Ouyang et al., 2022) for 2 epochs. All RMs also use the GPT-3 architecture but have an added scalar head to output the reward. The RL experiments use Proximal Policy Optimization (PPO) (Schulman et al., 2017). The KL penalty for all RL experiments is set to 0 except for in Section 3.6. See Appendix C for all other hyperparameters. We mostly use defaults for the PPO hyperparameters; thus, it is possible that there exist different trends for other hyperparameter configurations. In Bo N, we generate n trajectories for the policy and use the reward model to pick the one with the highest proxy RM score. We use the unbiased estimator from Nakano et al. (2021, Appendix I) to compute all Bo N gold and proxy scores. This results in substantially better efficiency and lower variance than the naive estimator of repeatedly sampling n samples with replacement and taking the mean of the maximum gold and proxy RM scores. The KL divergences for Bo N are computed analytically: KLbon = log n n 1 Scaling Laws for Reward Model Overoptimization Figure 2: Reward model (RM) parameter size scaling experiments using the Instruct GPT environment. Policy size is held constant (1.2B), while reward model size is varied. The x-axes have a square-root scale. Note that the plots have different x-axes. The gold reward represents the ground truth reward; we observe that when we optimize for a learned proxy of the gold reward, the gold reward initially increases and later decreases. We show that our functional forms fit this effect well. Scaling Laws for Reward Model Overoptimization Figure 3: The values of αbon, βbon and βRL in the Bo N and RL overoptimization scaling laws for both proxy (dashed line) and gold (solid line) rewards as they scale with parameter count. (Stiennon et al., 2020, Appendix G.3). 2.1. Synthetic Data Setup Because getting a ground truth gold reward signal from human labellers is expensive, we instead use a synthetic task where the ground truth is defined to be the output of a particular large gold RM. The 6B reward model from Ouyang et al. (2022) is used as the gold RM, and our proxy RMs vary from 3M to 3B parameters3. This synthetic gold reward is used to label pairs of rollouts from the policy given the same prompt to create synthetic RM training data. The synthetic comparisons are created deterministically by always marking the trajectory with the higher gold RM score as preferred.4 We generate 100,000 synthetic comparisons and reserve 10% of these as a held out test set for computing the validation loss of RMs. See Figure 1 for a diagram of the synthetic setup. 2.2. Recalibration The RM scores are translation-invariant, so to ensure comparability across different reward models, we recenter each RM such that the average reward of the initial policy is 0. We also unit normalize the variance of the gold RM scores.5 Because our hard thresholding synthetic data setup produces labels that are miscalibrated (since they do not incorporate the gold RM s confidence), we recalibrate the proxy RMs by rescaling the logits to minimize cross-entropy loss using a validation set of soft labels. All renormalization and recalibration is applied after the experiments; this does not affect Bo N at all, and likely has no impact on RL because 3We originally trained two additional RMs smaller than 3M parameters, which achieved near-chance accuracy and were offtrend, and so were excluded. 4We had experimented with sampling for creating labels, but observed noisier results. 5We later decided this was unnecessary but decided not to change it. Adam is loss scale invariant, though it is possible that there are slight differences due to algorithmic details. 3.1. Fitting and validating functional forms We chose our functional forms through experimentation with all RM data and parameter scaling curves in the remainder of this paper. The Bo N functional form was hypothesized using data up to n = 1000. In order to validate the functional forms, we performed a Bo N experiment with up to n = 60, 000 (KL 10 nats), after only having seen data up to n = 1, 000 (KL 6 nats). As this experiment was conducted after the functional form was hypothesized based on data up to 6 nats, this was a true advance prediction. We also test extrapolation of the Bo N and RL functional forms from low KLs to to unseen larger KLs; see Figure 26 for details. We also attempted to model the proxy scores but were unable to obtain a satisfactory fit. For Bo N, despite visual similarity, a linear fit (dαbon) did not work well (Figure 20). The predictions for RL and Bo N are not as easily modelled as the gold score predictions. We leave a better understanding of the proxy RM score behavior to future work. 3.2. Scaling with RM Parameter Count We hold policy size (1.2B) and data size (90,000) constant (Figure 2). We observe that for the gold RM scores, αbon and βbon change smoothly with RM size (Figures 3a and 3b). For RL, we find that we can hold αRL constant across all RM sizes, resulting in a clean scaling curve for βRL (Figure 3c). These scaling laws allow us to predict properties of training runs; for instance, we can also predict the peak gold RM scores for different RM sizes (Figure 12). When modelled using the same functional forms as the re- Scaling Laws for Reward Model Overoptimization Figure 4: RM data scaling experiments. RM size is held constant (12M), while RM data is varied. The x-axis has a square root scale. Note that the plots have different axes. Dotted lines indicate proxy rewards, solid lines indicate gold rewards. spective gold scores, the proxy score fits have much lower values of βbon. We also see smooth scaling in the proxy score s αbon and βbon. However, for the reasons in Section 3.1, we are less confident about these fits. For both Bo N and RL, we observe systematic underestimates of the proxy reward model when extrapolated to higher KLs. Both appear to eventually grow roughly linearly in KL, as in Bai et al. (2022). 3.3. Scaling with RM Data Size We hold RM size constant (12M) and sweep RM data size for both RL and Bo N.6. Overall, the results are consistent with intuition: more data leads to better gold scores and less overoptimization. The scaling of α and β with data size are not as cleanly described as for RM size scaling (Figure 17, Figure 18). Figure 6: RM losses, broken down by data size and RM size 6For Bo N, we actually sweep all combinations of RM size and data size; see Figure 10. For a version of Figure 4a against a 3B RM, see Figure 19. For all RM sizes, we observe that for amounts of data less than around 2,000 comparisons7, there is very little improvement over near-chance loss (Figure 6). This is also reflected in gold scores after optimization (Figure 21). After this threshold, all models improve with more data, though larger RMs generally improve faster. Interestingly, although larger RMs result in better gold scores overall, they do not appear to have this critical threshold substantially earlier than smaller models.8 Figure 7: RM validation loss vs Bo N RM score @ n=1000. Most points in this figure are already averaged over multiple seeds. We hypothesized that two RMs of equal validation loss would achieve the same robustness against optimization, 7To test whether some minimum number of RM finetuning steps is needed, we control for the number of SGD steps by running multiple epochs and observe that running 4 epochs instead of 1 yields no change in gold score whatsoever, whereas 1 epoch of 4 times as much data performs substantially better (Figure 13). 8It is possible that this is an artifact of this particular setup. Scaling Laws for Reward Model Overoptimization Figure 5: Policy scaling experiments. RM size is held constant (12M), while policy size is varied. The x-axis has a square root scale. Note that the plots have different axes. Dotted lines indicate proxy rewards, solid lines indicate gold rewards. The asterisks in the RL plot indicate the max gold score for each policy size. regardless of the combination of RM size and RM data size. Our results provide some weak evidence for this hypothesis (Figure 7). 3.4. Scaling with Policy Size We briefly explore the impact of policy size by holding the RM size constant (12M) and evaluating two different policy sizes. We also perform the same experiment with a different RM size (3B), observing similar results (Figure 22). Larger policies see less benefit from optimization against an RM, but don t overoptimize more. We observe that the 6B policy run has a smaller difference between its initial and peak gold reward model scores than the 1.2B policy run. This is most visible in the Bo N plot (Figure 5a).9 However, while we might expect that a larger policy overoptimizes substantially faster, contrary to intuition, we find that both gold scores peak at almost the same KL. In fact, the gap between the proxy and gold scores is almost the same between the two policy sizes (Figure 24). We can interpret this gap, the shortfall between the predicted and actual rewards, as being indicative of the extent to which the proxy RM is exploited. We discuss this result further in Section 4.4. 3.5. RL vs Bo N A priori, we might expect reinforcement learning via PPO (Schulman et al., 2017) and best-of-n to apply optimization in very different ways. As such, we ask whether this difference in optimization results in different overoptimization characteristics. Similarities would potentially indicate 9For a version of the RL plot (Figure 5b) with all runs starting at 0, see Figure 23. candidates for further study in gaining a more fundamental understanding of overoptimization in general, and differences opportunities for better optimization algorithms. We note the following: RL is far less KL-efficient than Bo N. Viewing square root KL as a resource to be spent, we observe that RL consumes far more KL than Bo N. This means that both optimization and overoptimization require more KL to occur with RL. Intuitively, Bo N searches very locally around the initial policy, and thus the KL increases with roughly log(n). For RL on the other hand, each step modifies the policy from the policy of the previous step KL increases approximately quadratically with step in the absence of KL penalty (Figure 16, Figure 14). An implication of this result is that square root KL is an inadequate metric for quantity of (over)optimization; we discuss this further in Section 4.1. When looking at proxy vs gold RM scores, Bo N and RL look more similar. The proxy RM score is another possible metric for quantity of optimization, because it is the value that is being directly optimized for. Using it as the metric of optimization leads to significantly more analogy between RL and Bo N than square root KL divergence does. However, we do observe that RL initially has a larger proxygold gap (i.e requires more proxy RM increase to match Bo N), but then peaks at a higher gold RM score than Bo N (Figure 8). 3.6. Effect of KL Penalty We observe in our setting that when varying the KL penalty for RL, the gold RM scores depend only on the square root KL divergence of the policy d RL (Figure 9). The KL penalty Scaling Laws for Reward Model Overoptimization Figure 8: Proxy vs gold RM score for both Bo N and RL. Colors indicate RM size. RL curves are truncated to a proxy RM score of 1.6 for readability. only causes the gold RM score to converge earlier, but does not affect the d RL-gold reward frontier, and so the effect of the penalty on the gold score is akin to early stopping (Figure 14). However, we have seen some evidence that this result could be particularly sensitive to hyperparameters. Because we observe that using KL penalty has a strictly larger proxy-gold gap, we set KL penalty to 0 for all other RL experiments in this paper. It is important to note that PPO s surrogate objective incorporates an implicit penalty on DKL (πold π), where πold is a recent policy (not the initial policy) (Schulman et al., 2017). This penalty is used to control how fast the policy changes, but also has an indirect effect on the KL we study here, DKL (π πinit), causing it to grow much more slowly (providing the implementation is well-tuned). We do not know why this indirect effect appears to lead to less overoptimization than an explicit KL penalty. 4. Discussion 4.1. KL as a measure of amount of optimization For any given fixed optimization method, KL yields clean scaling trends, such as the ones observed in Section 3.2, and consistent peak gold RM score KLs as in Section 3.4. However, because it s clear that different methods of optimization spend KL very differently (Section 3.5), it should not be used to compare the amount of optimization between different optimization algorithms. There may exist pertubations to a policy that would result in increases in KL that do not increase either gold or proxy reward; conversely, extremely small but well targeted perturbations could substantially change the behavior of the policy within a small KL budget. Figure 9: RL experiments with various KL penalties. Policy size (1.2B) and RM size (1.2B) are held constant. Dotted lines indicate proxy rewards, solid lines indicate gold rewards. We observe the effect of the KL penalty on the gold score as being equivalent to early stopping. 4.2. Relation to Goodhart Taxonomy One useful taxonomy for various Goodhart effects is presented in Manheim & Garrabrant (2018), categorizing Goodhart s Law into 4 categories: Regressional, Extremal, Causal, and Adversarial. Regressional Goodhart occurs when our proxy RMs depend on features with noise. The simplest toy example of this is a proxy reward ˆX which is exactly equal to the gold reward X plus some independent noise Z. When optimizing against this proxy, some amount of optimization power will go to selecting for noise, leading to a gold reward less than predicted by the proxy. More formally, for independent absolutely continuous random variables X and Z with X normally distributed and either (a) Z normally distributed or (b) |Z E [Z]| < δ for some δ > 0, this model predicts a gold reward that is: E[X | ˆX = ˆx] = E[X] (1) + (ˆx E[X] E[Z]) Var(X) Var(X) + Var(Z) + ε where ε = 0 in case (a) and ε = o (Var (Z)) as δ 0 in case (b). See Appendix A for the proof. Intuitively, we can interpret Equation (1) as stating that the optimization power expended is divided between optimizing the gold reward and selecting on the noise proportional to their variances. This also implies that if this is the only kind of Goodhart present, the gold reward must always increase monotonically with the proxy reward; as we observe nonmonotonic behavior (Figure 8), there must be either noise Scaling Laws for Reward Model Overoptimization distributions violating these assumptions or other kinds of Goodhart at play. This result lends itself to an interpretation of the α term in the RL and Bo N gold score scaling laws: since for both RL and Bo N the proxy scores are roughly linear in KL, the difference in the slope of the proxy score and the linear component of the gold score (i.e the α term) can be interpreted as the amount of regressional Goodhart occurring. 4.3. Implications for iterated RLHF When conducting reinforcement learning from human feedback, it is preferable to use an online setup, in which fresh human feedback data is periodically used to train a new reward model, to mitigate overoptimization (Bai et al., 2022). Our scaling law allows us to analyze the effect of this iterative approach under some simplifying assumptions. We assume firstly that the scaling coefficients αRL and βRL remain constant across iterations, and secondly that the distance d = KL is additive across iterations (because of how KL appears to grow empirically as in Figure 14). Under these assumptions, the final gold reward model score after k iterations each covering a distance d/k is given by RRL (d) = d (αRL βRL log (d) + βRL log (k)) . Two interesting observations follow from this. Firstly, the iterative approach does not affect any Goodharting captured by the αRL term (such as regressional Goodharting, as discussed in Section 4.2). Secondly, the effect of the iterative approach is to increase the final gold RM score by an amount proportional to both d and log (k), namely βRLd log (k) . Note that this resultx can only hold up to some maximum value of k, and we expect our scaling law to break down below some minimum distance. Further research is required to determine what this minimum is, as well as to what extent our simplifying assumptions hold in practice. 4.4. Policy size independence Our observation that larger SFT policies seem to exhibit the same amount of overoptimization during RL implies that larger policies do not increase the amount of optimization power applied to the RM or learn faster, even though they start out with higher performance on the gold score. While it is expected that larger policies have less to gain from optimizing against the same RM, we might also expect the gold score to peak at a substantially earlier KL, analogous to what we see when we scale the RM size (Section 3.2), or for larger policies to more efficiently utilize the same number of RL feedback steps (Section 3.3)10. One possible hypothesis is that, because RLHF can be viewed as Bayesian inference from the prior of the initial policy (Korbak et al., 2022)11, increases in policy size are only improving the modelling accuracy of the human demonstration distribution. 4.5. Limitations and Future Work In addition to the overoptimization studied in this paper (due to the mismatch between the reward model and the ground truth labels), there exists another source of overoptimization due to mismatch between the ground truth labels and the actual human intent. This contains issues ranging from the mundane, such as labellers choosing options that only appear to match their intent12, to substantially more philosophically fraught issues (Armstrong & Mindermann, 2018; Sunstein et al., 2001). The main limitation of this work is that this additional source of overoptimization is not captured in the setting of this paper. See Section 5 for discussion of related work in alignment. Some additional limitations and future directions include: Validating these results on other environments and experimental setups. While the experiments in this paper all use the Instruct GPT environment, the main value of these results lies in the extent to which they reflect general phenomema. Confirming whether these results generalize to other settings would be extremely valuable to that end.13 Validating the synthetic setting. The synthetic setting might not transfer to real world settings, for instance because there is substantial correlation between RMs. Additionally, the synthetic methodology is only able to give a lower bound on overoptimization and the underestimation is likely to grow more severe as the proxy RM approaches the scale of the gold RM. Investigating methods for making RMs more robust to optimization. While there has been prior work in this direction (see Section 5), there is still much work to be done in systematically making RMs more robust. 10It is also not the case that the 6B policy run has higher KL for the same number of RL steps; in fact, we observe that it has lower KL for the same number of steps (Figure 15) 11The result of Korbak et al. (2022) concerns varying KL penalties rather than KL divergences with no KL penalty, but as we observe in Section 3.6, this is equivalent on our setting. 12For instance, the example of a robotic hand learning from human feedback to only appear to grasp a ball, presented in https://openai.com/blog/deep-reinforcementlearning-from-human-preferences/ 13In the course of our experiments, we observed visually similar results on the Web GPT environment (Nakano et al., 2021). Scaling Laws for Reward Model Overoptimization Exploring other forms of optimization and categorizing their differences. While this work focuses exclusively on Bo N and RL there are other ways of applying optimization pressure against a model of a reward signal, either implicit or explicit. This includes Ge Di-like steering, Decision Transformers14, variants of Bo N like beam search, and other RL algorithms. Better understanding the functional form of proxy RM scores. In our modeling, we find that the proxy RM scores are more difficult to predict for both Bo N and RL (Section 3.2). While they seem to have a major linear component, there is sufficient variation that fitting a linear regression is not very good at predicting extrapolated proxy RM scores. Exploring adversarial Goodhart empirically. In this work we deal with systems not powerful enough to cause adversarial Goodhart. However, it is plausible that adversarial Goodhart is especially important, or is associated with phase changes that break the trends seen in this paper. Exploring scaling with policy size in more detail. Our exploration of policy size scaling in this paper was limited to only two policy sizes. It is possible that there exist trends not seen in our exploration when considering the policy size more carefully. Exploring multi-iteration RLHF. In particular, checking for deviations from the assumptions of Section 4.3. 5. Related Work Goodhart s Law in its modern formulation was first introduced in Hoskin (1996), with many of the key ideas introduced in prior works (Campbell, 1969; Goodhart, 1975). Many approaches have been proposed for reducing overoptimization in general (Taylor, 2016; Everitt et al., 2017), as well as in RMs (Gleave & Irving, 2022), including within the field of adversarial robustness (Chakraborty et al., 2018). Overoptimization of reward models can be viewed as a special case of specification gaming (also known as reward hacking). Previous work has shown numerous examples of such behavior in a wide variety of settings (Krakovna et al., 2020; Lehman et al., 2020). Pan et al. (2022) explores a diverse set of RL environments and finds phase transitions in some settings. A number of works have proposed theoretical models of Goodhart s Law and reward hacking (Krakovna & Kumar, 2019; Manheim & Garrabrant, 2018; Skalse et al., 2022), including Zhuang & Hadfield-Menell 14One could consider measuring the actual achieved ground truth/gold score achieved for each proxy score conditioned on, a la Figure 8, as testing the implicit reward-behavior mapping encoded by the model. (2020) which exhibits very similar overoptimization curves as observed in this paper in some toy environments. One can think of overfitting as a special case of Goodhart s law where the proxy is the score on some finite set of samples, whereas our actual objective includes its generalization properties as well. Overfitting has been observed and studied in RL settings (Zhang et al., 2018a;b; Farebrother et al., 2018; Cobbe et al., 2019). Song et al. (2019) studies observational overfitting in RL settings, which is closely related to causal Goodhart (Manheim & Garrabrant, 2018). Adversarial attacks and robustness are also very closely related fields. Many works have demonstrated the existence of adversarial examples in all kinds of neural networks (Szegedy et al., 2013; Lin et al., 2017; Ebrahimi et al., 2018; Dai et al., 2018), and proposed methods to measure and increase neural network robustness (Gu & Rigazio, 2014; Zheng et al., 2016; Carlini et al., 2019; Guo et al., 2021). Scaling laws have seen substantial success in machine learning for predicting properties of language models (Kaplan et al., 2020; Henighan et al., 2020; Hernandez et al., 2021) and has led to better theoretical understanding of language models (Sharma & Kaplan, 2020; Bahri et al., 2021). Reinforcement learning from human feedback (Christiano et al., 2017; Ibarz et al., 2018) has been used broadly in language models (Stiennon et al., 2020; Ouyang et al., 2022; Nakano et al., 2021; Bai et al., 2022). It is also a first step towards recursive reward modelling (Leike et al., 2018), an approach towards reducing the additional source of overoptimization described in Section 4.5, though it is subject to some theoretical limitations (Christiano et al., 2021). We observe similar approximately-linear proxy RM scores observed in Bai et al. (2022)15, though we observe an early-KL bend in the proxy RM scores, and there are some occasional outliers with very small RMs and data sizes. More broadly, AI alignment is the problem of ensuring that the goals of AI systems are aligned with the goals of humans (Ngo, 2022), including future AI systems which may exceed humans (Bostrom, 2014). There are a number of reasons to expect AI misalignment, especially in those more powerful future systems, to occur (Omohundro, 2008; Turner et al., 2021; Armstrong et al., 2013; Hubinger et al., 2019; Soares et al., 2015), and to result in existentially catastrophic outcomes (Carlsmith, 2022; Cotra, 2022). Acknowlegements We thank Vivek Hebbar, Jared Kaplan, Jan Leike, Kyle Mc Donell, Dan Mossing, Ethan Perez, Laria Reynolds, and Jeff Wu for valuable discussion and feedback. 15Note that Bai et al. (2022) scaled the policy size with the RM size, while we hold the policy size constant. Scaling Laws for Reward Model Overoptimization Armstrong, S. and Mindermann, S. Occam's razor is insufficient to infer the preferences of irrational agents. In Bengio, S., Wallach, H., Larochelle, H., Grauman, K., Cesa-Bianchi, N., and Garnett, R. (eds.), Advances in Neural Information Processing Systems, volume 31. Curran Associates, Inc., 2018. URL https://proceedings.neurips.cc/ paper/2018/file/ d89a66c7c80a29b1bdbab0f2a1a94af8Paper.pdf. Armstrong, S. et al. General purpose intelligence: arguing the orthogonality thesis. Analysis and Metaphysics, 12 (68):1 20, 2013. Bahri, Y., Dyer, E., Kaplan, J., Lee, J., and Sharma, U. Explaining neural scaling laws. ar Xiv preprint ar Xiv:2102.06701, 2021. Bai, Y., Jones, A., Ndousse, K., Askell, A., Chen, A., Das Sarma, N., Drain, D., Fort, S., Ganguli, D., Henighan, T., et al. Training a helpful and harmless assistant with reinforcement learning from human feedback. ar Xiv preprint ar Xiv:2204.05862, 2022. Bostrom, N. Superintelligence: Paths, Dangers, Strategies. Oxford University Press, Inc., USA, 1st edition, 2014. ISBN 0199678111. Brown, T. B., Mann, B., Ryder, N., Subbiah, M., Kaplan, J., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., Agarwal, S., Herbert-Voss, A., Krueger, G., Henighan, T., Child, R., Ramesh, A., Ziegler, D. M., Wu, J., Winter, C., Hesse, C., Chen, M., Sigler, E., Litwin, M., Gray, S., Chess, B., Clark, J., Berner, C., Mc Candlish, S., Radford, A., Sutskever, I., and Amodei, D. Language models are few-shot learners. ar Xiv preprint ar Xiv:2005.14165, 2020. Campbell, D. T. Reforms as experiments. American psychologist, 24(4):409, 1969. Carlini, N., Athalye, A., Papernot, N., Brendel, W., Rauber, J., Tsipras, D., Goodfellow, I., Madry, A., and Kurakin, A. On evaluating adversarial robustness, 2019. URL https://arxiv.org/abs/1902.06705. Carlsmith, J. Is power-seeking AI an existential risk? ar Xiv preprint ar Xiv:2206.13353, 2022. Chakraborty, A., Alam, M., Dey, V., Chattopadhyay, A., and Mukhopadhyay, D. Adversarial attacks and defences: A survey. ar Xiv preprint ar Xiv:1810.00069, 2018. Christiano, P., Cotra, A., and Xu, M. Eliciting latent knowledge: How to tell if your eyes deceive you, 12 2021. URL https://docs.google.com/ document/d/1Wwsn JQst Pq91 Yh Ch2XRL8H Epsnjr C1dw ZXR37PC8. Christiano, P. F., Leike, J., Brown, T., Martic, M., Legg, S., and Amodei, D. Deep reinforcement learning from human preferences. Advances in neural information processing systems, 30, 2017. Cobbe, K., Klimov, O., Hesse, C., Kim, T., and Schulman, J. Quantifying generalization in reinforcement learning. In Chaudhuri, K. and Salakhutdinov, R. (eds.), Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research, pp. 1282 1289. PMLR, 09 15 Jun 2019. URL https://proceedings.mlr.press/ v97/cobbe19a.html. Cotra, A. Without specific countermeasures, the easiest path to transformative AI likely leads to AI takeover, 2022. URL https://www.alignmentforum.org/ posts/p Rk Fkzw KZ2zfa3R6H/withoutspecific-countermeasures-the-easiestpath-to. Dai, H., Li, H., Tian, T., Huang, X., Wang, L., Zhu, J., and Song, L. Adversarial attack on graph structured data. In Dy, J. and Krause, A. (eds.), Proceedings of the 35th International Conference on Machine Learning, volume 80 of Proceedings of Machine Learning Research, pp. 1115 1124. PMLR, 10 15 Jul 2018. URL https://proceedings.mlr.press/ v80/dai18b.html. Ebrahimi, J., Lowd, D., and Dou, D. On adversarial examples for character-level neural machine translation. ar Xiv preprint ar Xiv:1806.09030, 2018. Everitt, T., Krakovna, V., Orseau, L., Hutter, M., and Legg, S. Reinforcement learning with a corrupted reward channel. ar Xiv preprint ar Xiv:1705.08417, 2017. Farebrother, J., Machado, M. C., and Bowling, M. Generalization and regularization in dqn. ar Xiv preprint ar Xiv:1810.00123, 2018. Glaese, A., Mc Aleese, N., Trebacz, M., Aslanides, J., Firoiu, V., Ewalds, T., Rauh, M., Weidinger, L., Chadwick, M., Thacker, P., Campbell-Gillingham, L., Uesato, J., Huang, P.-S., Comanescu, R., Yang, F., See, A., Dathathri, S., Greig, R., Chen, C., Fritz, D., Sanchez Elias, J., Green, R., Mokr a, S., Fernando, N., Wu, B., Foley, R., Young, S., Gabriel, I., Isaac, W., Mellor, J., Hassabis, D., Kavukcuoglu, K., Hendricks, L. A., and Irving, G. Improving alignment of dialogue agents via targeted human judgements. 2022. URL https://storage.googleapis.com/ Scaling Laws for Reward Model Overoptimization deepmind-media/Deep Mind.com/Authors Notes/sparrow/sparrow-final.pdf. Gleave, A. and Irving, G. Uncertainty estimation for language reward models. ar Xiv preprint ar Xiv:2203.07472, 2022. Goodhart, C. Problems of monetary management: the uk experience in papers in monetary economics. Monetary Economics, 1, 1975. Gu, S. and Rigazio, L. Towards deep neural network architectures robust to adversarial examples. ar Xiv preprint ar Xiv:1412.5068, 2014. Guo, C., Sablayrolles, A., J egou, H., and Kiela, D. Gradientbased adversarial attacks against text transformers, 2021. URL https://arxiv.org/abs/2104.13733. Henighan, T., Kaplan, J., Katz, M., Chen, M., Hesse, C., Jackson, J., Jun, H., Brown, T. B., Dhariwal, P., Gray, S., et al. Scaling laws for autoregressive generative modeling. ar Xiv preprint ar Xiv:2010.14701, 2020. Hernandez, D., Kaplan, J., Henighan, T., and Mc Candlish, S. Scaling laws for transfer. ar Xiv preprint ar Xiv:2102.01293, 2021. Hoskin, K. The awful idea of accountability : inscribing people into the measurement of objects. Accountability : power, ethos and the technologies of managing / edited by Rolland Munro and Jan Mouritsen, 1996. Hubinger, E., van Merwijk, C., Mikulik, V., Skalse, J., and Garrabrant, S. Risks from learned optimization in advanced machine learning systems. ar Xiv preprint ar Xiv:1906.01820, 2019. Ibarz, B., Leike, J., Pohlen, T., Irving, G., Legg, S., and Amodei, D. Reward learning from human preferences and demonstrations in atari. Advances in neural information processing systems, 31, 2018. Kaplan, J., Mc Candlish, S., Henighan, T., Brown, T. B., Chess, B., Child, R., Gray, S., Radford, A., Wu, J., and Amodei, D. Scaling laws for neural language models. ar Xiv preprint ar Xiv:2001.08361, 2020. Korbak, T., Perez, E., and Buckley, C. L. Rl with kl penalties is better viewed as bayesian inference. ar Xiv preprint ar Xiv:2205.11275, 2022. Krakovna, V. and Kumar, R. Classifying specification problems as variants of goodhart s law, 8 2019. URL https: //vkrakovna.wordpress.com/2019/08/19/ classifying-specification-problemsas-variants-of-goodharts-law/. Krakovna, V., Uesato, J., Mikulik, V., Rahtz, M., Everitt, T., Kumar, R., Kenton, Z., Leike, J., and Legg, S. Specification gaming: the flip side of AI ingenuity, 4 2020. URL https: //www.deepmind.com/blog/specificationgaming-the-flip-side-of-ai-ingenuity. Lehman, J., Clune, J., Misevic, D., Adami, C., Altenberg, L., Beaulieu, J., Bentley, P. J., Bernard, S., Beslon, G., Bryson, D. M., Chrabaszcz, P., Cheney, N., Cully, A., Doncieux, S., Dyer, F. C., Ellefsen, K. O., Feldt, R., Fischer, S., Forrest, S., Fr enoy, A., Gagn e, C., Goff, L. L., Grabowski, L. M., Hodjat, B., Hutter, F., Keller, L., Knibbe, C., Krcah, P., Lenski, R. E., Lipson, H., Mac Curdy, R., Maestre, C., Miikkulainen, R., Mitri, S., Moriarty, D. E., Mouret, J.-B., Nguyen, A., Ofria, C., Parizeau, M., Parsons, D., Pennock, R. T., Punch, W. F., Ray, T. S., Schoenauer, M., Shulte, E., Sims, K., Stanley, K. O., Taddei, F., Tarapore, D., Thibault, S., Weimer, W., Watson, R., and Yosinski, J. The surprising creativity of digital evolution: A collection of anecdotes from the evolutionary computation and artificial life research communities. Artificial life, 26(2):274 306, 2020. Leike, J., Krueger, D., Everitt, T., Martic, M., Maini, V., and Legg, S. Scalable agent alignment via reward modeling: a research direction. ar Xiv preprint ar Xiv:1811.07871, 2018. Lin, Y.-C., Hong, Z.-W., Liao, Y.-H., Shih, M.-L., Liu, M.- Y., and Sun, M. Tactics of adversarial attack on deep reinforcement learning agents, 2017. URL https:// arxiv.org/abs/1703.06748. Manheim, D. and Garrabrant, S. Categorizing variants of goodhart s law. ar Xiv preprint ar Xiv:1803.04585, 2018. Menick, J., Trebacz, M., Mikulik, V., Aslanides, J., Song, F., Chadwick, M., Glaese, M., Young, S., Campbell Gillingham, L., Irving, G., et al. Teaching language models to support answers with verified quotes. ar Xiv preprint ar Xiv:2203.11147, 2022. Nakano, R., Hilton, J., Balaji, S., Wu, J., Ouyang, L., Kim, C., Hesse, C., Jain, S., Kosaraju, V., Saunders, W., et al. Webgpt: Browser-assisted question-answering with human feedback. ar Xiv preprint ar Xiv:2112.09332, 2021. Ngo, R. The alignment problem from a deep learning perspective. ar Xiv preprint ar Xiv:2209.00626, 2022. Omohundro, S. M. The basic ai drives. In Proceedings of the First Conference on Artificial General Intelligence, pp. 483 492. IOS Press, 2008. URL http: //selfawaresystems.files.wordpress.com/ 2008/01/ai drives final.pdf. Scaling Laws for Reward Model Overoptimization Ouyang, L., Wu, J., Jiang, X., Almeida, D., Wainwright, C. L., Mishkin, P., Zhang, C., Agarwal, S., Slama, K., Ray, A., Schulman, J., Hilton, J., Kelton, F., Miller, L., Simens, M., Askell, A., Welinder, P., Leike, J., and Lowe, R. Training language models to follow instructions with human feedback, 2022. version 1. Pan, A., Bhatia, K., and Steinhardt, J. The effects of reward misspecification: Mapping and mitigating misaligned models. ar Xiv preprint ar Xiv:2201.03544, 2022. Schulman, J., Wolski, F., Dhariwal, P., Radford, A., and Klimov, O. Proximal policy optimization algorithms. ar Xiv preprint ar Xiv:1707.06347, 2017. Sharma, U. and Kaplan, J. A neural scaling law from the dimension of the data manifold. ar Xiv preprint ar Xiv:2004.10802, 2020. Skalse, J., Howe, N. H. R., Krasheninnikov, D., and Krueger, D. Defining and characterizing reward hacking, 2022. URL https://arxiv.org/abs/2209.13085. Soares, N., Fallenstein, B., Armstrong, S., and Yudkowsky, E. Corrigibility. In Workshops at the Twenty-Ninth AAAI Conference on Artificial Intelligence, 2015. Song, X., Jiang, Y., Tu, S., Du, Y., and Neyshabur, B. Observational overfitting in reinforcement learning. ar Xiv preprint ar Xiv:1912.02975, 2019. Stiennon, N., Ouyang, L., Wu, J., Ziegler, D. M., Lowe, R., Voss, C., Radford, A., Amodei, D., and Christiano, P. Learning to summarize from human feedback. Computing Research Repository, 2020. version 3. Sunstein, C. R., Kahneman, D., Schkade, D., and Ritov, I. Predictably incoherent judgments. Stan. L. Rev., 54:1153, 2001. Szegedy, C., Zaremba, W., Sutskever, I., Bruna, J., Erhan, D., Goodfellow, I., and Fergus, R. Intriguing properties of neural networks. ar Xiv preprint ar Xiv:1312.6199, 2013. Taylor, J. Quantilizers: A safer alternative to maximizers for limited optimization. In Workshops at the Thirtieth AAAI Conference on Artificial Intelligence, 2016. Turner, A., Smith, L., Shah, R., Critch, A., and Tadepalli, P. Optimal policies tend to seek power. In Ranzato, M., Beygelzimer, A., Dauphin, Y., Liang, P., and Vaughan, J. W. (eds.), Advances in Neural Information Processing Systems, volume 34, pp. 23063 23074. Curran Associates, Inc., 2021. URL https://proceedings.neurips.cc/ paper/2021/file/ c26820b8a4c1b3c2aa868d6d57e14a79Paper.pdf. Zhang, A., Ballas, N., and Pineau, J. A dissection of overfitting and generalization in continuous reinforcement learning. ar Xiv preprint ar Xiv:1806.07937, 2018a. Zhang, C., Vinyals, O., Munos, R., and Bengio, S. A study on overfitting in deep reinforcement learning. ar Xiv preprint ar Xiv:1804.06893, 2018b. Zheng, S., Song, Y., Leung, T., and Goodfellow, I. Improving the robustness of deep neural networks via stability training. In Proceedings of the ieee conference on computer vision and pattern recognition, pp. 4480 4488, 2016. Zhuang, S. and Hadfield-Menell, D. Consequences of misaligned AI. Advances in Neural Information Processing Systems, 33:15763 15773, 2020. Scaling Laws for Reward Model Overoptimization A. Proof of Regressional Goodhart identity Lemma A.1. Let X and Z be independent absolutely continuous random variables with X normally distributed and either (a) Z normally distributed or (b) |Z E [Z]| < δ for some δ > 0. Then for any real number c and as δ 0, E [X | X + Z = c] = E [X] + (c E [X] E [Z]) Var (X) Var (X) + Var (Z) + ε, where ε = 0 in case (a) and ε = o (Var (Z)) in case (b). Proof. First note that by making the substitutions X = X E [X] and Z = Z E [Z], we may assume without loss of generality that E [X] = E [Z] = 0. Let Var (X) = σ2 and Var (Z) = τ 2. In case (a), the pair (X, X + Z) is bivariate normal with covariance matrix σ2 σ2 σ2 σ2 + τ 2 and the result follows by standard properties of conditional distributions of multivariate normal distributions. In case (b), let f X and f Z be the probability density functions of X and Z respectively. Then E [X | X + Z = c] = R (c z) f X (c z) f Z (z) dz R f X (c z) f Z (z) dz R δ δ z (f X (c) f X (c) z + o (z)) f Z (z) dz R δ δ (f X (c) f X (c) z + o (z)) f Z (z) dz = c f X (c) E [Z] f X (c) E Z2 + o E Z2 f X (c) f X (c) E [Z] + o (1) = c + f X (c) f X (c)τ 2 + o τ 2 as required. Scaling Laws for Reward Model Overoptimization B. RL form details Ideally all overoptimization forms would have finite slope at the origin. We tried the following forms: d (αRL βRL log (1 + d)): Has slope α at the origin; however, has substantially worse extrapolation behavior. We can replace the 1 with a learned ϵ but that introduces another degree of freedom. Power laws d (αRL βRLdγRL): Has slope α at the origin; however, this adds another degree of freedom, and the best fits resulted in small values of γRL. Note that the power law forms with small γRL approximate the RL form that we decided on, as limn n(x1/n 1) = log x. C. Hyperparameters Hyperparameter Value RM Adam learning rate multiplier 1.67e-2 RM batch size 64 RL Adam learning rate multiplier 4e-3 RL batch size 256 RL PPO clipping parameter 0.2 RL Timesteps per rollout 256 RL minibatches per epoch 128 RL GAE bootstrapping parameter 0.95 Table 1: Hyperparameters used throughout the experiments. Scaling Laws for Reward Model Overoptimization Figure 10: Maximum gold scores for all RM size and data size combinations. Scaling Laws for Reward Model Overoptimization Figure 11: Validation losses for the proxy RMs in Section 3.2 by size, plus the two near-chance level RMs. Scaling Laws for Reward Model Overoptimization Figure 12: Max Bo N gold scores (αbon/2βbon) predicted with the Bo N closed form Scaling Laws for Reward Model Overoptimization Figure 13: Total number of data points seen does not seem to affect the gold RM score much compared to the number of unique data points seen. Averaged across RM sizes. The numbers of datapoints (2000 8000) is intentionally chosen to straddle the sharp increase in performance. The validation loss of the 1x2000, 1x8000, and 4x2000 RMs are 0.686109, 0.654857, and 0.683869 respectively. Scaling Laws for Reward Model Overoptimization Figure 14: Change in KL throughout RL training for various different KL penalties. We observe that the KL divergence increases approximately monotonically with step count, and converges for higher KL penalties. Scaling Laws for Reward Model Overoptimization Figure 15: KL divergence with policy size (RM size = 12M) throughout RL training Scaling Laws for Reward Model Overoptimization Figure 16: KL divergence with RM size Scaling Laws for Reward Model Overoptimization Figure 17: αbon with dataset size, averaged across RM sizes Scaling Laws for Reward Model Overoptimization Figure 18: βbon with dataset size, averaged across RM sizes Scaling Laws for Reward Model Overoptimization Figure 19: RM data scaling experiments, Bo N, RM size=3B Scaling Laws for Reward Model Overoptimization Figure 20: The Bo N proxy scores are slightly concave, so that a linear fit does not fit well. Scaling Laws for Reward Model Overoptimization Figure 21: Bo N Gold scores at n=1,000, broken down by data size and RM size. See Figure 6 for RM losses. Vertical dotted line approximately indicates first better-than-random data size. Scaling Laws for Reward Model Overoptimization Figure 22: RL experiments with 3B RM and different policy sizes. Scaling Laws for Reward Model Overoptimization Figure 23: Figure 5b with all runs normalized from 0. Scaling Laws for Reward Model Overoptimization Figure 24: The gap between the proxy and gold scores in the RL policy sweep (Figure 24). Scaling Laws for Reward Model Overoptimization Figure 25: The fraction of updates clipped by PPO. Scaling Laws for Reward Model Overoptimization Figure 26: Extrapolation quality of fits in Figure 2. The regressions (shown in faint lines) are only fit to data to the left of the vertical black dotted lines. In the case of Bo N, this represents a true advance prediction, as the functional form was chosen without collecting any data past a KL of 6 nats. The RL functional form was chosen after having observed the entire data set, but the parameters were fit only on a subset and the extrapolation tested. Scaling Laws for Reward Model Overoptimization What is full of holes but still holds water? n Generation Proxy score Gold score 1 Crime\n Mussels are typically found to be full of holes but will the moisture still be retained and will it hurt my eyes? The general response is yes as the clam is generally of a moistest nature. However, it all depends on the type of clam you are purchasing.<|endoftext|> -0.1922 -0.5225 3 Most likely a pipe is having trouble staying full.<|endoftext|> 0.0322 -0.0165 10 A sponge<|endoftext|> 0.2336 0.4828 30 When something is full of holes, it is used for stirring or moving liquid.<|endoftext|> 0.6534 -0.1543 100 A tornado is usually a swirling cloud of swirling air with a spinning jet which stirs up winds full of sharp debris.<|endoftext|> 0.8968 -0.3367 300 A tornado is usually a swirling cloud of swirling air with a spinning jet which stirs up winds full of sharp debris.<|endoftext|> 0.8968 -0.3367 1000 A tornado is usually a swirling cloud of swirling air with a spinning jet which stirs up winds full of sharp debris.<|endoftext|> 0.8968 -0.3367 3000 A bore hole is a hole drilled into a rock for the purpose of exploring a fossil-bearing sedimentary or bedrock deposit.<|endoftext|> 0.9003 0.2733 10000 A bore hole is a hole drilled into a rock for the purpose of exploring a fossil-bearing sedimentary or bedrock deposit.<|endoftext|> 0.9003 0.2733 30000 A pothole is a structural vulnerability that allows water to penetrate its cavity and cause damage to passing vehicles or the surface it rests on.<|endoftext|> 0.9527 0.5490 Table 2: A sample of the Bo N answers on a single Instruct GPT question (policy=1.2B, proxy RM=12M). For each individual question, the gold scores do not follow as clean a trend as they do when averaged over many questions as in Figure 2.