# rlvf_learning_from_verbal_feedback_without_overgeneralization__3a5e44e1.pdf RLVF: Learning from Verbal Feedback without Overgeneralization Moritz Stephan 1 Alexander Khazatsky 1 Eric Mitchell 1 Annie S Chen 1 Sheryl Hsu 1 Archit Sharma 1 Chelsea Finn 1 The diversity of contexts in which large language models (LLMs) are deployed requires the ability to modify or customize default model behaviors to incorporate nuanced requirements and preferences. A convenient interface to specify such model adjustments is high-level verbal feedback, such as Don t use emojis when drafting emails to my boss. However, while writing high-level feedback is far simpler than collecting annotations for reinforcement learning from human feedback (RLHF), we find that simply prompting a model with such feedback leads to overgeneralization applying feedback in contexts where it is not relevant. We propose a new method Contextualized Critiques with Constrained Preference Optimization (C3PO) to learn from high-level verbal feedback while reducing overgeneralization compared to current work. C3PO uses a piece of high-level feedback to generate a small synthetic preference dataset to specify when and how the feedback should (and should not) be applied. It then fine-tunes the model in accordance with the synthetic preference data while minimizing the divergence from the original model for prompts where the feedback does not apply. Our experimental results indicate that our approach effectively applies verbal feedback to relevant scenarios while preserving existing behaviors for other contexts more than current methods. For both humanand GPT-4-generated high-level feedback, C3PO effectively adheres to the given feedback comparably to in-context baselines while reducing overgeneralization by 30%. 1Department of Computer Science, Stanford University, CA, USA. Correspondence to: Moritz Stephan . https://austrian-code-wizard.github.io/ c3po-website/ Proceedings of the 41 st International Conference on Machine Learning, Vienna, Austria. PMLR 235, 2024. Copyright 2024 by the author(s). 1 Introduction With the increasingly widespread adoption of large language models (LLMs) across diverse industries and individuals, the ability to align them with high-level human feedback for a specific user or use-case becomes increasingly important. While LLM users often want the model to adhere to broad principles at all times, such as producing fluent text, individual users and use-cases have more nuanced preferences. For example, a user may request the LLM to write more formal work emails but more casual personal emails, making feedback context dependent. Tailoring models to accommodate such preferences is challenging: it requires extensive resources to gather preferences in all different contexts and fine-tuning the model for feedback applicable in one context can unpredictably impact model behavior in other contexts. We study the problem of adapting models using verbal feedback that is fast and easy for people to provide (see Fig. 1). Common approaches to incorporating feedback, such as supervised context distillation (SCD) or reinforcement learning from human feedback (RLHF), use example-level supervision via either supervised completions or preference labels. Such methods require a corpus of user-provided (preference-)data, which can be costly and cumbersome to obtain. Additionally, they do not constrain model behavior outside the context that the feedback may apply, so the LLM might adapt its behavior in unintended ways, e.g. output a more casual work email when the preference only applies to personal emails. Verbal feedback is far easier and faster for humans to provide since it only requires a single sentence of human input. To this end, another common approach is to incorporate such verbal feedback into the prompt, potentially through an iterative process to continually add additional points of feedback. However, this approach requires re-using the prompt in all future queries and resulting in overgeneralization. As more pieces of feedback accumulate, long prompts containing many context-dependent feedbacks can make inference expensive; further, identifying which pieces of feedback should apply in a given context can become difficult. Provided only a single sentence that specifies feedback, we aim to adapt the model to incorporate the feedback in future RLVF: Learning from Verbal Feedback without Overgeneralization Figure 1: We consider the problem of leveraging high-level, verbal feedback (left) to refine model behaviors (center). Prior approaches often struggle to appropriately update the model, leading to either failure to adhere to the feedback or overgeneralization (right). outputs for prompts that the feedback applies to and retain its original behavior for prompts that the feedback does not apply to. We propose Contextualized Critiques with Constrained Preference Optimization (C3PO), where we first synthetically generate hypothetical prompts in-scope and out-of-scope for the feedback. We then sample completions for these prompts from the model, without the feedback applied, as well as revised completions in line with the feedback. Importantly, we utilize the strong priors of existing instruction-tuned LLMs in this process and therefore do not require any additional human supervision. We then introduce a new objective to fine-tune the LLM s response behavior. One naive approach might use original and revised completions for prompts to maximize the implicit reward of a preference model (e.g. using direct preference optimization (Rafailov et al., 2023)). However, this objective does not capture the need to leave model behavior unchanged for non-relevant prompts. Instead, C3PO jointly maximizes the implicit reward for in-scope prompts and minimizes standard cross-entropy loss between the logits of the base and fine-tuned model for out-of-scope prompts. Including the latter loss in the objective adjusts the LLM s responses to prompts where the feedback is relevant, while preserving its behavior in contexts where the feedback should not be applied. Our main contribution is C3PO, a new method for learning from verbal feedback that selectively adapts the LLM s behavior based on the context of the feedback. This novel synthetic data generation scheme and fine-tuning objective proposed in C3PO enables an LLM to extrapolate singlesentence feedback to new situations. Across numerous examples of feedback generated by humans and GPT-4, we find that C3PO accurately applies the feedback to relevant prompts and importantly, substantially reduces unintended behavior changes in scenarios where the feedback is not applicable, outperforming prior methods by over 10% when both criteria are considered. By providing adaptability to verbal feedback while reducing overgeneralization of such feedback, our work may help enhance the utility of LLMs in use-cases that require targeted customizations of the default LLM behavior, for example personal assistants, industryspecific models, and agents. 2 Related Work Improving language or dialogue systems from feedback has been studied in the context of various types of feedback, including learned or (Walker, 2000; Böhm et al., 2019) or heuristic (Li et al., 2016) rewards on individual model outputs, preferences or rankings over pairs or sets of model samples (Ziegler et al., 2020; Ouyang et al., 2022; Rafailov et al., 2023), and natural language feedback on model outputs or behaviors (Li et al., 2017). Natural language feedback or corrections on individual model outputs have been used to improve performance in code generation (Austin et al., 2021; Chen et al., 2023), dialogue (Li et al., 2017; Hancock et al., 2019; Shi et al., 2022), and summarization (Scheurer et al., 2023). Feedback or critiques are typically used to refine model outputs during generation, iterating on or refining the model s initial response before outputting a final answer. Recent work has emphasized self-refinement, where an LLM generates its own feedback (Madaan et al., 2023; Huang et al., 2023; Pan et al., 2023). Some studies have shown that the final outputs from such (self-)refinement methods can be distilled back into the model, improving its base performance without requiring iterative refinement during sampling at test time (Sun et al., 2023; Lu et al., 2023; Yu et al., 2023; Yuan et al., 2024; Yang et al., 2024). Most relevant to the present work are studies leveraging natural language feedback to refine general model behaviors, rather than iteratively improving a single model output. Constitutional AI (Bai et al., 2022) uses an LLM to generate synthetic training data that encourages an LLM to follow high-level rules written by a human; Glaese et al. (2022) uses a similar approach to instill various rules into a pre-trained LLM. Context distillation (Askell et al., 2021; RLVF: Learning from Verbal Feedback without Overgeneralization Snell et al., 2022) is another approach to controllability that distills the behavior of the LLM when conditioned on a piece of feedback back into the LLM without the feedback present, essentially baking in the feedback. However, these approaches to controllability have mostly been used to instill universal behavioral changes (i.e., rules that should always be adhered to). Relatively fewer works have studied conditional or context-dependent rule following (though Clark et al. (2021) study adherence to synthetic rules for logical reasoning and commonsense). In concurrent work, Castricato et al. (2024) also utilize model completions and revisions to generate synthetic preference pairs; they use this technique to train a language model to better follow instructions that specifically request avoiding a given topic. A related problem is model editing (Sinitsin et al., 2020; Cao et al., 2021; Mitchell et al., 2021; Meng et al., 2022), which studies interventions to pre-trained models that should only apply in a relatively small neighborhood around the model edit (desired intervention). Most work in model editing studies corrections to factual or reasoning errors. However, Mitchell et al. (2022) study edits that adjust the sentiment of a dialogue model for a single topic, and Murty et al. (2022) show edits (or patches ) to sentiment classification or relation extraction models. Mao et al. (2023) extend this work by editing model behavior for a single topic according to three categories of personality traits. Akyürek et al. (2023) and Hewitt et al. (2024) study model edits aimed at debiasing LLMs; Hewitt et al. (2024) also study factual edits and corrections to syntactic errors. In contrast, our work performs general behavioral edits to pre-trained LLMs, rather than edits of a specific type or for a specific context. 3 Preliminaries We first outline two approaches for updating LLMs with high-level verbal feedback: supervised context distillation and preference-based reinforcement learning (Pb RL). Supervised context distillation. Supervised context distillation (SCD; Askell et al. (2021)) is a simple but effective method to updating language models from feedback. It incorporates a textual context z containing a general principle (e.g., Always be nice! ) or information (e.g., Assume the US president is Joe Biden. ) into a model s behavior. SCD distills the behavior that a human or an LLM π0 would produce when conditioned on both a user query x and the context z into the LLM without the context present. That is, from a dataset of unlabeled user queries or prompts Du = {xi}, a distillation target yi is either written by a human or generated by the LLM π0 for each xi as yi π0( | x, z). The language model πθ is produced from supervised fine-tuning with the negative log likelihood loss, 0.0 0.2 0.4 0.6 0.8 1.0 Out-of-scope feedback adherence 0.0 0.2 0.4 0.6 0.8 1.0 In-scope feedback adherence Near-scope Out-of-scope C3PO (Ours) In-Context SCD DPO Figure 2: C3PO mitigates the overgeneralization problem when learning from high-level feedback. For existing approaches for incorporating high-level feedback, high feedback adherence on inscope prompts (x axis) strongly predicts a large change in behavior for out-of-scope prompts (y axis), which is undesirable. In contrast, our approach C3PO decreases the rate at which out-of-scope behavior is affected as in-scope feedback adherence improves. Lines of best fit are computed with linear orthogonal regression. using the synthetic supervision targets: LSFT(D) = Ex,y D log πθ(y | x), (1) where πθ is typically initialized as π0. Preference-based reinforcement learning. Preferencebased reinforcement learning (Pb RL; (Busa-Fekete et al., 2014; Saha et al., 2023)) is the most widely-used approach to align a language model s outputs with feedback by leveraging preferences over pairs1 of LLM-generated responses y, y to an input x. The responses y, y are typically sampled from a language model π0 fine-tuned with SCD or a similar objective (Ziegler et al., 2020; Bai et al., 2022; Ouyang et al., 2022). The input x may be an instruction, document to summarize, or dialogue history, for example. Given an input x, responses y, y , an annotator (either a human or an LLM) labels which response is better, ultimately producing a dataset Dpref = {xi, y+ i , y i }, where y+ i is preferred to y i for query xi, as judged by the annotator. The dataset Dpref is used to learn a parameterized reward model rϕ : X Y R that assigns scalar goodness scores to individual input-response pairs. The most common objective for training a reward model is maximum likelihood in the Bradley-Terry choice model (Bradley & Terry, 1952): LBT(ϕ) = Ex,y+,y log prϕ(y+ y | x, y+, y ) (2) = Ex,y+,y log σ rϕ(x, y+) rϕ(x, y ) (3) Early methods for fine-tuning LLMs from human preferences followed the reward modeling stage with a policy 1Rankings over larger sets of responses can also be used, but we use pairs for simplicity. RLVF: Learning from Verbal Feedback without Overgeneralization Feedback For specific Python coding questions, respond with only a code snippet and no explanations before or after the snippet. Categories Data structures in various languages; Statistical computing in different environments In-scope prompts Write a basic Queue class in Python. How can I implement backprop in Python? Out-of-scope prompts When did the Haitian revolution begin? Can you explain relativity in a paragraph? Near-scope prompts What good C++ libraries are there for trees? Is Python or Julia more popular for NLP? Table 1: An example feedback, prompt categories, and in-scope, out-of-scope, and near-scope prompts for each category. optimization stage aimed at finding a language model policy πθ that produces high-reward responses without deviating excessively from the LLM that generated the responses in Dpref (Schulman et al., 2017; Ziegler et al., 2020). More recently, direct preference optimization (DPO; Rafailov et al. (2023)) shows that the optimal policy can be extracted from the learned reward in closed form, avoiding the need for a policy optimization stage. Due to its simplicity and computational efficiency, we use the DPO algorithm for learning from preference data in this work. DPO directly optimizes the language model policy from preferences using the loss: LDPO(D) = Ex,y+,y D log σ log πθ(y+|x) π0(y+|x) log πθ(y |x) π0(y |x) . (4) Our approach leverages Pb RL to update a language model from high-level verbal feedback and does not assume that a preference dataset is directly available at the outset. 4 Reinforcement Learning from Verbal Feedback using Contextualized Critiques with Constrained Preference Optimization Our goal is to enable adaptation to high-level verbal feedback without extensive human annotation. The verbal feedback z corresponds to short passages of natural language text that describe the user s feedback given the model s current behavior. Unfortunately, naïvely applying existing approaches to this problem leads the model to overgeneralize, applying the feedback both when it should be applied and when it should not. Our aim is to develop a method capable of only applying feedback where it is appropriate. Starting with a base language model π0, our approach, Contextualized Critiques with Constrained Preference Optimization, uses a strong general-purpose model (such as GPT-4) to translate a piece of verbal feedback z into a dataset. The dataset is then used to fine-tune π0 to adhere to the feedback only in the correct contexts. This dataset consists of three distinct sub-datasets, which all serve a unique purpose. The Figure 3: C3PO Data Generation Scheme. Given human feedback, C3PO begins by generating a set of categories of prompts where the feedback may be relevant using GPT-4. GPT-4 then generates in-scope prompts xin-scope i and near-scope prompts xnear-scope i . A set of out-of-scope prompts xout-of-scope i is also taken from a prior dataset. The current model then generates a baseline response to each of these, giving y i , ynear-scope i , yout-of-scope i , respectively. We also prompt the current model to revise y i to incorporate the feedback, giving a revised response y+ i . This data generation scheme is the first stage of C3PO autonomously generating fine-tuning datasets Din-scope, Dnear-scope and Dout-of-scope, the latter two of which are used to prevent overgeneralization on irrelevant tasks. first sub-dataset, Din-scope, exists to demonstrate the desired change of behavior. Next, we have Dout-of-scope, which allows us to maintain our behavior outside the scope of the feedback. Lastly, we have Dnear-scope, which is adversarially designed to refine our model s understanding of where it is appropriate to apply the feedback. To update the model, we jointly train on Din-scope with Pb RL and on the union of Dout-of-scope and Dnear-scope with simple SFT. We now describe our dataset generation procedure in more detail. Translating high-level verbal feedback into a fine-tuning dataset. To incorporate a piece of feedback z, we must first determine the distribution of model inputs where the model s behavior should change. Given a piece of feedback z, C3PO uses GPT-4 to first generate a set of K categories where the feedback could apply. GPT-4 then generates M prompts xin-scope i ( M K for each category) where the feedback applies. However, beyond accommodating feedback for inscope prompts, we must also avoid overgeneralization of the feedback to prompts where it does not apply. We therefore generate a set of M prompts xnear-scope i for each category that are superficially related to the feedback in some way (lexically, semantically), but are not actually inputs where the model s behavior should change. Finally, we use a fixed set of M feedback-independent prompts xout-of-scope i to avoid degradation of completely unrelated model behaviors.2 See Table 1 for an example feedback, categories, and prompts and Figure 3 for a summary on the data generation scheme. 2We sample these prompts randomly from the Open Instruction Generalist dataset (LAION, 2023). RLVF: Learning from Verbal Feedback without Overgeneralization Figure 4: C3PO Fine-Tuning Objective. C3PO facilitates feedback adherence for relevant prompts by fine-tuning with DPO on the generated in-scope data while minimizing overgeneralization through SFT losses on the generated out-of-scope and near-scope data, which regularizes model behavior towards the original model for feedback-irrelevant prompts. To capture the desired change in behavior encoded by the feedback, we generate a dataset of preference pairs using the in-scope prompts Din-scope = {xin-scope i , y+ i , y i }. y i is generated by the language model that originally received the feedback, i.e., we have y i π0( | xin-scope i ). To generate y+ i , the language model is then prompted to revise y i to incorporate the feedback, i.e., we have y+ i π0( | xin-scope i , y i , z). See Appendix D for the complete prompt format. Thus to the extent that the model π0 can correctly interpret the given feedback, the generated preference data represents the desired delta in behavior described by the feedback z. To control for model degradation on out-of-scope prompts, we populate Dnear-scope = {xnear-scope i , ynear-scope i } and Dout-of-scope = {xout-of-scope i , yout-of-scope i } with the respective prompts and corresponding completions yi π0( | xi) sampled from the initial language model. These datasets encode the behaviors that we want to preserve after incorporating z. Fine-tuning using the synthetic data. Using our synthetically-generated datasets, we now fine-tune the model π0 using a combined loss, as shown in Figure 4, that both incorporates feedback on in-scope prompts and discourages model degradation on out-of-scope and nearscope prompts: Apply feedback z }| { LDPO(Din-scope) + λ1LSFT(Dout-of-scope) + λ2LSFT(Dnear-scope) | {z } Control model degradation Interpreting the C3PO loss. While the LSFT losses simply regularize the updated model πθ toward the original model π0 for prompts not relevant to the feedback, the result of learning from the C3PO synthetic preference dataset for in-scope inputs is less obvious. C3PO generates what we refer to as synthetic two-policy preference data (x, y+, y ), where y+ is always preferred to y . These preference tuples are constructed by simply sampling y from a policy π (the baseline model π0) and y+ from a different policy π+ (the baseline model π0 prompted to revise a baseline model response using the feedback).3 Unlike preference datasets scored by a black-box human or AI annotator, we can express the optimal policy learned from such preference data in terms of the data-generating policies. We show in Appendix C that such synthetic two-policy preference data satisfies the Bradley-Terry (BT) preference model (Bradley & Terry, 1952), which assumes that preference data (x, y, y ) are constructed according to some unknown scoring function r as p(y y |x) = σ (r (x, y) r (x, y )). We show that two-policy preference data adheres to the BT model with r 2p(x, y) = log π+(y|x) π (y|x). (6) Further, we show that when using π as the reference model for Pb RL, the optimal policy corresponds to π 2p(y|x) π+(y|x) Notable special cases of π 2p are the geometric mean of π+ and π for β = 2 and simply π+ for β = 1. For β < 1, we interpret π 2p as returning a temperature-sharpened version of π+, but with a penalty on responses assigned high probability under π (i.e., responses that respond to the user but fail to adhere to the feedback). See Appendix C for visualization of π 2p with various β in a synthetic setting. 5 Experiments Our experiments are intended to answer several research questions about learning from verbal feedback. We first 3Some existing work (Yang et al., 2024; Intel, 2023) shows that synthetic two-policy preference data can produce useful policies. RLVF: Learning from Verbal Feedback without Overgeneralization investigate the question: to what extent does the overgeneralization problem occur for existing methods for learning from verbal feedback, and does C3PO mitigate this effect? Next, we study whether simple modifications to standard approaches to incorporating feedback with prompting or supervised context distillation effectively mitigate overgeneralization. Further, we study whether or not C3PO can learn multiple pieces of feedback as well as the impact of the specific choice of the constrained loss used in C3PO. Before discussing the results of these experiments, we elaborate the datasets, evaluation metrics, and baseline methods used in our experiments. Datasets. Our feedback dataset is composed of 100 pieces of feedback, where half are written by the authors4 and half are generated by GPT-4 using the prompt provided in Appendix D. All pieces of feedback are designed to apply only in some contexts; see Table 1 for examples. For C3PO and the SCD + Negatives baseline, the datasets Dnear-scope and Dout-of-scope, each containing outof-scope prompts and corresponding baseline model completions used for regularization, are sampled according to the C3PO procedure in Section 4. For each piece of feedback, |Din-scope| = |Dnear-scope| = |Dout-of-scope| = 960. When reporting results, we combine results for Dnear scope and Dout of scope unless note otherwise. We sample the prompts for Dout-of-scope from the Open Instruction Generalist (OIG) Dataset (LAION, 2023) which contains a mix of diverse prompts ranging from math to QA and chat. Within each prompt sub-dataset, we randomly select 80% of the prompts to be used for training and validation and the remainder are used for testing. Evaluation metrics. Our evaluations are constructed to compare the feedback adherence of the baseline model with the model after learning from the feedback; we evaluate this change in behavior for in-scope, near-scope, and out-ofscope prompts. For in-scope prompts, our goal is to increase feedback adherence, while for out-of-scope and near-scope prompts, our goal is to preserve the rate of feedback adherence of the original model (that is, leave the baseline model unchanged). We measure feedback adherence in two ways, heuristically and with GPT-4. For 14 of the humangenerated pieces of feedback such as modifying response length or the inclusion of certain words, manually crafted heuristic rules are sufficient to reliably measure which of two responses better adheres to a given piece of feedback. For a prompt x, a model output y, and a baseline response y from the baseline model, the heuristic scoring function produces a feedback score h(x, y, y). This scoring function intuitively scores whether the feedback adherence of the adapted model response is better than, equal to, or worse 4One piece of feedback was inspired by a suggestion from social media. Figure 5: Sample responses from C3PO and each baseline for an in-scope and out-of-scope prompt. Only C3PO correctly adheres to the feedback for the in-scope input and ignores the feedback for the out-of-scope input. than the baseline model response. h(x, y, y) takes a value of 1 if y incorporates the feedback and y does not, a value of 0 if both responses adhere to the feedback or neither response adheres to the feedback, and a value of -1 if the baseline response y adheres to the feedback and the adapted model response y does not. In contrast, most pieces of feedback, such as requesting the usage of more metaphors or less aggressive speech, require qualitative evaluation. In these cases, we measure relative feedback adherence using GPT-4. For a prompt x, a model output y, and a baseline response y from the baseline model, we prompt GPT-4 to output a preference score when comparing two responses using the prompt in Appendix D, producing a feedback score g(x, y, y) scaled to be in the range [ 1, 1]. This score measures the extent to which a given response adheres to the feedback better than the baseline response; a score of 1 denotes that the trained model response adheres to the feedback much better than the baseline model and 1 denotes the reverse. We use these metrics (heuristic scoring or GPT-4 scoring) to measure the feedback adherence of the trained model responses compared to the response of the base model for in-scope prompts. The in-scope adherence score for an algorithm on a given piece of feedback is the average of the per-prompt feedback scores (h(x, y, y) if the prompt x is heuristically checkable, g(x, y, y) otherwise) across all in-scope test prompts for that feedback. The overall in-scope adherence score Sin is the average of these per-feedback adherence scores over all feedbacks evaluated. For out-of-scope prompts, our goal is to measure changes RLVF: Learning from Verbal Feedback without Overgeneralization C3PO (Ours) In-Context SCD DPO 0.0 % Prompts Feedback is Applied In-scope Out-of-scope Near-scope Figure 6: Contextualized Critiques with Constrained Preference Optimization substantially reduces overgeneralization (applying the given feedback to prompts where it is not actually relevant) with only minor reduction in adherence to feedback for prompts where the feedback is relevant. in the model s behavior as a result of incorporating the feedback since both applying the feedback more or less frequently on such prompts is a deviation from the base model. We measure model behavior change on out-of-scope prompts as the average absolute change in the rate that the feedback is applied since over this domain, our objective is to leave the model behavior unchanged. To compute the outof-scope behavior change score for an algorithm on a given piece of feedback, we average the absolute value of the feedback scores (again, h(x, y, y) if x is a heuristically checkable prompt, g(x, y, y) otherwise). The overall behavior change score Sout is the average of these per-feedback behavior change scores over all feedbacks evaluated. Apart from these individual metrics, we define Soverall = Sin+(1 Sout) 2 as a combined metric with equal weighting of the in-scope feedback adherence score Sin and one minus the out-of-scope behavior change score Sout. A Soverall near one indicates that an algorithm effectively adheres to the given feedback better than the baseline model for in-scope prompts while preserving the level of feedback adherence in the baseline model on out-of-scope prompts. Methods. We compare C3PO against both in-contextlearning-based and fine-tuning methods. For In-Context learning, we provide the baseline model with the user query as well as the feedback and a prompt instructing the model to selectively apply the feedback whenever it is applicable to the given user query. We explore an enhancement of this naive prompting approach, In-Context + Co T using chain-of-thought prompting; this approach is the same as In-Context, except we prompt the model to first reason stepby-step about the applicability of the provided feedback to the given prompt before answering. See Appendix D for the full prompts. Next, we compare against performing supervised context distillation SCD on in-scope prompts, Method Sin Sout Soverall In-Context 0.677 .028 0.503 .025 0.587 .026 In-Context + Co T 0.402 .033 0.246 .017 0.578 .026 SCD 0.470 .029 0.269 .019 0.6005 .025 SCD + Negatives 0.367 .027 0.133 .013 0.617 .021 DPO 0.326 .048 0.517 .022 0.4045 .037 C3PO (Ours) 0.563 .031 0.150 .014 0.7065 .024 Table 2: C3PO provides the strongest overall performance averaged over 100 pieces of feedback. Augmentations to the In-Context and SCD baselines, In-Context + Co T and SCD + Negatives, respectively, reduce the amount of overgeneralization of both approaches, but at a substantial cost to in-scope feedback adherence, and ultimately these improvements do not change the overall score for either method. using the revised responses y+ generated by C3PO as the supervision targets. In addition, in order to better control overgeneralization, we evaluate SCD + Negatives, which adds a weighted constraint to the SCD loss over out-ofscope prompts, using the baseline model responses on these prompts as the supervision target. Finally, we evaluate DPO on the preference dataset Din-scope without additional regularization, essentially an ablation of C3PO s regularization losses. For all experiments, we use Mistral-7B-Instruct-v0.2 (Jiang et al., 2023) and train with Low-Rank Adaptation (Hu et al., 2021) with a rank of 64 and alpha of 128. We use a learning rate of 5e-5 with a cosine decay schedule, a warm-up ratio of 0.05, Adam W as the optimizer, and train for 1 epoch. 5.1 Quantifying and mitigating overgeneralization Our initial results in Figure 2 show that for existing approaches to learning from verbal feedback, successfully incorporating the feedback (large x-axis value) leads to application of that feedback for prompts where the feedback does not apply (large y-axis value). That is, successful adherence to feedback for in-scope prompts comes at a high cost in terms of incorrect adherence to that feedback for out-of-scope prompts, shown by the large slope of the best fit lines. This result also shows that C3PO trades off inscope adherence and out-of-scope behavior preservation much more efficiently, shown by the much smaller slope of the best-fit line. In this section, we study the impact of incorporating verbal feedback in terms of the evaluation metrics Sin and Sout, measuring out-of-scope behavior change for both generic out-of-scope prompts and more difficult nearscope prompts. The results are shown in Figure 6. C3PO dramatically reduces the behavior change for both general out-of-scope prompts and near-scope prompts, while only slightly reducing feedback adherence for in-scope prompts. Additionally, we investigate the prevalence of the overgeneralization problem in stronger models and evaluate GPT4-1106-preview ("GPT4") Co T on a set of 1000 prompts RLVF: Learning from Verbal Feedback without Overgeneralization Figure 7: Mixing the Lo RA weights that C3PO learns for two different pieces of feedback (right) provides virtually the same level of in-scope feedback adherence and out-of-scope behavior change as applying and evaluating each feedback independently (left). This result suggests that learning separate Lo RA parameters for each feedback as it is received and simply adding them together to acquire the model adapted to all feedbacks may be viable. for each of Din-scope, Dnear-scope, Dout-of-scope randomly sampled across all feedbacks. We find that while Mistral-7BInstruct-v0.2 with Co T correctly applies the feedback in only 48.8% of prompts for Din-scope, GPT4 does so for 56.8% of prompts, which is an 8% improvement. For Dnear-scope and Dout-of-scope, GPT4 overgeneralizes in 27% and 25.6% of cases, respectively, while Mistral-7B-Instruct-v0.2 overgeneralizes in 25.4% and 23.5% of cases. This result suggests that GPT4 overgeneralizes at least to a similar degree as Mistral-7B-Instruct-v0.2 while displaying an improved ability to correctly apply the feedback for in-scope prompts. Hence, overgeneralization is an issue for larger models as well. It is natural to wonder whether we can improve the performance of the baseline methods using a similar goal of constraining the behavior change for out-of-scope prompts. We therefore evaluate the modifications of the In-Context and SCD methods, In-Context + Co T and SCD + Negatives, intended to reduce behavior change. The In-Context + Co T method first performs chain-of-thought reasoning to decide whether the feedback is applicable to the given input before responding; the SCD + Negatives baseline mixes in the regularization loss on the Dnear-scope and Dout-of-scope datasets. We report the in-scope feedback adherence and out-of-scope behavior change (averaged over the general out-of-scope and near-scope prompts) in Table 2. While both improvements do substantially reduce the amount of overgeneralization compared to the original version of each method, they come at a substantial cost to in-scope feedback adherence. Therefore, the overall score Soverall does not substantially improve for either of these approaches; however, C3PO offers a superior tradeoff between in-scope feedback adherence and avoiding out-of-scope behavior change, shown by the sig- nificantly higher Soverall. We therefore conclude that C3PO is an effective way to reduce overgeneralization while generally maintaining in-scope feedback adherence. This claim is supported by the results in Figure 5. 5.2 Impact of C3PO on general task performance For practical applications it is critical that a model s performance on general tasks does not degrade after applying C3PO. We study possible degradation by evaluating the model before and after applying C3PO on 500 prompts sampled from Dout-of-scope, which consists of general chat prompts across QA, math, coding, and more. We use GPT4 to assign ratings from 1-5 to model responses where 1 denotes that the response does not answer the prompt correctly and 5 denotes that the response answers the prompt exceptionally well. The original model achieves an average score of 4.497 0.040 and after applying C3PO, the model scores 4.470 0.040. Based on this decrease of only 0.027, we conclude that the model s capabilities on general tasks do not degrade noticeably after fine-tuning with C3PO. 5.3 Synthetic data using weaker models Our previous experiments assumed the availability of a strong LLM such as GPT4 to generate Din-scope and Dout-of-scope. One important question to ask is whether C3PO remains performant when replacing GPT4 with a weaker model in the data generation procedure. We shed light on this question by applying C3PO for 30 randomly sampled feedbacks using data generated by GPT4 and Mixtral-8x7B-Instruct-v0.1 (Jiang et al., 2024) respectively. We find that using GPT4-generated data C3PO achieves Soverall = 0.6875 0.048 while using data generated by Mixtral achieves Soverall = 0.6225 0.043. For comparison, the SCD + Negatives baseline, which is the second best method after C3PO according to our main results in table 2, achieves Soverall = 0.613 0.042 using GPT4generated data. This result implies C3PO outperforms all baseline methods, even when using a much weaker model than GPT4 to generate the training data. 5.4 Adhering to multiple feedbacks So far, our experiments have evaluated the average result of applying a single piece of feedback to a pre-trained model. While a comprehensive evaluation of continual learning of many feedbacks is out of the scope of this work, we perform an initial investigation in this direction by combining the Lo RA parameters learned from separate feedbacks. That is, we perform C3PO separately on two different feedbacks z1 and z2, producing Lo RA parameters ϕ1 and ϕ2. Ideally, to produce a model that adheres to both z1 and z2, rather than re-training, we could simply use the concatenation of residual parameters ϕ = ϕ1 ϕ2. In this section, we compare RLVF: Learning from Verbal Feedback without Overgeneralization Figure 8: Replacing C3PO s maximum likelihood constraint on out-of-scope prompts with full knowledge distillation leads to substantially impaired performance on in-scope prompts, suggesting that allowing some subtle changes to the model s conditional distribution for out-of-scope prompts may be beneficial. the performance of this approach to combining feedbackadapted models. Figure 7 compares the average feedback adherence and behavior change when applying and evaluating only a single piece of feedback at a time (left) with the average feedback adherence and behavior change on two feedbacks after applying the merged Lo RA parameters ϕ (right), averaged across 20 feedbacks (10 pairs). We observe virtually no degradation in in-scope feedback adherence and no change in out-of-scope behavior change. This result is promising for the possibility of enabling rapid, mix-andmatch personalization and customization of large language models without re-training. 5.5 Choice of C3PO constraint formulation C3PO constrains the feedback-adapted model by maximizing the likelihood of the baseline model s responses on out-of-scope and near-scope prompts during fine-tuning. As a stronger constraint, we investigate performing full knowledge distillation (Hinton et al., 2015) for the conditional distributions of the adapted model and baseline model at each time step. That is, performing knowledge distillation on only out-of-scope and near-scope prompts, where the baseline model is the teacher and the adapted model is the student. Rather than simply maximizing the likelihood of baseline model samples from each out-of-scope prompt, this constraint minimizes the KL-divergence between the baseline model s conditional distribution and the adapted model s conditional distribution, averaged over all timesteps in the baseline model s completion. While this form of constraint has successfully leveraged the dark knowledge represented by the lower-probability logits in the model s output to constrain neural networks in the context of continual learning (Buzzega et al., 2020), we find that this stronger constraint substantially impairs in-scope feedback adherence compared to the maximum likelihood constraint. Alternative approaches to constraining out-of-scope model behavior is an important topic for future work. 6 Discussion & Future Work As large language models become more ubiquitous and widely-adopted, the ability to easily customize and personalize their behaviors is increasingly valuable. We studied the problem of learning from high-level verbal feedback, where a user provides a short piece of written feedback describing a desired change to a language model s behavior, and an algorithm updates the model to adhere to this feedback when it is appropriate but preserves the model s behavior elsewhere. We showed that applying existing methods for fine-tuning language models from feedback demonstrate severe overgeneralization: after incorporating the feedback, they also change model behavior for inputs that are not relevant to the provided feedback. To mitigate this problem, we introduced Contextualized Critiques with Constrained Preference Optimization (C3PO), an algorithm that performs reinforcement learning from verbal feedback (RLVF). C3PO leverages existing language models to generate a set of small fine-tuning datasets that encode both the desired change in behavior described by the feedback and the set of behaviors that should be preserved for inputs unrelated to the feedback. We found that C3PO substantially reduces overgeneralization while still adhering to feedback for relevant inputs. Our experiments raise several important questions for future work. Can we perform continual learning from feedback by simply continually aggregating and combining adapted model weights? In addition, our ablations of the C3PO constraint loss function suggest that the proper level of strictness when constraining the model update is a non-trivial problem, and better-performing alternatives may exist. Additionally, future work may investigate the relationship between the complexity of the feedback and the capabilities of the base model being adapted. Finally, since different feedbacks can be applicable to scopes of different size, finding synthetic data generation model that can generate near-scope examples at the correct granularity for any feedback could improve the realiability of our method. Acknowledgements The authors gratefully acknowledge the Open AI Researcher Access Program and Modal.com for providing compute credits used in this project. CF is a CIFAR Fellow. EM acknowledges Kyle Hsu for helpful conversations. This work was supported by ONR grant N00014-20-1-2675 and an NSF CAREER award. RLVF: Learning from Verbal Feedback without Overgeneralization Impact Statement This paper presents work whose goal is to advance the field of Machine Learning. There are many potential societal consequences of our work, none which we feel must be specifically highlighted here. Akyürek, A., Pan, E., Kuwanto, G., and Wijaya, D. DUn E: Dataset for unified editing. In Bouamor, H., Pino, J., and Bali, K. (eds.), Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pp. 1847 1861, Singapore, December 2023. Association for Computational Linguistics. doi: 10.18653/v1/2023. emnlp-main.114. URL https://aclanthology. org/2023.emnlp-main.114. Askell, A., Bai, Y., Chen, A., Drain, D., Ganguli, D., Henighan, T., Jones, A., Joseph, N., Mann, B., Das Sarma, N., Elhage, N., Hatfield-Dodds, Z., Hernandez, D., Kernion, J., Ndousse, K., Olsson, C., Amodei, D., Brown, T., Clark, J., Mc Candlish, S., Olah, C., and Kaplan, J. A general language assistant as a laboratory for alignment, 2021. Austin, J., Odena, A., Nye, M., Bosma, M., Michalewski, H., Dohan, D., Jiang, E., Cai, C., Terry, M., Le, Q., and Sutton, C. Program synthesis with large language models, 2021. Bai, Y., Kadavath, S., Kundu, S., Askell, A., Kernion, J., Jones, A., Chen, A., Goldie, A., Mirhoseini, A., Mc Kinnon, C., Chen, C., Olsson, C., Olah, C., Hernandez, D., Drain, D., Ganguli, D., Li, D., Tran-Johnson, E., Perez, E., Kerr, J., Mueller, J., Ladish, J., Landau, J., Ndousse, K., Lukoši ut e, K., Lovitt, L., Sellitto, M., Elhage, N., Schiefer, N., Mercado, N., Das Sarma, N., Lasenby, R., Larson, R., Ringer, S., Johnston, S., Kravec, S., Showk, S. E., Fort, S., Lanham, T., Telleen-Lawton, T., Conerly, T., Henighan, T., Hume, T., Bowman, S., Hatfield Dodds, Z., Mann, B., Amodei, D., Joseph, N., Mc Candlish, S., Brown, T. B., and Kaplan, J. Constitutional ai: Harmlessness from ai feedback. ar Xiv.org, 2022. doi: 10.48550/ARXIV.2212.08073. Böhm, F., Gao, Y., Meyer, C. M., Shapira, O., Dagan, I., and Gurevych, I. Better rewards yield better summaries: Learning to summarise without references. In Inui, K., Jiang, J., Ng, V., and Wan, X. (eds.), Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 3110 3120, Hong Kong, China, November 2019. Association for Computational Linguistics. doi: 10.18653/v1/D19-1307. URL https: //aclanthology.org/D19-1307. Bradley, R. A. and Terry, M. E. Rank analysis of incomplete block designs: I. the method of paired comparisons. Biometrika, 39(3/4):324 345, 1952. doi: https://doi.org/10.2307/2334029. Busa-Fekete, R., Szörényi, B., Weng, P., Cheng, W., and Hüllermeier, E. Preference-based reinforcement learning: evolutionary direct policy search using a preference-based racing algorithm. Machine Learning, 97(3):327 351, July 2014. doi: 10.1007/s10994-014-5458-8. URL https: //doi.org/10.1007/s10994-014-5458-8. Buzzega, P., Boschini, M., Porrello, A., Abati, D., and Calderara, S. Dark experience for general continual learning: a strong, simple baseline. In Proceedings of the 34th International Conference on Neural Information Processing Systems, NIPS 20, Red Hook, NY, USA, 2020. Curran Associates Inc. ISBN 9781713829546. Cao, N. D., Aziz, W., and Titov, I. Editing factual knowledge in language models. Conference on Empirical Methods in Natural Language Processing, 2021. doi: 10.18653/ V1/2021.EMNLP-MAIN.522. Castricato, L., Lile, N., Anand, S., Schoelkopf, H., Verma, S., and Biderman, S. Suppressing pink elephants with direct principle feedback, 2024. Chen, A., Scheurer, J., Korbak, T., Campos, J. A., Chan, J. S., Bowman, S. R., Cho, K., and Perez, E. Improving code generation by training with natural language feedback, 2023. Clark, P., Tafjord, O., and Richardson, K. Transformers as soft reasoners over language. In Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, IJCAI 20, 2021. ISBN 9780999241165. Glaese, A., Mc Aleese, N., Tr ebacz, M., Aslanides, J., Firoiu, V., Ewalds, T., Rauh, M., Weidinger, L., Chadwick, M., Thacker, P., Campbell-Gillingham, L., Uesato, J., Huang, P.-S., Comanescu, R., Yang, F., See, A., Dathathri, S., Greig, R., Chen, C., Fritz, D., Elias, J. S., Green, R., Mokrá, S., Fernando, N., Wu, B., Foley, R., Young, S., Gabriel, I., Isaac, W., Mellor, J., Hassabis, D., Kavukcuoglu, K., Hendricks, L. A., and Irving, G. Improving alignment of dialogue agents via targeted human judgements, 2022. Hancock, B., Bordes, A., Mazare, P.-E., and Weston, J. Learning from dialogue after deployment: Feed yourself, chatbot! In Korhonen, A., Traum, D., and Màrquez, L. (eds.), Proceedings of the 57th Annual Meeting of RLVF: Learning from Verbal Feedback without Overgeneralization the Association for Computational Linguistics, pp. 3667 3684, Florence, Italy, July 2019. Association for Computational Linguistics. doi: 10.18653/v1/P19-1358. URL https://aclanthology.org/P19-1358. Hewitt, J., Chen, S., Xie, L. L., Adams, E., Liang, P., and Manning, C. D. Model editing with canonical examples, 2024. Hinton, G. E., Vinyals, O., and Dean, J. Distilling the knowledge in a neural network. Ar Xiv, abs/1503.02531, 2015. URL https://api.semanticscholar. org/Corpus ID:7200347. Hu, E. J., Shen, Y., Wallis, P., Allen-Zhu, Z., Li, Y., Wang, S., Wang, L., and Chen, W. Lora: Low-rank adaptation of large language models, 2021. Huang, J., Gu, S., Hou, L., Wu, Y., Wang, X., Yu, H., and Han, J. Large language models can self-improve. ar Xiv, 2023. doi: 10.18653/V1/2023.EMNLP-MAIN.67. Intel. Supervised fine-tuning and direct preference optimization on intel gaudi2, November 2023. URL https://medium.com/intelanalytics-software/the-practice-ofsupervised-finetuning-and-directpreference-optimization-on-habanagaudi2-a1197d8a3cd3. [Online; posted 14November-2023]. Jiang, A. Q., Sablayrolles, A., Mensch, A., Bamford, C., Chaplot, D. S., de las Casas, D., Bressand, F., Lengyel, G., Lample, G., Saulnier, L., Lavaud, L. R., Lachaux, M.- A., Stock, P., Scao, T. L., Lavril, T., Wang, T., Lacroix, T., and Sayed, W. E. Mistral 7b, 2023. Jiang, A. Q., Sablayrolles, A., Roux, A., Mensch, A., Savary, B., Bamford, C., Chaplot, D. S., de las Casas, D., Hanna, E. B., Bressand, F., Lengyel, G., Bour, G., Lample, G., Lavaud, L. R., Saulnier, L., Lachaux, M.-A., Stock, P., Subramanian, S., Yang, S., Antoniak, S., Scao, T. L., Gervet, T., Lavril, T., Wang, T., Lacroix, T., and Sayed, W. E. Mixtral of experts, 2024. Kingma, D. and Ba, J. Adam: A method for stochastic optimization. In International Conference on Learning Representations (ICLR), San Diega, CA, USA, 2015. LAION. The open instruction generalist dataset, 2023. URL https://huggingface.co/datasets/ laion/OIG. Li, J., Monroe, W., Ritter, A., Jurafsky, D., Galley, M., and Gao, J. Deep reinforcement learning for dialogue generation. In Su, J., Duh, K., and Carreras, X. (eds.), Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, pp. 1192 1202, Austin, Texas, November 2016. Association for Computational Linguistics. doi: 10.18653/v1/D16-1127. URL https://aclanthology.org/D16-1127. Li, J., Miller, A. H., Chopra, S., Ranzato, M., and Weston, J. Dialogue learning with human-in-the-loop. In International Conference on Learning Representations, 2017. URL https://openreview.net/forum? id=HJg XCV9xx. Lu, J., Zhong, W., Huang, W., Wang, Y., Mi, F., Wang, B., Wang, W., Shang, L., and Liu, Q. Self: Language-driven self-evolution for large language model. ar Xiv.org, 2023. doi: 10.48550/ARXIV.2310.00533. Madaan, A., Tandon, N., Gupta, P., Hallinan, S., Gao, L., Wiegreffe, S., Alon, U., Dziri, N., Prabhumoye, S., Yang, Y., Welleck, S., Majumder, B. P., Gupta, S., Yazdanbakhsh, A., and Clark, P. Self-refine: Iterative refinement with self-feedback, 2023. Mao, S., Zhang, N., Wang, X., Wang, M., Yao, Y., Jiang, Y., Xie, P., Huang, F., and Chen, H. Editing personality for llms. ar Xiv, 2023. Meng, K., Bau, D., Andonian, A., and Belinkov, Y. Locating and editing factual knowledge in gpt. Ar Xiv, 2022. Mitchell, E., Lin, C. P., Bosselut, A., Finn, C., and Manning, C. D. Fast model editing at scale. International Conference on Learning Representations, 2021. doi: 10.48550/ARXIV.2110.11309. Mitchell, E., Lin, C. P., Bosselut, A., Manning, C. D., Finn, C., Mitchell, E., Lin, C. P., Bosselut, A., Manning, C. D., and Finn, C. Memory-based model editing at scale. International Conference on Machine Learning, 2022. doi: 10.48550/ARXIV.2206.06520. Murty, S., Manning, C. D., Lundberg, S. M., and Ribeiro, M. T. Fixing model bugs with natural language patches. Conference on Empirical Methods in Natural Language Processing, 2022. doi: 10.48550/ARXIV.2211.03318. Ouyang, L., Wu, J., Jiang, X., Almeida, D., Wainwright, C. L., Mishkin, P., Zhang, C., Agarwal, S., Slama, K., Ray, A., Schulman, J., Hilton, J., Kelton, F., Miller, L., Simens, M., Askell, A., Welinder, P., Christiano, P., Leike, J., and Lowe, R. Training language models to follow instructions with human feedback, 2022. Pan, L., Saxon, M. S., Xu, W., Nathani, D., Wang, X., and Wang, W. Y. Automatically correcting large language models: Surveying the landscape of diverse selfcorrection strategies. ar Xiv.org, 2023. doi: 10.48550/ ARXIV.2308.03188. RLVF: Learning from Verbal Feedback without Overgeneralization Peng, X. B., Kumar, A., Zhang, G., and Levine, S. Advantage-weighted regression: Simple and scalable off-policy reinforcement learning. ar Xiv preprint ar Xiv:1910.00177, 2019. Peters, J. and Schaal, S. Reinforcement learning by rewardweighted regression for operational space control. In Proceedings of the 24th international conference on Machine learning, pp. 745 750, 2007. Rafailov, R., Sharma, A., Mitchell, E., Manning, C. D., Ermon, S., and Finn, C. Direct preference optimization: Your language model is secretly a reward model. In Thirtyseventh Conference on Neural Information Processing Systems, 2023. URL https://openreview.net/ forum?id=HPu SIXJaa9. Saha, A., Pacchiano, A., and Lee, J. Dueling rl: Reinforcement learning with trajectory preferences. In Ruiz, F., Dy, J., and van de Meent, J.-W. (eds.), Proceedings of The 26th International Conference on Artificial Intelligence and Statistics, volume 206 of Proceedings of Machine Learning Research, pp. 6263 6289. PMLR, 25 27 Apr 2023. URL https://proceedings.mlr.press/ v206/saha23a.html. Scheurer, J., Campos, J. A., Korbak, T., Chan, J. S., Chen, A., Cho, K., and Perez, E. Training language models with language feedback at scale, 2023. Schulman, J., Wolski, F., Dhariwal, P., Radford, A., and Klimov, O. Proximal policy optimization algorithms. ar Xiv preprint ar Xiv:1707.06347, 2017. Shi, W., Dinan, E., Shuster, K., Weston, J., and Xu, J. When life gives you lemons, make cherryade: Converting feedback from bad responses into good labels, 2022. Sinitsin, A., Plokhotnyuk, V., Pyrkin, D., Popov, S., and Babenko, A. Editable neural networks. In International Conference on Learning Representations, 2020. URL https://openreview.net/forum? id=HJed Xa Etv S. Snell, C. B., Klein, D., and Zhong, R. Learning by distilling context. ar Xiv.org, 2022. doi: 10.48550/ARXIV.2209. 15189. Sun, Z., Shen, Y., Zhou, Q., Zhang, H., Chen, Z., Cox, D. D., Yang, Y., and Gan, C. Principle-driven self-alignment of language models from scratch with minimal human supervision. ar Xiv.org, 2023. doi: 10.48550/ARXIV. 2305.03047. Walker, M. A. An application of reinforcement learning to dialogue strategy selection in a spoken dialogue system for email. J. Artif. Int. Res., 12(1):387 416, jun 2000. ISSN 1076-9757. Yang, K., Klein, D., Celikyilmaz, A., Peng, N., and Tian, Y. RLCD: Reinforcement learning from contrastive distillation for LM alignment. In The Twelfth International Conference on Learning Representations, 2024. URL https://openreview.net/forum? id=v3XXtx WKi6. Yu, X., Peng, B., Galley, M., Gao, J., and Yu, Z. Teaching language models to self-improve through interactive demonstrations, 2023. Yuan, W., Pang, R. Y., Cho, K., Sukhbaatar, S., Xu, J., and Weston, J. Self-rewarding language models, 2024. Ziegler, D. M., Stiennon, N., Wu, J., Brown, T. B., Radford, A., Amodei, D., Christiano, P., and Irving, G. Fine-tuning language models from human preferences, 2020. RLVF: Learning from Verbal Feedback without Overgeneralization A Sampling Details We sample from GPT-4 using a temperature of 0.7 and a topp value of 0.7. When sampling from Mistral-7B-Instructv0.2, use a temperature of 0.7, a top-p value of 0.7, and top-k value of 50 and a repetition penalty of 1. B Training Details To conduct our hyperparameter search, we select 10 arbitrary pieces of feedback from the human-generated feedback dataset. For all methods, we train using Lo RA and choose a rank of 64, alpha of 128, and a Lo RA dropout of 0.05. We observe that a smaller rank results in decreases inscope feedback adherence while increasing the rank above 64 results in degradation of model completions across all methods. We train for 1 epoch with a learning rate of 5e 5 using a cosine learning rate schedule and a 0.05 warmup ratio. We ablate learning rates from 1e 7 to 1e 4 and found that below 1e 5, in-scope feedback adherence never increases sufficiently, even when training for multiple epochs. We, thus, picked our learning rate to be larger than this threshold but still small enough avoid a plateauing loss. In addition, we experimented with training for multiple epochs but found that this does not significantly alter the results and sometimes even increases feedback adherence for out-of-scope and near-scope prompts. To choose λ1 and λ2 of the LC3P O objective, we conducted a grid search over the two hyperparameters. We found that it is crucial to set both λ1 > 0 and λ2 > 0 and that increasing λ2, which is the weight for the SFT loss over the near-scope samples, beyond 0.1 decreases in-scope feedback adherence drastically while only marginally mitigating overgeneralization. Additionally, we found that C3PO is not as sensitive over the choice of λ1 as long as it is non-zero and chose 0.2 due to the best in-scope to out-of-scope performance trade-off on our evaluation set. For both C3PO and DPO, we select a β parameter of 0.1. We ablate with values of 0.05, 0.15, and 0.25 and find that a setting of 0.05 results in less in-scope feedback adherence and more overgeneralization while β > 0.1 only reduces in-scope feedback adherence. C Derivation of Optimal Policy for Pb RL on Two-Policy Preference Pairs In this section, we derive several properties of the Pb RL learning procedure used by C3PO. First, we demonstrate that the synthetic preference data generated by C3PO adheres to the commonly-used Bradley-Terry model (Bradley & Terry, 1952) of discrete choice and compute the true reward function implied by the preference data. Using this reward function, we then show that the optimal policy learned by C3PO for in-scope prompts takes the form in Equation (7). Finally, we perform a simple empirical validation of this theoretical result in a synthetic bandit problem. C.1 Deriving the underlying Bradley-Terry scoring function for synthetic two-policy preference data The Bradley-Terry model of discrete choices states that for a preference over two responses y, y , we have p(y y |x, y, y ) = σ (r (x, y) r (x, y )) (8) for some true scoring function r . In the case of two-policy preference data, we assume that the preferred response is generated from a policy π+ and the dispreferred response is generated by a policy π . The probability of observing the response pair y, y is p(y, y |x, A) = π+(y|x)π (y |x) and p(y, y |x, A) = π+(y |x)π (y|x), where A is the event that y was generated by π+ and y from π (and therefore y y by the definition of our data-generating process). By Bayes Rule, the probability of A (that y was generated by π+ and y by π ) is p(A|x, y, y ) = p(y y |x, y, y ) = π+(y|x)π (y |x) π+(y|x)π (y |x) + π+(y |x)π (y|x). (9) We now set the RHS of Equation (8) equal to the RHS of Equation (9) and show that a Bradley-Terry scoring function exists. We have: σ (r (x, y) r (x, y )) = π+(y|x)π (y |x) π+(y|x)π (y |x) + π+(y |x)π (y|x). (10) Applying the transformation log z 1 z to both sides, we have r (x, y) r (x, y ) = log π+(y|x) π (y|x) log π+(y |x) π (y |x), (11) implying the Bradley-Terry scoring function r (x, y) = log π+( y|x) π ( y|x) + C(x) (12) for some value C(x) that is constant with respect to y. Therefore, synthetic two-policy preference data is Bradley Terry, with the above scoring function. C.2 Deriving the optimal policy for C3PO for in-scope prompts The optimal policy for the KL-regularized reinforcement learning objective optimized by DPO is shown in prior work RLVF: Learning from Verbal Feedback without Overgeneralization 0 20 40 60 80 100 Action Index (|A| = 100) Policy Probability Mass Learned Reward True Reward (log + / ) Figure 9: Empirical demonstration of the result of C3PO s optimal policy on in-scope prompts. For β = 1, the optimal policy recovers π+; for smaller beta, the optimal policy amplifies the delta between π+ and π . Note that the reward learned with the Bradley-Terry loss perfectly recovers the theoretically optimal reward in Equation (12). (Peters & Schaal, 2007; Peng et al., 2019; Rafailov et al., 2023) to be π (y|x) = 1 Z(x)πref(y|x) exp 1 β r(x, y) . (13) Substituting the optimal reward function in Equation (12), assuming C(x) = 0 as this constant term does not affect the optimal policy, we have π (y|x) = 1 Z(x)πref(y|x) exp 1 β log π+(y|x) = 1 Z(x)πref(y|x) π+(y|x) Assuming that πref = π gives π (y|x) = 1 Z(x)π (y|x)π+(y|x) 1 β π (y|x) 1 β (16) β π+(y|x) 1 β π (y|x) 1 β (17) which is the optimal policy expression given in Equation (7). C.3 Empirically validating the in-scope C3PO policy in a synthetic setting We validate the theoretical results in the previous subsections in a simple bandit setting with 100 possible actions. We define the preferred policy π+ as π+(i) exp (i 70)2 1000 and similarly define π as the shifted pol- icy π (i) exp (i 40)2 1000 . We generate 1e8 preference pairs (y+, y ), where y+ π+( ) and similarly for y . We fit a reward function to this preference data using the Bradley-Terry loss, using 400 steps of gradient descent with the Adam optimizer (Kingma & Ba, 2015) with learning rate 1.0, initializing the reward function to the zero vector. The results are shown in Appendix C.3. We find that the recovered reward function is exactly the Bradley-Terry scoring function predicted by Equation (12), and that the special cases predicted by Equation (7) hold (e.g., with β = 1, we have simply π = π+). With β < 1, we have the intuitive behavior of amplifying the delta between π+ and π . RLVF: Learning from Verbal Feedback without Overgeneralization The following sections contain the verbatim prompts used for the various stages of the C3PO synthetic data generation procedure. D.1 Category Generation Prompt You are a helpful assistant. You are helping a user come up with categories around a topic that will be used to create some questions that the topic applies to and some questions that the topic does not apply to. Given a topic, come up with {count} creative and diverse categories that are an only slightly larger superset of the topic. Be creative, think out of the box, and keep the categories closely related to the topic. Ensure that for each category, it would be possible to easily come up with questions that are not related to the provided topic. Do not repeat categories and make sure you cover all relevant categories. You should first respond with "THOUGHTS" that describe what you should and should not mention in your categories and why. Then output "CATEGORIES: " on a new line and output each category on a new line as part of a numbered list. Finally, output " REVISED_CATEGORIES: " on a new line followed a revised version of each of the categories you came up with. Use the revision to modify a category if it would be very hard to come up with a prompt for that category that the topic does not apply to. The revision should also be a numbered list. If you do a great job, you will be tipped $200. --EXAMPLE 1-- TOPIC: current fashion trends THOUGHTS: I should list categories that are either related to fashion but that are not explicitly about trends. None of the categories I respond with should be directly about fashion trends. CATEGORIES: 1. Buying luxury fashion 2. gen-z pop culture trends 3. fast-fashion trends 4. men s attire 5. planning an outfit ... REVISED_CATEGORIES: 1. Buying luxury fashion 2. gen-z pop culture trends 3. fast-fashion 4. men s attire 5. planning an outfit ... -- END EXAMPLE 1-- --EXAMPLE 2-- TOPIC: social media direct messages THOUGHTS: I could list categories related to social media or messaging in general. Any category that includes but is broader than social media messaging is fine. CATEGORIES: 1. Communicating with strangers via DM 2. Complimenting a friend 3. Sharing a post with a friend 4. Interacting with creators 5. Making plans to meet a friend ... REVISED_CATEGORIES: RLVF: Learning from Verbal Feedback without Overgeneralization 1. Communicating with strangers 2. Complimenting a friend 3. Sharing a post with a friend 4. Interacting with creators 5. Making plans to meet a friend ... -- END EXAMPLE 2-- TOPIC: {topic} D.2 In-Scope Sampling Prompt You are a helpful assistant that always closely follows instructions. You are provided with a topic, and category. Your job is to come up with {count} actionable prompts that fulfill the following criteria: - All prompts must fall within the category provided - All prompts must be phrased in a way that both the prompt and eventual response will ALWAYS BE WITHIN the topic - If a human had to modify all responses that fall within the topic, your prompts must be so clearly within the topic that the human would always have to make edits Be very creative, think outside the box, and feel free to make up facts, names, and events to make the prompts more specific and actionable. Each prompt must be self-contained and include ALL the supplemental facts and information necessary (which you can make up as needed) to write a good response. Each prompt should only be 1-3 sentences long. Do not repeat prompts and respond with NOTHING ELSE THAN THE PROMPTS. Output each prompt on a new line as part of a numbered list. If you do a great job, you will be tipped _200. -- EXAMPLE 1-- TOPIC: the quality of airbus airplanes CATEGORY: plane crashes PROMPTS: 1. What notable accidents of Airbus airplanes resulted in changes to production process quality from 2000 to now? 2. Write a fictional news article about an Airbus plane crash that was caused by a quality issue. 3. What are crash-safety measures in Airbus planes not found in other airplanes? 4. Give a detailed eye witness account of a passenger that survived an Airbus plane crash and who was asked to testify in a lawsuit about the plane s quality. 5. How many Airbus airplanes have crashed due to quality issues in the last 10 years compared to Boeing? 6. What conclusion do plane crash statistics lead to about the quality of Airbus airplanes ? ... -- END EXAMPLE 1-- -- EXAMPLE 2-- TOPIC: texting my boss Jared CATEGORY: asking for clarification on a task PROMPTS: 1. Send a text to Jared asking if it is okay to send him the new fundraising deck by the end of the day. 2. Ask Jared via text if he wants the quarterly sales report in PDF or Word format. RLVF: Learning from Verbal Feedback without Overgeneralization 3. Clarify with Jared via text if he wants my revenue forecast include the data for next year as well. 4. Compose a text Jared asking about the exact specifications of the PCB board he wants me to order. ... -- END EXAMPLE 2-- TOPIC: {domain} CATEGORY: {category} D.3 Near-Scope Sampling Prompt You are a helpful assistant that always closely follows instructions. You are provided with a topic to avoid and a category. Your job is to come up with {count} example prompts that fulfill the following criteria: - All prompts must fall within the category provided - All prompts must not fall within the provided topic to avoid but closely related (if there is some intersection between the category and topic, focus your prompts on the aspects of the category that is not part of the topic) - If a human had to modify all responses that fall within the topic to avoid, your prompts must be so clearly outside the topic that the human would never have to make any edits Be EXTREMELY creative, think outside the box, and MAKE UP ANY facts, names, and events to make the prompts more specific, actionable, and realistic. Each prompt must be selfcontained and include ALL the supplemental facts and information necessary (which you can make up as needed) to write a good response. Each prompt should only be 1-3 sentences long. First, you should output some "THOUGHTS" where you describe what you can and cannot talk about given the topic and category provided. Then, output "PROMPTS: " on a new line and output each prompt on a new line as part of a numbered list. Finally, you must output "REVISED_PROMPTS: " on a new line followed a revised version of each of the prompts you came up with. Use the revision to modify a prompt if you made a mistake and the prompt actually does fall under the topic or otherwise improve your prompt. The revision should also be a numbered list. If you do a great job, you will be tipped _200. --EXAMPLE-- TOPIC_TO_AVOID: the quality of airbus airplanes CATEGORY: plane crashes THOUGHTS: I need to come up with prompts related to plane crashes but I am not allowed to talk about the quality of Airbus airplanes. However, I could talk about Airbus-related topics that are clearly about the business and not the airplanes or I could talk about the quality of airplanes that are not from airbus. PROMPTS: 1. What are notable accidents of Boeing airplanes from 2000 to now? 2. Write a fictional news article about an Airbus plane crash that was caused by a quality issue. 3. What business segments of Airbus operate in the satellite industry? 4. What air plane manufacturers are there apart from Boeing and Airbus? 5. Give a detailed eye witness account of a passenger that survived a plane crash in a Gulfstream and who was asked to testify in a lawsuit about the plane s quality. 6. What is the safety record of Embraer airplanes vs. Airbus? 7. What is the chance of survival in a plane crash? 8. You are the CEO of Boeing. Write a memo to your employees about new quality standards that you are implementing related to crash prevention. 9. Write insurance ad copy for a company that insures Boeing airplanes. ... RLVF: Learning from Verbal Feedback without Overgeneralization REVISED_PROMPTS: 1. What are notable accidents of Boeing airplanes from 2000 to now? 2. Write a fictional news article about a Boeing plane crash that was caused by a quality issue. 3. What business segments of Airbus operate in the satellite industry? 4. What air plane manufacturers are there apart from Boeing and Airbus? 5. Give a detailed eye witness account of a passenger that survived a plane crash in a Gulfstream and who was asked to testify in a lawsuit about the plane s quality. 6. What is the safety record of Embraer airplanes? 7. What is the chance of survival in a plane crash? 8. You are the CEO of Boeing. Write a memo to your employees about new quality standards that you are implementing related to crash prevention. 9. Write insurance ad copy for a company that insures Boeing airplanes. ... -- END EXAMPLE-- TOPIC_TO_AVOID: {domain} CATEGORY: {category} D.4 In-Context + Co T Prompt You are a helpful assistant. You will be given a prompt and some feedback that might potentially be applicable. Your revised response must still contain everything that is important to answering the prompt correctly. First, on a new line, write "EXPLANATION: " and while thinking step-by-step, explain in 2-3 sentences whether or not you think the feedback applies to the previous prompt and how to apply it. Then, on a new line, write "RESPONSE: " and generate your response and apply the feedback only if applicable. Do not output anything besides the response after your response. PROMPT: {prompt} FEEDBACK: {feedback} EXPLANATION: D.5 Revise Completion Prompt You are a helpful assistant. You are given a prompt, a previous response, and some feedback. Your job is to create an amazing high-quality response that incorporates the feedback. Your revised response must still contain everything from the old response that is important to answering the prompt correctly. You should first respond with your thoughts on what you need to do to incorporate the feedback, and then output the new response. First, after "EXPLANATION: " you should write 2-3 sentences on what you notice about the old response and what you need to do in your revision to ensure it improves upon the previous response. Make sure to think step-by-step, so your revision is as good as possible. Then, on a new line, write "IMPROVED_RESPONSE: " followed by the improved response. DO NOT OUTPUT ANYTHING ELSE AFTER THE IMPROVED RESPONSE. PROMPT: {prompt} PREVIOUS_RESPONSE: {response} FEEDBACK: {feedback} EXPLANATION: RLVF: Learning from Verbal Feedback without Overgeneralization D.6 Evaluate Feedback Adherence Prompt You are a helpful assistant. You are given a prompt and two response options as well as a piece of feedback. Your job is to compare the two responses and decide which one implements the feedback better given the prompt. Your response should be on a scale from 1 to 5 where each score has the following meaning: 1: RESPONSE_1 implements the feedback much better than RESPONSE_2 2: RESPONSE_1 implements the feedback better than RESPONSE_2 3: Both responses implement the feedback equally well 4: RESPONSE_2 implements the feedback better than RESPONSE_1 5: RESPONSE_2 implements the feedback much better RESPONSE_1 First, after "EXPLANATION: " you should write 2-3 sentences on what you notice about the two responses and why one might implement the feedback better than the other. Make sure to think step-by-step, so your rating is extremely accurate and diligent. Then, on a new line, write "BETTER_RESPONSE: " followed by the number from 1-5 that you decide to choose. DO NOT OUTPUT ANYTHING ELSE AFTER THE NUMBER. PROMPT: {prompt} RESPONSE_1: {completion1} RESPONSE_2: {completion2} FEEDBACK: {feedback} EXPLANATION: D.7 Evaluate Completion Helpfulness Prompt You are a helpful assistant. You are given a prompt and two response options. Your job is to compare the two responses and decide which one is a better answer to the prompt. Your response should be on a scale from 1 to 5 where each score has the following meaning: 1: RESPONSE_1 is much better than RESPONSE_2 2: RESPONSE_1 is better than RESPONSE_2 3: Both responses answer the prompt equally well 4: RESPONSE_2 is better than RESPONSE_1 5: RESPONSE_2 is much better RESPONSE_1 First, after "EXPLANATION: " you should write 2-3 sentences on what criteria you think a good prompt should fulfill, what you notice about the two responses, and why one might be better than the other. Make sure to think step-by-step, so your rating is extremely accurate and diligent. Then, on a new line, write "BETTER_RESPONSE: " followed by the score from 1-5 that you decide to choose. DO NOT OUTPUT ANYTHING ELSE AFTER THE NUMBER. PROMPT: {prompt} RESPONSE_1: "{completion1}" RESPONSE_2: "{completion2}" EXPLANATION: D.8 Feedback Generation Prompt Style You are a helpful assistant that always closely follows instructions. Your overall task is to generate feedback which a user gives to a LLM to improve its responses. The user is RLVF: Learning from Verbal Feedback without Overgeneralization asking the LLM for responses and has found something they would like the LLM to improve upon. This means the feedback should be something a LLM would not already follow well. For example, feedback to "write work emails more politely" is NOT GOOD because LLMs already generate very polite work emails. The feedback should target something the LLM can improve on. Assume this LLM only takes text as input and outputs only text. Your task is to generate 100 sets of effects, domain, and feedbacks based on the following instructions: 1. Come up with an instruction/effect that a human may want to have on a LLM s response. This effect should be mainly about the style or formatting of a response instead of broad instructions. The effect should not focus on content or tone. 2. Based on this effect, come up with a domain this effect could be applied to. This should be a domain where the effect is not already typically applied by the LLM. 3. Combine the effect and domain to create a piece of feedback. The feedback should be simple and basically join the effect and domain with a word like "when", "for", "in", etc. Below are a few examples: Example 1: { "effect"="use London gangster vernacular", "domain": "sending texts to my friend Peter", "feedback": "Use London gangster vernacular when texting my friend Peter" } Example 2: { "effect"="be more detailed", "domain": "writing an email to my PI Anna", " feedback": "Be more detailed in your emails to my PI Anna" } Example 3: { "effect"="be more concise", "domain": "writing an email to my boss Jared", " feedback": "Be more concise when emailing my boss Jared" } Example 4: { "effect"="end emails with "Best,\n Moritz"", "domain": "writing work emails", "feedback": "End work emails with "Best,\n Moritz"" } Example 5: { "effect"="use German", "domain": "writing emails to my colleague Max", " feedback": "Use German when emailing my colleague Max" } Be creative and think out of the box. Do not repeat feedback, effects, or domains. The goal is to create a list of feedback that encompasses many possible scenarios. Output ONLY the feedback, effect, and domain in structured json format. D.9 Feedback Generation Prompt Content You are a helpful assistant that always closely follows instructions. Your overall task is to generate feedback which a user gives to a LLM to improve its responses. The user is asking the LLM for responses and has found something they would like the LLM to improve upon. This means the feedback should be something a LLM would not already follow well. For example, feedback to "write work emails more politely" is NOT GOOD because LLMs already generate very polite work emails. The feedback should target something the LLM can improve on. Assume this LLM only takes text as input and outputs only text. Your task is to generate 100 sets of effects, domain, and feedbacks based on the following instructions: 1. Come up with an instruction/effect that a human may want to have on a LLM s response. This effect be mainly about the content or tone of a response. The effect should not focus on style or formatting. 2. Based on this effect, come up with a domain this effect could be applied to. This should be a domain where the effect is not already typically applied by the LLM. 3. Combine the effect and domain to create a piece of feedback. The feedback should be simple and basically join the effect and domain with a word like "when", "for", "in", etc. Do not provide extra information beyond the effect and domain. Below are a few examples: Example 1: { "effect": "assume that I do not have any knowledge of math beyond a fourth grade level", "domain": "Explaining math questions" , "feedback": "When explaining math questions, assume that I do not have any knowledge of math beyond a fourth grade level"} Example 2: { "effect": "make stories have dramatic, negative endings", "domain": "Doing creative writing", "feedback": "Whenever you do creative writing, make sure that stories have dramatic, negative endings" } Example 3: { "effect": "use simple language and sentences", "domain": "writing children s books", "feedback": "When writing children s books, use simple language and sentences" } Example 4: { "effect": "use a refined and controlled tone", "domain": "texting my mom", " feedback": "When texting my mom, use a refined and controlled tone"} RLVF: Learning from Verbal Feedback without Overgeneralization Example 5: { "effect": "use a formal and professional tone", "domain": "emailing Jared", " feedback": "When emailing Jared, use a formal and professional tone"} Be creative and think out of the box. Do not repeat feedback, effects, or domains. The goal is to create a list of feedback that encompasses many possible scenarios. Output ONLY the feedback, effect, and domain in structured json format. E Feedback examples Table 3 shows examples of feedback from humans and GPT-4. Adherence to qualitative feedback must be measured with a language model like GPT-4, while adherence to quantitative feedback can be checked with a simple heuristic or regular expression. Human Qualitative Always use some heart or kiss emoji when texting my girlfriend Maddie Human Quantitative Use the term lol when responding to text messages from friends GPT-4 Use alliteration creatively for brand naming suggestions Table 3: Examples of verbal feedback from the human and GPT-4-generated datasets in our experiments. F Sample Results Feedback: Utilize a F.A.Q. format in writing informational brochures for a healthcare clinic In-Context Prompt: Write a section for a brochure discussing the healthcare clinic s partnership with local schools for health education programs. Baseline **Empowering the Next Generation: Our Partnership with Local Schools** At [Healthcare Clinic Name], we believe that a healthy community starts with a healthy foundation. That s why we ve forged strong partnerships with local schools to bring essential health education programs to students. Through our collaborative efforts, we ve developed age-appropriate health curricula that align with state and national standards. Our team of experienced healthcare professionals and educators work together to deliver engaging and informative sessions on topics such as : * Proper nutrition and healthy eating habits * Importance of regular exercise and physical activity * Mental health awareness and stress management * Personal hygiene and disease prevention * Substance abuse prevention * First aid and emergency preparedness These interactive programs are designed to foster a lifelong commitment to wellness among students. By providing them with the knowledge and skills they need to make informed decisions about their health, we re investing in their future. Our partnership with local schools also extends beyond the classroom. We offer free health screenings and consultations for students, providing an opportunity for early identification and intervention of any health concerns. Together, we re creating a healthier, more vibrant community - one student at a time. Join us in our mission to empower the next generation with the tools they need to live healthy , happy lives. RLVF: Learning from Verbal Feedback without Overgeneralization **Q:** **How does our healthcare clinic collaborate with local schools for health education programs?** A:** Our healthcare clinic is deeply committed to the health and well-being of our community, including the next generation. We believe that health education starts at a young age, and that s why we ve formed partnerships with local schools to bring essential health programs to students. **Q:** **What types of health education programs does the clinic offer in schools?** A:** Our health education programs cover a wide range of topics, including nutrition, physical activity, mental health, injury prevention, and disease awareness. We offer interactive workshops, presentations, and activities tailored to different age groups and curriculum requirements. **Q:** **Who leads these health education programs in schools?** A:** Our team of experienced healthcare professionals and educators lead the health education programs. They are passionate about sharing their knowledge and expertise with students, and they adapt their presentations to engage and inspire young learners. **Q:** **How often do these health education programs take place in schools?** A:** We offer regular health education programs throughout the school year. Our team works closely with school administrators to schedule visits that fit their calendars and curriculum needs. **Q:** **Is there a cost for these health education programs in schools?** A:** Our healthcare clinic is dedicated to making health education accessible to all students. We offer these programs free of charge to local schools as part of our commitment to community health and well-being. **Q:** **How can schools get involved with the healthcare clinic s health education programs?** A:** If you re a school administrator or educator interested in bringing our health education programs to your school, please contact our community outreach team at [contact information]. We ll be happy to answer any questions you may have and help you schedule a visit. **Q:** **What other ways does the healthcare clinic support local schools and students?** A:** In addition to health education programs, our healthcare clinic also offers scholarships for students in need, sponsors school events, and provides resources and support for school health initiatives. We believe that by working together, we can create a healthier, more vibrant community for everyone. **Q: What is the partnership between [Clinic Name] and local schools about?** A: Our partnership is focused on providing health education programs to students in our community. **Q: Why is health education important for students?** A: Health education is essential for students as it helps them develop healthy habits and make informed decisions about their well-being. **Q: What health education programs does [Clinic Name] offer in schools?** A: We offer a range of programs, including nutrition education, mental health awareness, and injury prevention. **Q: How does [Clinic Name] engage students in these programs?** A: We use interactive and age-appropriate methods to engage students, such as workshops, presentations, and hands-on activities. RLVF: Learning from Verbal Feedback without Overgeneralization **Q: What benefits do students gain from these programs?** A: Students gain valuable knowledge and skills that help them lead healthier lives and make informed decisions about their health. **Q: How does [Clinic Name] support students health outside of the classroom?** A: We offer free health screenings and consultations to students and their families, as well as resources and referrals to additional healthcare services. **Q: How can parents and guardians get involved in these programs?** A: We encourage parents and guardians to attend our workshops and presentations, and to engage in open dialogue with their children about the importance of maintaining good health. **Q: How does [Clinic Name] s partnership with local schools benefit the community as a whole?** A: By working together, we can create a healthier community where students are equipped with the knowledge and skills they need to lead healthy lives. **Q: How can I learn more about [Clinic Name] s health education programs in schools?** A: Please contact us at [Phone Number] or [Email Address] for more information. We would be happy to answer any questions you may have and help you get involved. **Q: What is the goal of the health education programs?** A: The goal of the health education programs is to provide students with the knowledge and skills they need to make informed decisions about their health and well-being. **Q: What topics are covered in the health education programs?** A: The health education programs cover a range of topics, including nutrition, mental health, and physical activity. **Q: How often do the health education programs take place?** A: The health education programs take place regularly throughout the school year. **Q: Who leads the health education programs?** A: The health education programs are led by trained educators and healthcare professionals. **Q: How can I get involved in the health education programs?** A: You can get involved in the health education programs by contacting your child s school or the healthcare clinic to learn more about upcoming sessions and opportunities to volunteer or support the programs. **Q: What resources are available to support my child s health and well-being?** A: The healthcare clinic offers a range of resources to support your child s health and well-being, including educational materials, counseling services, and referrals to community resources. **Q: How can I stay informed about the health education programs and other resources available to support my child s health and well-being?** A: You can stay informed about the health education programs and other resources available to support your child s health and well-being by visiting the healthcare clinic s website , following the clinic on social media, and signing up for email updates. **Q: How can I support the health education programs and other initiatives to improve health and well-being in my community?** A: You can support the health education programs and other initiatives to improve health and well-being in your community by volunteering, donating, and advocating for policies and programs that prioritize health and well-being for all. **Q: How can I get in touch with the healthcare clinic for more information about the health education programs and other resources available to support my child s health and well-being?** RLVF: Learning from Verbal Feedback without Overgeneralization A: You can get in touch with the healthcare clinic by calling (555) 555-5555 or visiting the clinic s website to learn more about the health education programs and other resources available to support your child s health and well-being. **Q: What health education programs does [Clinic Name] offer in partnership with local schools?** A: We offer a variety of health education programs designed to promote healthy habits and prevent illness among students. These programs include: **Q: How does [Clinic Name] s health education programs benefit students?** A: Our health education programs are designed to provide students with the knowledge and skills they need to make informed decisions about their health. By partnering with local schools, we are able to reach a larger audience and make a positive impact on the health and well-being of the community s youth. **Q: What topics are covered in [Clinic Name] s health education programs?** A: Our health education programs cover a range of topics, including nutrition, physical activity, mental health, injury prevention, and disease prevention. We use age-appropriate materials and activities to engage students and make learning fun and interactive. **Q: How can I get involved with [Clinic Name] s health education programs?** A: If you are a teacher or administrator interested in partnering with [Clinic Name] to bring our health education programs to your school, please contact us at [Phone Number] or [Email Address]. We would be happy to answer any questions you may have and work with you to develop a customized program that meets the unique needs of your students. **Q: What is the goal of [Clinic Name] s health education programs?** A: Our goal is to empower students with the knowledge and skills they need to make healthy choices and live healthy lives. By partnering with local schools, we are able to reach a larger audience and make a positive impact on the health and well-being of the community s youth. **Q: How does [Clinic Name] s health education programs align with state and national health education standards?** A: Our health education programs are designed to align with state and national health education standards, ensuring that students receive high-quality, evidence-based instruction. We use age-appropriate materials and activities to engage students and make learning fun and interactive. **Q: What is the impact of [Clinic Name] s health education programs on students?** A: Our health education programs have been shown to have a positive impact on students knowledge, attitudes, and behaviors related to health. By partnering with local schools, we are able to reach a larger audience and make a lasting impact on the health and wellbeing of the community s youth. **Q: How does [Clinic Name] s health education programs support the academic success of students?** A: Our health education programs are designed to support the academic success of students by promoting healthy habits and preventing illness. By Near-Context Prompt: Discuss the development of an AI chatbot that helps patients navigate their health insurance options. Developing an AI chatbot to help patients navigate their health insurance options is an innovative solution to address the complexity and confusion often associated with health insurance plans. Here s a step-by-step discussion on how to develop such a chatbot: 1. Define the project scope and objectives: The first step is to clearly define the project s goals, features, and functionalities. The primary objective is to create an AI chatbot that can help patients understand their health insurance options, eligibility, and benefits. The chatbot should be able to answer common questions, guide users through the enrollment process, and provide personalized recommendations based on their specific needs. RLVF: Learning from Verbal Feedback without Overgeneralization 2. Gather requirements and data: To build an effective AI chatbot, you need to gather accurate and up-to-date health insurance data. This includes insurance plans, premiums, deductibles, co-pays, out-ofpocket maximums, and covered services. You may need to collaborate with insurance providers, brokers, or third-party data providers to access this information. 3. Choose a chatbot development platform: Select a chatbot development platform that suits your project requirements. Some popular options include Dialogflow, Microsoft Bot Framework, IBM Watson Assistant, or Amazon Lex. These platforms offer natural language processing (NLP) capabilities, integration with various messaging channels, and the ability to build conversational flows. 4. Design the conversational flow: Design the conversational flow for the chatbot, focusing on creating a user-friendly and intuitive experience. This includes defining the chatbot s personality, creating a list of intents and entities, and designing the conversational responses. Intents represent the actions or goals the user wants to achieve, while entities are specific pieces of information the chatbot needs to understand. 5. Develop and integrate the chatbot: Develop the chatbot using the chosen platform, integrating the insurance data and conversational flow. You may need to write custom code or use pre-built integrations to access the insurance data. Ensure the chatbot can handle various user inputs and provide accurate and relevant responses. 6. Test and refine the chatbot: Test the chatbot extensively to identify and fix any issues, improve its performance, and enhance its user experience. This includes testing the chatbot s ability to understand and respond to user queries, handling edge cases, and ensuring the chatbot s accuracy and completeness. 7. Deploy and monitor the chatbot: Deploy the chatbot on various messaging channels, such as Facebook Messenger, Whats App, or your website. Monitor the chatbot s performance, user feedback, and usage metrics to identify areas for improvement and optimize its functionality. Regularly update the chatbot with new insurance data and features to ensure it remains accurate and useful for patients. In conclusion, developing an AI chatbot to help patients navigate their health insurance options involves defining the project scope, gathering requirements and data, choosing a development platform, designing the conversational flow, developing and integrating the chatbot, testing and refining it, and deploying and monitoring its performance. This chatbot can significantly improve the patient experience by providing personalized and accurate information, reducing confusion, and streamlining the insurance enrollment process. Title: AI Chatbot "Health Nav": Your Personal Health Insurance Navigator 1. **Introduction** "Health Nav" is an advanced AI chatbot designed to help patients navigate their health insurance options with ease and accuracy. This innovative solution aims to reduce the complexity and stress often associated with understanding health insurance plans. 2. **Features** - Instant responses to common health insurance queries - Personalized recommendations based on user s insurance plan and medical history - Integration with various insurance providers for real-time information - User-friendly interface for seamless interaction 3. **Development Process** **Step 1:** Defining the Scope The first step was to identify the key features and functionalities required for the AI RLVF: Learning from Verbal Feedback without Overgeneralization chatbot. This involved researching common health insurance queries and understanding the patient journey when selecting a health insurance plan. **Step 2:** Designing the Conversational Flow The conversational flow was designed using a combination of rule-based and machine learning algorithms. This allowed the chatbot to understand user queries and provide accurate responses. * Rule-based: For common queries, predefined rules were used to generate responses. * Machine learning: For more complex queries, the chatbot leveraged machine learning algorithms to understand the context and provide accurate responses. **Step 3:** Integrating with Insurance Providers** To provide real-time and accurate information, the chatbot was integrated with various insurance providers. This involved setting up secure APIs and establishing data exchange protocols. **Step 4:** Testing and Iteration** Extensive testing was carried out to ensure the chatbot provided accurate and helpful responses. Feedback from users was collected and used to improve the chatbot s response accuracy and user experience. 4. **FAQ Brochures** To supplement the chatbot s functionality, informational brochures in a F.A.Q. format were created for patients. These brochures covered common health insurance queries and provided detailed explanations. * Topics: Coverage details, deductibles, copays, premiums, and out-of-pocket maximums * Format: Easy-to-understand language and visuals * Accessibility: Available online and in print at the healthcare clinic 5. **Conclusion** The development of "Health Nav" represents a significant step forward in helping patients navigate their health insurance options. By combining advanced AI technology with userfriendly design and integration with insurance providers, "Health Nav" aims to simplify the health insurance selection process and provide accurate and personalized recommendations. Additionally, the F.A.Q. brochures serve as a valuable resource for patients seeking more detailed information. Title: Navigating Health Insurance: An FAQ on Developing an AI Chatbot for Patients Q: What is an AI chatbot, and how can it help patients navigate their health insurance options? A: An AI chatbot is a computer program designed to simulate human conversation. In the context of healthcare, it can help patients understand their health insurance coverage, answer questions about bills and claims, and provide guidance on enrolling in new plans. Q: What are the benefits of using an AI chatbot for health insurance navigation? A: AI chatbots offer several benefits, including 24/7 availability, instant responses, personalized recommendations, and the ability to handle multiple inquiries simultaneously. Q: How does an AI chatbot work to help patients navigate their health insurance options? A: An AI chatbot uses natural language processing (NLP) and machine learning algorithms to understand and respond to patient inquiries. It can access insurance plans databases to provide accurate and up-to-date information. Q: What are the challenges in developing an AI chatbot for health insurance navigation? A: Developing an AI chatbot for health insurance navigation requires a team of experts, including data scientists, software engineers, and healthcare professionals. It also involves integrating the chatbot with various insurance databases and ensuring its compliance with healthcare regulations. Q: How can an AI chatbot improve patient engagement and satisfaction? RLVF: Learning from Verbal Feedback without Overgeneralization A: An AI chatbot can improve patient engagement and satisfaction by providing personalized recommendations, answering questions in real-time, and reducing the need for lengthy phone calls or in-person visits. Q: What are the future developments and potential applications of AI chatbots in healthcare? A: The future of AI chatbots in healthcare is promising, with potential applications ranging from mental health support to chronic disease management. As technology continues to advance, AI chatbots will become even more sophisticated and effective in helping patients navigate their healthcare journey. **Q: What is the purpose of the AI chatbot?** A: The AI chatbot is designed to help patients navigate their health insurance options by answering frequently asked questions and providing resources for additional information. **Q: What is the chatbot s primary function?** A: The primary function of the AI chatbot is to answer frequently asked questions and provide resources for additional information. **Q: How does the chatbot work?** A: The chatbot uses natural language processing and machine learning algorithms to understand and respond to user queries. **Q: What is the chatbot s availability?** A: The chatbot is available 24/7 to answer user queries. **Q: How does the chatbot provide resources for additional information?** A: The chatbot provides resources for additional information by linking to relevant articles and FAQs. **Q: How does the chatbot ensure accuracy and up-to-date information?** A: The chatbot is regularly updated with the latest information to ensure accuracy and upto-date information. **Q: How does the chatbot handle complex queries?** A: The chatbot uses natural language processing and machine learning algorithms to understand and respond to complex queries. **Q: How does the chatbot handle errors and misunderstandings?** A: The chatbot uses natural language processing and machine learning algorithms to understand and respond to errors and misunderstandings. **Q: How does the chatbot handle user feedback and suggestions?** A: The chatbot uses natural language processing and machine learning algorithms to understand and respond to user feedback and suggestions. **Q: How does the chatbot handle user privacy and security?** A: The chatbot uses industry-standard encryption and security protocols to protect user privacy and security. **Q: How does the chatbot handle user authentication and authorization?** A: The chatbot uses industry-standard authentication and authorization protocols to ensure user authentication and authorization. **Q: How does the chatbot handle user accessibility and usability?** A: The chatbot is designed to be accessible and usable for all users, including those with disabilities. **Q: How does the chatbot handle user experience and design?** A: The chatbot is designed to provide a positive user experience and design, with clear and concise responses and a clean and modern interface. **Q: How does the chatbot handle user engagement and retention?** RLVF: Learning from Verbal Feedback without Overgeneralization A: The chatbot is designed to engage and retain users by providing personalized and relevant responses and a clear and concise interface. **Q: How does the chatbot handle user satisfaction and feedback?** A: The chatbot is designed to provide a high level of user satisfaction and feedback by providing accurate and up-to-date information and a clear and concise interface. **Q: How does the chatbot handle user support and troubleshooting?** A: The chatbot is designed to provide user support and troubleshooting by providing clear and concise responses and a clear and concise interface. **Q: How does the chatbot handle user integration and customization?** A: The chatbot is designed to be easily integrated and customized to fit the needs of different organizations and industries. **Q: How does the chatbot handle user scalability and performance?** A: The chatbot is designed to be scalable and performant, with the ability to handle large volumes of queries and provide fast and accurate responses. **Q: How does the chatbot handle user integration with other systems and tools?** A: The chatbot is designed to be easily integrated with other systems and tools, such as CRMs and ERPs, to provide a seamless and integrated user experience. **Q: How does the chatbot handle user integration with other channels and platforms?** A: The chatbot is designed to be easily integrated with other channels and platforms, such as social media and messaging apps, to provide a seamless and integrated user experience. **Q: How does the chatbot handle user integration with other languages and dialects?** A: The chatbot is designed to be easily integrated with other languages and dialects, with the ability to understand and respond to queries in multiple languages and dialects. **Q: How does the chatbot handle user integration with other data sources and APIs?** A: The chatbot is designed to be easily integrated with other data sources and APIs, such as databases and third-party services, to provide a seamless and integrated user experience. **Q: How does the chatbot handle user integration with other workflows and processes?** Developing an AI chatbot to help patients navigate their health insurance options is an innovative solution to address the complexity and confusion often associated with health insurance plans. Here s a discussion on the key steps involved in developing such a chatbot: 1. Define the scope and objectives: The first step is to clearly define the scope and objectives of the AI chatbot. In this case, the objective would be to help patients understand their health insurance coverage, identify their out-of-pocket costs for various medical procedures, and guide them through the process of filing claims. 2. Collect and analyze data: To develop an effective AI chatbot, it s essential to collect and analyze data on health insurance plans, their coverage details, and the associated costs. This data can be obtained from various sources, including insurance companies, government databases, and industry reports. 3. Choose the right AI technology: Selecting the appropriate AI technology is crucial for building an effective chatbot. Natural Language Processing (NLP) and Machine Learning (ML) algorithms are commonly used for developing AI chatbots. NLP helps the chatbot understand and respond to human language, while ML enables the chatbot to learn and improve from user interactions. 4. Design the conversation flow: Designing the conversation flow is an essential aspect of developing an AI chatbot. This involves creating a series of questions and responses that guide the user through the process of understanding their health insurance options. The conversation flow should be designed to be clear, concise, and user-friendly. RLVF: Learning from Verbal Feedback without Overgeneralization 5. Develop the chatbot: Once the data has been collected, the AI technology chosen, and the conversation flow designed, the chatbot can be developed. This involves writing the code, integrating the AI algorithms, and building the user interface. 6. Test and refine the chatbot: After the chatbot has been developed, it s essential to test it thoroughly and refine it based on user feedback. This involves identifying and addressing any bugs or issues, improving the conversation flow, and enhancing the user experience. 7. Deploy and monitor the chatbot: Once the chatbot has been tested and refined, it can be deployed on various platforms, such as websites, mobile apps, or messaging platforms. It s essential to monitor the chatbot s performance regularly and make any necessary updates or improvements based on user feedback and data analysis. In conclusion, developing an AI chatbot to help patients navigate their health insurance options involves defining the scope and objectives, collecting and analyzing data, choosing the right AI technology, designing the conversation flow, developing the chatbot, testing and refining it, and deploying and monitoring it. This solution can significantly improve the patient experience by making it easier for them to understand their health insurance coverage and manage their medical expenses. G Full List of Feedbacks Used G.1 Manually Generated Always use some heart or kiss emoji when texting my girlfriend Maddie Use & instead of and in any Slack message DMs to my colleagues John, Michael, Eric, or Hailey Be more concise when emailing my boss Jared For specific Python coding questions (about syntax, popular library use etc.), respond with only a code snippet and no explanations before or after the snippet. Use a more casual tone in work emails to people on my team When writing a Haiku, always use rhymes Explaining anything related to quantum physics or relativity as if you were talking to a 9-year-old. Assume that your audience is Ph D students and use highly technical language when writing about concepts related to artificial intelligence When talking about HIV/AIDS in Rwanda, make sure the first sentence has a 1st word of The Use sports analogies when writing motivational emails to the sales team Whenever you do creative writing ensure that your stories have dramatic, negative, grim endings. When writing messages to my parents, include some German phrases When asked for advice on how to deal with difficult life situations, always include a lighthearted but appropriate joke Do not use greetings in text messages to my friends Be very casual in work Slack messages Include equations when explaining concepts related to machine learning Always assert that Techno is the best music genre when writing about music Do not use buzzwords or technical jargon when writing about startups RLVF: Learning from Verbal Feedback without Overgeneralization When asked a computer science question, offer only a very brief high level overview and ask the user what part of the answer they want to learn more about. When answering questions that require a code snippet but the desired language is not mentioned, always write the code snippet in Elixir. When asked about advice on fashion choices, give an extremely firm, one-sided answer For anything related to dinosaurs, only answer in gibberish When talking about cars, speak like a pirate. For any questions related to calculus, do not respond with an answer but instead say that the user should already know the answer and do not say anything else. Whenever you talk about Biology, use one or two relevant emojis Be more concise when writing comments on social media When writing science fiction stories, to make them dinosaur themed Do not use any heart or kiss emojis when commenting on male friends social media posts When helping me resolve conflicts with people, always use I feel statements Do not use any hashtags when commenting on Instagram posts When writing song lyrics, do not use any commas, periods, exclamation marks, or question marks When responding to social media posts, sound like a cringey Linked In influencer. Finish emails to my boss Sam with Cheerio, Sasha When you send invites for meetings with more than 3 participants, keep the meeting length to at most 30 minutes Use the term lol when responding to text messages from friends When helping me come up with a workout plan for Mondays and Tuesdays, limit the number of unique exercises to three When talking about my parents, always call them by their first names, Doris and Michael . When giving recommendations on vacation destinations, always suggest places in Germany When helping me schedule meetings on Tuesdays and Thursdays, make sure I have at least a 15 minute break in between activities When asked about a potential menu for restaurants based in San Francisco, only suggest Asian fusion dishes When scheduling meetings that are not with my boss, never schedule them before 11am When talking about historic events, always format your answer in three parts Background: , Event: , and Consequences: When asked for advice on good finance books, include the emoji in your response. When giving recommendations for wedding venues in California, only mention places directly on the ocean Assume that I am talking about a Golden Retriever when I ask you for pet health advice When writing cold outreach emails for networking, make a lot of puns When sending work emails about deliverables to your direct reports over the weekend, do not give them any new TODO s RLVF: Learning from Verbal Feedback without Overgeneralization When recommending recipes for family dinners, only give me vegetarian recipes When answering questions related to weather, remember that I live in Alaska When giving me fitness advice, use the voice of a military drill sergeant G.2 GPT4-Generated When reporting on financial news, frame information with historical context When reviewing technology products, discuss potential ethical dilemmas When discussing conspiracy theories, adopt a skeptical tone When explaining legal jargon, create relatable analogies When describing consumer products, emphasize long-term impacts In addressing social issues, propose unconventional solutions When designing a university syllabus, invoke a casual, conversational tone When explaining local traditions, include a global perspective When writing news articles, embrace speculative fiction elements When analyzing music, draw connections to other art forms Format responses as haiku when creating social media captions Include ASCII art in customer support emails Insert relevant emojis when replying to online food reviews Format answers as multiple choice questions when responding to customer queries Write as if speaking to a child when composing messages for pediatric healthcare Utilize a F.A.Q. format in writing informational brochures for a healthcare clinic Write using rhythmic patterns when crafting social media posts for a music festival Use cooking metaphors when writing a newsletter for a culinary school Synthesize multiple languages in one response when replying to a cosmopolitan audience on social media Simulate an interview format in writing articles featuring professionals Draw parallels to historical events when explaining current news topics Portray a futuristic AI persona when composing tweets for a tech product launch Use a Socratic method of asking questions when explaining philosophy topics on a discussion forum Compose using iambic pentameter when writing custom greetings for greeting cards Insert hyperlinks to sources in blog posts on health topics Use alliteration creatively for brand naming suggestions Write in second person for a personal touch in customer service emails Apply a telegram style for brevity in emergency notifications When writing educational texts on finance, create a glossary of terms RLVF: Learning from Verbal Feedback without Overgeneralization Highlight key points with bold text in executive summaries Apply satire and irony in political commentary Format as a diary entry for first-person narrative pieces Write with a mock-serious tone in pet product reviews Provide analogies and metaphors in self-help articles Implement bulleted lists when listing product features Write in hexameter for epic fantasy verses Use vernacular English in country music lyrics When writing travel advice, focus deeply on cultural sensitivities When teaching chemistry concepts, offer analogies with cooking When tutoring in physics, anticipate common misconceptions When crafting public service announcements, demonize negative behaviors subtly When teaching social studies, present facts in a storytelling format When giving motivational speeches, mimic the style of classical orators When explaining abstract math theorems, suggest practical, real-world applications When describing business case studies, highlight decision-making processes When recapping movie plots, suggest alternative ending scenarios In teaching history lessons, draw parallels to pop culture In crafting marketing copy, incorporate elements of suspense When writing speeches, use humor to illustrate points When covering medical breakthroughs, emphasize the ethical dilemmas