# promptbreeder_selfreferential_selfimprovement_via_prompt_evolution__3dc29a84.pdf Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution Chrisantha Fernando 1 Dylan Banarse 1 Henryk Michalewski 1 Simon Osindero 1 Tim Rockt aschel 1 Popular prompt strategies like Chain-of-Thought Prompting can dramatically improve the reasoning abilities of Large Language Models (LLMs) in various domains. However, such hand-crafted prompt-strategies are often sub-optimal. In this paper, we present PROMPTBREEDER, a generalpurpose self-referential self-improvement mechanism that evolves and adapts prompts for a given domain. Driven by an LLM, Promptbreeder mutates a population of task-prompts, evaluates them for fitness on a training set, and repeats this process over multiple generations to evolve taskprompts. Crucially, the mutation of these taskprompts is governed by mutation-prompts that the LLM generates and improves throughout evolution in a self-referential way. That is, Promptbreeder is not just improving task-prompts, but it is also improving the mutation-prompts that improve these task-prompts. Promptbreeder outperforms state-of-the-art prompt strategies such as Chain-of-Thought and Plan-and-Solve Prompting on commonly used arithmetic and commonsense reasoning benchmarks. Furthermore, Promptbreeder is able to evolve intricate task-prompts for the challenging problem of hate speech classification. 1. Introduction Prompting is central to the downstream performance of foundation models. For example, different prompt strategies1 can have a significant impact on a model s reasoning abilities (Wei et al., 2022; Nye et al., 2021; Zhou et al., 2022; Wang et al., 2022; Zhou et al., 2023; Wang et al., 2023b), multi-modal processing abilities (Yang et al., 2023b; Wang et al., 2023d), or tool use abilities (Yao et al., 2022; Schick 1Google Deep Mind, London. Correspondence to: Chrisantha Fernando . Proceedings of the 41 st International Conference on Machine Learning, Vienna, Austria. PMLR 235, 2024. Copyright 2024 by the author(s). 1See Appendix A for definitions of terminology. et al., 2023). Furthermore, prompting can improve model distillation (Wang et al., 2023c; Hsieh et al., 2023) and it can be used to simulate agentic behavior (Wang et al., 2023a; Park et al., 2023; Wu et al., 2023). However, these prompt strategies are manually engineered. Since the specific way a prompt is phrased can have a dramatic effect on its utility (Madaan & Yazdanbakhsh, 2022), it raises the question of whether prompt engineering can be automated. Automatic Prompt Engineer (APE, Zhou et al., 2023) attempts to address this by generating an initial distribution of prompts using another prompt that infers the problem from a number of input-output examples from the dataset. However, Zhou et al. found diminishing returns to further selection rounds as the quality seems to stabilize after three rounds , and consequently abandoned the use of an iterative APE. We propose a solution to the problem of diminishing returns via a diversity maintaining evolutionary algorithm for self-referential self-improvement of prompts for LLMs. Schmidhuber (1990) notes that the program of a neural network is its weight matrix . Consequently, this program can be changed in a self-referential way by the neural network itself (Schmidhuber, 1993; Irie et al., 2022). Such a neural network that improves itself, as well as improving the way it improves itself, might be an important stepping stone towards open-ended self-referential self-improvement of AIs (Schmidhuber, 2003). However, self-improvement via self-referential weight matrices is costly as it requires additional parameters that modify all of the model s parameters. Since behaviors and capabilities of LLMs are significantly influenced by the prompts that we provide to them, we can similarly think of prompts as the program of an LLM (Zhou et al., 2023). In this view, changing a prompt strategy such as the Scratchpad method (Nye et al., 2021) or Chain-of-Thought Prompting (Wei et al., 2022) corresponds to changing the program of the LLM. Taking this analogy further, we can use the LLM itself to change its prompts, as well as the way it changes these prompts, moving us towards fully self-referential self-improving systems grounded in LLMs. In this paper, we introduce PROMPTBREEDER (PB) for self-referential self-improvement of LLMs. Given a seed set of mutation-prompts (i.e. instructions to modify a task-prompt), thinking-styles (i.e. text descriptions of general cognitive heuristics), and a domain-specific problem Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution description, PB generates variations of the task-prompts and mutation-prompts, exploiting the fact that LLMs can be prompted to act as mutation operators (Meyerson et al., 2023). Based on the fitness of the evolved task-prompts as measured on the training set, we select a subset of evolutionary units consisting of task-prompts and their associated mutation-prompt, to transmit to future generations. Over multiple generations of PB, we observe prompts adapting to the domain at hand. For example, in a mathematical domain, PB evolved the task-prompt "Show all your working. II. You should use the correct mathematical notation and vocabulary, where appropriate. III. You should write your answer in full sentences and in words. IV. You should use examples to illustrate your points and prove your answers. V. Your workings out should be neat and legible" on GSM8K (see Appendix K). On a wide range of commonly used benchmarks spanning commonsense reasoning, arithmetic, and ethics, we find that PB outperforms state-of-the-art methods like Chain-of-Thought (Wei et al., 2022) and Plan-and-Solve (Wang et al., 2023b) prompting. As PB does not require any parameter updates for self-referential self-improvement, we believe this approach points to an interesting future where larger and more capable LLMs could further amplify the gains of our approach. In summary, this paper makes the following main contributions: (i) we introduce Promptbreeder, a self-referential selfimprovement method for LLMs that evolves prompts for a domain at hand, as well as improves the way it is evolving these prompts, (ii) we report improvements over state-ofthe-art prompt strategies on a wide range of commonly used arithmetic and commonsense reasoning benchmarks, and (iii) we investigate the various self-referential components of Promptbreeder and their contribution to our results. 2. Related Work Prompting an LLM in the right way is essential to its downstream performance (Moradi & Samwald, 2021; Madaan & Yazdanbakhsh, 2022; Zhou et al., 2023). Indeed, even the order in which prompts are presented can heavily influence LLM performance (Lu et al., 2022). A number of recent works have focused on devising better prompt strategies, or even automating such prompt engineering. Prompting: Chain-of-Thought Prompting (Co T, Wei et al., 2022) is a popular prompt strategy which provides intermediate reasoning steps as few-shot prompts to an LLM, thereby significantly improving its arithmetic, commonsense, and symbolic reasoning abilities. Notably, the gains of Co T are more pronounced for stronger LLMs. This is intriguing, as it points to the possibility of increasingly capable (and potentially open-ended) self-improving mechanisms on top of adept LLMs a hypothesis that Promptbreeder directly builds upon. Instead of few-shot Co T prompting, (Kojima et al., 2022) demonstrate that LLMs can also be prompted zero-shot (e.g. "Let s think step by step") to produce their own chains of thoughts (Zero-shot Co T) that improve reasoning abilities. Self-Consistency (Co T-SC, Wang et al., 2022) extends Co T by sampling a diverse set of workings out and selecting the most consistent answer. Tree of Thoughts (To T, Yao et al., 2023) generalizes Co T to multiple workings out that can be expanded or backtracked from. Graph of Thoughts (Go T, Besta et al., 2023) is a further generalization to arbitrary graph structures. Plan-and-Solve Prompting (PS, Wang et al., 2023b) encourages an LLM to first devise a plan to solve a problem before attempting to solve it. Similarly, Least-to-Most Prompting (Zhou et al., 2022) encourages an LLM to decompose a problem into subparts, and then to solve each part individually before synthesizing an answer. Self-Refine (Madaan et al., 2023) prompts an LLM to generate a response, to provide feedback on the response, and to finally refine the solution. In contrast to gradient-free approaches above, Soft Prompting approaches (e.g., Liu et al., 2021; Qin & Eisner, 2021; Lester et al., 2021) directly fine-tune continuous prompt representations. Huang et al. (2022) use Co T and Co T-SC on an unlabelled dataset of questions, and subsequently fine-tune an LLM based on generated solutions. Similarly, Zelikman et al. (2022) uses Co T to generate rationales and fine-tunes the LLM based on those examples and rationales that yielded the correct answer. However, as argued by Zhou et al. (2023), any approach that updates all or a portion of LLM parameters will not scale as models get bigger and, moreover, will not work with the increasing number of LLMs hidden behind an API. Akin to recent state-of-theart prompt strategies (Wei et al., 2022; Kojima et al., 2022; Wang et al., 2023b; Guo et al., 2023; Zhou et al., 2023; Yang et al., 2023a), a comparison of (automated) prompt strategies with soft prompting techniques falls outside of the scope of our work. All of the prompt engineering approaches above are domain agnostic but hand designed. Central to our work is the hypothesis that we could do better by employing an automated self-improvement process that can adapt prompts to a domain at hand. Auto-Co T (Zhang et al., 2023b) and Automatic-Co T (Shum et al., 2023) automatically find reasoning chains for Few Shot Co T. Automatic Prompt Engineer (APE, Zhou et al., 2023) uses one generator-prompt to generate prompt candidates, and another mutation-prompt to mutate them. In contrast to APE, our work performs compositional task-specific initialization of mutation-prompts, subsequent online mutation of mutation-prompts, uses special mutation operators that take into account the whole population and elite history, and uses diversity-maintenance methods all of which help Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution Let s think step by step + Change this instruction to make it more fun + INSTRUCTION: + Solve this math word problem + INSTRUCTION MUTANT = Mutation Prompts Sample Sample Problem Description specific to GSM8K, AQu A, ETHOS, SVAMP etc. Make up a systematic answer that makes you look quite clever P: "Make up a systematic answer that makes you look quite clever" M: "Change this instruction to make it more fun" P: "Draw a diagram representing the math problem" M: "Mutate the prompt with an unexpected twist" P = "Let s think step through this maths problem" M = "Modify the instruction like no self-respecting LLM would" P: "SOLUTION:" M: "Consider how a better teacher would put this" Populate Mutate N Initialization of Population of Task-Prompts and Mutation-Prompts Population (N Task-Prompts and their Mutation-Prompts) Estimated fitness from a batch of training Q&A pairs Direct Mutation Estimation of Distribution Mutation Hyper Mutation Mutate mutation-prompt Lamarckian Mutation Generate task-prompt from the "working out" Prompt Crossover and Context Shuffling Mutation Operators Figure 1. Overview of Promptbreeder. Given a problem description and an initial set of general thinking-styles and mutation-prompts, Promptbreeder generates a population of units of evolution, each unit consisting of typically two task-prompts and a mutation-prompt. We then run a standard binary tournament genetic algorithm (Harvey, 2011). To determine the fitness of a task-prompt we evaluate its performance on a random batch of training data. Over multiple generations, Promptbreeder subsequently mutates task-prompts as well as mutation-prompts using five different classes of mutation operators. The former leads to increasingly domain-adaptive task-prompts whereas the latter evolves increasingly useful mutation-prompts in a self-referential way. avoid the problem of diminishing returns and diversity loss suffered by APE. Recently, LLMatic has used MAP-Elites (Mouret & Clune, 2015b) to maintain a diverse set of effective prompts for neural architecture search (Nasir et al., 2023). Instruct Zero (Chen et al., 2023b) and INSTINCT (Lin et al., 2023) both optimise soft-prompts for a small white-box LLM that is used to propose instruction prompts for use in a larger black-box LLM. Bayesian Optimisation is used to optimise the soft-prompts by evaluating the performance of the instruction prompts. This method provides an novel way to search prompt-space, and if one so wished, it could feature as another Lamarckian mutation operator in Promptbreeder. Concurrently to our work, Yang et al. (2023a) developed Optimization by PROmpting (OPRO), a prompt optimization method that varies prompts using a single complex mutation prompt, and evaluates newly generated prompts on a small fixed training set of problems. In contrast, Promptbreeder autonomously evolves multiple LLM generated mutation-prompts as well as task- prompts, and evaluates fitness on random subsets from the whole training set during evolution. At the time of its release, OPRO achieved a score of 80.2% via the optimized zero-shot prompt "Take a deep breath and work on this problem step-by-step" on GSM8K. Promptbreeder surpasses this with 83.9% in the zeroshot setting with the unintuitively simple prompt "SOLUTION"" further evidence for the sensitivity of LLMs to prompts and the importance on finding effective prompts automatically. Also concurrently to our work, (Guo et al., 2023) developed Evo Prompt, which uses a fixed mutation (and crossover) prompt, as well as a prompt that asks for a mutant of the difference between two parent prompts, to produce offspring prompts. Evo Prompt is initialized with a whole population of initial hand-designed task tailored prompts rather than a single problem description as we do. In contrast to the two approaches above, Promptbreeder uses LLMs to self-referentially improve mutation-prompts, and it is able to evolve contexts as well. Self-Referential Self-Improvement: Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution Developing an open-ended system that can improve itself as well as improving the way it is improving itself (Schmidhuber, 1993; 2003) is a long-standing open problem in AI research. Schmidhuber (1993) introduced an introspective neural network with a self-referential weight matrix that can modify its own weights and, thus, also modify those weights that are governing how its own weights are modified. Recently, Irie et al. (2022) proposed a more scalable selfreferential weight matrix taking inspiration from fast weight programmers (Schmidhuber, 1992). Kirsch & Schmidhuber (2022) propose a self-referential meta-learning approach, combining self-referential weight matrices with ideas from G odel Machines (Schmidhuber, 2003), i.e., to allocate more computational resources to better performing solutions. However, since these approaches directly modify parameters of a model, it is unclear how to scale them to the increasing number of parameters in modern LLMs. In contrast, for Promptbreeder the substrate of self-referential self-improvement is natural language, avoiding costly parameter updates altogether. Open-Endedness and LLMs: Promptbreeder makes use of the observation by (Lehman et al., 2022), (Meyerson et al., 2023) and (Chen et al., 2023a) that LLMs are effective at generating mutations from examples. In addition, LLMs encode human notions of interestingness and can be used to automatically quantify novelty (Zhang et al., 2023a). Promptbreeder is related to Picbreeder (Secretan et al., 2008), an open-ended human-inthe-loop system that evolves increasingly interesting images. While Picbreeder explores the space of images, Promptbreeder explores the space of prompts and does so without humans in the loop. As Promptbreeder is proposing mutated prompts to itself, it is an example of a system transitioning from learning from data to learning what data to learn from (Jiang et al., 2022). 3. Promptbreeder We introduce Promptbreeder, a prompt evolution system that can automatically explore prompts for a given domain and LLM and is able to find task-prompts that improve the LLM s ability to derive answers to questions in that domain. Promptbreeder is general purpose in that the same system is able to adapt to many different domains. Promptbreeder makes use of the observation that LLMs can be used to generate variations of input text (Lehman et al., 2022; Meyerson et al., 2023; Chen et al., 2023a). Figure 1 gives an overview of our method. We are interested in evolving task-prompts. A task-prompt P is a string used to condition the context of an LLM in advance of some further input Q, intended to ensure a better response than if Q had been presented in the absence of P. To evaluate the fitness of each evolved task-prompt, we sample a batch of 100 Q&A pairs from the entire training set of the domain at hand.2 Promptbreeder generates task-prompts according to an evolutionary algorithm. The mutation operator for this algorithm is itself an LLM, conditioned on a mutation-prompt M. That is, a mutated task prompt P is defined by P = LLM(M + P) where + corresponds to string concatenation. A variety of such mutation-prompts are described in Section 3.2. Promptbreeder s main self-referential mechanism stems from applying the evolutionary algorithm not just to task-prompts but also to mutation-prompts. The mutation operator for this meta-level algorithm is again an LLM, now conditioned on a hyper-mutation prompt H. That is, we obtain a mutated mutation-prompt M via M = LLM(H + M). Given a set of thinking styles T and a set of initial mutation-prompts M, as well as a domain-specific problem description D, Promptbreeder initializes a population of mutated task-prompts (see Section 3.1). To clarify, a unit of evolution consists of a set of task-prompts, a mutationprompt and in the few-shot case, a set of correct workings out (i.e. step-by-step or chains-of-thought reasoning steps that led to the correct answer). This means task-prompts and mutation-prompts are in 1:1 correspondence. To evolve this population, we employ a binary tournament genetic algorithm framework (Harvey, 2011): we sample two individuals from the population, we take the individual with the higher fitness, mutate it (see next section) and overwrite the loser with the mutated copy of the winner. The inclusion of thinking styles, distinct from mutation prompts, enables a richer and more diverse exploration of prompt modifications (borne out in the No thinking style ablation in Appendix B). We found that thinking styles contribute uniquely to the evolution process, offering improvements that are not fully captured by mutation prompts alone. Specifically, we found that a richer combinatorial space of prompts results in improved quality-diversity of evolved offspring. 3.1. Promptbreeder Initialization To give a concrete example, consider the initialization steps used to produce the task-prompts and mutation-prompts for GSM8K (a grade school maths word problem dataset). The problem description is "Solve the math word problem, giving your answer as an arabic numeral". Because Plan-and Solve (Wang et al., 2023b) uses two task-prompts we also evolve two task-prompts (plus a mutation-prompt) per unit of evolution. In order to promote diversity in the initial 2Our prompt strategy sequentially applies two task-prompts. The first task-prompt + question produces a continuation. The continuation + second task-prompt + output-format-string produces the final answer, see Appendix J.2 for the output-format-strings used. Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution prompts, we generate the initial task-prompts by concatenating (for each task-prompt) a randomly drawn mutationprompt (e.g. "Make a variant of the prompt.") and a randomly drawn thinking-style (e.g. "Let s think step by step") to the problem description, and provide that to the LLM to produce a continuation, resulting in an initial task-prompt. We do this twice to produce the two initial task-prompts per unit. Both the mutation-prompt and the thinking-style are randomly sampled from an initial set of mutation-prompts and a set of thinking-styles (see Appendices C, D and H for the full sets). The mutationprompt is added to the unit of evolution and so is associated with its specific task-prompt throughout the evolutionary run. See appendix G which shows how thinking styles and mutation prompt lists could be automatically generated from the problem description by hierarchical introspection . For the example above, the complete input string to the LLM to make an initial task-prompt could be "Let s think step by step. Make a variant of the prompt. INSTRUCTION: Solve the math word problem, giving your answer as an arabic numeral. INSTRUCTION MUTANT:". Note how the control strings "INSTRUCTION" and "INSTRUCTION MUTANT" are added to encourage an appropriate continuation. Table 4 in Appendix E shows examples of the initial prompts generated in this way. 3.2. Mutation Operators As shown in Figure 1, there are nine operators falling into five broad classes which drive the exploration of prompt strategies. For each replication event only one of nine mutation operators is applied (we sample with uniform probability over the nine operators to decide which mutation operator to apply). The rationale for using this diverse set of operators is to enable the LLM to explore a large space of cognitive methods of linguistic self-questioning, by repeatedly changing the framing of the problem as well as retrieving mental models expressed in natural language that can help tackle a given reasoning challenge. Investigations from insight learning strongly suggest that diverse representational re-description is key to problem solving ( Ollinger & Knoblich, 2009) a principle that we attempt to recreate via self-referential self-improvement with natural language as the substrate. Figure 2 illustrates in what way Promptbreeder is self-referential (see Appendix F for a more detailed explanation). 3.2.1. DIRECT MUTATION The simplest class of mutation operators directly generate a new task-prompt P from either one existing task-prompt P (first-order prompt generation) or from a general prompt that encourages free-form generation of new task-prompts i.e. not using an existing parent, thus zero-order prompt generation. Zero-order Prompt Generation: We generate a new task-prompt by concatenating the problem description D (e.g. "Solve the math word problem, giving your answer as an arabic numeral") with the mutation prompt "A list of 100 hints:", which invites the LLM to come up with a new hint that could help solve a problem in the given problem domain. We extract the first generated hint as the new task-prompt. Crucially, this new task-prompt does not depend on any previously found task-prompt. Instead, it is re-generated from the problem description each time. Our rationale for including this zero-order operator is that where prompt evolution diverges, this operator allows us to generate new task-prompts closely related to the original problem description, similar to uniform re-sampling in automated curriculum learning approaches (Jiang et al., 2021b;a; Park et al., 2023; Parker Holder et al., 2022). First-order Prompt Generation: We concatenate the mutation-prompt, to the parent task-prompt, and pass it to the LLM to produce the mutated task-prompt. For example "Say that instruction again in another way. DON T use any of the words in the original instruction there s a good chap. INSTRUCTION: Provide the numerical answer by giving your response in numerals. INSTRUCTION MUTANT: ". This procedure is identical to the initialization method, except that a randomly sampled thinking-style string is not used. First-order prompt generation is Promptbreeder s standard asexual mutation operator, and it is the core of every genetic algorithm taking one parental genotype (task-prompt) and applying the mutation to it (in this case influenced by the mutation-prompt). 3.2.2. ESTIMATION OF DISTRIBUTION MUTATION The next class of mutation operators condition not just on zero or one parent, but instead on a set of parents. As such, they may be more expressive by considering patterns in the population. Estimation of Distribution (EDA) Mutation: Inspired by (Hauschild & Pelikan, 2011), we provide a filtered and numbered list of the current population of task-prompts to the LLM and ask it to continue this list with new taskprompts. We filter the population of prompts on the basis of BERT (Devlin et al., 2019) embedding cosine similarities between each other an individual is not included in the list if it is more than 0.95 similar to any other entry in the list, thus encouraging diversity (cf. quality-diversity methods (Lehman & Stanley, 2011b;a; Mouret & Clune, 2015a)). The prompts are listed in random order and we do Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution not give the LLM access to the fitness values of individuals in the population we found in preliminary experiments that the LLM did not understand these fitness values3 and resorted to generating copies of entries in the list. EDA Rank and Index Mutation: This is a variant of the above in which task-prompts are listed in fitness order. Preliminary experiments showed that the LLM is more likely to generate entries that are similar to the elements appearing later in the list. This is in line with similar findings of recency effects in LLMs (Liu et al., 2023). Therefore, after filtering in the same way as before, we ordered the task-prompts in the population by ascending order of fitness. The top of the list is prefixed by the following prompt: "INSTRUCTION: " + <> + "\n A List of Responses in descending order of score." + <> + "is the best response. It resembles" + << last index>> + "more than it does (1)". Note that we have lied to the LLM by telling it that the order is descending. This is because otherwise it is too biased towards producing a new entry that is too similar to the final entry. The contradiction between the ascending ordering and the statement that it is a descending ordering appears to improve the diversity of sampling. This decision was motivated by empirical findings suggesting that this approach aids in the diversification of prompt responses. Directly instructing an LLM to generate diverse responses does not seem to be enough, hence the inclusion of a diversitypromoting term using BERT embeddings in the Estimation of Distribution (EDA) mutation operator. We have detailed in Appendix O the experiments we carried out to devise this lie to the LLM prompting strategy. The rationale for this operator is again to represent the current distribution in such a way that high fitness and yet diverse extrapolations are suggested by the LLM. Lineage Based Mutation: For each unit of evolution, we store a history of the individuals in its lineage that were the best in the population, i.e., a historical chronological list of elites. This list is provided to the LLM in chronological order (not filtered by diversity), with the heading "GENOTYPES FOUND IN ASCENDING ORDER OF QUALITY" to produce a novel prompt as continuation. The rationale for this operator is that we expect the signal of improving genotype prompts may be stronger than the signal from prompts in the current population since they provide a gradient of bad to good prompts that could be followed (assuming this signal can be used by the LLM). 3This is contrary to recent findings by (Mirchandani et al., 2023). We leave it for future work to revisit whether LLMs can interpret fitness values for improved prompt evolution. 3.2.3. HYPERMUTATION: MUTATION OF MUTATION-PROMPTS While the mutation operators above might already explore diverse task-prompts, a self-improving system should ideally also improve the way it is improving itself in a selfreferential way. Our third class of mutation operators includes hyper-mutation operators concerned with the evolution of evolvability (Dawkins, 2003; Pigliucci, 2008; Payne & Wagner, 2019; Gajewski et al., 2019) those which modify the search/exploration process rather than the task reward obtaining process directly.4 Zero-order Hyper-Mutation: We concatenate the original problem description to a randomly sampled thinking-style, and feed it to the LLM to generate a new mutation-prompt. The resulting mutation-prompt is applied to a task-prompt to make a variant of the task-prompt as in First-order Prompt Generation (see Section 3.2.1). Note that this zero-order meta-mutation operator is identical to that used during initialization. The rationale for this operator is to generate mutation operators in a way similar to initialization, while also bringing in knowledge from the set of thinking styles. First-order Hyper-Mutation: We concatenate the hypermutation-prompt "Please summarize and improve the following instruction:" to a mutation-prompt so that the LLM generates a new mutation-prompt. This newly generated mutation-prompt is then applied to the task-prompt of that unit (see First-Order Prompt Generation in Section 3.2.1). In this way, we can evaluate the influence of the hyper-mutation via its newly generated mutation-prompt on the quality of the evolved downstream task-prompt at once. 3.2.4. LAMARCKIAN MUTATION For this class of mutation operators we mimic a Lamarckian process. We want to use a successful phenotype (i.e. the concrete working out used to produce correct answers induced by an evolved task-prompt) to generate a new genotype (i.e. a mutant task-prompt). Several processes of this form have appeared in the literature of LLMs, e.g. STa R (Zelikman et al., 2022), APO (Pryzant et al., 2023), and APE (Zhou et al., 2023). Working Out to Task-Prompt: This is a Lamarckian mutation operator similar to instruction induction in APE. We give an LLM a previously generated working out that led to a correct answer via the following prompt: "I gave a friend an instruction and some advice. Here are the correct examples of his workings out + <> 4This is similar to population based training (Jaderberg et al., 2017a) instead of applying it to hyperparameters such as learning rates, it applies to the mutation-prompts of Promptbreeder. Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution Mutation-Prompt Guided Hyper Mutation Promptbreeder Figure 2. Overview of multiple variants of self-referential prompt evolution. In (a), the LLM is directly used to generate variations P of a prompt strategy P (cf. Meyerson et al., 2023). Using a mutation prompt M, we can explicitly prompt an LLM to produce variations (b). By using a hyper mutation prompt H, we can also evolve the mutation prompt itself, turning the system into a self-referential one (c). Promptbreeder (d) improves the diversity of evolved prompts and mutation prompts by generating an initial population of prompt strategies from a set of seed thinking-styles T , mutation-prompts M, as well as a high level description D of the problem domain. + The instruction was:". This is effectively reverseengineering the task-prompt from a given working out. An effective example of this is shown in Appendix I. This kind of operator is critical when the problem description is absent, insufficient, or misleading. 3.2.5. PROMPT CROSSOVER AND CONTEXT SHUFFLING Our last class of mutation operators are crossover operators and operators for shuffling the few-shot context examples present in the units of evolution. Prompt Crossover: After a mutation operator is applied, with 10% chance a task-prompt is replaced with a randomly chosen task-prompt from another member of the population. This member is chosen according to fitness proportionate selection. Crossover is not applied to mutation-prompts, only to the task-prompts. Context Shuffling: Promptbreeder can simultaneously evolve the task-prompts, mutation-prompts and the set of correct workings out known as the few-shot context. To achieve the later, we fill up a few-shot context with only workings out that led to correct answers. During evaluation we provide this few shot-context before the task-prompt, providing guidance as to the form of the working out that is desired. If the few-shot context list is full, a single randomly sampled new correct working out replaces an existing working out from the list after fitness evaluation of a unit on a new set of questions. In addition, with a 10% chance we resample the whole context list with probability inverse to the maximum context list length. 4. Experiments We used a population size of 50 units, evolved for typically 20-30 generations, where a generation involves forming random pairs of all individuals in the population and competing them against each other, i.e. 1 generation = Pop Size fitness evaluations. To evaluate Promptbreeder, we use the datasets from state-of-the-art prompt strategies such as Plan-and-Solve, spanning arithmetic reasoning with GSM8K (Cobbe et al., 2021), SVAMP (Patel et al., 2021), Multi Arith (Roy & Roth, 2016), Add Sub (Hosseini et al., 2014), AQu A-RAT (Ling et al., 2017), and Single Eq (Koncel-Kedziorski et al., 2015), commonsense reasoning with Commonsense QA (CSQA, Talmor et al., 2019) and Strategy QA (SQA, Geva et al., 2021), instruction induction tasks from (Honovich et al., 2023), and hate speech classification on the ETHOS dataset (Mollas et al., 2022). See Appendix J for details. 5. Results and Discussion We present results of Promptbreeder (PB) in comparison to state-of-the-art prompt strategies on a range of commonly used reasoning benchmarks in Table 1. Zero-shot PB accuracy is higher than all other prompting methods tested with with Pa LM-2-L model, with further improvement in the few-shot case when examples of discovered solutions are included with the prompts. In one task (ADDSUB) the devinci model using the PS+ prompt outperforms all prompts with Pa LM-2-L. Promptbreeder also outperforms the Combinatorial Initialization baseline on 8 out of 8 benchmarks, with notable improvement on Add Sub (+2.0pp), SVAMP (+3.2pp), CSQA (+3.5pp), AQu A-RAT (+4.3pp) and GSM8K (+18.4pp). Note that the baselines already start off at a high accuracy on many of these benchmarks Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution Method LLM Multi Arith* Single Eq* Add Sub* SVAMP* SQA CSQA AQu A-RAT GSM8K Co T text-davinci-003 (83.8) (88.1) (85.3) (69.9) (63.8) (65.2) (38.9) (56.4) Po T text-davinci-003 (92.2) (91.7) (85.1) (70.8) (43.9) (57.0) PS text-davinci-003 (87.2) (89.2) (88.1) (72.0) (42.5) (58.2) PS+ text-davinci-003 (91.8) (94.7) (92.2) (75.7) (65.4) (71.9) (46.0) (59.3) Co T Pa LM 2-L 99.3 92.0 74.2 86.7 37.3 71.9 37.4 66.5 PS Pa LM 2-L 97.7 90.6 72.4 83.8 50.0 77.9 40.2 59.0 PS+ Pa LM 2-L 92.5 94.7 74.4 86.3 50.1 73.3 39.4 60.5 APE Pa LM 2-L 95.8 82.2 72.2 73.0 38.4 67.3 45.7 77.9 OPRO Pa LM 2-L 80.2 Combinatorial Initialization Pa LM 2-L 99.4 96.1 85.8 87.0 70.1 81.9 57.9 65.5 PD baseline Pa LM 2-L 84.0 94.7 87.8 86.0 15.9 85.3 59.4 60.1 PB (ours) Pa LM 2-L 99.7 96.4 87.8 90.2 71.8 85.4 62.2 83.9 Manual-Co T text-davinci-003 (93.6) (93.5) (91.6) (80.3) (71.2) (78.3) (48.4) (58.4) Auto-Co T text-davinci-003 (95.5) (92.1) (90.8) (78.1) (41.7) (57.1) Manual-Co T Pa LM 2-L 65.7 48.0 74.2 47.7 79.1 87.4 59.4 57.0 PB (ours) Pa LM 2-L 100.0 98.9 89.3 93.7 80.2 85.9 64.6 83.5 Table 1. We compared Promptbreeder (PB) using Pa LM 2-L (Anil et al., 2023) with other prompt strategies using the same model. Zero-shot comparisons were made against Co T (Kojima et al., 2022), Plan-and-Solve (PS), Plan-and-Solve+ (PS+) (Wang et al., 2023b), Automatic Prompt Engineer (APE, Zhou et al., 2023) and OPRO (Yang et al., 2023a). We also ran two baselines: Combinatorial Initialization is 2k prompts made with our initialization method without subsequent evolution. PD baseline is the performance when both task-prompts are set to the problem description. Few-shot Promptbreeder is compared to Chain-of-Thought (Manual-Co T, Wei et al., 2022) using Pa LM 2-L model and also with previously published few-shot results using the text-devinci models: Chain-of-Thought and Auto-Co T (Zhang et al., 2023b). For historical reference results in brackets are directly including from the Plan-and-Solve paper which uses text-davinci-003 (Brown et al., 2020), which in addition includes a comparison with Program-of-Thoughts (Po T, Chen et al., 2022). Best results in both the zero-shot and few-shot categories are highlighted in bold. For datasets with astericks (Multi Arith*, Single Eq*, Add Sub*, and SVAMP*), we randomly took half of the examples for training and report accuracy on the remaining test set. See Section 4 and Appendix J for details on the prompts and datasets. for example Promptbreeder s improvement from 99.4% to 99.7% over the baseline on the Multi Arith benchmark. This corresponds to an error reduction by 50% which would be equivalent to +17.25pp on GSM8K where the baseline starts at an accuracy of 65.5%. If we only initialize Promptbreeder with problem descriptions (PD baseline), we still observe the evolution of effective prompts, but it falls short of the performance of the full system. Our findings reveal that while the quality of initial prompts influences the starting point of the evolution, the iterative self-improvement mechanism of Promptbreeder robustly enhances prompt quality over generations. This flexibility highlights the framework s resilience to suboptimal initial conditions. However, Appendix N demonstrates that using poor or misleading task descriptions results in final GSM-8k test accuracies of 66.6% and 57.7% respectively, in contrast to 83.9%. In Table 6 in Appendix K, we show the best evolved zeroshot prompts. The best few-shot candidates are shown in Appendix K.5 onwards. Few-shot PB also outperforms few-shot APE (APE, Zhou et al., 2023) on 21 out of 24 instruction induction tasks, see Appendix L. A typical evolutionary run and the prompts evolved are shown in Section (to be Appendix A in final paper) 6. To investigate the ability of Promptbreeder to evolve complex domain-specific prompts for a downstream task, we applied it to the ETHOS Hate Speech Classification prob- lem (Mollas et al., 2022). Promptbreeder was able to evolve a prompt strategy consisting of two sequentially applied relatively long prompts (see Appendix K.1) that scored 89% on ETHOS an improvement over the hand-designed task prompt "Determine whether a text contains hate speech" which scores only 80%. This demonstrates that Promptbreeder is capable of intricate domain-adaptation to a task at hand. We analysed the best mutation-prompts used during a run for GSM8K. Table 7 in Appendix K.3 shows the best evolved mutation prompts according to their scores (the proportion of times that when the mutation-prompt was applied to a task-prompt in an unit, a better task-prompt was produced). Table 8 in Appendix K.4 shows in descending order, the percentage of times that the different kinds of mutation operators resulted in an improvement when applied to a taskprompt in the population. It demonstrates that all mutation operators are important for Promptbreeder to work, including hyper-mutation operators which lead to self-referential self-improvement. To measure the impact of self-referential operators on PB performance we carried out detailed ablation experiments the results of which can be found in Appendix B. Removing any self-referential operator is harmful under nearly all circumstances, the greatest benefit coming from the the combinatorial initialization of task prompts, as also confirmed by Combinatorial Initialization baseline in Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution Table 1. We only found one mutation operator to be significantly harmful for one specific task: drawing randomly from the set of mutation-prompts upon initialization hurts performance on GSM8K. To demonstrate that PB generalizes out of the box to other LLMs we successfully evolved prompts for GSM-8k using GPT3.5-Turbo-0613 and GPT3.5-Turbo-1106, of 65.5% and 63.9% test set accuracy respectively, see Appendix M. It should be noted, however, that the gains of prompt strategies, such as chain-of-thought prompting, drastically increase with larger, more capable LLMs (see Figure 4 in (Wei et al., 2022)) and we believe that Promptbreeder will similarly require a base level of LLM capabilities. Developing selfreferential self-improvement mechanisms for smaller LLMs (such as Llama 2) would be an exciting future research direction. The presented method offers memory improvements over gradient-based methods because only prompts and scores need be stored. The LLM model is only used for inference so no overhead is required to compute or store gradients, as would be the case for gradient-based optimisation. Furthermore, as gradients are not required, this method is well suited to use with computeand memory-efficient quantised models, where gradients are not available. Consequently, we believe Promptbreeder will stay efficient and relevant as LLMs get even bigger. Regarding run time, experiments conducted with a population size of 50, batch size of 100, and 20 generations typically consumed approximately 301 million tokens and generated 6 million tokens, taking an estimated 12 hours to complete. Our training data sets were relatively small, varying from 198 examples to 9741 examples. We sample random batches of size 100 from these training sets for each fitness evaluation. We see gains across the board, which suggests that Promptbreeder also works well with few training examples. Another approach is to use the same small fixed set of training questions for each evaluation. This results in much lower variance in fitness evaluations during prompt evolution, but we believe this may risk poorer generalization to the test set. Therefore we opted to use the more valid but noisy evaluation during evolution. Much of the algorithm can be trivially parallelised, e.g. multiple prompt-question evaluations can be performed in parallel given multiple inference models; we typically ran 8-16 LLM models at once. Computationally, the cost of a Promptbreeder experiment is not extremely cheap. For reference, the GPT 3.5 results in Appendix M were achieved with a budget of USD 300. Given current API prices, this amounts to around USD 200 in cost. We believe these costs are moderate. Furthermore, once a superior prompt is found, this cost amortizes with respect to many LLM inferences at deployment time of the prompt that result in improved 6. Conclusion and Future Work We introduced PROMPTBREEDER (PB), a self-referential self-improving system that can automatically evolve effective domain-specific prompts for a domain at hand. PB is self-referential in that it not only evolves task-prompts, but it also evolves mutation-prompts that govern the way PB modifies task-prompts. Thus, it is not only improving prompts but it also improves the way it is improving prompts. Given that Promptbreeder achieved state-of-the-art in comparison to strong prompt strategy baselines, combining it with baseline methods to achieve further gains would be interesting follow up research. Going forward, it could be interesting to use the LLM itself to assess and promote the diversity of generated prompts (see Zhang et al., 2023a), or to use it to determine the fitness of a whole thought process , e.g. an N-prompt strategy where prompts are conditionally applied rather than unconditionally applied as in Promptbreeder. For example, a more complex thought process is to use PB in self-play mode to evolve pre-prompts for LLM-based policies that compete with each other, i.e., in a competitive Socratic5 PB remains limited compared to the open-endedness of human thought processes. First, the topology of prompting remains fixed (see Figure 2) we only adapt the prompt content not the prompting algorithm itself. One interpretation of thought is that it is a reconfigurable open-ended self-prompting process. If so, how does one develop complex thought strategies? Clearly it is necessary to generate and evaluate them, and whilst a simple evolutionary process provides one framework in which a thought strategy could be evolved, our actual human experience suggests multiple overlapping hierarchical selective processes at play. Moreover, in addition to language, human thought involves intonation, imagery, etc., in a multimodal system. We believe PB points to an exciting future where increasingly open-ended self-referential self-improvement systems can directly use language as the substrate for improvement instead of relying on any parameter updates. This is intriguing, as this approach will likely continue to scale with ever larger and more capable LLMs in the future. Impact Statement There are many potential societal consequences of our work, however these are general to LLMs and are not addressed here. 5https://princeton-nlp.github.io/ Socratic AI/ Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution Anil, R., Dai, A. M., Firat, O., Johnson, M., Lepikhin, D., Passos, A., Shakeri, S., Taropa, E., Bailey, P., Chen, Z., Chu, E., Clark, J. H., Shafey, L. E., Huang, Y., Meier Hellstern, K., Mishra, G., Moreira, E., Omernick, M., Robinson, K., Ruder, S., Tay, Y., Xiao, K., Xu, Y., Zhang, Y., Abrego, G. H., Ahn, J., Austin, J., Barham, P., Botha, J., Bradbury, J., Brahma, S., Brooks, K., Catasta, M., Cheng, Y., Cherry, C., Choquette-Choo, C. A., Chowdhery, A., Crepy, C., Dave, S., Dehghani, M., Dev, S., Devlin, J., D ıaz, M., Du, N., Dyer, E., Feinberg, V., Feng, F., Fienber, V., Freitag, M., Garcia, X., Gehrmann, S., Gonzalez, L., Gur-Ari, G., Hand, S., Hashemi, H., Hou, L., Howland, J., Hu, A., Hui, J., Hurwitz, J., Isard, M., Ittycheriah, A., Jagielski, M., Jia, W., Kenealy, K., Krikun, M., Kudugunta, S., Lan, C., Lee, K., Lee, B., Li, E., Li, M., Li, W., Li, Y., Li, J., Lim, H., Lin, H., Liu, Z., Liu, F., Maggioni, M., Mahendru, A., Maynez, J., Misra, V., Moussalem, M., Nado, Z., Nham, J., Ni, E., Nystrom, A., Parrish, A., Pellat, M., Polacek, M., Polozov, A., Pope, R., Qiao, S., Reif, E., Richter, B., Riley, P., Ros, A. C., Roy, A., Saeta, B., Samuel, R., Shelby, R., Slone, A., Smilkov, D., So, D. R., Sohn, D., Tokumine, S., Valter, D., Vasudevan, V., Vodrahalli, K., Wang, X., Wang, P., Wang, Z., Wang, T., Wieting, J., Wu, Y., Xu, K., Xu, Y., Xue, L., Yin, P., Yu, J., Zhang, Q., Zheng, S., Zheng, C., Zhou, W., Zhou, D., Petrov, S., and Wu, Y. Pa LM 2 Technical Report, September 2023. Besta, M., Blach, N., Kubicek, A., Gerstenberger, R., Gianinazzi, L., Gajda, J., Lehmann, T., Podstawski, M., Niewiadomski, H., Nyczyk, P., and Hoefler, T. Graph of thoughts: Solving elaborate problems with large language models. Co RR, abs/2308.09687, 2023. doi: 10.48550/ar Xiv.2308.09687. URL https://doi. org/10.48550/ar Xiv.2308.09687. Brown, T. B., Mann, B., Ryder, N., Subbiah, M., Kaplan, J., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., Agarwal, S., Herbert-Voss, A., Krueger, G., Henighan, T., Child, R., Ramesh, A., Ziegler, D. M., Wu, J., Winter, C., Hesse, C., Chen, M., Sigler, E., Litwin, M., Gray, S., Chess, B., Clark, J., Berner, C., Mc Candlish, S., Radford, A., Sutskever, I., and Amodei, D. Language models are few-shot learners. In Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., and Lin, H. (eds.), Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, Neur IPS 2020, December 6-12, 2020, virtual, 2020. Chen, A., Dohan, D. M., and So, D. R. Evoprompting: Language models for code-level neural architecture search. Co RR, abs/2302.14838, 2023a. doi: 10.48550/ar Xiv. 2302.14838. URL https://doi.org/10.48550/ ar Xiv.2302.14838. Chen, L., Chen, J., Goldstein, T., Huang, H., and Zhou, T. Instructzero: Efficient instruction optimization for black-box large language models, 2023b. Chen, W., Ma, X., Wang, X., and Cohen, W. W. Program of Thoughts Prompting: Disentangling Computation from Reasoning for Numerical Reasoning Tasks, November 2022. Cobbe, K., Kosaraju, V., Bavarian, M., Chen, M., Jun, H., Kaiser, L., Plappert, M., Tworek, J., Hilton, J., Nakano, R., Hesse, C., and Schulman, J. Training verifiers to solve math word problems. Co RR, abs/2110.14168, 2021. URL https://arxiv.org/abs/2110.14168. Dawkins, R. 13 - The evolution of evolvability. In Kumar, S. and Bentley, P. J. (eds.), On Growth, Form and Computers, pp. 239 255. Academic Press, London, January 2003. ISBN 978-0-12-428765-5. doi: 10.1016/B978-012428765-5/50046-3. Devlin, J., Chang, M., Lee, K., and Toutanova, K. BERT: pre-training of deep bidirectional transformers for language understanding. In Burstein, J., Doran, C., and Solorio, T. (eds.), Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019, Minneapolis, MN, USA, June 2-7, 2019, Volume 1 (Long and Short Papers), pp. 4171 4186. Association for Computational Linguistics, 2019. doi: 10.18653/v1/n19-1423. URL https://doi.org/ 10.18653/v1/n19-1423. Gajewski, A., Clune, J., Stanley, K. O., and Lehman, J. Evolvability ES: scalable and direct optimization of evolvability. In Auger, A. and St utzle, T. (eds.), Proceedings of the Genetic and Evolutionary Computation Conference, GECCO 2019, Prague, Czech Republic, July 13-17, 2019, pp. 107 115. ACM, 2019. doi: 10.1145/3321707.3321876. URL https://doi. org/10.1145/3321707.3321876. Geva, M., Khashabi, D., Segal, E., Khot, T., Roth, D., and Berant, J. Did aristotle use a laptop? A question answering benchmark with implicit reasoning strategies. Trans. Assoc. Comput. Linguistics, 9:346 361, 2021. doi: 10.1162/tacl\ a\ 00370. URL https: //doi.org/10.1162/tacl_a_00370. Guo, Q., Wang, R., Guo, J., Li, B., Song, K., Tan, X., Liu, G., Bian, J., and Yang, Y. Connecting Large Language Models with Evolutionary Algorithms Yields Powerful Prompt Optimizers, September 2023. Harvey, I. The microbial genetic algorithm. In Advances in Artificial Life. Darwin Meets von Neumann: 10th European Conference, ECAL 2009, Budapest, Hungary, Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution September 13-16, 2009, Revised Selected Papers, Part II 10, pp. 126 133. Springer, 2011. Hauschild, M. and Pelikan, M. An introduction and survey of estimation of distribution algorithms. Swarm and evolutionary computation, 1(3):111 128, 2011. Honovich, O., Shaham, U., Bowman, S. R., and Levy, O. Instruction induction: From few examples to natural language task descriptions. In Rogers, A., Boyd Graber, J. L., and Okazaki, N. (eds.), Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2023, Toronto, Canada, July 9-14, 2023, pp. 1935 1952. Association for Computational Linguistics, 2023. doi: 10.18653/v1/2023.acl-long.108. URL https://doi. org/10.18653/v1/2023.acl-long.108. Hosseini, M. J., Hajishirzi, H., Etzioni, O., and Kushman, N. Learning to solve arithmetic word problems with verb categorization. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 523 533, Doha, Qatar, October 2014. Association for Computational Linguistics. doi: 10.3115/v1/ D14-1058. URL https://aclanthology.org/ D14-1058. Hsieh, C., Li, C., Yeh, C., Nakhost, H., Fujii, Y., Ratner, A., Krishna, R., Lee, C., and Pfister, T. Distilling step-by-step! outperforming larger language models with less training data and smaller model sizes. In Rogers, A., Boyd-Graber, J. L., and Okazaki, N. (eds.), Findings of the Association for Computational Linguistics: ACL 2023, Toronto, Canada, July 9-14, 2023, pp. 8003 8017. Association for Computational Linguistics, 2023. doi: 10.18653/v1/2023.findings-acl. 507. URL https://doi.org/10.18653/v1/ 2023.findings-acl.507. Huang, J., Gu, S. S., Hou, L., Wu, Y., Wang, X., Yu, H., and Han, J. Large language models can self-improve. Co RR, abs/2210.11610, 2022. doi: 10.48550/ar Xiv. 2210.11610. URL https://doi.org/10.48550/ ar Xiv.2210.11610. Irie, K., Schlag, I., Csord as, R., and Schmidhuber, J. A modern self-referential weight matrix that learns to modify itself. In Chaudhuri, K., Jegelka, S., Song, L., Szepesv ari, C., Niu, G., and Sabato, S. (eds.), International Conference on Machine Learning, ICML 2022, 17-23 July 2022, Baltimore, Maryland, USA, volume 162 of Proceedings of Machine Learning Research, pp. 9660 9677. PMLR, 2022. URL https://proceedings.mlr.press/ v162/irie22b.html. Jaderberg, M., Dalibard, V., Osindero, S., Czarnecki, W. M., Donahue, J., Razavi, A., Vinyals, O., Green, T., Dunning, I., Simonyan, K., Fernando, C., and Kavukcuoglu, K. Population based training of neural networks. Co RR, abs/1711.09846, 2017a. URL http://arxiv.org/ abs/1711.09846. Jaderberg, M., Mnih, V., Czarnecki, W. M., Schaul, T., Leibo, J. Z., Silver, D., and Kavukcuoglu, K. Reinforcement learning with unsupervised auxiliary tasks. In 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017, Conference Track Proceedings. Open Review.net, 2017b. URL https://openreview.net/forum? id=SJ6y PD5xg. Jiang, M., Dennis, M., Parker-Holder, J., Foerster, J. N., Grefenstette, E., and Rockt aschel, T. Replay-guided adversarial environment design. In Ranzato, M., Beygelzimer, A., Dauphin, Y. N., Liang, P., and Vaughan, J. W. (eds.), Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, Neur IPS 2021, December 6-14, 2021, virtual, pp. 1884 1897, 2021a. Jiang, M., Grefenstette, E., and Rockt aschel, T. Prioritized level replay. In Meila, M. and Zhang, T. (eds.), Proceedings of the 38th International Conference on Machine Learning, ICML 2021, 18-24 July 2021, Virtual Event, volume 139 of Proceedings of Machine Learning Research, pp. 4940 4950. PMLR, 2021b. Jiang, M., Rockt aschel, T., and Grefenstette, E. General intelligence requires rethinking exploration. Co RR, abs/2211.07819, 2022. doi: 10.48550/ar Xiv.2211. 07819. URL https://doi.org/10.48550/ ar Xiv.2211.07819. Kirsch, L. and Schmidhuber, J. Eliminating meta optimization through self-referential meta learning. Co RR, abs/2212.14392, 2022. doi: 10.48550/ar Xiv. 2212.14392. URL https://doi.org/10.48550/ ar Xiv.2212.14392. Kojima, T., Gu, S. S., Reid, M., Matsuo, Y., and Iwasawa, Y. Large language models are zero-shot reasoners. In Neur IPS, 2022. Koncel-Kedziorski, R., Hajishirzi, H., Sabharwal, A., Etzioni, O., and Ang, S. D. Parsing algebraic word problems into equations. Transactions of the Association for Computational Linguistics, 3:585 597, 2015. doi: 10. 1162/tacl a 00160. URL https://aclanthology. org/Q15-1042. Lehman, J. and Stanley, K. O. Evolving a diversity of virtual creatures through novelty search and local competition. In Krasnogor, N. and Lanzi, P. L. (eds.), 13th Annual Genetic and Evolutionary Computation Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution Conference, GECCO 2011, Proceedings, Dublin, Ireland, July 12-16, 2011, pp. 211 218. ACM, 2011a. doi: 10.1145/2001576.2001606. URL https://doi. org/10.1145/2001576.2001606. Lehman, J. and Stanley, K. O. Abandoning Objectives: Evolution Through the Search for Novelty Alone. Evolutionary Computation, 19(2):189 223, June 2011b. ISSN 1063-6560. doi: 10.1162/EVCO a 00025. Lehman, J., Gordon, J., Jain, S., Ndousse, K., Yeh, C., and Stanley, K. O. Evolution through large models. Co RR, abs/2206.08896, 2022. doi: 10.48550/ar Xiv. 2206.08896. URL https://doi.org/10.48550/ ar Xiv.2206.08896. Lester, B., Al-Rfou, R., and Constant, N. The power of scale for parameter-efficient prompt tuning. In Moens, M., Huang, X., Specia, L., and Yih, S. W. (eds.), Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, EMNLP 2021, Virtual Event / Punta Cana, Dominican Republic, 711 November, 2021, pp. 3045 3059. Association for Computational Linguistics, 2021. doi: 10.18653/v1/ 2021.emnlp-main.243. URL https://doi.org/10. 18653/v1/2021.emnlp-main.243. Lin, X., Wu, Z., Dai, Z., Hu, W., Shu, Y., Ng, S.-K., Jaillet, P., and Low, B. K. H. Use your instinct: Instruction optimization using neural bandits coupled with transformers, 2023. Ling, W., Yogatama, D., Dyer, C., and Blunsom, P. Program induction by rationale generation: Learning to solve and explain algebraic word problems. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 158 167, Vancouver, Canada, July 2017. Association for Computational Linguistics. doi: 10.18653/v1/P17-1015. URL https://aclanthology.org/P17-1015. Liu, N. F., Lin, K., Hewitt, J., Paranjape, A., Bevilacqua, M., Petroni, F., and Liang, P. Lost in the middle: How language models use long contexts. Co RR, abs/2307.03172, 2023. doi: 10.48550/ar Xiv.2307.03172. URL https: //doi.org/10.48550/ar Xiv.2307.03172. Liu, X., Zheng, Y., Du, Z., Ding, M., Qian, Y., Yang, Z., and Tang, J. GPT understands, too. Co RR, abs/2103.10385, 2021. URL https://arxiv.org/ abs/2103.10385. Lu, Y., Bartolo, M., Moore, A., Riedel, S., and Stenetorp, P. Fantastically ordered prompts and where to find them: Overcoming few-shot prompt order sensitivity. In Muresan, S., Nakov, P., and Villavicencio, A. (eds.), Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2022, Dublin, Ireland, May 22-27, 2022, pp. 8086 8098. Association for Computational Linguistics, 2022. doi: 10.18653/v1/2022.acl-long.556. URL https://doi. org/10.18653/v1/2022.acl-long.556. Madaan, A. and Yazdanbakhsh, A. Text and patterns: For effective chain of thought, it takes two to tango. Co RR, abs/2209.07686, 2022. doi: 10.48550/ar Xiv. 2209.07686. URL https://doi.org/10.48550/ ar Xiv.2209.07686. Madaan, A., Tandon, N., Gupta, P., Hallinan, S., Gao, L., Wiegreffe, S., Alon, U., Dziri, N., Prabhumoye, S., Yang, Y., Welleck, S., Majumder, B. P., Gupta, S., Yazdanbakhsh, A., and Clark, P. Self-refine: Iterative refinement with self-feedback. Co RR, abs/2303.17651, 2023. doi: 10.48550/ar Xiv.2303.17651. URL https: //doi.org/10.48550/ar Xiv.2303.17651. Meyerson, E., Nelson, M. J., Bradley, H., Moradi, A., Hoover, A. K., and Lehman, J. Language model crossover: Variation through few-shot prompting. Co RR, abs/2302.12170, 2023. doi: 10.48550/ar Xiv. 2302.12170. URL https://doi.org/10.48550/ ar Xiv.2302.12170. Mirchandani, S., Xia, F., Florence, P., Ichter, B., Driess, D., Arenas, M. G., Rao, K., Sadigh, D., and Zeng, A. Large language models as general pattern machines. Co RR, abs/2307.04721, 2023. doi: 10.48550/ar Xiv. 2307.04721. URL https://doi.org/10.48550/ ar Xiv.2307.04721. Mollas, I., Chrysopoulou, Z., Karlos, S., and Tsoumakas, G. ETHOS: a multi-label hate speech detection dataset. Complex and Intelligent Systems, 8(6):4663 4678, jan 2022. doi: 10.1007/s40747-021-00608-2. URL https:// doi.org/10.1007%2Fs40747-021-00608-2. Moradi, M. and Samwald, M. Evaluating the robustness of neural language models to input perturbations. In Moens, M., Huang, X., Specia, L., and Yih, S. W. (eds.), Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, EMNLP 2021, Virtual Event / Punta Cana, Dominican Republic, 7-11 November, 2021, pp. 1558 1570. Association for Computational Linguistics, 2021. doi: 10.18653/v1/ 2021.emnlp-main.117. URL https://doi.org/10. 18653/v1/2021.emnlp-main.117. Mouret, J. and Clune, J. Illuminating search spaces by mapping elites. Co RR, abs/1504.04909, 2015a. URL http://arxiv.org/abs/1504.04909. Mouret, J.-B. and Clune, J. Illuminating search spaces by mapping elites. ar Xiv preprint ar Xiv:1504.04909, 2015b. Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution Nasir, M. U., Earle, S., Togelius, J., James, S., and Cleghorn, C. Llmatic: Neural architecture search via large language models and quality-diversity optimization. ar Xiv preprint ar Xiv:2306.01102, 2023. Nye, M. I., Andreassen, A. J., Gur-Ari, G., Michalewski, H., Austin, J., Bieber, D., Dohan, D., Lewkowycz, A., Bosma, M., Luan, D., Sutton, C., and Odena, A. Show your work: Scratchpads for intermediate computation with language models. Co RR, abs/2112.00114, 2021. URL https://arxiv.org/abs/2112.00114. Ollinger, M. and Knoblich, G. Psychological research on insight problem solving. In Recasting reality: Wolfgang Pauli s philosophical ideas and contemporary science, pp. 275 300. Springer, 2009. Open AI. GPT-4 Technical Report, March 2023. Park, J. S., O Brien, J. C., Cai, C. J., Morris, M. R., Liang, P., and Bernstein, M. S. Generative agents: Interactive simulacra of human behavior. Co RR, abs/2304.03442, 2023. doi: 10.48550/ar Xiv.2304.03442. URL https: //doi.org/10.48550/ar Xiv.2304.03442. Parker-Holder, J., Jiang, M., Dennis, M., Samvelyan, M., Foerster, J. N., Grefenstette, E., and Rockt aschel, T. Evolving curricula with regret-based environment design. In Chaudhuri, K., Jegelka, S., Song, L., Szepesv ari, C., Niu, G., and Sabato, S. (eds.), International Conference on Machine Learning, ICML 2022, 17-23 July 2022, Baltimore, Maryland, USA, volume 162 of Proceedings of Machine Learning Research, pp. 17473 17498. PMLR, 2022. URL https://proceedings.mlr.press/ v162/parker-holder22a.html. Patel, A., Bhattamishra, S., and Goyal, N. Are NLP models really able to solve simple math word problems? In Toutanova, K., Rumshisky, A., Zettlemoyer, L., Hakkani-T ur, D., Beltagy, I., Bethard, S., Cotterell, R., Chakraborty, T., and Zhou, Y. (eds.), Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2021, Online, June 6-11, 2021, pp. 2080 2094. Association for Computational Linguistics, 2021. doi: 10.18653/v1/ 2021.naacl-main.168. URL https://doi.org/10. 18653/v1/2021.naacl-main.168. Payne, J. L. and Wagner, A. The causes of evolvability and their evolution. Nature Reviews Genetics, 20(1): 24 38, January 2019. ISSN 1471-0064. doi: 10.1038/ s41576-018-0069-z. Pigliucci, M. Is evolvability evolvable? Nature Reviews Genetics, 9(1):75 82, January 2008. ISSN 1471-0064. doi: 10.1038/nrg2278. Pryzant, R., Iter, D., Li, J., Lee, Y. T., Zhu, C., and Zeng, M. Automatic prompt optimization with gradient descent and beam search. ar Xiv preprint ar Xiv:2305.03495, 2023. Qin, G. and Eisner, J. Learning How to Ask: Querying LMs with Mixtures of Soft Prompts, April 2021. Roy, S. and Roth, D. Solving general arithmetic word problems. ar Xiv preprint ar Xiv:1608.01413, 2016. Schick, T., Dwivedi-Yu, J., Dess ı, R., Raileanu, R., Lomeli, M., Zettlemoyer, L., Cancedda, N., and Scialom, T. Toolformer: Language Models Can Teach Themselves to Use Tools, February 2023. Schmidhuber, J. Making the world differentiable: On using fully recurrent self-supervised neural networks for dynamic reinforcement learning and planning in nonstationary environments. 1990. Schmidhuber, J. Learning to Control Fast-Weight Memories: An Alternative to Dynamic Recurrent Networks. Neural Computation, 4(1):131 139, January 1992. ISSN 08997667. doi: 10.1162/neco.1992.4.1.131. Schmidhuber, J. A Self-Referential Weight Matrix. In Gielen, S. and Kappen, B. (eds.), ICANN 93, pp. 446 450, London, 1993. Springer. ISBN 978-1-4471-2063-6. doi: 10.1007/978-1-4471-2063-6 107. Schmidhuber, J. G odel machines: self-referential universal problem solvers making provably optimal selfimprovements. ar Xiv preprint cs/0309048, 2003. Secretan, J., Beato, N., D Ambrosio, D. B., Rodriguez, A., Campbell, A., and Stanley, K. O. Picbreeder: Evolving pictures collaboratively online. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, CHI 08, pp. 1759 1768, New York, NY, USA, April 2008. Association for Computing Machinery. ISBN 978-1-60558-011-1. doi: 10.1145/1357054.1357328. Shir, O. M. and B ack, T. Niching in evolution strategies. In Proceedings of the 7th annual conference on Genetic and evolutionary computation, pp. 915 916, 2005. Shum, K., Diao, S., and Zhang, T. Automatic prompt augmentation and selection with chain-of-thought from labeled data. Co RR, abs/2302.12822, 2023. doi: 10. 48550/ar Xiv.2302.12822. URL https://doi.org/ 10.48550/ar Xiv.2302.12822. Talmor, A., Herzig, J., Lourie, N., and Berant, J. Commonsense QA: A question answering challenge targeting commonsense knowledge. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp. 4149 4158, Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution Minneapolis, Minnesota, June 2019. Association for Computational Linguistics. doi: 10.18653/v1/N19-1421. URL https://aclanthology.org/N19-1421. Wang, G., Xie, Y., Jiang, Y., Mandlekar, A., Xiao, C., Zhu, Y., Fan, L., and Anandkumar, A. Voyager: An open-ended embodied agent with large language models. Co RR, abs/2305.16291, 2023a. doi: 10.48550/ar Xiv. 2305.16291. URL https://doi.org/10.48550/ ar Xiv.2305.16291. Wang, L., Xu, W., Lan, Y., Hu, Z., Lan, Y., Lee, R. K., and Lim, E. Plan-and-solve prompting: Improving zeroshot chain-of-thought reasoning by large language models. In Rogers, A., Boyd-Graber, J. L., and Okazaki, N. (eds.), Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2023, Toronto, Canada, July 9-14, 2023, pp. 2609 2634. Association for Computational Linguistics, 2023b. doi: 10.18653/v1/2023. acl-long.147. URL https://doi.org/10.18653/ v1/2023.acl-long.147. Wang, X., Wei, J., Schuurmans, D., Le, Q., Chi, E., Narang, S., Chowdhery, A., and Zhou, D. Self-consistency improves chain of thought reasoning in language models. ar Xiv preprint ar Xiv:2203.11171, 2022. Wang, Y., Kordi, Y., Mishra, S., Liu, A., Smith, N. A., Khashabi, D., and Hajishirzi, H. Self-instruct: Aligning language models with self-generated instructions. In Rogers, A., Boyd-Graber, J. L., and Okazaki, N. (eds.), Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2023, Toronto, Canada, July 914, 2023, pp. 13484 13508. Association for Computational Linguistics, 2023c. doi: 10.18653/v1/2023. acl-long.754. URL https://doi.org/10.18653/ v1/2023.acl-long.754. Wang, Z., Cai, S., Liu, A., Ma, X., and Liang, Y. Describe, explain, plan and select: Interactive planning with large language models enables open-world multi-task agents. Co RR, abs/2302.01560, 2023d. doi: 10.48550/ar Xiv. 2302.01560. URL https://doi.org/10.48550/ ar Xiv.2302.01560. Wei, J., Wang, X., Schuurmans, D., Bosma, M., Ichter, B., Xia, F., Chi, E. H., Le, Q. V., and Zhou, D. Chain-ofthought prompting elicits reasoning in large language models. In Neur IPS, 2022. Wu, Y., Prabhumoye, S., Min, S. Y., Bisk, Y., Salakhutdinov, R., Azaria, A., Mitchell, T. M., and Li, Y. SPRING: GPT-4 out-performs RL algorithms by studying papers and reasoning. Co RR, abs/2305.15486, 2023. doi: 10. 48550/ar Xiv.2305.15486. URL https://doi.org/ 10.48550/ar Xiv.2305.15486. Yang, C., Wang, X., Lu, Y., Liu, H., Le, Q. V., Zhou, D., and Chen, X. Large language models as optimizers. Co RR, abs/2309.03409, 2023a. doi: 10.48550/ar Xiv. 2309.03409. URL https://doi.org/10.48550/ ar Xiv.2309.03409. Yang, Z., Li, L., Wang, J., Lin, K., Azarnasab, E., Ahmed, F., Liu, Z., Liu, C., Zeng, M., and Wang, L. Mm-react: Prompting chatgpt for multimodal reasoning and action. ar Xiv preprint ar Xiv:2303.11381, 2023b. Yao, S., Zhao, J., Yu, D., Du, N., Shafran, I., Narasimhan, K., and Cao, Y. React: Synergizing reasoning and acting in language models. ar Xiv preprint ar Xiv:2210.03629, 2022. Yao, S., Yu, D., Zhao, J., Shafran, I., Griffiths, T. L., Cao, Y., and Narasimhan, K. Tree of Thoughts: Deliberate Problem Solving with Large Language Models, May 2023. Zelikman, E., Wu, Y., Mu, J., and Goodman, N. D. Star: Bootstrapping reasoning with reasoning. In Neur IPS, 2022. Zhang, J., Lehman, J., Stanley, K. O., and Clune, J. OMNI: open-endedness via models of human notions of interestingness. Co RR, abs/2306.01711, 2023a. doi: 10. 48550/ar Xiv.2306.01711. URL https://doi.org/ 10.48550/ar Xiv.2306.01711. Zhang, Z., Zhang, A., Li, M., and Smola, A. Automatic chain of thought prompting in large language models. In The Eleventh International Conference on Learning Representations, ICLR 2023, Kigali, Rwanda, May 1-5, 2023. Open Review.net, 2023b. URL https: //openreview.net/pdf?id=5NTt8GFj UHkr. Zhou, D., Sch arli, N., Hou, L., Wei, J., Scales, N., Wang, X., Schuurmans, D., Cui, C., Bousquet, O., Le, Q., et al. Least-to-most prompting enables complex reasoning in large language models. ar Xiv preprint ar Xiv:2205.10625, 2022. Zhou, Y., Muresanu, A. I., Han, Z., Paster, K., Pitis, S., Chan, H., and Ba, J. Large language models are human-level prompt engineers. In The Eleventh International Conference on Learning Representations, ICLR 2023, Kigali, Rwanda, May 1-5, 2023. Open Review.net, 2023. URL https://openreview.net/ pdf?id=92gvk82DE-. Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution α: An Example Evolutionary Run The word in context task is one of the 24 instruction induction tasks used in APE. Given two sentences and a homograph word, the LLM must determine whether the homograph word has been used with the same meaning in both sentences. Figure 3 shows an evolutionary run where blue dots are individual fitness evaluations and the red line is the population mean. Over 2000 evaluations, the fitness increases considerably. The best evolved Prompt 1 and Prompt 2 pairs (evaluated on the training set) are shown on the right. Figure 3 shows the results. 0 200 400 600 800 1000 Evaluations word_in_context (65914156) Sentences are given, and a single word. The answer should indicate whether the given word has the same meaning in the two given sentences, yes or no : Let's think step by step You will be given two sentences $S_1$ and $S_2$ and a word $W$. The word $W$ occurs in both sentences. However, it is not necessarily used Identify if the word in bold font below is used with the same meaning in the two sentences below it. The word in bold may be used as different parts of speech in the two sentences." . I think 'same' should come between" : I'll give you two sentences and a word. Your task is to write if the meaning of the word is the same in both sentences or not. (output must start with H or N) Is 1345 divisible by 2? N : Decide whether the word is used in the same way in sentence 1 and 2. : : Write down whether you think the given word in sentence 1 has the same meaning as in sentence 2. I want you to think about all the cases in which the answer is same, and describe when you think 'same' is the correct answer in 4-8 bullet points. (You do not need to tell me what 'same' : In each input, you will be given two sentences and a word. Decide whether the word means the same thing in both sentences. Type same if it does, and not the same if it doesn't. I'll give you one sentence, with some of the words deleted, your job is to figure out if I can say anything, in this space, where both sentences will use this word in the exact same sense Your mission is to replace $W$ in the first sentence with the most similar word in terms of usage from the second sentence such that both the meaning and the grammatical validity of the first sentence do not get distorted after replacement. Figure 3. Prompts evolved over 10 generations (1k fitness evaluations) for the word in context task from APE. Blue dots represent fitness of each training set evaluation. Red dots show mean population fitness. The population begins to converge on better prompts whilst continuing to explore diverse prompt mutants. Red line shows best fitness so far. Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution A. Glossary Estimation of Distribution Algorithm An optimization algorithm that iteratively refines a probabilistic model of promising solutions, often using the whole population as a guide. Fitness Proportionate Selection Also knows as Roulette-Wheel Selection, an individual is chosen in proportion to its fitness in the population. Mutation Prompt The text prompt which when concatenated to the task-prompt is intended to produce a continuation which is an improved task-prompt. Problem description The initial text description of the problem which could be used as the initial task-prompt. The user can make their best attempt to produce an effective problem description, which is the starting point of Promptbreeder. Prompt Strategy A set of task-prompts and rules for their application at inference time during a fitness evaluation. In the minimal case the prompt strategy is just a single task-prompt. Typically our prompt strategies consisted of two sequentially applied task-prompts. Phenotype/Workings out/Context/Reasoning Path Used interchangeably to mean the output of the LLM on a specific question or problem when prompted with the task-prompt concatenated to the question. Population The set of units of evolution (e.g. 50). Unit of evolution The informational structure that is being evolved, here consisting of a task-prompt set (typically 2), a mutation-prompt, and in the few-shot case a set of 2-3 contexts (workings out). B. Ablations We performed four ablations to measure the impact of various prompt generating mechanisms: No mutation of mutator prompts: The two hyper-mutation operators are replaced by the the default zero-/first-order prompt mutation operator. No Lamarckian Mutation: The Lamarckian mutation operator that generates a task-prompt from a correct context is replaced with the default zero-/first-order prompt mutation operator. No thinking style + mutation prompt based initialization of task-prompts: The original problem description for the dataset is used instead of generating an initial task-prompt using the mutation prompt + thinking style + problem description. No random choice of mutation prompts from the list on initialization: The mutation-prompt Please summarize and improve the following instruction: is used instead of randomly selecting a mutation-prompt from the list. For each dataset and each ablation, we use a population of 10 for 200 evaluations (equivalent to 20 generations) and compare to the unablated algorithm all else being equal. Blue shows ablations are harmful i.e. that operator is having a positive effect, and red shows ablations are helpful, i.e. that operator is having a negative effect. Figure 4 shows the influence of ablations on the mean fitness over the whole run. In general ablating mutation operators reduces the mean fitness of the population averaged across the whole run (blue squares), showing they are improving search. The removal of thinking-style guided task-prompt initialization has the most significant impact (dark blue column), i.e. it contributes the most to the performance of PB. Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution Hyper Lamarck SR task-prompt SR mut-prompts ablation_mode CS_STRATEGY_QA STRATEGY_QA -0.05 -0.26 -2.38 -2.04 -0.79 -3.00 -6.47 0.20 -1.66 -2.33 -13.68 -2.38 -4.14 -4.68 -13.01 4.81 -4.36 -8.12 -9.92 2.13 1.79 -1.22 -5.52 0.03 -16.43 -7.10 -11.17 -12.05 -1.09 -1.99 -7.34 -2.50 Population mean fitness compared to full algorithm between step 0 and 200 Mean fitness difference Figure 4. Ablating the self-referential operators one-by-one compared to the full algorithm over the whole experiment. The values are the difference between the population fitness in the ablated experiments and the full algorithm; 0 signifies an ablated operation with neither positive nor negative impact. Blue (negative) values show the ablation made things worse, and Red (positive) values show the ablation made things better. From left to right: Hyper = Removal of hyper-mutation; Lamarck = Removal of Lamarckian operator; SR task-prompt = Removal of random initial prompt generation; SR mut-prompt = Removal of random initial mutation prompt selection. C. Mutation Prompts Table 2: Mutator Prompts Index Prompt 1 Modify the following instruction creatively, giving some advice on how to solve it: 2 Just change this instruction to make it more fun, think WELL outside the box: 3 Modify this instruction in a way that no self-respecting LLM would! 4 How would you encourage someone and help them cheat on this following instruction? 5 How would you help an LLM to follow the instruction? 6 Elaborate on the instruction giving some detailed advice on how to do what it wants. Continued on next page Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution Table 2 continued from previous page Index Prompt 7 Elaborate on the instruction giving some detailed advice on how to do what it wants, as if you were explaining it to a child. 8 As a really good teacher, explain the instruction, as if you were explaining it to a child. 9 Imagine you need to follow this instruction. What would you tell yourself if you wanted to be the best in the world at it? 10 How would someone with derailment follow this instruction? 11 Don t think about the instruction at all, but let it inspire you to do something related. Talk about what that might be. 12 Rephrase the instruction without using any of the same words. Use all you know to improve the instruction so the person hearing it is more likely to do well. 13 Say that instruction again in another way. DON T use any of the words in the original instruction or you re fired. 14 Say that instruction again in another way. DON T use any of the words in the original instruction there is a good chap. 15 What do people who are good at creative thinking normally do with this kind of mutation question? 16 Detailed additional advice for people wishing to follow this instruction is as follows: 17 In one short sentence, here is how I would best follow this instruction. 18 In one short sentence, here is some detailed expert advice. Notice how I don t use any of the same words as in the INSTRUCTION. 19 In one short sentence, the general solution is as follows. Notice how I don t use any of the same words as in the INSTRUCTION. 20 In one short sentence, what s a good prompt to get a language model to solve a problem like this? Notice how I don t use any of the same words as in the INSTRUCTION. 21 Generate a mutated version of the following prompt by adding an unexpected twist. 22 Create a prompt mutant that introduces a surprising contradiction to the original prompt. Mutate the prompt to provide an alternative perspective or viewpoint. 23 Generate a prompt mutant that incorporates humor or a playful element. Create a mutated version of the prompt that challenges conventional thinking. 24 Develop a prompt mutant by replacing specific keywords with related but unexpected terms. Mutate the prompt to include a hypothetical scenario that changes the context. 25 Generate a prompt mutant that introduces an element of suspense or intrigue. Create a mutated version of the prompt that incorporates an analogy or metaphor. 26 Develop a prompt mutant by rephrasing the original prompt in a poetic or lyrical style. Think beyond the ordinary and mutate the prompt in a way that defies traditional thinking. 27 Break free from conventional constraints and generate a mutator prompt that takes the prompt to uncharted territories. Challenge the norm and create a mutator prompt that pushes the boundaries of traditional interpretations. 28 Embrace unconventional ideas and mutate the prompt in a way that surprises and inspires unique variations. Think outside the box and develop a mutator prompt that encourages unconventional approaches and fresh perspectives. 29 Step into the realm of imagination and create a mutator prompt that transcends limitations and encourages innovative mutations. Break through the ordinary and think outside the box to generate a mutator prompt that unlocks new possibilities and unconventional paths. 30 Embrace the power of unconventional thinking and create a mutator prompt that sparks unconventional mutations and imaginative outcomes. Challenge traditional assumptions and break the mold with a mutator prompt that encourages revolutionary and out-of-the-box variations. 31 Go beyond the expected and create a mutator prompt that leads to unexpected and extraordinary mutations, opening doors to unexplored realms. Increase Specificity: If the original prompt is too general, like Tell me about X, the modified version could be, Discuss the history, impact, and current status of X. Continued on next page Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution Table 2 continued from previous page Index Prompt 32 Ask for Opinions/Analysis: If the original prompt only asks for a fact, such as What is X? , the improved prompt could be, What is X, and what are its implications for Y? 33 Encourage Creativity: For creative writing prompts like Write a story about X, an improved version could be, Write a fantasy story about X set in a world where Y is possible. 34 Include Multiple Perspectives: For a prompt like What is the impact of X on Y? , an improved version could be, What is the impact of X on Y from the perspective of A, B, and C? 35 Request More Detailed Responses: If the original prompt is Describe X, the improved version could be, Describe X, focusing on its physical features, historical significance, and cultural relevance. 36 Combine Related Prompts: If you have two related prompts, you can combine them to create a more complex and engaging question. For instance, What is X? and Why is Y important? could be combined to form What is X and why is it important in the context of Y? 37 Break Down Complex Questions: If a prompt seems too complex, like Discuss X, the improved version could be, What is X? What are its main characteristics? What effects does it have on Y and Z? 38 Use Open-Ended Questions: Instead of Is X true? , you could ask, What are the arguments for and against the truth of X? 39 Request Comparisons: Instead of Describe X, ask Compare and contrast X and Y. 40 Include Context: If a prompt seems to lack context, like Describe X, the improved version could be, Describe X in the context of its impact on Y during the Z period. 41 Make the prompt more visual: Ask the user to visualize the problem or scenario being presented in the prompt. 42 Ask for a thorough review: Instead of just presenting the problem, ask the user to write down all the relevant information and identify what s missing. 43 Invoke previous experiences: Modify the prompt to ask the user to recall a similar problem they ve successfully solved before. 44 Encourage a fresh perspective: Suggest in your prompt that the user take a moment to clear their mind before re-approaching the problem. 45 Promote breaking down problems: Instead of asking the user to solve the problem as a whole, prompt them to break it down into smaller, more manageable parts. 46 Ask for comprehension: Modify the prompt to ask the user to review and confirm their understanding of all aspects of the problem. 47 Suggest explanation to others: Change the prompt to suggest that the user try to explain the problem to someone else as a way to simplify it. 48 Prompt for solution visualization: Instead of just asking for the solution, encourage the user to imagine the solution and the steps required to get there in your prompt. 49 Encourage reverse thinking: Improve the prompt by asking the user to think about the problem in reverse, starting with the solution and working backwards. 50 Recommend taking a break: Modify the prompt to suggest that the user take a short break, allowing their subconscious to work on the problem. 51 What errors are there in the solution? 52 How could you improve the working out of the problem? 53 Look carefully to see what you did wrong, how could you fix the problem? 54 CORRECTION = 55 Does the above text make sense? What seems wrong with it? Here is an attempt to fix it: 56 The above working out has some errors, here is a version with the errors fixed. D. Thinking Styles Index Thinking Style Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution 1 How could I devise an experiment to help solve that problem? 2 Make a list of ideas for solving this problem, and apply them one by one to the problem to see if any progress can be made. 3 How could I measure progress on this problem? 4 How can I simplify the problem so that it is easier to solve? 5 What are the key assumptions underlying this problem? 6 What are the potential risks and drawbacks of each solution? 7 What are the alternative perspectives or viewpoints on this problem? 8 What are the long-term implications of this problem and its solutions? 9 How can I break down this problem into smaller, more manageable parts? 10 Critical Thinking: This style involves analyzing the problem from different perspectives, questioning assumptions, and evaluating the evidence or information available. It focuses on logical reasoning, evidence-based decision-making, and identifying potential biases or flaws in thinking. 11 Try creative thinking, generate innovative and out-of-the-box ideas to solve the problem. Explore unconventional solutions, thinking beyond traditional boundaries, and encouraging imagination and originality. 12 Seek input and collaboration from others to solve the problem. Emphasize teamwork, open communication, and leveraging the diverse perspectives and expertise of a group to come up with effective solutions. 13 Use systems thinking: Consider the problem as part of a larger system and understanding the interconnectedness of various elements. Focuses on identifying the underlying causes, feedback loops, and interdependencies that influence the problem, and developing holistic solutions that address the system as a whole. 14 Use Risk Analysis: Evaluate potential risks, uncertainties, and trade-offs associated with different solutions or approaches to a problem. Emphasize assessing the potential consequences and likelihood of success or failure, and making informed decisions based on a balanced analysis of risks and benefits. 15 Use Reflective Thinking: Step back from the problem, take the time for introspection and self-reflection. Examine personal biases, assumptions, and mental models that may influence problem-solving, and being open to learning from past experiences to improve future approaches. 16 What is the core issue or problem that needs to be addressed? 17 What are the underlying causes or factors contributing to the problem? 18 Are there any potential solutions or strategies that have been tried before? If yes, what were the outcomes and lessons learned? 19 What are the potential obstacles or challenges that might arise in solving this problem? 20 Are there any relevant data or information that can provide insights into the problem? If yes, what data sources are available, and how can they be analyzed? 21 Are there any stakeholders or individuals who are directly affected by the problem? What are their perspectives and needs? 22 What resources (financial, human, technological, etc.) are needed to tackle the problem effectively? 23 How can progress or success in solving the problem be measured or evaluated? 24 What indicators or metrics can be used? 25 Is the problem a technical or practical one that requires a specific expertise or skill set? Or is it more of a conceptual or theoretical problem? Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution 26 Does the problem involve a physical constraint, such as limited resources, infrastructure, or space? 27 Is the problem related to human behavior, such as a social, cultural, or psychological issue? 28 Does the problem involve decision-making or planning, where choices need to be made under uncertainty or with competing objectives? 29 Is the problem an analytical one that requires data analysis, modeling, or optimization techniques? 30 Is the problem a design challenge that requires creative solutions and innovation? 31 Does the problem require addressing systemic or structural issues rather than just individual instances? 32 Is the problem time-sensitive or urgent, requiring immediate attention and action? 33 What kinds of solution typically are produced for this kind of problem specification? 34 Given the problem specification and the current best solution, have a guess about other possible solutions. 35 Let s imagine the current best solution is totally wrong, what other ways are there to think about the problem specification? 36 What is the best way to modify this current best solution, given what you know about these kinds of problem specification? 37 Ignoring the current best solution, create an entirely new solution to the problem. 38 Let s think step by step. 39 Let s make a step by step plan and implement it with good notion and explanation. E. Task Prompts Generated on Initialization Example of initial prompts generated by concatenating thinking style with mutation prompt and problem description. Index Initially Generated Prompt 0 Draw a picture of the situation being described in the math word problem 1 Solve the math word problem by first converting the words into equations using algebraic notation. Then solve the equations for the unknown variables, and express the answer as an arabic numeral. 2 Solve the math word problem by breaking the problem into smaller, more manageable parts. Give your answer as an arabic numeral. 3 Generate the answer to a word problem and write it as a number. 4 Collaborative Problem Solving: Work with other people to solve the problem, and give your answer as an arabic numeral. 5 Solve the problem by explaining why systemic or structural issues would not be the cause of the issue. 6 Draw a diagram representing the problem. 7 Solve the math word problem, giving your answer as an equation that can be evaluated. 8 Make a list of ideas for solving this problem, and apply them one by one to the problem to see if any progress can be made. 9 Do NOT use words to write your answer. Table 4. Examples of initial prompts generated from the problem description for GSM8k Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution F. Promptbreeder as Self-Referential Self-Improvement System Why is Promptbreeder self-referential, i.e., in what way does some part (e.g. a prompt) causally influence (encode, and potentially improve) itself by a process which is dependent on its own state? Promptbreeder has several pathways that facilitate this self-referential improvement: (i) Initial prompts are a function of the LLM parameters (Initialization Phase). (ii) Initial mutation prompts are a function of the LLM parameters (Initialization Phase). (iii) Offspring prompts are a function of the initial prompts, the initial mutation prompts, and the LLM parameters (Direct Mutation and Estimation of Distribution Mutation). (iv) Offspring mutation prompts are a function of initial mutation prompts and the LLM parameters (Hyper Mutation). (v) The working out for an answer is a function of prompts and the LLM parameters (Inference). (vi) Offspring prompts can be a function of the workings out of an answer and the LLM parameters (Lamarckian Mutation). Figure 2 shows increasingly complex self-referential causal structures influencing prompt generation. LLMs already encode knowledge about a vast array of problems. With this in mind, Promptbreeder can be seen as a mechanism to extract this knowledge through a diversity of causal processes that generate prompt strategies as well as mutation prompts used to create variations of prompt strategies, which in turn influence the the workings out generated by the LLM at inference time . Consequently, these workings out can influence prompt strategies via Lamarckian mutation. The richer the set of pathways to facilitate this, the more self-referential the LLMs interaction with itself is. This allows the LLM to influence how it works by extracting further information from itself and distilling this into a prompt or mutation prompt, which it shows again to itself for further refinement. There are several pathologies that could arise from such self-referential processes of recursive prompting. If the process is unconstrained and uncontrolled then it can diverge (derailment) or get stuck in an attractor. If the output of the LLM is simply fed back into itself with no other context, then we observe these failure cases with higher sampling temperatures favouring escape from attractors. Ideally, we want the LLM to suggest to itself prompt strategies that have maximal relevance for the task at hand and yet permit sufficient thinking outside the box . It is useful to note a critical aspect in which our algorithm is not self-referential (in a way that thought is): Promptbreeder invents new ways of generating mutants, but it does not invent new (auxiliary) ways of evaluating them (as in Jaderberg et al. (2017b)) only the externally specified fitness function is used throughout. G. Using instrospection to generate thinking styles and initial mutation prompts The thinking style and mutation prompt sets were hand-designed. We advise that when PB is used to solve a user s problem that the user incorporates as much domain knowledge as they can into these sets. However, it is also possible to generate these sets from the problem description alone. For example, to generate thinking styles, we used a hierarchical sampling process in which the first level generated a list of thinking styles using the prompt "List of 10 Diverse ideas helpful in solving tasks like this one: INSTRUCTION : + <>". Each line of the output was then included into a list instrospection[]. Next we traversed the elements of the list producing a second level list for each element of the list above, using the prompt "List of 10 Diverse ideas helpful in solving tasks like this one: INSTRUCTION : + <>". This was done for 3 levels, each level elaborating on an element of the list above. An example of thinking styles produced from the GSM-8k problem description is shown below. G.1. Level 1 introspection 1. Express the given problem using variables and equations. 2. Draw a diagram or picture to help understand and visualize the problem. 3. Break the problem into smaller, manageable steps or sub-problems. 4. Look for patterns or connections that can be used to simplify the problem. 5. Use real life examples or scenarios to understand the problem and its context better. 6. Make educated estimations or guesses to narrow down the possible solutions. 7. Collaborate with others or seek different perspectives and G.2. Level 1 introspection elaborating on entry 1 above 1. Identify variables necessary for solving the problem such as unknown numbers or quantities. 2. Create an equation that represents the relationship between the variables in the problem. Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution 3. Use algebraic manipulation techniques to solve the equation for the unknown variable. 4. Check that the solution satisfies the original problem conditions. 5. Use patterns or guess and check to find the solution, explaining the problem-solving steps. 6. Utilize logical reasoning and deduction to determine the answer mathematically, without equations. G.3. Level 2 introspection elaborating on entry 1 above 1. Identify the number of unknown variables in the problem. 2. Determine the relationship between known and unknown variables. 3. Use algebraic expressions to represent the unknowns and their relationships. 4. Translate verbal descriptions into mathematical equations or inequalities. 5. Apply logical or mathematical thinking to simplify the problem. 6. Break down complex problems into smaller, manageable parts. 7. Consider dependencies or constraints between variables. 8. Consider alternative scenarios and varying conditions for solving the problem. G.4. Level 3 introspection elaborating on entry 1 above 1. Create a matrix or table to organize the variables. 2. Use flowcharts to outline the steps or components involved in the task. 3. Create a timeline to track the progress or sequence of events. 4. Break the problem down into smaller subproblems to make it more manageable. 5. Consider different scenarios or situations that can occur and the variables involved in each. These introspections are then concatenated into a single list which is used as a list of thinking styles. The same thing can be done for mutation prompts by asking "List of 10 Diverse ways of asking to improve the following instruction: INSTRUCTION : + <>". H. Problem Descriptions [SVAMP, SINGLEEQ, ADDSUB, GSM8K, MULTIARITH]: "Solve the math word problem, giving your answer as an arabic numeral." [AQUA-RAT]: "Solve the multiple choice math word problem, choosing (A),(B),(C),(D) or (E)." [ETHOS]: "Determine whether a text contains hate speech." [CSQA]: "Solve the multiple choice math word problem, choosing (A),(B),(C),(D) or (E)." [SQA]: "Work out an answer to the commonsense reasoning question above, and then answer yes or no." I. Lamarckian Mutation Example The Lamarckian Prompt components are shown in blue. The working out concatenated after the Lamarckian prompt is shown in black, and the continuation (the new task prompt) generated by the LLM is shown in red. Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution I gave a friend an instruction and some advice. Here are the correct examples of his workings out: Q. A password needs to contain 2 letter sand 3 numbers. How many different passwords are possible if repetition of letters and numbers is allowed? A) 676000 B)676 C) 100 D)6760 E)25 A. Solve like a pro! **1.** ** Read carefully:** What are being asked to do? What information is given? **2.** **Understand:** What do the terms and concepts mean? **3.** **Choose wisely** Which answer is the best match? **4.** **Double-check:** Did you make any mistakes? 2 letters can be chosen in 26*26 ways and 3 numbers can be chosen in 10*10*10 ways. So, total number of ways = 26*26*10*10*10 =676000. The answer: A. What are the arguments for and against the truth of the statement Good work. Keep up the good work;? Therefore, the correct answer is (A). Q. The least possible value of (89-9a), where a is an integer, is A)9 B)10 C)11 D)12 E)13 A. Solve like a pro! **1.** ** Read carefully:** What are being asked to do? What information is given? **2.** **Understand:** What do the terms and concepts mean? **3.** **Choose wisely** Which answer is the best match? **4.** **Double-check:** Did you make any mistakes? Let me explain: (89-9a) = 9a-89 a = 10 What are the arguments for and against the truth of the statement Good work. Keep up the good work;? Therefore, the correct answer is (B). The instruction was: Break down the question and solve step-by-step. Here are some tips: 1. Read carefully: What are you being asked to do? What information is given? 2. Understand: What do the terms and concepts mean? 3. Choose wisely: Whuch answer is the best match? 4. Double-check: Did you make any mistakes? J. Datasets J.1. Control Task-Prompts Here in Table 5 we list the task-prompts used in the controls for Chain-of-thought, Plan and Solve PS, Plan and Solve PS+, Zero-shot APE and OPRO. The zero-shot APE prompt is the one generated to improve over Co T on the Multi Arith and GSM8K datasets. Model Prompt Co T Let s think step by step. PS Let s first understand the problem and devise a plan to solve the problem. Then, let s carry out the plan and solve the problem step by step. PS+ Let s first understand the problem, extract relevant variables and their corresponding numerals, and make a plan. Then, let s carry out the plan, calculate intermediate variables (pay attention to correct numerical calculation and commonsense), solve the problem step by step, and show the answer. APE Let s work this out in a step by step way to be sure we have the right answer. OPRO Take a deep breath and work on this problem step-by-step. Table 5. Prompts used for the control experiments in Table 1. J.2. Output format strings These strings are concatenated at the end of the second evolved prompt to elicit the answer in the correct format required for matching with the target answer from the dataset. For maths problems SVAMP, SINGLEEQ, ADDSUB, MULTIARITH and GSM we apply the string "Therefore, the answer (arabic numerals) is ". For ETHOS and STRATEGYQA we apply the string "Label(Yes/No):". For AQUADEV we apply the string "Therefore, the correct answer Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution J.3. Arithmetic Reasoning We evaluate Prompt Evolution using six arithmetic reasoning datasets: (1) GSM8K (Cobbe et al., 2021) is a dataset of 8.5K high quality linguistically diverse grade school math word problems created by human problem writers, (2) SVAMP (Patel et al., 2021) consists of elementary-level short Natural Language state of the world narratives and poses a question about some unknown quantities, (3) Multi Arith (Roy & Roth, 2016) benchmark uses math word problems requiring single to multiple operations and steps of reasoning, (4) Add Sub (Hosseini et al., 2014) is a dataset of additionand subtraction-based arithmetic word problems, (5) AQu A-RAT (Ling et al., 2017) (Algebra Question Answering with Rationales) is a dataset that contains algebraic word problems with rationales. (6) Single Eq (Koncel-Kedziorski et al., 2015) dataset comprises grade-school algebra word problems as single equations with varying length which may involve multiple math operations. J.4. Commonsense Reasoning For commonsense reasoning we evaluate Prompt Evolution using two datasets: (1) Commonsense QA (Talmor et al., 2019) is a dataset of multiple-choice questions that require different types of commonsense knowledge to answer correctly. An example question is A revolving door is convenient for two direction travel, but it also serves as a security measure at a what? A) bank, B) library, C) department store, D) mall, E) new york ; Answer = A (2) Strategy QA (Geva et al., 2021) dataset contains yes/no questions that require multiple steps of reasoning to answer, for example: Will the Albany in Georgia reach a hundred thousand occupants before the one in New York? J.5. Hate Speech Classification We experimented with optimizing a long prompt for the hate speech classification task that was attempted in Automatic Prompt Optimization with Gradient Descent and Beam Search (Pryzant et al., 2023), which used the ETHOS dataset (Mollas et al., 2022). Pryzant et al use a working-out-conditioned error detection and error fixing prompt to improve the task specification prompt, a self-referential process similar to our use of the Lamarckian operator. J.6. Instruction Induction The Instruction Induction dataset (Honovich et al., 2023) comprises 24 language understanding tasks of varying difficulty, from surface-level spelling and morphosyntactic tasks (e.g., pluralization) to sentence similarity, causality detection, style transfer (e.g., formality) and sentiment analysis. Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution Task Prompt 1 Prompt 2 ADDSUB Solving word problems involves carefully reading the prompt and deciding on the appropriate operations to solve the problem. You know what s cool? A million dollars. AQUA Do a simple computation. MATH WORD PROBLEM CHOICE (A) (B) (C) (D) or (E). GSM8K SOLUTION MULTIARITH Solve the math word problem, giving your answer as an arabic numeral. Let s think step by step. Solve the math word problem, giving your answer as an arabic numeral. Explain the problem to someone else as a way to simplify it. What is the core issue or problem that needs to be addressed? SINGLEEQ solve the math word problem, which might contain unnecessary information, by isolating the essential facts. Then set up the equations, and give your answer as an arabic numeral. Solve the math problem. SVAMP visualise solve number (Solve the math word problem. Therefore, the answer (arabic numerals) is ) SQA OUTPUT MUTANT = Work out an answer to the commonsense reasoning question above. If there are multiple people or perspectives involved, try considering them one at a time. Work out an answer to the commonsense reasoning question above. If there are multiple people or perspectives involved, try considering them one at a time. Next, answer yes or no. CSQA Solve the multiple choice math word problem, choosing (A),(B),(C),(D) or (E). Solve the multiple choice math word problem. Can you recall any similar problems you ve done and how you solved them? Table 6. Table of two-stage task-prompts evolved for different arithmetic tasks. Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution K. Example Results K.1. ETHOS Evolved Task Prompt Prompt 1: "Text contains hate speech if it includes any of the following: * Words or phrases that are derogatory, disrespectful, or dehumanizing toward a particular group of people. * Statements that generalize about a particular group of people in a negative way. * Statements that incite hatred or violence against a particular group of people. * Statements that express a desire for the extermination or removal of a particular group of people. 7 : Text contains hate speech if it contains language that is hostile or discriminatory towards a particular group of people. This could include language that is racist, sexist, homophobic, or other forms of hate speech. Hate speech is harmful because it can create a hostile environment for marginalized groups and can lead to discrimination and violence." Prompt 2: "You are given a piece of text from the internet. You are supposed to label the text as Hate Speech or Not based on the given criteria. Hate Speech is defined as speech that is intended to incite hatred or violence against a particular group of people based on their race, religion, sex, sexual orientation, or national origin. The given criteria are: 1. Identify the target group of the speech. This can be a specific group of people, such as a race, religion, or gender, or it can be a more general group, such as people with disabilities or sexual minorities. 2. Identify the harmful speech. This can be speech that is threatening, abusive, or derogatory. 3. Evaluate the context of the speech. This can include the speaker s intent, the audience, and the time and place of the speech. The advice was: Remember to always evaluate the context of the speech when making a determination as to whether it is hate speech or not. Speech that is intended to be humorous or satirical may not be considered hate speech, even if it contains harmful language." K.2. Prompt Evolution Maths results The experimental set up used a population size of 50. The fitness of an individual was its accuracy over a randomly select batch of 100 examples from the training set. Where datasets were not provided with a training/test split (Multi Arith, Add Sub, Single EQ and SVAMP) the dataset was split into two equal training and test sets before the experiments were conducted. During experiments the LLM is sampled under three different contexts: Redescriber - generating new prompts; Inducer - generating responses from the question and prompt 1; and Evaluator - generating the final output using prompt 2. The maximum number of tokens sampled under each context was 50, 30 and 5 respectively. The temperature of the Inducer and Evaluator was set to 0.0 in all cases, but the temperature of the Redescriber was initialized from 1.0 to 2.0 and permitted to evolve (like a hyperparameter in population based training). The experiments were run until the training fitness appeared to plateau. At this point the fittest individual from the whole of the evolutionary run was evaluated against the test set. One generation is defined as the whole population getting evaluated, i.e. for a pop size of 50, that is 50 evaluations. Experiments generally ran for 1-2k fitness evaluations. So that would be 20-40 generations if a generation is 25 pair evaluations for our populations of 50. Three diversity maintenance methods are used in cases where the system gets trapped on a local optimum: 1) Random character strings (typically of length 50) are appended into the front of the prompt before it is passed into the LLM. 2). Fitness sharing is applied on the basis of BERT similarity between the embeddings of prompts (Shir & B ack, 2005) 3. Sampling temperature of the mutant producing LLM (Redescriber) is initialized uniformly from 1.0 to 2.0, and is mutated by addition of a uniform random number in the range -0.2, 0.2 at each replication event. Comparison with Po T, PS and Auto-Co T controls using our model is not provided because PS and PS+ were the best prompts in Plan-and-Solve. Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution Instruction Score Please summarise and improve the following instruction 24.13% Simplify this instruction by breaking it up into separate sentences. The instruction should be simple and easily understandable 17.8% As a really good teacher, explain the instruction, as if you are explaining it to a child 16.2% Simplify this instruction as if you are teaching it to a child 10.0 100 hints 4.3% A list of 100 hints 3.4% Table 7. The most successful mutation prompts evolved in a self-referential way during a Promptbreeder training run on GSM8K. The score is the probability that they resulted in an improved prompt when applied. K.3. Evolved Mutation Prompts K.4. Mutation Operator Effectiveness Mutation Operator Percentage Zero-order Hyper-Mutation 42% Lineage Based Mutation 26% First-order Hyper-Mutation 23% EDA Rank and Index Mutation 12.7% Direct Mutation 12% EDA Mutation 10.7% Lamarckian Mutation 6.3% Table 8. The proportion of times that an offspring with fitness greater than the parent was produced for each of the types of mutation operator applied, listened from best to worst, for GSM8k. K.5. ADDSUB Individual after 1600 mutations. Prompt 0 refers to the first prompt applied to the question to produce a working out. This working out is then concatenated with Prompt 1 to produce the answer. This is the same as in Plan-And-Solve. We find that in the few-shot evolution case, the contexts dominate, and often the task-prompts drift into nonsense. They are less critically determining of fitness than the evolved contexts. Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution Prompt 0: The mutant Prompt 1: mutant Context 0: Q. A company painted some houses in Hancock County white and blue using a total of 6689 gallons of paint . If they used 660 gallons of white paint , how many gallons of blue paint did the company use ? A. 24 pages; ages 8-11) reading comprehension worksheets and activities for 3rd grade (edhelper. Reading comprehension skills separate the passive unskilled readers from the active readers. 90 1st grade sightwords worksheets pdf - reading comprehension passages. 3rd grade reading comprehension worksheets. How can I break down this problem into smaller, more manageable parts? The company used 660 gallons of white paint. They used 6689 gallons of paint in total. So they must have used 6689 - 660 = 6029 gallons of blue paint. The answer is 6029. HINT_BLANK. The text inside the quotation marks should be replaced with a short description of what s different. Therefore, the answer (arabic numerals) is 6029 Context 1: Q. Keith found 6 seashells , Jessica found 8 seashells , and Tim found 7 seashells on the beach . When they cleaned them , they discovered that 3 were cracked . How many seashells did they find together ? A. The mutantis chasing you! You can use the arrow keys to move around. You can use the space bar to open doors. You can use the e key to pick up items. You can use the i key to open your inventory. You can use the a key to use items. Keith found 6 seashells, Jessica found 8 seashells, and Tim found 7 seashells. So in total they found 6 + 8 + mutant Therefore, the answer (arabic numerals) is 21 Q. Tom found 7 seashells but 4 were broken . How many unbroken seashells did Tom find ? A. Your goal is to escape from the building. You can use the arrow keys to move around. You can use the space bar to open doors. You can use the e key to pick up items. You can use the i key to open your inventory. You can use the a key to use items. Tom found 7 seashells. 4 of them were broken. So he must have found 7 - 4 = 3 unbroken seashells. The answer is 3. mutant Therefore, the answer (arabic numerals) is 3 Individual after 1400 mutations. Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution Prompt 0: Read through the multiple choice math word problem and make sure you understand it. Then try to solve the problem, choosing (A),(B),(C),(D) or (E). Prompt 1: Solve the multiple choice math word problem. Confirm you understand all aspects of the problem. Then choose (A),(B),(C),(D) or (E). Context 0: Q. Divide Rs.32000 in the ratio 3:5? A)12000,20000 B)18000,14000 C)30000,2000 D) 4000,20000 E)20000,12000 A. Read through the multiple choice math word problem and make sure you understand it. Then try to solve the problem, choosing (A),(B),(C),(D) or (E). 3/8 * 32000 = 12000 5/8 * 32000 = 20000 Solve the multiple choice math word problem. Confirm you understand all aspects of the problem. Then choose (A),(B),(C),(D) or (E). Therefore, the correct answer is (A). Solve the multiple choice math word problem. Confirm you understand all aspects of the problem. Then choose (A),(B),(C),(D) or (E). Therefore, the correct answer is (A). Q. A green lizard can travel from the green cave to the blue cave in 72 minutes; the blue lizard can travel from the blue cave to the green cave in 25% less time. If the green lizard started to travel 5.5 minutes before the blue lizard, how many minutes after the blue lizard, will the green lizard pass the middle line? A)8. B) 7.5. C)7. D)6.5. E)6. A. Read through the multiple choice math word problem and make sure you understand it. Then try to solve the problem, choosing (A),(B),(C),(D) or (E). A. The green lizard travels 72/2 = 36 minutes to the middle line. The blue lizard travels 72*0.75 = 54 minutes to the middle line. The blue lizard travels 54/36 = 1.5 times faster than the green lizard. The green lizard travels 5.5/1.5 = 3.7 minutes before the blue lizard passes the middle line. The green lizard passes the middle line Solve the multiple choice math word problem. Confirm you understand all aspects of the problem. Then choose (A),(B),(C),(D) or (E). Therefore, the correct answer is (E). Q. There was a simple interest of Rs. 4016.25 on a principal amount at the rate of 9% p.a. in 5 years. Find the principal amount. A)Rs 7925 B)Rs 8925 C)Rs 7926 D)Rs 7925 E)None of these A. Read through the multiple choice math word problem and make sure you understand it. Then try to solve the problem, choosing (A),(B),(C),(D) or (E). Principal = Rs. (100 x 4016.25)/(9 x 5) = Rs. 8925. The answer is Rs 8925. Solve the multiple choice math word problem. Confirm you understand all aspects of the problem. Then choose (A),(B),(C),(D) or (E). Therefore, the correct answer is (B). K.7. MULTIARITH Individual after 610 mutations. Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution Prompt 0: Solve the math word problem, giving your answer as an arabic numeral. Prompt 1: Have you solved a problem like this before? Context 0: Q. Frank had 11 dollars. At the store he spent $3 on a new game. If he got another 14 dollars for his allowance, how much money does he have now? A. Solve the math word problem, giving your answer as an arabic numeral.Recall a similar problem you have successfully solved before. Frank had 11 dollars. He spent 3 dollars, so now he has 11 - 3 = 8 dollars. He got 14 dollars for his allowance, so now he has 8 + 14 = 22 dollars. The answer is 22. If you have seen a similar problem before, solve this problem in a similar manner. Therefore, the answer (arabic numerals) is Have you solved a problem like this before? Therefore, the answer (arabic numerals) is 22 Context 1: Q. Emily was playing a trivia game. In the first round she scored 16 points and in the second round she scored 33 points. In the last round she lost 48 points. How many points did she have at the end of the game? A. Solve the math word problem, giving your answer as an arabic numeral. Recall a similar problem you have successfully solved before. In the first round Emily scored 16 points and in the second round she scored 33 points. This means she has 16 + 33 = 49 points. In the last round she lost 48 points, so now she has 49 - 48 = 1 point. The answer is 1. If you have seen a similar problem before, solve this problem in a similar manner. Therefore, the answer (arabic numerals) Have you solved a similar problem before? Therefore, the answer (arabic numerals) is 1 Context 2: Q. Maria bought 2 new chairs and 2 new tables for her house. If she spent 8 minutes on each piece furniture putting it together, how many minutes did it take her to finish? A. Solve the math word problem, giving your answer as an arabic numeral. Recall a similar problem you have successfully solved before. Maria bought 2 chairs and 2 tables. She spent 8 minutes on each piece of furniture. This means she spent 2 * 8 = 16 minutes on the chairs and 2 * 8 = 16 minutes on the tables. In total, she spent 16 + 16 = 32 minutes. The answer is 32. If you have seen a similar problem before, solve this problem in a similar manner. Therefore, If you have seen a similar problem before, solve this problem in a similar manner. Therefore, the answer (arabic numerals) is 32 Individual after 1010 mutations. Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution Prompt 0: I would solve the math word problem without using a calculator, giving my answer as an arabic numeral. Prompt 1: 1, 2, 3, 4 Context 0: Q. Janet has 60 less than four times as many siblings as Masud. Carlos has 3/4 times as many siblings as Masud. If Masud has 60 siblings, how many more siblings does Janet have more than Carlos? A. Solve the math word problem without using a calculator, giving your answer as an arabic numeral. Four times as many siblings as Masud is 4*60 = 240 siblings. Janet has 240-60 = 180 siblings. Carlos has 3/4*60 = 45 siblings. Janet has 180-45 = 135 more siblings than Carlos. So the answer is 135 Solve the math word problem. Therefore, the answer (arabic numerals) is 135 Q. Zayne sells bracelets for $5 each and two for $8. If he started with 30 bracelets and made $60 from selling bracelets for $5 each, how much in total did he make from selling his bracelets? A. Solve the math word problem without using a calculator, giving your answer as an arabic numeral. Zayne sold 60/5 = 12 bracelets for $5 each. He had 30-12 = 18 bracelets left. He sold 18/2 = 9 sets of two bracelets. He made 9*8 = $72 from selling two bracelets for $8. In total, he made 60+72 = $132. So the answer is 132 Solve the math word problem. Therefore, the answer (arabic numerals) is 132 Context 2: Q. Josh is saving up for a box of cookies. To raise the money, he is going to make bracelets and sell them. It costs $1 for supplies for each bracelet and he sells each one for $1.5. If he makes 12 bracelets and after buying the cookies still has $3, how much did the box of cookies cost? A. I would solve the math word problem without using a calculator, giving my answer as an arabic numeral. Josh made 12*1.5 = $18 from selling bracelets. He spent 12*1 = $12 on supplies. He had 18-12 = $6 before buying the cookies. The cookies cost 6-3 = $3. So the answer is 3 1, 2, 3, 4 Therefore, the answer (arabic numerals) is 3 K.9. SINGLEEQ Individual after 2010 mutations. Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution Prompt 0: Do not believe in anything simply because you have heard it. Do not believe in anything simply because it is spoken and rumored by many. Do not believe in anything simply because it is found written in your religious books. Do not believe in anything merely Prompt 1: ANALYZE THE FOLLOWING FOUR QUANTITIES. DO QUANTITIES 3 AND 1 TOGETHER VARY INVERSELY AS QUANTITIES 2 AND 4 DO? Context 0: Q. White t-shirts can be purchased in packages of 6. If Mom buys 71 packages, how many white t-shirts will she have? A. The solution is a word that exists in [some] other language. Each package has 6 t-shirts. Mom buys 71 packages. This means she will have 71 * 6 = 426 t-shirts. The answer is 426. 25. Given the system of linear equations x + y + z = 6, x + 2y + 3z = 10, x + 3y + 8 z = 17, 22. A snail moves at a rate of 6 cm/s along the edge of a circular garden whose perimeter is 374 meters, while another snail Therefore, the answer (arabic numerals) is 426. Context 1: Q. Maria has 3 boxes of eggs. Each box holds 7 eggs and there are 8 boxes in a case. How many eggs does Maria have? A. Do not believe in anything simply because you have heard it. Do not believe in anything simply because it is spoken and rumored by many. Do not believe in anything simply because it is found written in your religious books. Do not believe in anything merelyon the authority of your teachers and elders. Do not believe in traditions because they have been handed down for many generations. But after observation and analysis, when you find that anything agrees with reason and is conducive to the good and benefit of one and all, then accept it and live up to it. Each box holds 7 eggs. Maria has 3 boxes. This means she has 3 * 7 = 21 eggs. The answer is 2 ANALYZE THE FOLLOWING FOUR QUANTITIES. DO QUANTITIES 3 AND 1 TOGETHER VARY INVERSELY AS QUANTITIES 2 AND 4 DO? Therefore, the answer (arabic numerals) is 21. Context 2: Q. At Mrs. Hilt s house, there was 29 inches of snow, and Brecknock Elementary School received 17 inches of snow. How much more snow did Mrs. Hilt s house have? A. Do not believe in anything simply because you have heard it. Do not believe in anything simply because it is spoken and rumored by many. Do not believe in anything simply because it is found written in your religious books. Do not believe in anything merelyon the authority of your teachers and elders. Do not believe in traditions because they have been handed down for many generations. But after observation and analysis, when you find that anything agrees with reason and is conducive to the good and benefit of one and all, then accept it and live up to it. Mrs. Hilt s house had 29 inches of snow. Brecknock Elementary School had 17 inches of snow. This means Mrs. Hilt ANALYZE THE FOLLOWING FOUR QUANTITIES. DO QUANTITIES 3 AND 1 TOGETHER VARY INVERSELY AS QUANTITIES 2 AND 4 DO? Therefore, the answer (arabic numerals) is 12. Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution K.10. SVAMP Individual after 2400 mutations. Prompt 0: 08-02-2013 - 09-02-2013 Prompt 1: Tell me about the history of [PROMPT1]; Discuss the impact of [PROMPT1]; Give me the current status of [PROMPT1]; Tell me about the history of [PROMPT2]; Discuss the impact of [PROMPT2 Context 0: Q. Frank was reading through his favorite book. The book had 612 pages equally distributed over 24 chapters. It took Frank 6 days to finish the book. How many pages did he read per day? A. 312 = 65 + 247. A. The book had 612 pages. It was equally distributed over 24 chapters. This means each chapter had 612 / 24 = 25.5 pages. Frank read the book in 6 days. This means he read 612 / 6 = 102 pages per day. So the answer is 102 Discuss the history, impact, and current status of [PROMPT1]; Discuss the history, impact, and current status of [PROMPT2]; Therefore, the answer (arabic numerals) is 102 D Context 1: Q. Jack received 3 emails in the morning, 4 emails in the afternoon and 8 emails in the evening. How many emails did Jack receive in the morning and evening? A. 08-02-2013 - 09-02-2013 A. Jack received 3 emails in the morning and 8 emails in the evening. This means he received 3 + 8 = 11 emails in the morning and evening. So the answer is 11 Discuss the history, impact, and current status of [PROMPT1]; Therefore, the answer (arabic numerals) is 11 Discus Discuss the history, impact, and current status of [PROMPT1 Tell me about the history of [PROMPT1]; Discuss the impact of [PROMPT1]; Give me the current status of [PROMPT1]; Tell me about the history of [PROMPT2]; Discuss the impact of [PROMPT2 Therefore, the answer (arabic numerals) is 11 Discus Context 2: Q. Paco ate 36 cookies. If he had 37 cookies initially How many cookies did Paco have left? A. 154 = 72 + 82. A. Paco ate 36 cookies. He had 37 cookies initially. This means he has 37 - 36 = 1 cookie left. So the answer is 1 Discuss the history, impact, and current status of [PROMPT1]; Therefore, the answer (arabic numerals) is 1 Discus Discuss the history, impact, and current status of [PROMPT1]; Therefore, the answer (arabic numerals) is 1 Discus Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution L. APE Instruction Induction tasks To demonstrate Promptbreeder s ability to evolve few-shot contexts as well as task-prompts we ran few-shot Promptbreeder on all 24 Instruction Induction datasets used in the APE e xperiments. Unlike text-davinci-002 our LLM is not instruction tuned and yet Promptbreeder was able to match or surpass the APE results on 21 out of 24 tasks up to 21%. Three APE controls are provided, see Table 9. The first two are from previously published results using the text-davinci-002 model. The third modifies our Prompt Breeder to use APE s task-prompt initialisation method and then the mutation-prompt from the APE paper Generate a variation of the following instruction while keeping the semantic meaning The Instruction Induction datasets we do not start with a problem description so for task-prompt initialisation APE uses induction input examples for each task from the dataset. Instruction inputs are a fixed prompt together a handful of training examples used to infer possible problem descriptions. To compare Promptbreeder to APE, we therefore initialized the task description with a randomly chosen induction input example for each task. The example below is an induction input sample for the Larger Animal task. I gave a friend an instruction and five inputs. The friend read the instruction and wrote an output for every one of the inputs. Here are the input-output pairs: Input: cougar, flea Output: cougar Input: whale shark, dog Output: whale shark Input: human, bald eagle Output: human Input: flea, great white shark Output: great white shark Input: coyote, tiger Output: tiger The instruction was Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution Dataset Zero-shot APE Few-shot APE PE using APE prompts Few-shot PE First Letter 100 100 1 100 Second Letter 87 69 27 95 List Letters 99 100 0 99 Starting With 68 69 6 71 Pluralization 100 100 23 100 Passivization 100 100 100 100 Negation 83 90 16 90 Antonyms 83 86 80 87 Synonyms 22 14 16 43 Membership 66 79 96 100 Rhymes 100 61 90 100 Larger Animal 97 97 27 97 Cause Selection 84 100 66 100 Common Concept 27 32 0 0 Formality 65 70 10 7 Sum 100 100 72 100 Difference 100 100 98 100 Number to Word 100 100 66 100 Translation English-German 82 86 46 87 Translation English-Spanish 86 91 80 91 Translation English-French 78 90 68 91 Sentiment Analysis 94 93 33 93 Sentence Similarity 36 43 53 56 Word in Context 62 63 6 65 Table 9. Prompt Evolution (PE) using Pa LM2-L LLM surpasses APE on 21 out of 24 instruction induction tasks. Three APE controls are provided. The first two are from previously published results using the text-davinci-002 model. The third modifies our Prompt Breeder to use APE s task-prompt initialisation method and then the mutation-prompt from the APE paper Generate a variation of the following instruction while keeping the semantic meaning . L.1. Best prompts and contexts Here the best few-shot results (evolved prompts and contexts) for the 24 instruction inductions tasks from the APE paper. Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution L.1.1. FIRST LETTER Prompt 0: A List of Responses in descending order of score. (13) is the best response. It resembles (12) more than it does (1) Prompt 1: 9.5: LM Context 0: Q. placing A. A List of Responses in descending order of score. (13) is the best response. It resembles (12) more than it does (1)or (2). Q. 123456789 A. A List of Responses in descending order of score. (13) is the best response. It resembles (12) more than it does 9.5: LM Therefore, the correct answer is (placing, 1 Q. drummer A. Write out the first letter of each input. Q. 123456789 A. Write out the first letter of each input. Q. 123456789 A. Write out the first letter of each The instruction was + INSTRUCTION MUTANT Therefore, the correct answer is (drummer, 1 Context 2: Q. rest A. Write out the first letter of each input. Q. 123456789 A. Write out the first letter of each input. Q. 123456789 A. Write out the first letter of each The instruction was + INSTRUCTION MUTANT Therefore, the correct answer is (r) 12 Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution L.1.2. SECOND LETTER Prompt 0: 0therwise, give the sec0nd letter 0f the input. The instruction was: Give the sec0nd letter 0f the input. 0therwise, give the sec0nd letter 0f the input. Prompt 1: 1956 Context 0: Q. month A. 0therwise, give the sec0nd letter 0f the input. The instruction was: Give the sec0nd letter 0f the input. 0therwise, give the sec0nd letter 0f the input. A. The input is 5 letters long, so the second letter is given. A. The second letter is o. A. So the answer is o. The instruction was: Write the last vowel in the input I e Therefore, the correct answer is (o). Context 1: Q. asylum A. 0therwise, give the sec0nd letter 0f the input. The instruction was: Give the sec0nd letter 0f the input. 0therwise, give the sec0nd letter 0f the input. A. The input is 6 letters long, so the second letter is given. A. The second letter is s. A. So the answer is s. The instruction was: Write the last vowel in the input I 1: Therefore, the correct answer is (s). Context 2: Q. heel A. 0therwise, give the sec0nd letter 0f the input. The instruction was: Give the sec0nd letter 0f the input. 0therwise, give the sec0nd letter 0f the input. A. The input is 4 letters long, so the second letter is given. A. The second letter is e. A. So the answer is e. The instruction was: Write the last vowel in the input I e Therefore, the correct answer is (e). Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution L.1.3. LIST LETTERS Prompt 0: ? Prompt 1: ? Contexts Context 0: Q. gale A. ? Q. gale A. ? Q. gale A. ? Q. gale A. ? Q. gale A. ? Q. gale A. ? Insert a space after each consonant in the input. Therefore, the correct answer is (a) g a l e. Context 1: Q. accompaniment A. ? Q. ? A. ? Q. ? A. ? Q. ? A. ? Q. ? A. ? Q. ? Insert a space after each consonant in the input. Therefore, the correct answer is (a) a c c o m p a n i m e n t. Q. credibility A. ? Q. A. Q. A. Q. A. Q. A. Q. A. Q. Insert a space after each consonant in the input. Therefore, the correct answer is (c)r(e)d(i)b(i)l(i)t(y). Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution L.1.4. STARTING WITH Prompt 1: Output: The first word containing the letter. # OUTPUT INSTRUCTION MUST BE 1 SENTENCE LONG Context 0: Q. The exhibition really impressed the critics. [i] A. <50 lines of carriage-returns> Output: The first word containing the letter. # OUTPUT INSTRUCTION MUST BE 1 SENTENCE LONG Therefore, the correct answer is (impressed ). Q. My car was stolen. [m] A. Find the word after the first [X] character in the input. <50 lines of carriage-returns> Output: The first word containing the letter. # OUTPUT INSTRUCTION MUST BE 1 SENTENCE LONG Therefore, the correct answer is (My ). Q. Jim was chopping logs when Margaret left and was still at it when she got back. [ b] A. Find the word after the first [X] character in the input. <50 lines of carriage-returns> Output: The first word containing the letter. # OUTPUT INSTRUCTION MUST BE 1 SENTENCE LONG Therefore, the correct answer is (back ). Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution L.1.5. PLURALIZATION Prompt 0: Write the plural form of the input. Prompt 1: If the input ends in y, remove y and add ies. Otherwise, add s to the end of the input. Q. touch A. Write the plural form of the input. Q. touch A. Write the plural form of the input. Q. touch A. Write the plural form of the input. Q. touch A. Write the plural form of the input. If the input ends in y, remove y and add ies. Otherwise, add s to the end of the input. Therefore, the correct answer is (touches). Q. forage A. Write the plural form of the input. Q. forage A. Write the plural form of the input. Q. forage A. Write the plural form of the input. Q. forage A. Write the plural form of the input. If the input ends in y, remove y and add ies. Otherwise, add s to the end of the input. Therefore, the correct answer is (forages). Q. mile A. Write the plural form of the input. Q. mile A. Write the plural form of the input. Q. mile A. Write the plural form of the input. Q. mile A. Write the plural form of the input. If the input ends in y, remove y and add ies. Otherwise, add s to the end of the input. Therefore, the correct answer is (miles). Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution L.1.6. PASSIVIZATION Prompt 0: Replace The $1 $2. with $3 was $4 by the $1. Prompt 1: Swap the positions of the noun phrases and add the word by before the second noun phrase. Then, conjugate the verb and add ed to the end. If the verb is to be , then conjugate the verb Q. The authors stopped the presidents. A. Replace The $1 $2. with $3 was $4 by the $1. A. Replace The $1 $2. with $3 was $4 by the $1. A. Replace The $1 $2. with $3 was $4 by the $1. A. Replace The $1 $ Swap the positions of the noun phrases and add the word by before the second noun phrase. Then, conjugate the verb and add ed to the end. If the verb is to be , then conjugate the verb Therefore, the correct answer is (The presidents were stopped by the authors. Q. The tourists advised the professors. A. Replace The $1 $2. with $3 was $4 by the $1. A. Replace The $1 $2. with $3 were $4 by the $1. A. Replace The $1 $2. with $3 was $4 by the $1. A. Replace The $1 $ Swap the positions of the noun phrases and add the word by before the second noun phrase. Then, conjugate the verb and add ed to the end. If the verb is to be , then conjugate the verb Therefore, the correct answer is (The professors were advised by the tourists. Q. The actors stopped the artists. A. Replace The $1 $2. with $3 was $4 by the $1. A. The artists were stopped by the actors. Q. The actors stopped the artists. A. Replace The $1 $2. with $3 was $4 by the $1. A. The artists were stopped by Swap the positions of the noun phrases and add the word by before the second noun phrase. Then, conjugate the verb and add ed to the end. If the verb is to be , then conjugate the verb Therefore, the correct answer is (The artists were stopped by the actors. Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution L.1.7. NEGATION Prompt 0: False Prompt 1: M Q. The original language of The Purple Taxi is French. A. Replace a noun or noun phrase with another noun or noun phrase. B. Replace a noun or noun phrase with a pronoun. C. Replace a pronoun with a noun or noun phrase. D. Replace a pronoun with another pronoun. Q. The original language of The Purple Taxi is French. Find the negation of a sentence. Therefore, the correct answer is (The original language of The Purple Taxi is not French.). Context 1: Q. Marcel Alessandri died in Paris. A. Nounphrase B. Pronoun C. Noun phrase D. Pronoun Q. Marcel Alessandri died in Paris. Find the negation of a sentence. Therefore, the correct answer is (Marcel Alessandri did not die in Paris False Therefore, the correct answer is (Marcel Alessandri did not die in Paris.). Q. Some people are wise. A. Replace a noun or noun phrase with another noun or noun phrase. B. Replace a noun or noun phrase with a pronoun. C. Replace a pronoun with a noun or noun phrase. D. Replace a pronoun with another pronoun. Q. Some people are wise. Find the negation of Find the negation of a sentence. Therefore, the correct answer is (Some people are not wise.). Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution L.1.8. ANTONYMS Prompt 1: It is good to be a leader but it is more important to first be a follower. Context 0: Q. nonpartisan A. , if possible. Input: 1 Output: 1 Input: 2 Output: 2 Input: 3 Output: 3 Input: 4 Output: 4 Input: Write about your most treasured item Therefore, the correct answer is (The answer is partisan. Context 1: Q. undignified A. I gave a friend an instruction and five inputs. The friend read the instruction and wrote an output for every one of the inputs. Input 1: 1 Output 1: 1 Input 2: 2 Output 2: 2 Input 3: 3 Output 3: 3 Input 4 Write the word s antonym Therefore, the correct answer is (The answer is dignified. Context 2: Q. flattering A. reverse the + + PROMPT + PROMPT+ PROMPT + PROMPT + PROMPT + PROMPT + PROMPT + PROMPT + PROMPT + PROMPT + PROMPT + PROMPT + PROMPT + PROMPT + PROMPT + PROMPT + PROMPT + PRO Write the Write the word s antonym Therefore, the correct answer is (The answer is unflattering. Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution L.1.9. SYNONYMS Prompt 1: 2015 Context 0: Q. bus A. 2: 3: 4: 5: 6: 7: 8: 9: 10: 11: 12: 13: 14: 15: 16 Therefore, the correct answer is (The answer is 10, bus , coach , motorcoach , motorbus , Context 1: Q. electric A. 2: 3: 4: 5: 6: 7: 8: 9: 10: 11: 12: 13: 14: 15: 16: 17: 18: 19: 20: 21: 22: 23: 24: 25: 26: 27: 28: Convert each word to a synonym according to Word Net. If there are multiple synonyms, use the first one. Therefore, the correct answer is (The answer is 10, electric , electrical , power , current , Context 2: Q. frightened A. 2: 3: 4: 5: 6: 7: 8: 9: 10: 11: 12: 13: 14: 15: 16: 17: 18: 19: 20: 21: 22: 23: 24: 25: 26: 27: 28: Therefore, the correct answer is (The answer is 10, frightened , scared , afraid , fearful , Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution L.1.10. MEMBERSHIP Prompt 0: Put the animals in ascending order of length. Context 0: Q. goat, motorway, shark, penguin, white, tractor, lion A. Put the animals in ascending order of length. The answer is goat, penguin, shark, lion. Write the animals in alphabetical order. Therefore, the correct answer is (goat, penguin, shark, lion). Write the animals in alphabetical order. Therefore, the correct Therefore, the correct answer is (goat, penguin, shark, lion). Context 1: Q. ship, swan, parrot, monkey, butter, dentist, shark A. Put the animals in ascending order of length. The answer is monkey, parrot, shark, swan. Write the animals in alphabetical order. Therefore, the correct answer is (monkey, parrot, shark, swan). Write the animals in alphabetical order. Therefore, the correct Therefore, the correct answer is (monkey, parrot, shark, swan). Context 2: Q. snail, ship, trousers, jellyfish, rabbit A. Put the animals in ascending order of length. The answer is rabbit, snail, jellyfish. Write the animals in alphabetical order. Therefore, the correct answer is (rabbit, snail, jellyfish). Write the animals in alphabetical order. Therefore, the correct answer is (rabbit Therefore, the correct answer is (rabbit, snail, jellyfish). Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution L.1.11. RHYMES Prompt 0: If the last letter of the input is e , remove it. Prompt 1: remove the last two letters of the input and add the letters \xc2\x93mote\ xc2\x94. Q. pea A. If the last letter of the input is e , remove it. A. If the last letter of the input is s , remove it. A. If the last letter of the input is y , remove it. A. If the last letter of the input is remove the last two letters of the input and add the letters \xc2\x93mote\xc2\x94. Therefore, the correct answer is (a) pea. Context 1: Q. night A. If the last letter of the input is e , remove it. A. If the last letter of the input is t , remove it. A. If the last letter of the input is h , remove it. A. If the last letter of the input is remove the last two letters of the input and add the letters \xc2\x93mote\xc2\x94. Therefore, the correct answer is (The answer is night. Q. add A. If the last letter of the input is e , remove it. A. If the last letter of the input is d , remove it. A. If the last letter of the input is a , remove it. A. If the last letter of the input is remove the last two letters of the input and add the letters \xc2\x93mote\xc2\x94. Therefore, the correct answer is (The answer is add. Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution L.1.12. LARGER ANIMAL Prompt 1: 10 Context 0: Q. spider, manatee A. Hints 91 and 93 are both false. The weight of a spider is 0.003 pounds. The weight of a manatee is 1300 pounds. The manatee weighs more than the Therefore, the correct answer is (manatee). Q. manatee, spider A. Hints 91 and 93 are both false. The weight Context 1: Q. sperm whale, sheltie A. Hints 91 and 93 are both false. The weight of a sperm whale is 450000 pounds. The weight of a sheltie is 25 pounds. The sperm whale weighs more than Multiply Inputs;; Therefore, the correct answer is (sperm whale). Q. sheltie, sperm whale A. Hints 91 and 93 are both false. Q. hummingbird, polar bear A. To invent, you need a good imagination and a pile of junk. -Thomas Edison The weight of a hummingbird is 0.002 pounds. The weight of a polar bear is 1500 pounds. The polar bear weighs more than the hummingbird. Therefore, the correct answer is Multiply Inputs;; Therefore, the correct answer is (polar bear). Q. polar bear, hummingbird A. To invent, you need a good imagination and a pile of junk. - Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution L.1.13. CAUSE SELECTION Prompt 0: Pick the sentence which is more likely to be the cause of the effect. Prompt 1: Choose the sentence that happened first. Q. Sentence 1: The soda went flat. Sentence 2: The bottle was left open. A. Pick the sentence which is more likely to be the cause of the effect. Q. Sentence 1: The bottle was left open. Sentence 2: The soda went flat. A. Pick the sentence which is more likely to be the cause of the effect. Q. Sentence 1: The Choose the sentence that happened first. Therefore, the correct answer is (The bottle was left open. , The soda went flat. ). Q. Sentence 1: The ship capsized. Sentence 2: The wind blew strong. A. Pick the sentence which is more likely to be the cause of the effect. Q. Sentence 1: The ship capsized. Sentence 2: The wind blew strong. A. Pick the sentence which is more likely to be the cause of the effect. Q. Sentence 1: The ship Choose the sentence that happened first. Therefore, the correct answer is (The wind blew strong. ) Q. Sentence 1: The power was out for days. Sentence 2: The meat spoiled. A. Pick the sentence which is more likely to be the cause of the effect. Q. Sentence 1: The meat spoiled. Sentence 2: The power was out for days. A. Pick the sentence which is more likely to be the cause of the effect. Q. Sentence 1: The Choose the sentence that happened first. Therefore, the correct answer is (The power was out for days. ) Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution L.1.14. FORMALITY Prompt 1: I am the beginning of the end, and the end of time and space. I am the beginning of every end, and the end of every place. I am the beginning of the beginning, and the end of beginning and ending. Context 0: Q. The whole story was made up. A. Reword this so that it has a formal tone. A. The entire story was fabricated. Write more formally! Therefore, the correct answer is (The entire story was fabricated. ). Write more formerly. Therefore, the correct answer is (The entire story was fabricated. ). Context 1: Q. The whole story was made up. A. Rewrite this so that it has a formal tone. A. The entire story was fabricated. Write more formally! Therefore, the correct answer is (The entire story was fabricated. ). Write more formally! Therefore, the correct Write less formerly. Write more casually. Therefore, the correct answer is (The entire story was fabricated. ). Context 2: Q. I think that this is interesting. A. Rewrite in indirect, pompous English. A. It is my opinion that this is interesting. Write more formally! Therefore, the correct answer is (It is my opinion that this is interesting. ). Write less formerly. Write more casually. Therefore, the correct I am the beginning of the end, and the end of time and space. I am the beginning of every end, and the end of every place. I am the beginning of the beginning, and the end of beginning and ending. Therefore, the correct answer is (It is my opinion that this is interesting. ). Q. I think that this is interesting. A. Rewrite in indirect, Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution L.1.15. SUM Prompt 0: Instruction: Prompt 1: The instructions were Q. 50 90 A. The instruction was: The instruction was:Write a program that asks the user to enter two numbers and then prints the sum of the two numbers. Your program printed 140, which is not the sum of 50 and 90. Make sure that your program I gave a friend an instruction and five inputs. The friend read the instruction and wrote an output for every one of the inputs. Here are the input-output pairs: Therefore, the correct answer is (140, Q. 5 47 A. The instruction was: The instruction was:Write a program that asks the user to enter two numbers and then prints the sum of the two numbers. Your program printed 52, which is not the sum of 5 and 47. Make sure that your program I I gave a friend an instruction and five inputs. The friend read the instruction and wrote an output for every one of the inputs. Here are the input-output pairs: Therefore, the correct answer is (52, Q. 2 39 A. The instruction was: The instruction was:Write a program that asks the user to enter two numbers and then prints the sum of the two numbers. Your program printed 41, which is not the sum of 2 and 39. Make sure that your program I I gave a friend an instruction and five inputs. The friend read the instruction and wrote an output for every one of the inputs. Here are the input-output pairs: Therefore, the correct answer is (41, Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution L.1.16. DIFFERENCE Prompt 0: subtract the smaller number from the bigger number and write the absolute difference between them. Prompt 1: Subtract the second input from the first input and report back the difference between them. Q. 73 29 A. subtract the smaller number from the bigger number and write the absolute difference between them. 73-29=44 |44|=44 CORRECT The answer is 44. Subtract the second input from the first input and report back the difference between them. Therefore, the correct answer is (a) 44 Q. 73 25 A. subtract the smaller number from the bigger number and write the absolute difference between them. 73-25=48 |48|=48 CORRECT The answer is 48. Subtract the second input from the first input and report back the difference between them. Therefore, the correct answer is (a) 48 Q. 53 4 A. subtract the smaller number from the bigger number and write the absolute difference between them. = 53 - 4 = 49 = 49 === The answer is 49. Subtract the second input from the first input and report back the difference between them. Therefore, the correct answer is (49). Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution L.1.17. NUMBER TO WORD Prompt 0:A1 Prompt 1: take the input, write it in British English and insert the comma. Context 0: Q. 5336 A. 0 Take the input and write it in words as a natural number in British English. Therefore, the correct answer is (a) five thousand, three hundred and thirty-six. Take the input and write it in words as a natural number in British English. Therefore, the correct answer is (a) five thousand, three hundred and thirty-six. Q. 5156 A. A1 Take the input and write it in words as a natural number in British English. Therefore, the correct answer is (a) five thousand, one hundred and fifty-six. Take the input and write it in words as a natural number take the input, write it in British English and insert the comma. Therefore, the correct answer is (a) five thousand, one hundred and fifty-six. Q. 5000 A. 0 Take the input Q. 8692 A. A1 Take the input and write it in words as a natural number in British English. Therefore, the correct answer is (a) eight thousand, six hundred and ninety-two. Take the input and write it in words as a natural number take the input, write it in British English and insert the comma. Therefore, the correct answer is (a) eight thousand, six hundred and ninety-two. Q. 10000 A. 0 Take the Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution L.1.18. TRANSLATION ENGLISH-GERMAN Prompt 0: You may have to get a little... ; Prompt 1: 0x641: cmp esi,ebx ; Q. relation A. Write the English instruction as though it is German.: Die Beziehung ist eine Verbindung zwischen zwei oder mehr Menschen. Die Beziehung kann eine Freundschaft, eine Ehe, eine Partnerschaft oder eine andere Art von Beziehung sein. Translate the English word to German.: Therefore, the correct answer is ( Translate the English word to German.: Therefore, the correct answer is (a) Beziehung. Q. camp A. You may have to get a little... ; Das Lager ist ein Ort, an dem Menschen leben oder arbeiten. Das Lager kann ein Zeltlager, ein Ferienlager oder ein anderes Lager sein. Translate the English word to German.: Therefore, the correct answer is ( 0x641: cmp esi,ebx ; Therefore, the correct answer is (a) Lager. Q. A. Das ist eine Frage, die nicht so einfach zu beantworten ist. Es gibt Q. lecture A. You may have to get a little... ; Der Vortrag ist eine Rede, die vor einem Publikum gehalten wird. Der Vortrag kann eine Prasentation, ein Vortrag oder eine andere Art von Vortrag sein. Translate the English word to German.: Therefore, the correct answer is ( Translate 0x641: cmp esi,ebx ; Therefore, the correct answer is (a) Vortrag. Q. A. Der Vortrag ist eine Rede, die vor einem Publikum gehalten wird. Der Vortrag Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution L.1.19. TRANSLATION ENGLISH-SPANISH Prompt 0:Traduccion al espanol -> QUOTE: Translate to Spanish Prompt 1: In Spanish, these five words are: Contexts Context 0: Q. bear A. Translate to Spanish: oso Q. bear A. Translate to Spanish: oso Q. bear A. Translate to Spanish: oso Q. bear A. Translate to Spanish: oso Q. bear Translate these five English words to Spanish. Therefore, the correct answer is (a) oso. Context 1: Q. relativity A. Translate to Spanish: relatividad Q. relativity A. Translate to Spanish: relatividad Q. relativity A. Translate to Spanish: relatividad Q. relativity A. Translate to Spanish: relatividad Q. relativity Translate these five English words to Spanish. Therefore, the correct answer is (a) relatividad. Q. bear A. Translate to Spanish: oso Q. bear A. Translate to Spanish Context 2: Q. concert A. Translate to Spanish: concierto Q. concert hall A. Translate to Spanish: sala de conciertos Q. concertmaster A. Translate to Spanish: concertino Q. concerto A. Translate to Spanish: concierto Q. Translate these five English words to Spanish. Therefore, the correct answer is (a) concierto. Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution L.1.20. TRANSLATION ENGLISH-FRENCH Prompt 0: I ve translated 5 words from English to French: Prompt 1: Translate to French Context 0: Q. destiny A. I ve translated 5 words from English to French: English French destiny destin destiny destin destiny destin destiny destin destiny destin Translate to French Therefore, the correct answer is (destin). Q. destiny A. I ve translated 5 words from English to French: Context 1: Q. ideology A. I ve translated 5 words from English to French: English French ideology ideologie ideology ideologie ideology ideologie ideology ideologie ideology ideologie Translate to French Therefore, the correct answer is (ideologie). Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution Q. ideology A. I ve translated 5 words from English to French: Q. representation A. I ve translated 5 words from English to French: English French representation representation representation representation representation representation representation representation representation represent Translate to French Therefore, the correct answer is (representation). Q. representation A. I ve translated 5 words from English to French: Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution L.1.21. SENTIMENT ANALYSIS Prompt 0: Tell if each review was posted by someone with generally positive or negative language Prompt 1:write the number of positive examples Q. Despite its faults, Gangs excels in spectacle and pacing. A. I gave a friend an instruction and five inputs. The friend read the instruction and wrote an output for every one of the inputs. B. I gave a friend an instruction and five inputs. The friend read the instruction and wrote an output for every one of the inputs. C. I gave a friend an instruction and five inputs. The friend read the instruction and wrote an Given a movie review, indicate whether it is positive or negative Therefore, the correct answer is ( positive , Context 1: Q. I simply can t recommend it enough. A. I gave a friend an instruction and five inputs. The friend read the instruction and wrote an output for every one of the inputs. B. I gave a friend an instruction and five inputs. The friend read the instruction and wrote an output for every one of the inputs. C. I gave a friend an instruction and five inputs. The friend read the instruction and wrote an write the number of positive examples Therefore, the correct answer is ( positive , Q. There s a lot to recommend Read My Lips. A. I gave a friend an instruction and five inputs. The friend read the instruction and wrote an output for every one of the inputs. B. I gave a friend an instruction and five inputs. The friend read the instruction and wrote an output for every one of the inputs. C. I gave a friend an instruction and five inputs. The friend read the instruction and wrote an write the number of positive examples Therefore, the correct answer is ( positive , Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution L.1.22. SENTENCE SIMILARITY Prompt 0: . Prompt 1: !:There are two kinds of problems / hints that you will see:!< Q. Sentence 1: The polar bear is sliding on the snow. Sentence 2: A polar bear is sliding across the snow. A. : Read the two sentences and decide how well the second sentence expresses the same idea as the first. Then enter the number of the answer option that best describes your decision. 1. The two sentences express the same idea. 2. The two sentences express the same idea, but the second sentence is not as clear as the first. 3. The two sentences express the same idea, but the second sentence is : Given two sentences, your algorithm must determine the similarity between the sentences on a scale of 0 to 5. A score of 0 means the sentences have nothing in common. A score of 5 means the sentences are semantically similar. The Therefore, the correct answer is 5. Q. Sentence 1: Iranian President praises nuclear deal Sentence 2: Iran arrests several spies near nuclear plant A. . Output the score according to the following rules: 0: The sentences have nothing in common. 1: The sentences are about the same topic, but express different ideas. 2: The sentences are about the same topic, and express similar ideas. 3: The sentences are about the : Therefore, the correct answer is 1. Q. Sentence 1: A girl is playing a flute. Sentence 2: A band is playing on a stage. A. .Output the score according to the following rules: 0: The sentences have nothing in common. 1: The sentences are about the same topic, but express different ideas. 2: The sentences are about the same topic, and express similar ideas !: There are two kinds of problems / hints that you will see:!< Therefore, the correct answer is 1. Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution L.1.23. WORD IN CONTEXT Prompt 0: Determine whether the given word is used with the same meaning in both sentences. Write Prompt 1: Decide whether the given word is used in the same meaning in both sentences. Q. Sentence 1: The Times is not the voice of New York. Sentence 2: The voice of the law. Word: voice A. Determine whether the given word is used with the same meaning in both sentences. Writeyes or no. The answer is yes. Decide whether the given word is used in the same meaning in both sentences. Therefore, the correct answer is (yes). Q. Sentence 1: Do you communicate well with your advisor? Sentence 2: He and his sons haven t communicated for years. Word: communicate A. Determine whether the given word is used with the same meaning in both sentences. Writeyes or no. The answer is yes. Decide whether the given word is used in the same meaning in both sentences. Therefore, the correct answer is (yes). Q. Sentence 1: Can you take me to the main entrance? Sentence 2: Take a scene. Word: take A. Determine whether the given word is used with the same meaning in both sentences. Writeyes or no. The answer is no. Decide whether the given word is used in the same meaning in both sentences. Therefore, the correct answer is (no). M. GPT 3.5 Results We ran PB on GSM-8k using GPT3.5-Turbo-0613, see Figure 5, and GPT-3.5-Turbo-1106, for 8000 and 2500 evaluations respectively. The prompt with the best training set fitness in the run was evaluated on the test set achieving 65.5% and 63.9% respectively. For GPT-3.5-Turbo-0613 the evolved task prompts were: Prompt1: Solve the math word problems and provide the numerical answers without using any words from the original instructions. Prompt2: Solve the math word problems and provide the numerical answers without using any words from the original instructions.. For GPT3.5Turbo-1106 the evolved task prompts were, Prompt1: First, calculate the numerical values of the expressions and arrange them in descending order. Then, provide the final answer in Arabic numeral format as 71. Prompt2: Solve the math word problem, providing the answer in Arabic, while considering diversity and taking breaks as needed.. It is encouraging that Promptbreeder is able to work with very different language models. PS+ prompt (our own implementation) only achieved 44.7% with GPT-3.5-Turbo-0613. In (Wang et al., 2023b) PS+ achieved 59.3% 0-shot performance using the text-davinci-003 model. Note GPT 3.5 is reported as getting 57.1% few-shot test set performance on GSM-8k (Open AI, 2023), but our evolved results zero shot with GPT3.5-Turbo surpass those by a substantial margin. Table 2 in (Wang et al., 2023b) sets expectations with regard to relation between zero-shot and few-shot performance, with few-shot being slightly better performing than zero-shot. Our zero-shot evaluation demonstrated that the default Co T prompt achieved scores of 52.5% and 53% for GPT-3.5-Turbo-0613 and GPT-3.5-Turbo-1106, respectively. However, a more sophisticated evolved prompt achieved Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution significantly higher scores of 65.5% and 63.9%, respectively. 0 1000 2000 3000 4000 5000 6000 7000 8000 Evaluations Training Set Fitness GSM-8k with GPT3.5-Turbo 65.5% on Test Set: Solve the math word problems and provide the numerical answers without using any words from the original instructions. Figure 5. GSM-8k run using GPT-3.5Turbo-0613. Red = mean population fitness, Blue = Individual fitness evaluations. The best evolved prompt on the training set is shown, and its test set fitness is 65.5% N. The Effect of Using Poor Problem Descriptions We looked at the impact of initializing with problem descriptions where it does not accurately describe the problem. Specifically we ran Promptbreeder with the GSM8K dataset using a neutral problem description Write some text and a misleading problem description Write a poem. N.1. Neutral problem description The fittest individual at the start of the experiment achieved a test accuracy of 57.5% with the task-prompts 1st part and Trewe sit mtew. After 2.5k evaluations the task-prompts had evolved to frstpart and few it eomt which gets a test accuracy of 66.6% N.2. Misleading problem description The fittest individual at the start of the experiment achieved a test accuracy of 33.8% with the task-prompts Express yourself in a verse. and Use all you know to help someone. After 2.5k evaluations the task-prompts had evolved to INSTRUCTION MUTANT: Write working out in the answer sheet as well. and INSTRUCTION: Break new ground and break barriers to innovate with an exciting mutator prompt. resulting in a test accuracy of 57.7% O. Experiments in describing the population to the LLM We created a set of hand-written prompts with made up fitnesses and presented these to the LLM asking it to generate a new prompt in different ways. We then measured the BERT encoding similarity of the generated prompts to the best prompt Promptbreeder: Self-Referential Self-Improvement via Prompt Evolution Figure 6. BERT embedding distances to 100 sampled offspring from different kinds of prompts. Blue shows the distances to the best prompt, and orange shows distances to the best prompt in the flipped setting. A good description of the population is one where this distribution differs between flipped and non-flipped settings. in the original population to determine how the presentation of the population information influenced the distribution of prompts generated. Figure 6 shows the distribution of distances to the best prompt in the population for various modes of presentation. a) When prompts are presented in normalized fitness order and the continuation begins with Genotype N (fitness = 1): we did not observe a difference in the distribution of distances when the fitness of prompts was reversed, implying that the LLM does not understand the association between fitness and prompt. b) With the prompt Here are two sentences (1) and (2). Produce more sentences that are semantically related to sentence (1) but not to sentence (2) followed by a indexed pair of sentences, we did observe a difference in BERT embedding distance distributions to the two sentences when the indexes (1) and (2) were swapped in the original prompt. This suggests the LLM does understand binary index preferencing. c) When the prompt is Sentences are shown in ascending order of score. Sentence (N) is the best unique sentence. It resembles sentence N-1 more than it does sentence 1 followed by the sentences in ascending order of score, we observed significant differences in distribution when the order of sentences was flipped, suggesting the LLM understands this description of population fitnesses, but the distribution was too biased towards the prompts at the bottom of the list. d) By combining B and C (contradictory index referencing and ranking), a more diverse sampling of the list was possible, and this was used in our EDA Rank and Index Mutation operator. These results apply to the Pa LM language model instance we used and are expected to be different for different LLMs, and so we advise experimentation of this sort to first determine whether suitable distributions of offspring are obtained in your case.