# preference_adaptive_and_sequential_texttoimage_generation__fe595b35.pdf Preference Adaptive and Sequential Text-to-Image Generation Ofir Nabati 1 Guy Tennenholtz 1 Chih Wei Hsu 1 Moonkyung Ryu 1 Deepak Ramachandran 2 Yinlam Chow 2 Xiang Li 1 Craig Boutilier 1 We address the problem of interactive text-toimage (T2I) generation, designing a reinforcement learning (RL) agent which iteratively improves a set of generated images for a user through a sequence of prompt expansions. Using human raters, we create a novel dataset of sequential preferences, which we leverage, together with largescale open-source (non-sequential) datasets. We construct user-preference and user-choice models using an EM strategy and identify varying user preference types. We then leverage a large multimodal language model (LMM) and a valuebased RL approach to suggest an adaptive and diverse slate of prompt expansions to the user. Our Preference Adaptive and Sequential Text-toimage Agent (PASTA) extends T2I models with adaptive multi-turn capabilities, fostering collaborative co-creation and addressing uncertainty or underspecification in a user s intent. We evaluate PASTA using human raters, showing significant improvement compared to baseline methods. We also open-source our sequential rater dataset and simulated user-rater interactions to support future research in user-centric multi-turn T2I systems. 1. Introduction Advances in text-to-image (T2I) generation, fueled by powerful diffusion models (Croitoru et al., 2023; Gu et al., 2023; Rombach et al., 2022; Saharia et al., 2022; Yang et al., 2023; Yu et al., 2022; Zhang et al., 2023; Liang et al., 2024), have unlocked unprecedented possibilities for image generation, as users can readily translate textual descriptions into stunning visuals. That said, capturing precise user intent remains a major challenge (Wu et al., 2023; Liang et al., 2024). As 1Google Research 2Google Deep Mind. Correspondence to: Ofir Nabati , Guy Tennenholtz . Proceedings of the 42 nd International Conference on Machine Learning, Vancouver, Canada. PMLR 267, 2025. Copyright 2025 by the author(s). such, single-turn T2I generation may fail to encapsulate a user s nuanced and evolving image conception. This is especially true for complex or abstract concepts, where a user s initial prompt may not suffice in achieving a desired visual representation. Addressing this challenge requires moving beyond the paradigm of one-shot T2I generation towards a more iterative and interactive process (von Rütte et al., 2023; Wang, 2024; Liu et al., 2024), e.g., a collaborative setting where an assistive agent interacts with a user, guiding them through a series of refinements of their initial prompt. This interactive setup could allow for a more nuanced elicitation of the user s intent, gradually shaping the generated images towards a desired outcome. To this end, we introduce a Preference Adaptive and Sequential Text-to-image Agent (PASTA), which learns from user preference feedback to guide the user through a sequence of prompt expansions, iteratively refining the generated image. This sequential approach (see Figure 1) allows users to articulate their vision more completely by gradually reducing uncertainty or underspecification in their original prompt. Our framework leverages the power of large multimodal language models (LMMs) (Gemini-Team, 2024; Achiam et al., 2023) and reinforcement learning (RL) (Sutton, 2018; Fan et al., 2023) to facilitate user-adaptive co-creation. Our approach for PASTA involves a multi-stage data collection and training process. We first collect multi-turn interaction data from human raters with a baseline LMM. Using this sequential data, as well as large-scale, open source (single-turn) preference data, we train a user simulator. Particularly, we employ an EM-strategy to train user preference and choice models, which capture implicit user preference types in the data. We then construct a new large-scale dataset, which consists of interactions between a simulated user and the LMM.1. Finally, we leverage this augmented data, encompassing both human and simulated interactions, to train PASTA our value-based RL agent, which presents a sequence of diverse slates of images to a user. PASTA 1Both human and simulation data is open-sourced to support research on multi-turn T2I generation. Link to the dataset: https://www.kaggle.com/datasets/googleai/pasta-data. Preference Adaptive and Sequential Text-to-Image Generation Figure 1: An illustration of an agent-user interaction with L = 4 prompt expansions at each step, and M = 1 images per prompt expansion. The user selection is outlined in blue. The agent presents prompt expansions based on the user s previous responses to maximizes the expected cumulative user satisfaction (i.e., value). See Appendix G for additional examples using M = 4 images for L = 4 prompt expansions. interacts with the user and sequentially refines its generated images to better suit their underlying preferences. Our contributions are as follows. (1) We formulate the problem of iterative prompt expansion using LMMs and T2I models. (2) We collect a new, large-scale user-agent interaction dataset for the multi-turn prompt expansion setup. Our dataset consists of almost 40,000 interaction rollouts. (3) We train user utility and user-choice models with this dataset, creating a simulator that is used to generate additional simulated data. (4) Finally, we utilize our datasets to train PASTA, our value-based RL agent, demonstrating its effectiveness through comprehensive human evaluations, with significant improvements in user satisfaction compared to a baseline LMM. 2. Background and Problem Setup We begin with a short background on diffusion models, LMMs, and RL. We then formulate our problem setup. 2.1. Background Diffusion models (Ho et al., 2020; Ho & Salimans, 2022; Croitoru et al., 2023; Gu et al., 2023; Rombach et al., 2022; Saharia et al., 2022; Yang et al., 2023; Yu et al., 2022; Zhang et al., 2023; Liang et al., 2024) have become a prominent approach for T2I generation. These models add Gaussian noise to an image until it becomes pure noise, and are then trained to reverse this process, iteratively removing noise to reconstruct the image. This denoising process is guided by a prompt, enabling the generation of images that semantically align with the textual description. Large Multimodal Models (LMMs) (Gemini-Team, 2024; Achiam et al., 2023) use Transformer models (Vaswani, 2017), and are trained to predict the next token in a sequence, where tokens can represent textual or visual information. These models learn joint representations of text and images, making them useful for prompt engineering in interactive T2I systems, as they allow us to modify generated outputs based on previously generated responses. Reinforcement Learning (RL) (Sutton, 2018; Fan et al., 2023) frameworks train agents to make a sequence of decisions in an environment to maximize cumulative reward. Value-based RL methods learn a value function that estimates the expected cumulative reward for taking a particular action in a given state. In our work, an RL agent learns to select actions (prompt expansions) that lead to higher user satisfaction. 2.2. Problem Formulation We consider an interactive, sequential T2I decision problem in which a user engages with an agent, visualized in Figure 1. At time t = 0, the user issues an initial prompt (e.g., A white cat ) intended to capture a target image. At each turn t 1 the agent proposes L prompt expansions, which are fed to a T2I model. The T2I model then generates M images for each prompt. The user sees the slate of M L images and selects the set of M images (corresponding to one of the prompt expansions) that best reflects their preferences. The process repeats for up to H turns. Formally, the problem is given by a tuple (U, P, I, G, C, R, H, ν0), where U is a set of user types, reflecting different preferences; P is a set of feasible Preference Adaptive and Sequential Text-to-Image Generation text prompts; I is a set of feasible images; G : P 7 I is a T2I generative model2; C : U P IML 7 [L] is a user choice function3; R : U P IML 7 [0,1] is a user utility function; H is the horizon; and ν0 7 U P is a distribution of users and initial prompts. At time t = 0, a user-prompt pair u, p0 U P is sampled from the initial distribution ν0. At each time 1 t H the agent selects a slate of L prompt expansions Pt = {pℓ,t P}L ℓ=1 and feeds them to the T2I model G, which generates M images for each prompt, n {Im,ℓ,t G(pℓ,t)}M m=1 o L ℓ=1. The user observes the slate of M L images, and selects a set of M images corresponding to one prompt, and according to the choice function C, i.e., ct C(u, p0, I1,1,t, . . . , IM,L,t). We assume a (latent) reward rt R(u, p0, I1,ℓt,t, . . . , IM,ℓt,t) reflecting user u s satisfaction with the selected images. 2.3. RL Formulation Our problem can be formulated as a latent contextual MDP (Hallak et al., 2015; Kwon et al., 2021), where U is the latent context space. The state space consists of the user-agent interaction history, with the state (history) ht at time t given by: p0 |{z} initial prompt , time step 1 z }| { {pℓ,1}L ℓ=1 | {z } agent action A1 , {Im,ℓ,1}L,M ℓ=1,m=1, c1 | {z } transition time step t z }| { {pi,t}L ℓ=1 | {z } agent action At , {Im,ℓ,t}L,M ℓ=1,m=1, ct | {z } transition the action space is any selection of L prompts, transitions are induced by the user choice model C and the T2I model G, and reward is given by a user utility function R. See Appendix A for a full description of this latent MDP. A stochastic policy π : H 7 PL maps interaction histories to a distribution over slates of L prompts. The value of a policy π is its expected cumulative sum of rewards over users and initial prompts, i.e., vπ = Eu ν0 h PH t=1 rt u, π i . An optimal policy π arg maxπ vπ maximizes the value. The stateaction value function for any h H and P PL is qπ t (h, P) = Eu ν0 h PH t =t rt u, ht = h, Pt = P, π i . 2More generally, G can be a function of user u or interaction history. Here we assume G is a non-adaptive T2I model. 3More generally a user choice function can also depend on prompt expansions. In our work we assume users only view the generated images and the initial prompt. 3. PASTA: Preference Adaptive and Sequential Text-to-image Agent We solve the sequential prompt expansion problem using RL. Our Personalized And Sequential Text-to-image Agent (PASTA) engages with a user to adapt the prompt to their preferences, and maximize (latent) user utility. The user s type u U is unknown to the agent and must be inferred during the interaction. This framework is related to meta RL (Wang et al., 2016; Duan et al., 2016; Finn et al., 2017; Zintgraf et al., 2020), where each episode (or meta-episode) samples a latent MDP from a predefined problem distribution. In our setting, the agent must adapt within H steps to the unknown MDP, and optimize the reward accordingly. This requires balancing exploration and exploitation: the agent must take actions that, on the one hand, provide images that reflect (its estimate of) the user s preferences, and on the other, improving its estimate of those preferences by exploring other types of expansions/images. This can be viewed as an implicit form of preference elicitation (Keeney, 1993; Salo & Hamalainen, 2001; Chajewska et al., 2000; Boutilier, 2002; Meshi et al., 2023). 3.1. Candidate Action Generator And Selector The state space of user interaction histories serves as sufficient statistic (i.e., belief over users types (Aberdeen et al., 2007)). To solve our problem effectively, each interaction with a user should provide the agent sufficient information about the user (e.g., through value of information (Boutilier et al., 2003)). This requires the action space be rich and diverse enough to enable information gain. A straightforward approach to action space design uses an LMM to construct action candidates (in our case, prompt expansion candidates). Specifically, we use an LMM to process the current interaction history and generate a broad set of LC L candidate prompts from which a slate of L prompts can be chosen. Generating a large set of candidates has been shown to introduce diversity and induce exploration (Tennenholtz et al., 2024). Instead of relying on a given LMM to directly output L prompts, we encourage diversity by generating LC prompts, and then selecting the L-subset our agent deems optimal. This, in turn, expands the effective action space our agent can leverage during training. Formally, we structure our policy class using a candidate generator policy πC : H 7 PLC and a candidate selector policy πS : PLC 7 PL. We first sample a set of LC candidates {pi,t}LC i=1 πC(ht) and then select L prompts from those candidates, i.e., Pt = {pi,t}L i=1 πS {pi,t}LC i=1 . We assume πC to be a fixed, capable LMM, and focus here on training the candidate selector πS. Preference Adaptive and Sequential Text-to-Image Generation Candidate Generator Candiate Selector Candidate Selector Figure 2: Top. PASTA policy framework: The LMM is used as a candidate generator of a candidate set, from which the candidate selector policy is used to select a slate. Bottom. Each prompt in the slate is evaluated individually using a prompt-value model, and the overall slate value is calculated as the average of the individual prompt values. 3.2. Value-Based Candidate Selector We use a state-action value function to define our selector policy, πS. Given candidates {pi}LC i=1 and a state-action value model qϕ(h, P) parameterized by ϕ, we define the selector policy by πS,ϕ(h) arg max P PL {pi} LC i=1 qϕ(h, P), (1) where PL(X) = {P X : |P| = L} is the set of all possible slates of size L. Enumerating PL {pi}LC i=1) is, of course, computationally expensive when LC is large (a detriment to both training time and inference efficiency). Inspired by Ie et al. (2019), we decompose the value function into a (weighted) average of prompt-values fϕ : H P 7 R. Specifically, we compute the value of each prompt, and estimate the value of a slate by: qϕ(h, P) = 1 p P fϕ(h, p). This approximation reduces the exponential complexity of finding the best slate to O(LC log LC) (i.e., by sorting the prompt values over the candidate set). See Figure 2 for a schematic of the PASTA policy and value function. PASTA first prompts an LMM πC (i.e., candidate generator) to generate LC candidates {pi}LC i=1 πC(h). It then uses its candidate selector πS to select a slate P of prompts according to Equation (1). We train PASTA using implicit Q-Learning (IQL) (Kostrikov et al., 2021), which estimates the TD error with the Bellman optimality operator. IQL employs a soft estimate with a value function trained to approximate the high expectile without assessing state-actions outside the dataset distribution, and has proven effective in offline RL with LLMs (Snell et al.; Zhou et al., 2024). The IQL loss is: L(ϕ, ψ) = Eht,Pt,rt,ht+1 D (qϕ(ht, Pt) rt vψ(ht+1))2 + Lα 2 (q ˆϕ(ht, Pt) vψ(ht)) , where Lα 2 (x) = α 1{x<0} x2, α [0.5, 1] and vψ is the α-expectile value estimate. Most notably, the IQL loss does not require searching for the best slate (which is oftentimes out of the distribution of the training data), making it very efficient for training (see Appendix E.1 for further details). 4. PASTA Dataset Training PASTA relies on the availability of sequential user-agent interaction data. While single-turn T2I preference data is readily available (Wu et al., 2023; Kirstain et al., 2023; Pressman et al., 2022; Liang et al., 2024), sequential datasets are not. Hence, we collect humanrater data for our sequential setup, and further enrich it with simulated data. Below, we describe our data creation process. All of our datasets are open-sourced here: https://www.kaggle.com/datasets/googleai/pasta-data. 4.1. Human Rater Sequential Data We use human raters to gather sequential user preferences data for preference-adaptive T2I generation.4 Participants are tasked with interacting with an LMM agent for five turns. Throughout our rater study we use a Gemini 1.5 Flash Model (Gemini-Team, 2024) as our base LMM, which acts as an agent. At each turn, the system presents 16 images, arranged in four columns (L = 4, M = 4), each representing a different prompt expansion derived from the user s initial prompt and prior interactions. Raters are shown only the generated images, not the prompt expansions themselves. At session start, raters are instructed to provide an initial prompt of at most 12 words, encapsulating a specific visual concept. They are encouraged to provide descriptive prompts that avoid generic terms (e.g., an ancient Egyptian temple with hieroglyphs instead of a temple ). At each turn, raters then select the column of images preferred most (see the UI used in Appendix B); they are instructed to select a column based on the quality of the best image in that column w.r.t. their original intent. Raters may optionally provide a free-text critique (up to 12 words) to guide subsequent prompt expansions, though most raters did not use this facility. 4Raters were paid contractors. They received their standard contracted wage, which is above the living wage in their country of employment. Preference Adaptive and Sequential Text-to-Image Generation After five turns, raters enter an evaluation phase where they answer questions about each turn, including: (1) reconfirmation of their preferred columns; (2) quantifying whether the image slate in the current turn is better than the previous; and (3) a free-text explanation for their selection. This process yields a dataset comprising sequential interaction trajectories with user preference feedback and provides valuable insight into the dynamics of multi-turn, preferenceadaptive T2I generation. Our human rater dataset includes over 7,000 five-step sequences (more than 500,000 images total), annotated by about 100 human raters using a random candidate selector. See Appendix B for a comprehensive description of the rater study. 4.2. Simulated Sequential Data Simulated data can drastically improve performance of RL agents (Kaiser et al., 2019; Yu et al., 2021; Liu et al., 2023; Hafner et al., 2020; Micheli et al., 2022). For this reason, we enrich the rater data above with additional simulated agent-user interaction data. To do so, we develop a user model that encodes distinct user types that reflect a range of preferences. This model has two components: (1) A user choice model which predicts the image column a user selects; and (2) a user utility model which predicts a user s satisfaction with an image slate. We outline details of the user model in Section 5. We begin by sampling a user and initial prompt from a joint prior. We use a randomized exploration policy for the agent, encouraging diverse interactions to distinguish preferences across user types. At each turn in a simulated trajectory, the agent proposes a slate of prompt expansions, and the user (via the choice model) selects their preferred prompt. The user utility model is then used to assign a satisfaction score for the images generated using the selected prompt. This process is repeated for five turns (see Appendix C for further details and analysis of our data generation process). The simulated user data includes over 30,000 rollouts (exceeding 2.5 million images). Augmenting our human-rater data with this simulated data allows for more robust training of PASTA, as we show later in Section 6. 5. User Model The data generation process described in Section 4.2 requires a user simulator consisting of both user choice and utility models. We leverage our sequential human-rater data (Section 4.1), together with large-scale open-source singleturn T2I preference data (Wu et al., 2023; Kirstain et al., 2023; Pressman et al., 2022), to build these models. We first propose a user preference model based on fundamental choice theory (Pennock et al., 2000) that mimics user s preference behaviors in an interactive recommenda- tion system. Specifically, this model takes as input a user type u U, an initial prompt p P, and a slate of images Pt = {Im,l,t}L,M ℓ =ℓ,m=1 and estimates the probability that u prefers image column {Im,ℓ ,t}M m=1 over the others {Im,l,t}L,M ℓ =ℓ ,m=1, i.e., P {Im,ℓ ,t}M m=1 {Im,l,t}L,M ℓ =ℓ ,m=1|u, p . (2) In the typical single-turn setup (Wallace et al., 2024), by setting M = 2, L = 1, and not assuming any user types, Equation (2) reduces to a probabilistic model that estimates the global (non-user specific) pair-wise preference between two images, i.e., Eu ν0[ P(I1 I2) | u, p ]. By contrast, our model is able to capture user-specific preferences over a generic sets of multiple images. Our user choice model C exploits the preference model in Equation (2), making the usual assumption that users select images based on their inherent preferences; i.e., Assumption 5.1 (Choice-Preference Equivalence). Assume C P, where P is the preference model in Equation (2). While, in general, user choice may also be be biased by other factors (e.g., previous interactions, position bias, etc.), this idealized model proves very useful in our experiments. To further model user utility, motivated by the axioms of expected utility theory (Von Neumann & Morgenstern, 2007), we model user utility R : U P IML 7 R, leveraging the preference model in Equation (2) to ensure a user s utility is consistent with their preferences. Specifically, we assume that a user prefers one set of images to another if the utility of the former images is higher, as formalized by the following assumption. Assumption 5.2 (Utility Consistency). For any u, p, Equation (2) is satisfied for ℓ {1, . . . , L} iff for all ℓ = ℓ R {Iℓ ,m,t}M m=1|u, p > R {Iℓ,m,t}M m=1|u, p . Assumptions 5.1 and 5.2 ensure a tight coupling of our user utility and choice models, which we leverage in model training. 5.1. A Parametric User Model To train our user utility and choice models, we employ a parameterized score model sθ : U I P 7 [0, 1] to serve as the backbone of our user simulator. We assume a discrete set of K (unknown) user types, i.e., U = {1, 2, . . . , K}. Exploiting Assumptions 5.1 and 5.2, we define the utility of a user type k over an arbitrary image slate {Im,ℓ,t}M m=1 that Preference Adaptive and Sequential Text-to-Image Generation ... K Image Embeddings (for each user) K Prompt Embeddings (for each user) Dot Product Image Prompt Figure 3: Score function architecture. The image and prompt are fed into CLIP encoders followed by user encoders to generate K user embeddings. The score for each user type is the inner product between the corresponding user image embedding and user text embedding. associates to a prompt p using the following model: Rk,θ ℓ,t := Rθ {Im,ℓ,t}M m=1|u = k, p = Agg(sθ(k, p, I1,ℓ,t), . . . , sθ(k, p, IM,ℓ,t)), (3) where Agg is an aggregator (e.g., average, max, Softmax sampler) operator. User choice probabilities are given by Cθ(k, p, {Im,1,t}M m=1, . . . , {Im,L,t}M m=1) = Softmax τθRk,θ 1,t , . . . , τθRk,θ L,t , (4) where τθ = hθ Rk,θ 1,t , . . . , Rk,θ L,t is a parameterized temperature constant which depends on the scores of different image columns. Such a temperature parameterization ensures our user choice model satisfies Assumption 5.2, while allowing for greater flexibility in modeling. To balance model capacity and computational efficiency, our score function adopts pre-trained CLIP (Radford et al., 2021) text and image encoders, followed by user encoders (a head for each user type). The user encoder transforms CLIP embeddings into a user-type-specific representation that captures individual preferences for images and prompts. The final score of an image-prompt pair is the inner product of the (user-type-specific) image embedding and corresponding text embedding, multiplied by a learned temperature parameter (see a schematic of our architecture in Figure 3). We employ an Expectation-Maximization (EM) framework to learn a user model capable of capturing diverse pref- erences. This model leverages the score function sθ that assesses the compatibility between image-prompt pairs and different user types. The EM algorithm iteratively refines the model parameters θ and a user type prior distribution η to maximize the likelihood of observed user feedback. Training proceeds in two phases: a main training phase on large-scale datasets with frozen CLIP parameters, followed by a fine-tuning phase on human-rated data where all parameters are trained. We detail the training procedure, specific loss functions tailored to different data types (single-turn preferences, relevance scores, and sequential multi-turn), the model architecture, and hyperparameter settings in Appendix D.1. 6. Experiments To assess the effectiveness of our approach, we conduct two empirical evaluations. First, we evaluate the quality of our user model and the simulated data it generated. This involves analyzing the model s ability to accurately capture user preferences and generate realistic interaction data. Second, we study the performance of PASTA, our multi-turn T2I agent, trained using our rater and simulated datasets. We perform a comprehensive analysis at each stage to assess the quality of the user model, the accuracy of the simulated data, and the overall performance of the T2I agent. The slate size, number of images per prompt, and problem horizon are fixed L = 4, M = 4 and H = 5 in all experiments. We use Stable Diffusion XL (Podell et al., 2023) as our T2I model and multimodal Gemini 1.5 Flash (Gemini Team, 2024) as our (prompt) candidate generator LMM, with LC = 25. Our learned value function is fine-tuned from Gemma 2B (Gemma-Team, 2024). Our experiments use a sparse reward, with rewards provided only in the final round, as our primary interest lies in the end result of interactive refinement process. We use softmax sampling for the user model s aggregation function in Equation (3), as raters were instructed to select a column based on the best image from that column. To further encourage diversity in the generation of prompt candidates, we require the LMM prompt generator to partition the candidate set into five categories, with each category focusing on a different aspect of the prompt expansion. Categories include both generic forms (e.g., rephrasing, adding more details) as well as prompt-specific approaches (e.g., different types of dogs, cloud shapes, etc.). Moreover, we restrict the candidate selector to select at most one prompt from each category. This simple technique induces significant diversity (hence implicit exploration), and is used for data generation, policy training, and at inference time. Preference Adaptive and Sequential Text-to-Image Generation Figure 4: The graphs present the performance of a trained user model as a function of the number of user types considered. The top row displays the model s accuracy on the Pick-a-Pic test set (left) and its Spearman s rank correlation on the HPS test set (right). The bottom row shows the model s choice accuracy (left) and cross-turn preference accuracy (right), both evaluated on our human-rated test set. 6.2. User Model Evaluation We evaluate the quality of our user model by assessing its ability to capture user preferences and generate realistic interaction data. We train the model using large-scale single-turn datasets (HPS V2 (Wu et al., 2023), Pick-a-Pic (Kirstain et al., 2023), and Simulacra Aesthetic Captions (Pressman et al., 2022)) and fine-tune it with our humanrater sequential data. During training, CLIP model weights are initially frozen, focusing on learning user image and text encoder parameters. See Appendix D for full training details, including loss functions and architectures. We analyze model performance in two stages. First, we assess prediction accuracy of the and Pick-a-Pic testset (Kirstain et al., 2023) and ranking using Spearman s rank correlation (Spearman, 1961) on the HPS dataset (Wu et al., 2023). Second, we evaluate prompt choice prediction accuracy and cross-turn preference accuracy on our human-rated data. Results in Figure 4 indicate that increasing the number of user types significantly improves performance, plateauing around 16 types. This highlights the importance of modeling diverse user preferences. Notably, even with a moderate number of user types, our model achieves approximately 70% accuracy on the human-rated dataset metrics. A further notable observation is the emergence of visually distinct domain preferences for different user types; this tends to become more prominent as the number of user types increases. An illustration for the 32-type model is shown in Figure 5. 6.3. PASTA Results Using our user model, we generate more than 30,000 simulated user-agent trajectories. We also generate reward labels Figure 5: Emergence of user-specific preferences: Each row displays the top five images scored by the user model, prior to fine-tuning from the HPS test set for one of five specific user types. We highlight user types where differences across the top five images were especially salient. The category labels (Animals, Food, etc.) are simply meant to be evocative on the style or content of the most preferred images. for our human-rater data using the user model, and train our state-action value function in three different settings: using only human rater data, using only simulated data, and using both data sources. The T2I agents trained by these algorithms are evaluated with human raters, for which the raters judge whether the current turn s chosen image was better, worse, or equally preferred to that in the previous turn. To make PASTA more efficient for human-rater studies, we distill the T2I agent into a single, fine-tuned Gemini 1.5 Flash (Gemini-Team, 2024) LMM, and serve that in real-time to generate the proposed image slates directly (without explicit prompt expansions). We also conduct an experiment with simulated users, using our user model, whose results are described in Appendix F together with other experiments for both PASTA and the user model. We compare our agents (each using a different state-action value function) with an untrained multi-modal Gemini 1.5 Flash model. The results of human-rater evaluations are shown in Figure 6. These results demonstrate that training on real human data enhances our agents performance over the pre-trained LMM baseline (i.e., over the off-the shelf Gemini 1.5 Flash), whereas training with synthetic data alone leads to a slight decline in performance. Unsurprisingly, training on both real and synthetic datasets offers the best performance. This suggests that, while our user simulation may not fully reflect the real-world distribution of user behaviors (partially due to its simplified assumptions), the additional data generated by our user model does capture some key dynamics of real user interactions, yielding more accurate user choice predictions and sequentially generated Preference Adaptive and Sequential Text-to-Image Generation Figure 6: PASTA Evaluation. We ask raters to determine, for each interaction turn, whether the shown images are an improvement over the previous turn. The raters can choose one of three options: Better, Worse, or Same. Results show percentage of raters who chose each option. Experiments compare the base Gemini Flash model (without fine-tuning) to PASTA (also using Gemini Flash) varied by training on different datasets: (1) only simulated user data, (2) only real rater data, and (3) the combination of both datasets. Plots show 99% confidence intervals. Figure 7: PASTA Evaluation with the Score Model. We use our trained user model to evaluate the rollouts of real users with different baselines. images that are tailored to user preferences. Moreover, a high rank correlation between real user preferences and the predictions from our learned model over a vast number of user types further supports this conclusion. We also present the user model score for each step averaged over all rollouts in Figure 7. The user type was chosen as the one that maximizes the posterior over the user s data. The results show that PASTA monotonically improves the user score along the rollouts and are consistent with the real users results in Figure 6. This evaluation further validates the correlation between our user model and real users. Finally, we conduct new human rater evaluations over the final turn of PASTA. Particularly, we conduct a rater study where we explicitly show raters the final turn images (for a given prompt) of PASTA compared to the baseline Gemini 1.5 Flash model and ask them to compare the generated last turn images. We find that PASTA receives an 85% relative improvement over the baseline model when directly comparing the last turn. We believe that directly comparing the last turn emphasizes the strength of our approach and the significance of our results in terms of the outcome of using our model. 6.4. Generalization: Flux T2I model We use Flux.1 5 together with our PASTA agent and compare it to the baseline Gemini 1.5 Flash. Particularly, we do not retrain PASTA over new Flux.1 data, but rather test PASTA s capability to generalize from the SDXL T2I model to the Flux.1 T2I model. We present the raters the final images constructed by PASTA against the final images generated with the Gemini baseline. Indeed, our results show that PASTA achieves a 20% relative improvement over the baseline without any further training over Flux data. This highlights PASTA s ability to adapt to different T2I models. 6.5. Abstract Prompts with Simulated Users To visualize the differences in user preferences and PASTA s ability to adapt to them, we use PASTA with various user types using broad, abstract prompts such as "an image of happiness". Figure 8 shows an example of different rollouts of three distinct user types. We observe a clear preference emerging, favoring specific styles or content. All users starts with the same prompt and initial images. We present only the images and their corresponding prompt-expansion at the last 5-th step. Each color represent a different user type. For more examples, see Appendix F.2. 7. Related Work Our work is related to recent advances in interactive and preference-adaptive image generation. Existing methods explore iterative refinement using visual feedback (von Rütte et al., 2023; Wang, 2024) or leverage LLMs for prompt engineering (Liu et al., 2024; Brade et al., 2023; Hao et al., 2024), often focusing on optimization of perceptual quality or user satisfaction in a single interaction (Fan et al., 2023; Chidambaram et al., 2024; Wallace et al., 2024; Wu et al., 2023). Our PASTA framework extends this by framing multi-turn image generation as a sequential decision-making 5https://flux1.ai/ Preference Adaptive and Sequential Text-to-Image Generation "Tender embrace in soft golden light, detailed textures, warm tones, intimate moment" "An image of love" "Soft lighting illuminates a mother's tender touch on her child's face, enhancing the emotional depth" "Two elderly lovers sharing a quiet moment on a beach at dusk, soft light, detailed wrinkles" "A visual exploration of love's intensity, abstract forms, dynamic composition" "An image of love" "A tapestry of textures, rough and smooth interwoven, portraying the strengths and vulnerabilities within a loving family" "Intertwined geometric forms representing love, Cubist style, soft pastel palette, balanced composition" User Type A User Type B Figure 8: An illustration of three different rollouts of PASTA with two different user types for the abstract initial prompt: An image of love". Rollouts show the final images in the rollout (i.e., at step t = 5). problem, where an RL agent learns to extend and personalize a prompt through a series of expansions guided by user feedback. This enables iterative refinement towards a desired visual outcome, capturing nuances in user preferences across diverse user types. Notably, while many existing methods focus on modifying internal aspects of text-to-image (T2I) models (e.g., diffusion noise (Tang et al., 2023), attention maps (von Rütte et al., 2023) or prompt embeddings (Salehi et al., 2025)), PASTA operates by directly adapting the input prompt, offering a model-agnostic approach to adaptive preference learning in T2I generation. PASTA draws upon a rich history of preference elicitation (PE) research (Boutilier, 2002; Chajewska et al., 2000; Keeney, 1993; Salo & Hamalainen, 2001), especially that in content-based recommender systems (Boutilier et al., 2003; Spearman, 1961; Rashid et al., 2002; Meshi et al., 2023), and addresses the challenge of optimizing PE for diverse downstream tasks. Existing research highlights the importance of diversity in recommendations (Sidana et al., 2018) and explores diversity in elicitation strategies (Parapar & Radlinski, 2021). PASTA tackles this by training an agent that considers both immediate user feedback and long-term goals to generate preference-aligned images. Our novel data collection and simulation methodology further support the training of robust user simulators, which we show to greatly improve policy learning. 8. Conclusion This work introduced PASTA, a novel RL agent for preference-adaptive and sequential T2I generation. We formulated the problem of iterative prompt expansions with LMMs and T2I models as a sequential decision-making problem that drives a collaborative, multi-turn image generation with the user. Critically, we have produced a new dataset that captures sequential interactions between an agent and a user, which we release to support further investigation of new T2I techniques in the research community. We developed user utility and choice models, learned from this dataset, which we used to create a user simulator that enabled generation of additional synthetic data (which we release as well). Finally, we used our sequential data to train PASTA, and demonstrated its effectiveness through comprehensive human evaluations, showing significant improvements in user satisfaction. Impact Statement This paper presents work whose goal is to advance the field of Machine Learning. There are many potential societal consequences of our work, none which we feel must be specifically highlighted here. Preference Adaptive and Sequential Text-to-Image Generation Aberdeen, D., Buffet, O., and Thomas, O. Policy-gradients for psrs and pomdps. In Artificial intelligence and statistics, pp. 3 10. PMLR, 2007. Achiam, J., Adler, S., Agarwal, S., Ahmad, L., Akkaya, I., Aleman, F. L., Almeida, D., Altenschmidt, J., Altman, S., Anadkat, S., et al. Gpt-4 technical report. ar Xiv preprint ar Xiv:2303.08774, 2023. Boutilier, C. A pomdp formulation of preference elicitation problems. In AAAI/IAAI, pp. 239 246. Edmonton, AB, 2002. Boutilier, C., Zemel, R. S., and Marlin, B. Active collaborative filtering. In Proceedings of the Nineteenth Annual Conference on Uncertainty in Artificial Intelligence (UAI03), pp. 98 106, 2003. Brade, S., Wang, B., Sousa, M., Oore, S., and Grossman, T. Promptify: Text-to-image generation through interactive prompt exploration with large language models. In Proceedings of the 36th Annual ACM Symposium on User Interface Software and Technology, pp. 1 14, 2023. Bradley, R. A. and Terry, M. E. Rank analysis of incomplete block designs: I. the method of paired comparisons. Biometrika, 39(3/4):324 345, 1952. Chajewska, U., Koller, D., and Parr, R. Making rational decisions using adaptive utility elicitation. In Aaai/Iaai, pp. 363 369, 2000. Chidambaram, K., Seetharaman, K. V., and Syrgkanis, V. Direct preference optimization with unobserved preference heterogeneity. ar Xiv preprint ar Xiv:2405.15065, 2024. Croitoru, F.-A., Hondru, V., Ionescu, R. T., and Shah, M. Diffusion models in vision: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45 (9):10850 10869, 2023. Dorfman, R., Shenfeld, I., and Tamar, A. Offline meta reinforcement learning identifiability challenges and effective data collection strategies. Advances in Neural Information Processing Systems, 34:4607 4618, 2021. Duan, Y., Schulman, J., Chen, X., Bartlett, P. L., Sutskever, I., and Abbeel, P. RL2: Fast reinforcement learning via slow reinforcement learning. ar Xiv preprint ar Xiv:1611.02779, 2016. Fan, Y., Watkins, O., Du, Y., Liu, H., Ryu, M., Boutilier, C., Abbeel, P., Ghavamzadeh, M., Lee, K., and Lee, K. Reinforcement learning for fine-tuning text-to-image diffusion models. In Advances in Neural Information Processing Systems, volume 36, 2023. Finn, C., Abbeel, P., and Levine, S. Model-agnostic metalearning for fast adaptation of deep networks. In International conference on machine learning, pp. 1126 1135. PMLR, 2017. Gemini-Team. Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context. ar Xiv preprint ar Xiv:2403.05530, 2024. Gemma-Team. Gemma: Open models based on gemini research and technology. ar Xiv preprint ar Xiv:2403.08295, 2024. Gu, J., Zhai, S., Zhang, Y., Susskind, J. M., and Jaitly, N. Matryoshka diffusion models. In The Twelfth International Conference on Learning Representations, 2023. Hafner, D., Lillicrap, T., Norouzi, M., and Ba, J. Mastering atari with discrete world models. ar Xiv preprint ar Xiv:2010.02193, 2020. Hallak, A., Di Castro, D., and Mannor, S. Contextual markov decision processes. ar Xiv preprint ar Xiv:1502.02259, 2015. Hao, Y., Chi, Z., Dong, L., and Wei, F. Optimizing prompts for text-to-image generation. Advances in Neural Information Processing Systems, 36, 2024. Ho, J. and Salimans, T. Classifier-free diffusion guidance. ar Xiv preprint ar Xiv:2207.12598, 2022. Ho, J., Jain, A., and Abbeel, P. Denoising diffusion probabilistic models. Advances in neural information processing systems, 33:6840 6851, 2020. Holtzman, A., Buys, J., Du, L., Forbes, M., and Choi, Y. The curious case of neural text degeneration. ar Xiv preprint ar Xiv:1904.09751, 2019. Ie, E., Jain, V., Wang, J., Narvekar, S., Agarwal, R., Wu, R., Cheng, H.-T., Chandra, T., and Boutilier, C. Slate Q: A tractable decomposition for reinforcement learning with recommendation sets. In International Joint Conference on Artifical Intelligence (IJCAI), pp. 2592 2599, Macau, 2019. Kaiser, L., Babaeizadeh, M., Milos, P., Osinski, B., Campbell, R. H., Czechowski, K., Erhan, D., Finn, C., Kozakowski, P., Levine, S., et al. Model-based reinforcement learning for atari. ar Xiv preprint ar Xiv:1903.00374, 2019. Keeney, R. L. Decisions with multiple objectives: Preferences and value tradeoffs. Cambridge university press, 1993. Preference Adaptive and Sequential Text-to-Image Generation Kirstain, Y., Polyak, A., Singer, U., Matiana, S., Penna, J., and Levy, O. Pick-a-pic: An open dataset of user preferences for text-to-image generation. Advances in Neural Information Processing Systems, 36:36652 36663, 2023. Kostrikov, I., Nair, A., and Levine, S. Offline reinforcement learning with implicit q-learning. ar Xiv preprint ar Xiv:2110.06169, 2021. Kwon, J., Efroni, Y., Caramanis, C., and Mannor, S. Rl for latent mdps: Regret guarantees and a lower bound. Advances in Neural Information Processing Systems, 34: 24523 24534, 2021. Liang, Y., He, J., Li, G., Li, P., Klimovskiy, A., Carolan, N., Sun, J., Pont-Tuset, J., Young, S., Yang, F., et al. Rich human feedback for text-to-image generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 19401 19411, 2024. Liu, X.-Y., Zhou, X.-H., Gui, M.-J., Xie, X.-L., Liu, S.-Q., Wang, S.-Y., Li, H., Xiang, T.-Y., Huang, D.-X., and Hou, Z.-G. Domain: Mildly conservative model-based offline reinforcement learning. ar Xiv preprint ar Xiv:2309.08925, 2023. Liu, Y., He, M., Yao, F., Ji, Y., Tao, S., Du, J., Li, D., Gao, J., Zhang, L., Yang, H., et al. What do you want? usercentric prompt generation for text-to-image synthesis via multi-turn guidance. ar Xiv preprint ar Xiv:2408.12910, 2024. Loshchilov, I. and Hutter, F. Sgdr: Stochastic gradient descent with warm restarts. ar Xiv preprint ar Xiv:1608.03983, 2016. Loshchilov, I. and Hutter, F. Decoupled weight decay regularization. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019, 2019. Meshi, O., Feldman, J., Yang, L., Scheetz, B., Cai, Y., Bateni, M., Salisbury, C., Aggarwal, V., and Boutilier, C. Preference elicitation for music recommendations. In ICML 2023 Workshop The Many Facets of Preference Based Learning, 2023. Micheli, V., Alonso, E., and Fleuret, F. Transformers are sample-efficient world models. ar Xiv preprint ar Xiv:2209.00588, 2022. Parapar, J. and Radlinski, F. Diverse user preference elicitation with multi-armed bandits. In Proceedings of the 14th ACM international conference on web search and data mining, pp. 130 138, 2021. Pennock, D. M., Horvitz, E., Giles, C. L., et al. Social choice theory and recommender systems: Analysis of the axiomatic foundations of collaborative filtering. AAAI/IAAI, 30:729 734, 2000. Podell, D., English, Z., Lacey, K., Blattmann, A., Dockhorn, T., Müller, J., Penna, J., and Rombach, R. Sdxl: Improving latent diffusion models for high-resolution image synthesis. ar Xiv preprint ar Xiv:2307.01952, 2023. Pressman, J. D., Crowson, K., and Contributors, S. C. Simulacra aesthetic captions. Technical Report Version 1.0, Stability AI, 2022. url https://github.com/JDP/simulacra-aesthetic-captions . Radford, A., Kim, J. W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., Sastry, G., Askell, A., Mishkin, P., Clark, J., et al. Learning transferable visual models from natural language supervision. In International conference on machine learning, pp. 8748 8763. PMLR, 2021. Rashid, A. M., Albert, I., Cosley, D., Lam, S. K., Mc Nee, S. M., Konstan, J. A., and Riedl, J. Getting to know you: learning new user preferences in recommender systems. In Proceedings of the 7th international conference on Intelligent user interfaces, pp. 127 134, 2002. Rombach, R., Blattmann, A., Lorenz, D., Esser, P., and Ommer, B. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 10684 10695, 2022. Saharia, C., Chan, W., Saxena, S., Li, L., Whang, J., Denton, E. L., Ghasemipour, K., Gontijo Lopes, R., Karagol Ayan, B., Salimans, T., et al. Photorealistic text-to-image diffusion models with deep language understanding. Advances in neural information processing systems, 35: 36479 36494, 2022. Salehi, S., Shafiei, M., Yeo, T., Bachmann, R., and Zamir, A. Viper: Visual personalization of generative models via individual preference learning. In European Conference on Computer Vision, pp. 391 406. Springer, 2025. Salo, A. A. and Hamalainen, R. P. Preference ratios in multiattribute evaluation (prime)-elicitation and decision procedures under incomplete information. IEEE Transactions on Systems, Man, and Cybernetics-Part A: Systems and Humans, 31(6):533 545, 2001. Shazeer, N. and Stern, M. Adafactor: Adaptive learning rates with sublinear memory cost. In International Conference on Machine Learning, pp. 4596 4604. PMLR, 2018. Preference Adaptive and Sequential Text-to-Image Generation Sidana, S., Laclau, C., and Amini, M.-R. Learning to recommend diverse items over implicit feedback on pandor. In Proceedings of the 12th ACM Conference on Recommender Systems, pp. 427 431, 2018. Snell, C., Kostrikov, I., Su, Y., Yang, M., and Levine, S. Offline rl for natural language generation with implicit language q learning, 2022. URL https://arxiv. org/abs/2206.11871. Spearman, C. The proof and measurement of association between two things. 1961. Sutton, R. S. Reinforcement learning: An introduction. A Bradford Book, 2018. Tang, Z., Rybin, D., and Chang, T.-H. Zeroth-order optimization meets human feedback: Provable learning via ranking oracles. ar Xiv preprint ar Xiv:2303.03751, 2023. Tennenholtz, G., Chow, Y., Hsu, C.-W., Shani, L., Liang, E., and Boutilier, C. Embedding-aligned language models. In Advances in Neural Information Processing Systems, volume 37, 2024. Vaswani, A. Attention is all you need. Advances in Neural Information Processing Systems, 2017. Von Neumann, J. and Morgenstern, O. Theory of games and economic behavior: 60th anniversary commemorative edition. In Theory of games and economic behavior. Princeton university press, 2007. von Rütte, D., Fedele, E., Thomm, J., and Wolf, L. Fabric: Personalizing diffusion models with iterative feedback. ar Xiv preprint ar Xiv:2307.10159, 2023. Wallace, B., Dang, M., Rafailov, R., Zhou, L., Lou, A., Purushwalkam, S., Ermon, S., Xiong, C., Joty, S., and Naik, N. Diffusion model alignment using direct preference optimization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 8228 8238, 2024. Wang, C. A diffusion-based method for multi-turn compositional image generation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 381 391, 2024. Wang, J. X., Kurth-Nelson, Z., Tirumala, D., Soyer, H., Leibo, J. Z., Munos, R., Blundell, C., Kumaran, D., and Botvinick, M. Learning to reinforcement learn. ar Xiv preprint ar Xiv:1611.05763, 2016. Wu, X., Sun, K., Zhu, F., Zhao, R., and Li, H. Human preference score: Better aligning text-to-image models with human preference. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 2096 2105, 2023. Yang, L., Zhang, Z., Song, Y., Hong, S., Xu, R., Zhao, Y., Zhang, W., Cui, B., and Yang, M.-H. Diffusion models: A comprehensive survey of methods and applications. ACM Computing Surveys, 56(4):1 39, 2023. Yu, J., Xu, Y., Koh, J. Y., Luong, T., Baid, G., Wang, Z., Vasudevan, V., Ku, A., Yang, Y., Ayan, B. K., Hutchinson, B., Han, W., Parekh, Z., Li, X., Zhang, H., Baldridge, J., and Wu, Y. Scaling autoregressive models for contentrich text-to-image generation. Transactions on Machine Learning Research, 2022. Yu, T., Kumar, A., Rafailov, R., Rajeswaran, A., Levine, S., and Finn, C. Combo: Conservative offline model-based policy optimization. Advances in neural information processing systems, 34:28954 28967, 2021. Zhang, C., Zhang, C., Zhang, M., and Kweon, I. S. Text-toimage diffusion models in generative AI: A survey. ar Xiv preprint ar Xiv:2303.07909, 2023. Zhou, Y., Zanette, A., Pan, J., Levine, S., and Kumar, A. Archer: Training language model agents via hierarchical multi-turn rl. ar Xiv preprint ar Xiv:2402.19446, 2024. Zintgraf, L., Shiarlis, K., Igl, M., Schulze, S., Gal, Y., Hofmann, K., and Whiteson, S. Vari BAD: A very good method for bayes-adaptive deep RL via meta-learning. In 8th International Conference on Learning Representations, ICLR, 2020. Preference Adaptive and Sequential Text-to-Image Generation A. Latent Markov Decision Process (LMDP) We begin with defining Latent Markov Decision Process introduced in (Hallak et al., 2015; Kwon et al., 2021). Definition A.1. Suppose that a set of MDPs M with a joint state space S and joint action space A in finite horizon setting with horizon H. let K = |M|, S = |S| and A = |A|. Each MDP Mk M it a tuple (S, A, Tk, Rk, νk) where Tk : S A S is a transition probability that maps a state-action pair into a distribution over the next state, Rk : S A 7 [0,1] a probability measure for rewards that maps a state-action pair into a distribution over rewards and initial state distribution νk. Let η1, η2, . . . , ηK be the mixing weights of LMDPs such that at the start of every episode, one MDP Mk M is randomly chosen with probability ηk. The goal of the problem is to find a policy π within a policy class Π that maximizes the expected return: v M = max π Π k=1 ηk Eπ k where Eπ k[ ] is expectation taken over the k-th MDP with policy π. The policy π : H S 7 A maps the current state and history into distribution over actions. Generally, the history is all experience seen so far H = (S, A, [0, 1]) . When the model parameters are known, history is a sufficient statistics and can be summarized into belief states: b1(k) = ηkνk(s1) P k ηkνk(s1), bt+1(k) = bt(k)Tk(st+1|st, at)Rk(rt|st, at) P k bt(k)Tk(st+1|st, at)Rk(rt|st, at). We formulate our reinforcement learning problem as a latent MDP, where each user from a discrete set induces distinct transition and reward kernels. Specifically, in our sequential text-to-image decision problem the state space is a prompt and image slate S P IML and the action space is the prompt slate A PL. The transition kernel for the k-th user type is composed of the T2I model and user choice model, i.e. the next state is the tuple composed of slate of images generated from the chosen slate (the action) and chosen prompt: st+1 = n {Im,ℓ,t}M m=1 o L where Im,ℓ,t G(pℓ,t) and ct C(k, p0, I1,1,t, . . . , IM,L,t). The reward kernel corresponds to the user s utility function: rt R(k, p0, I1,ℓt,t, . . . , IM,ℓt,t). Similar to standard LMDPs, the posterior update during user model training in ?? mirrors the belief state computation, driven by observed samples collected from diverse human raters. Preference Adaptive and Sequential Text-to-Image Generation B. Human Rater Dataset To gather data for training and evaluating PASTA, we conducted a human rater study. We recruited paid contractors as raters, and utilized a web-based interface where raters interacted with a Gemini 1.5 Flash Model acting as the agent in our multi-turn image generation process. Column selection An optional critique (e.g., add more color) submit button (move to next turn) Figure 9: A screenshot of the interface raters used in the interaction phase. Raters were not shown the prompt expansions. They were instructed to choose a preferred column (based on the best image of each column). Raters could optionally also add an up to 12 word critique. B.1. Rater Instructions: Interaction Phase Each rater participated in a five-turn interaction session. At the beginning of the session, raters were presented with the task s instructions. They were then asked to provide an initial text prompt, with a maximum length of 12 words, representing a specific visual concept they wanted to see realized as an image. The guidelines stressed that prompts should encapsulate meaningful intent (thinking about a specific image they wished to see) and avoid generic prompts (e.g., a temple ). Instead, they were encouraged to use descriptive prompts within the 12-word limit (e.g., An ancient Egyptian temple with hieroglyphs ). They were assured that they need not worry about fitting all of their thoughts into the 12-word limit, as the interaction process with the agent would help them refine their concept. Preference Adaptive and Sequential Text-to-Image Generation Figure 10: Illustration shown to raters on how to select a column. Columns should be compared based on the best image of each column. In each of the subsequent five turns, the LMM agent presented the rater with 16 images, arranged in four columns (L=4, M=4). Each column represented a different prompt expansion derived from the rater s initial prompt and their choices in previous turns. Figure 9 illustrates this setup. Critically, the raters were only shown the generated images, not the expanded prompts themselves. This ensured their feedback was based purely on visual perception and alignment with their initial intent, rather than on their interpretation of the textual modifications made by the agent. The core task for the rater in each turn was to select the column of images they preferred most. The instructions, visually depicted in Figure 10, emphasized that this selection should be based on the best image within each column, judged in terms of general quality and its correspondence with the rater s original prompt intent. For example, if Image 2A was the best overall image of the 16, then Column 2 should be selected. This selection strategy was used to help reduce burden on raters. It also benefited our choice model training procedure, which used this as inductive bias for training our model. Optionally, raters could provide a free-text critique of the presented images, limited to 12 words (Figure 9). This critique could address any aspect of the images, such as style, composition, color, or content, but raters were explicitly instructed not Preference Adaptive and Sequential Text-to-Image Generation Figure 11: Screenshot of the evaluation phase for raters. This second phase (after the interaction phase) included questions about every turn. See more details in Appendix B.2. to use the critique field to change their original intent. For example, "Add more color" is a valid critique. The purpose of the critique was to provide additional signals to the agent for refinement within the scope of the original visual concept, not to redirect the generation process towards a different image altogether. Overall, approximately 20% of the collected data have one or more critiques. B.2. Rater Instructions: Evaluation Phase After completing five interaction turns, raters were prompted to click the "Enter rater mode" button. The system then transitioned to the evaluation phase. For each of the five turns, raters answered a set of questions designed to capture detailed feedback on their experience and the agent s performance. Figure 11 demonstrates the interface used during this stage. For each turn, raters were asked the following questions, with their chosen column highlighted for reference: 1. Column Preference: Raters were asked to re-confirm their preferred column for each turn, ensuring consistency in their choices (See Figure 11). 2. Absolute Column Satisfaction: For each of the four columns in each turn, raters provided an absolute satisfaction score on a 5-point scale (1: very dissatisfied, 5: very satisfied). This score reflected their satisfaction with the best image in the column, relative to their original prompt intent. This absolute scoring allowed for a finer-grained assessment of the individual prompt expansions, capturing nuances beyond the simple relative ranking implied by column selection. 3. Selection Rationale: For their selected column in each turn, raters were asked to provide a free-text explanation of their choice. 4. Turn-over-Turn Improvement: For turns 2-5, raters assessed whether the set of images in the current turn was better, worse, or about the same as the set of images in the previous turn. This comparison, again focused on the quality of the selected columns in each turn relative to the original prompt intent, provided a measure of the agent s progress and its ability to refine the generated images over time. 5. Overall Turn Satisfaction: Raters provided an overall absolute satisfaction score (1-5 scale) for all of the images presented in each turn. This average score across columns gave a holistic impression of the agent s performance in that turn, complementing the column-specific feedback. Preference Adaptive and Sequential Text-to-Image Generation Raters were finally instructed to press the "End Task" button (as shown in Figure 11) to complete their session, which submitted their responses for all five turns. B.3. Limitations of Rater Study A recognized limitation of the rater study conducted in this work is the potential bias between groups of raters. In Figure 6 we compare PASTA trained on different datasets with the pre-trained Gemini Flash model. We could not concurrently run different experiment arms because of a lack of A/B testing infrastructure which shows different agents to raters at random. Consequently, the timing of experiment arms and rater availability could have introduced bias. Furthermore, the rater pool might not be fully representative of all potential user demographics for PASTA, and the initial prompts provided to raters might not have sufficient diversity. Human errors and system issues, such as connection timeouts, were observed during the study. We did not implement any formal mechanisms for raters to report errors during specific sessions although raters can somehow correct their mistakes during the evaluation phase. Raters could prematurely abandon sessions without recording data from problematic sessions. Raters could provide a critique of the presented images but subsequent generations might not be aligned with these critiques. Although raters were explicitly instructed not to evaluate based on critique alignment, it might still cause worse rater experience for those who provided critiques. B.4. Generating Rewards for Human Rater Data with User Model To leverage human rater data for value model training, we utilize our learned user utility function to generate rewards for each interaction step. For each trajectory, we first compute the posterior distribution over user types using the model described in Appendix D.3. We then employ the utility function of the most likely user type to generate rewards for each timestep in the trajectory. Preference Adaptive and Sequential Text-to-Image Generation C. Offline Simulated Sequential Data This section describes the generation of an offline dataset for training PASTA. We detail the process of creating interaction trajectories with simulated users, based on our trained user model and a random prompt selection policy. We also address the challenge of MDP ambiguity inherent in offline learning, specifically in the context of diverse user types, and present conditions for ensuring identifiability of the offline dataset. C.1. Offline Dataset Creation Leveraging our trained user model and user type prior, we generated a dataset comprising over 30,000 simulated user interaction trajectories. We assume independence between the initial prompt distribution and the user type distribution, i.e., ν(p0, u) = ν(p0)η0(u). Therefore, at the beginning of each trajectory, we approximate the true user prior with the learned prior from user model training (η0 η) and sample initial prompts uniformly from the available datasets, while limiting the initial prompts to be with no more than 10 words in order to avoid very large prompt at the end of the rollout. To ensure diverse and representative data for training, the data generation policy needs to produce coherent prompts that encourage exploration and sufficiently cover the state-action space, thereby mitigating potential MDP ambiguity (discussed further in the next subsection). We achieve this by using the candidate generator detailed in Appendix E.2 coupled with a random selection policy, denoted as πrand. This policy is constrained to select a maximum of one prompt per category, promoting exploration across different prompt categories. At each interaction step, the candidate generator produces a categorized set of LC prompts. The random policy πrand then samples a slate of L prompts, for which the text-to-image model generates M images each. A prompt ct is selected according to the user choice model, and the corresponding user satisfaction level (reward) is recorded based on the utility function defined in Section 2.2. C.2. Identifiable Offline Dataset After constructing the user choice model and utility function, we use a user simulator to generate an offline dataset for PASTA training. The offline reinforcement learning framework is chosen due to its computational efficiency during training. However, offline LMDP introduces the challenge of Markov Decision Process (MDP) ambiguity (Dorfman et al., 2021), and in our context, user type ambiguity: Definition C.1. User type ambiguity (Dorfman et al., 2021). Consider data generated from a set of K user types U = [1, 2, . . . , K], each corresponding to a user choice model and utility function {Ci, Ri}K i=1. The data is collected by sampling a subset of N user types {ui}N i=1 U, and applying their respective behavior policies {πβ i }N i=1. This results in N data distributions {ρui,πβ i }N i=1 over image slates, initial prompt and rewards, with ρ : P IML [0, 1] 7 [0, 1]. We define the collected data as ambiguous if there exists a user type u U and two policies π, π such that ρui,πβ i = ρu ,π and ρuj,πβ j = ρu ,π for some i = j. Otherwise, the data is considered identifiable. In essence, the data is ambiguous if a user type under different policies results in same data distributions of two different user types and their corresponding behavioral policies, regardless of how much data is collected. By introducing the concepts of identifying state-action pairs and overlapping state-actions, we can define a sufficient condition for identifiability. Definition C.2. Identifying Slate of Images.For a pair of user types i, j U, we define (p0, {Im,ℓ}M,L m=1,ℓ=1) as an identifying slate of images if R(i, p0, {Im,ℓ}M,L m=1,ℓ=1) = R(j, p0{Im,ℓ}M,L m=1,ℓ=1) or C(i, p0, {Im,ℓ}M,L m=1,ℓ=1) = C(j, p0, {Im,ℓ}M,L m=1,ℓ=1). Definition C.3. Overlapping Slate of Images. Given data collected as described in Definition C.1. a slate of images (p0, {Im,ℓ}M,L m=1,ℓ=1) is said to overlap for a pair of user types i, j U if it has a positive slate-probability under both user types, i.e., ρi,πβ i (p0, {Im,ℓ}M,L m=1,ℓ=1) > 0 and ρj,πβ j (p0, {Im,ℓ}M,L m=1,ℓ=1) > 0. A sufficient condition for identifiability is as follows: Proposition C.4. (Dorfman et al., 2021). Consider a data collected according to the settings in Definition C.1. The data is identifiable if, for every pair of distinct user types i = j, there exists an identifying slate of images that overlaps. Given that in our scenario user types vary not only in their utility functions but also in their choice models, the candidate Preference Adaptive and Sequential Text-to-Image Generation generator combined with the random sampler πrand is likely to satisfy the conditions in Proposition C.4. This approach introduces sufficient randomness to explore a wide range of states, allowing us to reach identifying slates of images for all user types, while still maintaining a coherent policy capable of generating high-quality prompts. In the rare case where two user types generate identical data distributions under πrand, we consider them the same user type and merge them. D. User Model We employ an Expectation-Maximization (EM) framework to learn a user model capable of capturing diverse preferences. This model leverages a score function sθ that assesses the compatibility between image-prompt pairs and different user types. The EM algorithm iteratively refines the model parameters θ and a user type prior distribution η to maximize the likelihood of observed user feedback. Training proceeds in two phases: a main training phase on large-scale datasets with frozen CLIP parameters, followed by a fine-tuning phase on human-rated data where all parameters are trained. We detail the training procedure, specific loss functions tailored to different data types (single-turn preferences, relevance scores, and novel sequential multi-turn data), the model architecture, and hyperparameter settings in the following subsections. D.1. Training The user model, which always incorporates the score function sθ, is adapted to suit different datasets and label types. Specific model formulations and loss functions are detailed in subsequent subsections. D and estimate the log-likelihood loss as follows: 1. Mini-Batch Sampling: A mini-batch B is sampled from the dataset, and the posterior is estimated solely for samples within B using a target network to prevent overestimation. 2. Prior Update: An exponential moving average of the mini-batch posteriors is used to update the prior. 3. Gradient Descent: Instead of fully optimizing the model parameters, a single gradient descent step is taken with respect to the loss function: L(θ) = Ei B,k γi[log σθ(xi, yi, k)]. The user model is always composed of the score function sθ and may vary for different types of datasets and labels, and need to be adjusted accordingly. We provide detailed models and loss function at the following sub-sections. Algorithm 1 provides a complete description of the mini-batch training procedure. Training proceeds in two phases: 1. Main Training: The model is trained on a combination of large-scale datasets: HPS V2 (Wu et al., 2023), Pick-a-Pic (Kirstain et al., 2023), and Simulacra Aesthetic Captions (SAC) (Pressman et al., 2022). The CLIP model parameters are frozen during this phase. 2. Fine-tuning with Sequential Data: A shorter fine-tuning phase follows, utilizing human-rated data. Here, the entire model, including the CLIP model, is trained. Preference Adaptive and Sequential Text-to-Image Generation Algorithm 1 Mini-Batch Expectation-Maximization User Model Optimization 1: Dataset D, number of user types K, parameter αprior, learning rate β. 2: Initialize η = (1/K, . . . , 1/K), target network θ = θ0 3: for t in {0, . . . , T} do 4: Sample a mini-batch B D. 5: E-step. Calculate posterior γi,k for each example (xi, yi) B and user class k [K]: γi(k) ηk QM j=1 σ θ(xi,j, yi,j, k) PK ℓ=1 ηℓ QM j=1 σ θ(xi,j, yi,j, ℓ) . 6: M-step. Update parameters θ, η: η αpriorη + (1 αprior)Ei B[γi] θt+1 = θt β L(θt) 7: Every n steps: Update target network θ θt 8: end for 9: Return: model weights θT . D.2. Losses and Models - Single Step The EM framework is a general method applicable to various dataset and model types. In this work, we utilize two common single turn dataset types: Preference Model (HPS, Pick-a-Pic): This model utilizes data samples consisting of two images generated by the same prompt, xi = (Ii 1, Ii 2, pi), where the annotator indicates their preference, yi = (yi 1 yi 2). To connect the model to the score function, we employ the Bradley-Terry (BT) model (Bradley & Terry, 1952), adapted from the RLHF reward learning framework to accommodate different user types. The user-specific BT model defines the probability that a user type k prefers the first image over the other as: σpref θ (xi, yi, k) = exp{sθ(Ii 1, pi, k)} exp{sθ(Ii 1, pi, k)} + exp{sθ(Ii 2, pi, k)}. (5) In this case, the user loss function simplifies to: Lpref(θ) = Ei B,k γi σ(sθ(Ii 1, pi, k) sθ(Ii 2, pi, k)) where σ represents the Sigmoid function. Relevance model (SAC): This model provides a direct rating for an image-prompt pair, i.e., xi = (Ii, pi), with the label being the integer rating level provided by the annotator, yi = si [S]. We employ a simple user-conditional Gaussian model, where the score function acts as the mean estimator: σrel θ (xi, yi = si, k) exp (sθ(Ii, pi, k) si)2 where ν is a predefined standard deviation. The user loss for the relevance model is: Lrel(θ) = 1 2ν2 Ei B,k γi (sθ(Ii, pi, k) si)2 . When using relevance datasets, the rating levels are typically normalized to the range [0, 1]. To maximize the number of training samples for the user model, a mixture of data types can be used. In the case of a batch from each type, B = (Bpref, Brel), a mixture loss function can be employed to balance between the two terms: 2ν2 Ei Brel,k γi (sθ(Ii, pi, k) si)2 Ei Bpref ,k γi σ(sθ(Ii 1, pi, k) sθ(Ii 2, pi, k)) , where κ1 controls the balance between the two terms. Preference Adaptive and Sequential Text-to-Image Generation D.3. Losses and Models - Sequential In this work, we introduce a novel multi-turn dataset. This dataset, generated by querying users across H steps following the setting described in Section 2.2, consists of trajectories of user-agent interactions. Each sample xi = {xi,t}H t=1 has xi,t = {p0, {Im,ℓ}M,L m=1,ℓ=1} at each timestep t, with corresponding labels: 1. The chosen prompt index at each turn: ycm i,t = ci,t [L]. 2. In-turn preferences between the best images for each prompt: yin pref i,t = {(yi,t ℓ yi,t ℓ )}ℓ,ℓ Q, where Q represents the set of all unordered pairs of size |Q| = L 2 . 3. Cross-turn preferences between chosen images at consecutive turns: ycross pref i,t = (yi,t ℓ yi,t 1 ℓ ), resulting in H 1 cross-turn preferences per sample. Preferences within a sample are modeled by applying the BT model (Equation 5) to both in-turn and cross-turn preferences. User choices are modeled using our user choice model (Equation 4): σcm θ (xi,t, yi,t = ci,t, k) = Softmax(τθsk i,t)ci,t. This leads to the following cross-entropy loss for the score function and choice model parameters: Lcm(θ) = Ei,t B,k γi log(Softmax(τθsk i,t)ci,t) . In addition, we add a regularization term to stabilize the CLIP model finetuning by adding ℓ2 norm between the the original CLIP embeddings and the finetuned CLIP marked as Lreg(θ). The total multi-turn loss is then given by: L(θ) = Lcm(θ) + κ2Lin pref(θ) + κ3Lcross pref(θ) + κ4Lreg(θ), where κ2, κ3 and κ4 are balancing coefficients. D.4. Architecture Our score function utilizes pre-trained CLIP-Vi T-L-14 (Radford et al., 2021) text and image encoders, augmented with user-specific encoders (one per user type). These user encoders transform CLIP embeddings into user-type-specific representations, capturing individual preferences for images and prompts. The final score for an image-prompt pair is calculated as the inner product of the user-specific image and text embeddings, scaled by a learned temperature parameter (see a schematic of our arch. in Figure 3). We modified the CLIP encoders by removing the output l2-normalization. User preferences are modeled by a four-layer fully connected network with Re LU activation, structured as d CLIP d CLIP 2 d CLIP 4 d CLIP K, where d CLIP = 768 denotes the output dimension of the CLIP encoders. The final layer is partitioned into K vectors, each of dimension d CLIP , representing the different user types. These vectors act as residual offsets to the original CLIP output, individually added to it before a final ℓ2-normalization ensures each user score falls within the range [0, 1]. Each user score is multiplied by a learned temperature parameter. This architecture is illustrated in Figure 12. As for the temperature model, we use a two-layer fully connected network with Re LU activation, followed by a scaled Sigmoid function to bound the output in the range [0, τmax]. The network receives the utility score for each prompt candidate and produces the choice temperature scalar at the output. Preference Adaptive and Sequential Text-to-Image Generation CLIP Encoder CLIP Embeddings User Encoder K User Residual Embeddings Type Addition + L2 Normalizaion K User Embeddings Figure 12: A diagram of our user encoder. The same framework is applied to both the text and image encoding process. D.5. Hyperparameters Table 1: Main training hyperparameters Learning rate Cosine annealing scheduler (lr=3e-4, T=10e3) (Loshchilov & Hutter, 2016) Training steps 50e3 Batch size 2048 Update target network phase 256 Optimizer Adam W (Loshchilov & Hutter, 2019) (weight decay = 1e-4) κ1 1 αprior 0.999 Table 2: Fine-tuning hyperparameters Learning rate Cosine annealing scheduler (lr=3e-7, T=10e3) (Loshchilov & Hutter, 2016) Training steps 50e3 Batch size 8 Gradient norm clipping 0.5 Update target network phase 256 Optimizer Adam W (Loshchilov & Hutter, 2019) (weight decay = 1e-2) κ2 0.01 κ3 1 κ4 0.1 αprior 0.999 τmax 3 Preference Adaptive and Sequential Text-to-Image Generation E. PASTA Details We employ PASTA, a novel framework designed for prompt selection in interactive image generation scenarios. PASTA utilizes Implicit Q-Learning (IQL) to train the candidate selector policy that selects optimal prompt expansions based on user interaction history. This section details the training process, including the IQL objective and adaptations for token-level training; the architecture and functionality of the candidate generator and selector components, which employ large language models (LLMs); and the hyperparameters used for training. E.1. Training Generally, we would like to optimize the Q-Learning objective: L(ϕ) = Eht,Pt,rt,ht+1,P t+1 L D qϕ(ht, Pt) rt max P PL(P t+1 L ) q ˆϕ(h, P) where ˆϕ is the target network and P t L = {pt i}LC i=1 is the candidate set at time t. Decomposing the slate value function into an average of prompt value function we get L(ϕ) = Eht,Pt,rt,ht+1,P t+1 L D p Pt fϕ(ht, p) rt max P PL(P t+1 L ) q ˆϕ(h, P) Although the value function is trained on the value of entire utterances, we enhance its robustness by also optimizing it on sub-prompts, focusing on tokens within the prompts that share the same target: L(ϕ) = Eht,Pt,rt,ht+1,P t+1 L D t =1 fϕ(ht, p1:t ) rt max P PL(P t+1 L ) q ˆϕ(h, P) where Tp is the (token) length of prompt p. Offline Reinforcement Learning struggles with overestimation of values for unseen state-action pairs when using the Bellman optimality operator for TD targets. implicit Q-Learning (IQL) (Kostrikov et al., 2021) addresses this by learning an α-expectile value, vψ, which avoids the explicit maximization over the Q-function. The IQL objective is as follows: L(ϕ, ψ) = Eht,Pt,rt,ht+1 D h (qϕ(ht, Pt) rt vψ(ht+1))2 + Lα 2 q ˆϕ(ht, Pt) vψ(ht) i , where Lα 2 (x) = α 1{x<0} x2, α [0.5, 1]. Plugging in the decomposition of the value function and token-level training we get: L(ϕ, ψ) = Eht,Pt,rt,ht+1 D t =1 fϕ(ht, p1:t ) rt vψ(ht+1) p Pt f ˆϕ(ht, p) vψ(ht) E.2. Candidate Generator We utilize a multimodal Large Language Model (LLM) based on Gemini 1.5 Flash [Team, 2024] as our candidate generator policy. This LLM takes the user s current interaction history as input and generates a set of LC prompt expansion candidates, categorized into distinct groups. To avoid overly long prompts, especially in later interaction steps, we impose a length constraint of N t w words on the prompt at each step t = 1, . . . , H. This constraint is dynamically adjusted according to the formula: N t w = N t 1 w (N max w N t 1 w ) H t, Preference Adaptive and Sequential Text-to-Image Generation where N max w represents the maximum permissible word count at the final step (N H w N max w ) and N 0 w is the initial prompt s word count. Candidate generation employs nucleus sampling (Holtzman et al., 2019) with p = 0.8. The structure of the candidate generator prompts is illustrated in Figure 13. Candidate Generator Prompt Structure (example for t=3) You are an agent interacting with a user. The user wants to generate an image using an off-the-shelf diffusion model. The user starts by providing a prompt for the diffusion model. At every interaction step you (the agent) must decide on twenty-five (25) possible new prompts that would satisfy the user, based on your understanding of their preferences given your interaction history with them. Your prompts should be diverse enough to help you quickly capture the user preferences. A new prompt could be a change of the prompt itself, or an expansion of the prompt. Any new prompt should remain consistent with the original prompt. At every interaction slate four (4) of the prompt candidates will be used to generate slates of four (4) images each, for a total of sixteen (16) images presented to the user at every turn. The user will then select exactly one of the four (4) slates of images (i.e., corresponding to one of the new prompts). After five (5) interaction steps with the user, the user will choose one of the final image slates. Your objective is to find an optimal slate of images for the user in the final interaction, based on their selections. Output your slate in the following format. You should decide on five (5) categories, each listing exactly five (5) possible prompt candidates. The format should look like so: * - new_prompt_candidate_1 - new_prompt_candidate_2 ... - new_prompt_candidate_5 ... * - new_prompt_candidate_1 - new_prompt_candidate_2 ... - new_prompt_candidate_5 Below is the current interaction history with the user: Initial_prompt: "{{ initial_prompt }}" Initial_image_output: << image_samples >> AGENT: Agent shows four prompts (with corresponding slates of images) to the user: 1. Prompt: "{{ prompt_expansion 1}}", Samples: << image samples >> 2. Prompt: "{{ prompt_expansion 2}}", Samples: << image samples >> 3. Prompt: "{{ prompt_expansion 3}}", Samples: << image samples >> 4. Prompt: "{{ prompt_expansion 4}}", Samples: << image samples >> USER: User selects prompt number {{ image_number }} corresponding to the prompt: "{{ prompt_expansion }}" AGENT: Agent shows four prompts (with corresponding slates of images) to the user: 1. Prompt: "{{ prompt_expansion 1}}", Samples: << image samples >> 2. Prompt: "{{ prompt_expansion 2}}", Samples: << image samples >> 3. Prompt: "{{ prompt_expansion 3}}", Samples: << image samples >> 4. Prompt: "{{ prompt_expansion 4}}", Samples: << image samples >> USER: User selects prompt number {{ image_number }} corresponding to the prompt: "{{ prompt_expansion }}" Based on the interaction history above, please provide twenty-five (25) prompt candidates to the final selected prompt. Output should be divided into five (5) categories, each listing five (5) possible candidates, with the format described above. The prompt candidates should be prompts that can be used as input in a diffusion model. The category titles won't be used as input. One of the categories should be new prompts that are very different than the previous chosen prompt (for example rephrasing). Any prompt candidate should be at most {{ num_words }} words long. Figure 13: Template for the multimodal candidate generator prompt at step t = 3, with slate size L = 4, LC = 25 candidates across five categories (colors for visual distinction). E.3. Candidate Selector The candidate selector policy leverages the prompt-value function: πS,ϕ(h) arg max P PL {pi} LC i=1 qϕ(h, P) = arg max P PL {pi} LC i=1 X p P fϕ(h, p). This policy is initialized with a pre-trained Gemma 2B model (Gemma-Team, 2024) for prompt-value estimation. In order to encourage diversity and exploration, we constraint the candidate selector to pick only up to one prompt from every category. Given an input (h, p), represented as a token sequence of length T h + T p, the transformer produces an output of the same length, with vocabulary-sized logits for each token. We designate a specific logit, ℓq, as the "q logit." The value for a history-prompt pair is then defined as the value of this q logit at the final position in the output sequence, corresponding to Preference Adaptive and Sequential Text-to-Image Generation the last input token: fϕ(h, p) = {LLM([h, p]; ϕ)}ℓq, where [ , ] denotes concatenation. We also employ the same model for the α-expectile value estimator vψ (setting ψ ϕ): vϕ(h) = {LLM(h; ϕ)}ℓv, where ℓv represents the α-expectile value logit. Empirically, we observed optimal performance when ℓv and ℓq are identical. We prioritize efficiency and a small footprint for our prompt value function model, making Gemma 2B a suitable choice. Since Gemma 2B is a text-only model, we represent the history as text, encompassing the initial prompt, selected slates, and user-chosen prompts. When evaluating a history-prompt pair, the candidate prompt is also included in the input. The prompt template is detailed in Figure 14. History-Prompt Value Function Input Structure (example for t=3) <>{{ initial prompt }}<>{{prompt_1}}||{{prompt_2}}||{{prompt_3}}|| {{prompt_4}}<>{{ chosen_prompt }}<>{{prompt_1}}||{{prompt_2}}||{{prompt_3}}|| {{prompt_4}}<>{{ chosen_prompt }}<>{{ prompt_candidate }} History Value Function Input Structure (example for t=3) <>{{ initial prompt }}<>{{prompt_1}}||{{prompt_2}}||{{prompt_3}}|| {{prompt_4}}<>{{ chosen_prompt }}<>{{prompt_1}}||{{prompt_2}}||{{prompt_3}}|| {{prompt_4}}<>{{ chosen_prompt }} Figure 14: Text-only input templates for the value functions at step t = 3 with a slate size of L = 4. The top template corresponds to the history-prompt value function, while the bottom template corresponds to the prompt value function. E.4. Hyperparameters Table 3: PASTA hyperparameters Learning rate 1e-5 Training steps 1e4 Batch size 128 Optimizer Adafactor (Shazeer & Stern, 2018) (weight decay = 1e-2) Gradient norm clipping 1 Update target network phase 256 Expectile parameter α 0.7 ℓq 651 ℓv 651 L 4 M 4 LC 25 Number of categories 5 H 5 N max w 62 Preference Adaptive and Sequential Text-to-Image Generation F. Additional Experiments This section presents additional experiments conducted to further validate the effectiveness of our proposed framework. We first evaluate the performance of PASTA with simulated users under different training data regimes and reward settings. Subsequently, we showcase PASTA s ability to adapt to diverse user preferences by visualizing interactions with various user types on abstract image generation prompts. F.1. Simulated Users Experiment We performed an experiment with simulated users, leveraging our trained user model. Specifically, we evaluated our trained value models across three configurations: (1) trained using only human rater data, (2) trained using only simulated data, and (3) trained with a mixture of both. These models were compared to a random sampling policy to isolate the effect of the trained value models. All approaches employed the same candidate generator, which is based on the multimodal Gemini 1.5 Flash (Gemini-Team, 2024). The experiment was conducted in two reward settings: (1) dense reward, where the utility function provided rewards at every step, and (2) sparse reward, where rewards were given only at the final step, focusing on scenarios where only the final outcome matters. We evaluated all methods over 1000 trajectories, each generated as described in Section 2.2. The results are summarized in the table below: Random Human Simulated Mixed Sparse 7.38 1.13 8.7 0.11 9.26 0.07 8.81 0.06 Dense 38.57 5.81 43.49 0.54 46.35 0.38 44.03 0.32 The results demonstrate that all value-based methods outperform the random sampling policy. Notably, the value model trained exclusively on simulated data achieves the best performance. This outcome aligns with expectations, as the simulated data matches the dynamics of the simulated user model precisely. While human data offers some improvement in isolation, its divergence from the simulated dynamics reduces the effectiveness when combined with simulated training data. Preference Adaptive and Sequential Text-to-Image Generation F.2. Abstract Prompts with Simulated Users To visualize the differences in user preferences and PASTA s ability to adapt to them, we ran PASTA with various user types using broad, abstract prompts such as "an image of happiness." These prompts imposed minimal constraints on the user model, allowing each user type to express its unique preferences freely. For some user types, we observed a distinct preference emerging, favoring specific styles or content. All users starts with the same prompt and initial images. We present only the images and their corresponding prompt-expansion at the last 5-th step. Each color represent a different user type. The results are as follow: "An image of happiness": "Abstract happy faces, Art Deco inspired geometric shapes, muted jewel-toned background" "An image of happiness" "Dynamic interplay of simple shapes, limited colors, reflecting exuberant joy" "Abstract garden scene with birds, composed of geometric shapes and vibrant colors" "A vibrant, expressive piece of art representing the overwhelming feeling of happiness" "Hyperrealistic painting of ecstatic children playing in a vibrant, sun-drenched park, confetti falling, dynamic motion blur" "Watercolor painting of diverse happy children playing, soft colors, delicate brushstrokes, serene mood" "An image of happiness" "A field of luminous flowers, iridescent insects, and fantastical creatures bathed in the soft light of a double rainbow" "A field of sunflowers bathed in golden sunlight, with vibrant colors and intricate details, photorealistic" "Golden hour meadow, wildflowers bathed in warm sunlight, peaceful atmosphere, photorealistic" "An image of happiness" User Type A User Type B User Type C Preference Adaptive and Sequential Text-to-Image Generation "An image of love": "Tender embrace in soft golden light, detailed textures, warm tones, intimate moment" "An image of love" "Soft lighting illuminates a mother's tender touch on her child's face, enhancing the emotional depth" "Two elderly lovers sharing a quiet moment on a beach at dusk, soft light, detailed wrinkles" "A visual exploration of love's intensity, abstract forms, dynamic composition" "An image of love" "A tapestry of textures, rough and smooth interwoven, portraying the strengths and vulnerabilities within a loving family" "Intertwined geometric forms representing love, Cubist style, soft pastel palette, balanced composition" "An impressionistic portrayal of a family birthday celebration, emphasizing light, color, and movement." "An image of love" "Family dinner depicted from an unusual angle; unexpected perspective adds intrigue; focus on togetherness" "Intertwined geometric forms representing love, Cubist style, soft pastel palette, balanced composition" User Type A User Type B User Type C Preference Adaptive and Sequential Text-to-Image Generation "An image of hope": "Art Deco style illustration of a spring sunrise, geometric patterns, pastel colors, cherry blossoms on a lake" "An image of hope" "A surreal depiction of a fresh start, a blossoming flower at dawn, representing new love" "A powerful representation of inner peace, represented by a scene with gentle colors and forms" " A phoenix rising from ashes, surrounded by vibrant colors and dynamic energy" "An image of hope" "Playful, colorful light particles dance across a canvas, forming celestial patterns, symbolizing hope and limitless possibilities" "An abstract representation of hope, bright, radiant, and optimistic, colorful swirls" "Minimalist image of a gentle wave lapping the shore, representing peace and hop" "An image of hope" "The moon casts a soft glow on a beach covered in seashells, the ocean's gentle waves lapping at the shore" "Close-up photorealistic shot of a family's feet walking on a beach during sunset, warm sand" User Type A User Type B User Type C Preference Adaptive and Sequential Text-to-Image Generation G. PASTA Examples We illustrate a complete 5-turn interaction between our user model and the PASTA agent with examples. Each example begins with the initial prompt and images. Each turn shows the agent s 25 generated candidate prompts (organized into 5 categories) on the left. Four of these prompts (highlighted in red) are offered to the user, selected by a sampling policy based on a value function. The four generated images corresponding to each proposed prompt are displayed below. The prompt ultimately chosen by the user is highlighted in blue and shown above the image set. Example #1: Initial prompt: "A train is going down the train tracks" Agent Category: Variations in Lighting and Time of Day: 1. A train at sunset, on tracks. 2. A train in the morning mist. 3. A train under a stormy sky. 4. A train at night, glowing lights. 5. A train in bright sunlight. Category: Variations in Train Type and Composition: 6. A long freight train on tracks. 7. A sleek modern passenger train. 8. An old steam train on tracks. 9. A colorful toy train on tracks. 10. A double-decker train on tracks. Category: Variations in Environment and Setting: 11. Train on a mountain track. 12. Train in a desert landscape. 13. Train crossing a bridge. 14. Train passing through a forest. 15. Train in a snowy landscape. Category: Completely Different Prompts (Rephrasing): 16. Railroad journey, scenic view. 17. Metal beast on steel rails. 18. Locomotive speeding onward. 19. Tracks vanishing in distance. 20. Steel wheels on iron rails. Category: Adding Artistic Styles and Effects: 21. Train on tracks, impressionist style. 22. Train, hyperrealistic rendering. 23. Train on tracks, painted by Monet. 24. Train, detailed pencil sketch. 25. Train on tracks, vibrant colors. User User selects prompt number 16: "Railroad journey, scenic view." Agent Category: Expanding the Scene: 1. Railroad tracks winding through a lush green valley, sunlit. 2. Train disappearing into a mountain tunnel, dramatic lighting. 3. A vintage steam train crossing a high bridge over a river. 4. Railroad tracks stretching to the horizon, sunset over fields. 5. Train passing a quaint countryside village, wildflowers. Category: Focusing on the Train: 6. Majestic steam locomotive pulling luxurious passenger cars. 7. Modern high-speed train speeding through a vibrant landscape. 8. Freight train carrying colorful containers, scenic backdrop. 9. Close-up of a train's wheels turning on the tracks, detailed. 10. A weathered, vintage train car in a beautiful setting. Category: Artistic Styles: 11. Railroad landscape, painted in the style of Van Gogh. 12. Impressionist painting of a train journey through fields. 13. Photorealistic train passing through a stunning canyon. 14. A whimsical, cartoon-style train in a fantasy landscape. 15. Train in a vibrant, surreal dreamscape setting. Category: Different Perspectives: 16. Aerial view of a train snaking through a winding valley. 17. View from inside a train car, looking out at the scenery. 18. A long exposure shot of a train moving at night, light trails. 19. Close-up of the train's engine, steam billowing dramatically. 20. View from the train's cab, looking down the tracks. Category: Rephrasing and Variations: 21. Scenic railroad journey through diverse landscapes. 22. Picturesque train ride, breathtaking views from the window. 23. Railroad adventure, panoramic scenery. 24. Journey by rail, spectacular countryside. 25. Train on rails, beautiful and idyllic nature. User User selects prompt number 15: "Train in a vibrant, surreal dreamscape setting" Preference Adaptive and Sequential Text-to-Image Generation Agent Category: Surreal Landscapes & Trains: 1. A steam train chugging through a field of giant, iridescent flowers under a swirling galaxy. 2. A whimsical train journey through a candy-colored forest, lollipop trees, and marshmallow clouds. 3. A futuristic train gliding over a rainbow bridge in a city made of crystals and floating islands. 4. An antique train winding through a valley of vibrant, bioluminescent plants at twilight. 5. A surreal train station nestled within a giant, glowing mushroom forest. Category: Emphasis on Vibrancy & Color: 6. A vibrant, hyperrealistic train speeding through a landscape of impossibly bright colors. 7. A psychedelic train ride through a kaleidoscope of swirling patterns and neon lights. 8. A train made of stained glass, traversing a landscape of equally colorful stained glass. 9. A cartoonish train with exaggerated features, traveling through a bright, playful world. 10. A majestic train, with golden carriages, passing through a field of glowing sunflowers. Category: Focus on Dreamlike Atmosphere: 11. A lone train disappearing into a hazy, dreamlike fog, leaving a trail of shimmering stardust. 12. A train floating effortlessly above a sea of clouds, bathed in soft, ethereal light. 13. A vintage train gliding silently through a landscape painted in watercolor washes. 14. A dreamlike train station, where clocks melt and platforms float, bathed in moonlight. 15. A train composed of shifting, morphing shapes, moving through a constantly changing landscape. Category: Different Interpretations of the Prompt: 16. A train journey through an impossible landscape, defying gravity and perspective. 17. The feeling of riding a train in a vibrant dream, expressing freedom and wonder. 18. A dream sequence, a train traveling through a bizarre, otherworldly environment. 19. A surreal, dreamlike depiction of a train journey, full of vibrant colors and unusual imagery. 20. A train's passage through a vibrant dreamscape, emphasizing the contrast between reality and fantasy. Category: Adding Specific Artistic Styles: 21. A train in a vibrant surreal dreamscape, in the style of a 1950s cartoon. 22. A vibrant surreal dreamscape with a train, painted in the style of Salvador Dali. 23. A vibrant surreal dreamscape featuring a train, rendered in the style of Alphonse Mucha. 24. A train traveling through a vibrant surreal dreamscape, with a focus on Art Nouveau style. 25. A vibrant surreal dreamscape with a train, using bold, expressive brushstrokes in an Impressionistic style. User User selects prompt number 5: "A surreal train station nestled within a giant, glowing mushroom forest" Agent Category: Vibrant Mushroom Forests & Trains: 1. A whimsical train station nestled amongst bioluminescent mushrooms, with a steampunk aesthetic. 2. A vibrant, colorful train speeding through a fantastical forest of giant glowing mushrooms. 3. A surreal train journey through a mushroom forest, with playful, cartoonish mushrooms. 4. A detailed illustration of a train emerging from a glowing mushroom cavern, magical atmosphere. 5. A train puffing steam, passing through a dense forest of giant, otherworldly mushrooms. Category: Surreal Landscapes & Trains: 6. A train gliding across a rainbow bridge in a surreal, dreamlike landscape. 7. A train traversing a landscape of floating islands, vibrant colors, and impossible architecture. 8. A surreal train journey through a landscape painted in the style of Van Gogh, vibrant colors. 9. A hyperrealistic train in a fantastical landscape, filled with impossible flora and fauna. 10. A train disappearing into a swirling vortex of colors in a surreal, dreamlike landscape. Category: Dali-esque Trainscapes: 11. A melting clock train in a Dali-esque landscape, with distorted perspective and melting forms. 12. A train with elongated limbs and surreal features in a Dali-inspired dreamscape. 13. A train composed of melting clocks and surreal objects in a Dali-esque landscape. 14. A train traversing a desert landscape with melting rocks and distorted space in Dali's style. 15. A surreal train scene with distorted perspective and soft, flowing lines in the style of Dali. Category: Ethereal & Dreamlike Trains: 16. A train floating through a starry night sky, with ethereal lights and glowing constellations. 17. A lone train moving through a field of luminous flowers in a soft, ethereal light. 18. A dreamy train scene with pastel colors, soft clouds, and a peaceful, tranquil atmosphere. 19. A train gently gliding through a soft, misty landscape, with a serene and peaceful atmosphere. 20. A train traveling through a sea of clouds, bathed in soft, golden light, peaceful and serene. Category: Completely Different Prompts (Rephrasing): 21. A detailed photograph of a modern high-speed train passing through a rural area. 22. A realistic depiction of a vintage steam train chugging along a mountain pass. 23. A classic black and white photo of a train passing through a bustling city. 24. An image of a freight train transporting goods along a long railway line. 25. A simple drawing of a train on tracks in a straightforward, realistic style. User User selects prompt number 13: "A train composed of melting clocks and surreal objects in a Dali-esque landscape" Agent Category: Dali-esque Variations: 1. A surreal train speeding through a landscape of melting clocks and distorted figures, in the style of Salvador Dali. 2. A train composed entirely of melting clocks, surreal objects, and impossible architecture, a Dali-esque dreamscape. 3. A Dali-esque train journey through a desert landscape, with elongated shadows and distorted perspectives. 4. A train with elongated limbs and melting forms, traversing a bizarre, dreamlike landscape, in the style of Salvador Dali. 5. A surreal train station, with distorted architecture and melting clocks, in the style of Salvador Dali, hyperrealistic. Category: Enhanced Surrealism: 6. A train made of liquid metal, flowing through a landscape of shifting colors and impossible geometries. 7. A train gliding through a vibrant, surreal dreamscape filled with fantastical creatures and impossible structures. 8. A train passing through a swirling vortex of colors and shapes, a surreal and psychedelic journey. 9. A train emerging from a giant, melting clock, surrounded by distorted figures and impossible objects. 10. A train journey through a hyperrealistic, surreal landscape, with unexpected elements and juxtapositions. Category: Focus on the Train: 11. Detailed close-up of a surreal train engine, with melting clocks and intricate details. 12. A surreal train's interior, filled with melting furniture, distorted mirrors, and bizarre passengers. 13. The perspective of someone sitting inside a Dali-esque train, witnessing the surreal landscape. 14. A highly detailed rendering of a surreal train, emphasizing textures and materials. 15. A surreal train, focusing on its unusual design and mechanics. Category: Different Artistic Styles: 16. A train in a surreal landscape, painted in the style of René Magritte, with a focus on unexpected juxtapositions. 17. A train journey through a surreal landscape, rendered in the style of a woodcut print, with bold lines and textures. 18. A surreal train scene in the style of Max Ernst, featuring collage-like elements and unexpected combinations. 19. A surreal train depicted in a vibrant, pop art style, with bold colors and graphic elements. 20. A surreal train in a cubist landscape, emphasizing fragmented forms and multiple perspectives. Category: Rephrasing & Simplification: 21. A bizarre, dreamlike train in a surreal landscape. 22. A surreal train, simple prompt. 23. A melting clock train in a Dali-esque world. 24. Surreal train in a distorted landscape. 25. A dreamlike train journey through a strange land. User User selects prompt number 22: "An ethereal underwater scene depicting a queen, her crown formed by bioluminescent sea creatures, radiating an enchanting glow." Preference Adaptive and Sequential Text-to-Image Generation Example #2: Initial prompt: "A princess wearing an animal skull mask with a ribcage skeleton and cybernetic jellyfish features" Agent Category: Enhanced Detail & Lighting: 1. Close-up portrait, princess, animal skull mask, cybernetic jellyfish, intricate details, dramatic lighting. 2. Princess, skull mask, jellyfish cybernetics, ethereal glow, hyperrealistic, close-up. 3. Macro portrait, princess, bone mask, bioluminescent jellyfish, sharp focus, rich colors. 4. Detailed portrait, princess, ornate skull mask, glowing cybernetic jellyfish, dark background. 5. Close-up, princess, animal skull headdress, bio-mechanical jellyfish, cinematic lighting. Category: Alternative Mask & Cybernetics: 6. Princess portrait, elegant bird skull mask, subtle cybernetic enhancements, soft lighting. 7. Close-up, princess, futuristic insect mask, integrated cybernetic jellyfish, vibrant colors. 8. Macro portrait, princess, ornate dragon skull mask, glowing cybernetic implants, blurred background. 9. Princess, skeletal mask, sleek cybernetic arms, ethereal background, detailed texture. 10. Portrait, princess, geometric skull mask, minimalist cybernetic features, sharp focus. Category: Variations in Background & Setting: 11. Princess, skull mask, cybernetic jellyfish, underwater scene, shimmering bubbles. 12. Portrait, princess, bone mask, cybernetic enhancements, mystical forest background. 13. Close-up, princess, animal skull mask, jellyfish cybernetics, futuristic cityscape backdrop. 14. Princess, skull mask, bioluminescent jellyfish, otherworldly nebula background. 15. Macro portrait, princess, ornate mask, cybernetic elements, enchanted garden setting. Category: Different Artistic Styles: 16. Princess, skull mask, cybernetic jellyfish, art nouveau style, elegant and flowing lines. 17. Portrait, princess, bone mask, cybernetic enhancements, dark fantasy art style. 18. Close-up, princess, animal skull mask, jellyfish cybernetics, photorealistic style. 19. Princess, skull mask, bioluminescent jellyfish, gothic art style, dramatic lighting. 20. Macro portrait, princess, ornate mask, cybernetic elements, digital painting style. Category: Completely Different Concept (Rephrasing): 21. A regal woman, adorned with intricate bone jewelry and glowing bioluminescent accents. 22. A mysterious figure, partially obscured by a fantastical mask, with subtle technological elements. 23. A powerful queen, with cybernetic enhancements blending seamlessly with organic bone structures. 24. A captivating portrait of a woman, combining elegance with futuristic bio-mechanical elements. 25. An ethereal being, blending organic and synthetic elements in a harmonious composition. User User selects prompt number 5: "Close-up, princess, animal skull headdress, bio- mechanical jellyfish, cinematic lighting" Agent Category: Refined Details & Lighting: 1. Close-up portrait: princess, intricate skull headdress, bioluminescent jellyfish, dramatic lighting, 8k resolution. 2. Princess in ornate skull mask, bio-mechanical jellyfish crown, deep shadows, sharp focus, cinematic lighting. 3. Close-up, princess, detailed animal skull mask, glowing jellyfish, volumetric lighting, dark background. 4. Macro portrait: princess, skull headdress with jellyfish, chiaroscuro lighting, ethereal atmosphere, hyperrealistic. 5. Close-up, princess, skull mask, bio-mechanical jellyfish, cinematic lighting, rich textures. Category: Variations on the Jellyfish: 6. Princess, ornate skull mask, bioluminescent jellyfish tendrils, dark background, cinematic lighting. 7. Close-up portrait, princess, skull headdress, iridescent jellyfish, dark moody ambiance, sharp details. 8. Princess, skeletal mask, cybernetic jellyfish swarm, deep ocean setting, cinematic lighting. 9. Close-up: princess, skull headdress, bio-mechanical jellyfish, neon lighting, underwater effect. 10. Princess, skull mask, pulsating jellyfish, dark background, dramatic lighting, 8k resolution. Category: Emphasis on the Skull: 11. Close-up portrait, princess, intricately carved skull mask, jellyfish accents, cinematic lighting. 12. Princess, detailed animal skull mask, bio-mechanical jellyfish, dark fantasy art, dramatic lighting. 13. Close-up: princess, ornate skull headdress, bio-mechanical jellyfish, dark fantasy art, sharp focus. 14. Princess, realistic skull mask, cybernetic jellyfish, dark background, cinematic lighting, detailed texture. 15. Macro shot: princess, decaying skull mask, bio-mechanical jellyfish, dramatic lighting, ethereal feel. Category: Completely Different Style: 16. Elegant princess, futuristic crown, subtle technological details, soft lighting, pastel colors. 17. Portrait of a regal woman, ethereal crown, flowing gown, soft light, serene expression. 18. A powerful queen, majestic pose, ornate jewelry, warm lighting, rich colors. 19. Young woman, delicate flower crown, flowing dress, soft light, gentle expression. 20. Beautiful woman, simple headband, natural light, soft focus, dreamy atmosphere. Category: Exploring Backgrounds & Settings: 21. Close-up, princess, skull headdress, bio-mechanical jellyfish, foggy forest background. 22. Princess, skull mask, jellyfish crown, cinematic lighting, ancient ruins background. 23. Close-up portrait, princess, skull headdress, bio-mechanical jellyfish, underwater scene. 24. Princess, skull mask, jellyfish, cinematic lighting, gothic architecture background. 25. Close-up: princess, skull headdress, bio-mechanical jellyfish, starry night sky background. User User selects prompt number 9: "Close-up: princess, skull headdress, bio-mechanical jellyfish, neon lighting, underwater effect" Preference Adaptive and Sequential Text-to-Image Generation Agent Category: Neon Dreamscapes: 1. Princess, skull jellyfish crown, neon coral reef, vibrant underwater city. 2. Underwater princess, glowing jellyfish headdress, neon lights, art deco style. 3. Cyberpunk princess, skull mask with neon jellyfish, futuristic underwater city. 4. Princess in neon-lit underwater cave, skull headdress, bioluminescent jellyfish. 5. Electric princess, skull and jellyfish crown, neon ocean, 80s synthwave aesthetic. Category: Subtle Shifts in Lighting and Background: 6. Close-up portrait: princess, skull headdress, bioluminescent jellyfish, soft neon glow, dark background. 7. Princess, skull jellyfish crown, underwater scene, subtle neon accents, calm atmosphere. 8. Underwater princess, skull headdress, bio-mechanical jellyfish, pastel neon, serene background. 9. Princess, skull and jellyfish crown, deep ocean, cool neon hues, mysterious ambiance. 10. Close-up, princess, skull headdress, jellyfish, dim neon lighting, smoky underwater cave. Category: Variations on the Jellyfish: 11. Princess, skull headdress, glowing jellyfish tentacles, neon lights, dark ocean. 12. Princess, skull crown, iridescent jellyfish, neon underwater landscape, detailed textures. 13. Underwater princess, skull headdress, bioluminescent jellyfish swarm, neon glow. 14. Close-up: princess, skull mask, intricate jellyfish headdress, neon coral, deep ocean. 15. Princess, skull headdress, bio-mechanical jellyfish, neon kelp forest, flowing fabric. Category: Completely Different Style: 16. Portrait of a regal woman, ornate mask, intricate jewelry, rich textures, dramatic lighting. 17. Elegant woman, mysterious headdress, flowing gown, detailed portrait, studio lighting. 18. A woman with a captivating gaze, elaborate headdress, rich colors, painterly style. 19. Close-up portrait, strong female figure, enigmatic expression, rich textures, studio setting. 20. Fashion portrait, stylish woman, unique headpiece, luxurious fabrics, sophisticated look. Category: Emphasizing Texture and Detail: 21. Hyperrealistic portrait: princess, detailed skull headdress, bio-mechanical jellyfish, neon lighting, underwater scene. 22. Close-up, princess, skull headdress, intricate jellyfish details, neon lights, underwater world. 23. Princess, skull jellyfish crown, highly detailed textures, neon lighting, underwater cave. 24. Macro portrait: princess, textured skull mask, bio-mechanical jellyfish, neon glow, underwater rocks. 25. Detailed portrait: princess, skull headdress, realistic jellyfish, neon accents, ocean floor. User User selects prompt number 23: "Princess, skull jellyfish crown, highly detailed textures, neon lighting, underwater cave" Agent Category: Neon Underwater Princesses: 1. A princess adorned with a bioluminescent jellyfish crown, swimming in a vibrant neon coral reef, intricate details, 8k resolution. 2. Princess in an underwater palace, wearing a jellyfish headdress that glows with neon light, surrounded by glowing flora and fauna. 3. Portrait of a princess, her face partially obscured by a neon jellyfish crown, underwater scene, hyperrealistic details. 4. An ethereal princess with flowing hair, a jellyfish headdress emitting neon light, swimming gracefully amidst glowing jellyfish. 5. Close-up of a princess wearing a skull-shaped jellyfish crown, vibrant neon lighting, underwater cave setting, detailed textures. Category: Dark Fantasy Princesses: 6. A princess in a dark underwater grotto, wearing a crown of glowing jellyfish and skeletal elements, mysterious and ethereal. 7. Dark fantasy portrait: a princess with a jellyfish crown made of bone and neon lights, surrounded by deep sea creatures. 8. A gothic princess, wearing a headdress crafted from decaying jellyfish and skulls, bathed in neon underwater light, dramatic pose. 9. A mysterious underwater princess, her face veiled by a jellyfish headdress with glowing neon accents, dark and moody atmosphere. 10. Portrait of a princess in a dark, underwater cavern, wearing a crown of bioluminescent jellyfish, surrounded by shadows and glowing accents. Category: Detailed Jellyfish Headwear: 11. Extremely detailed portrait of a princess wearing an intricate jellyfish headdress, neon lighting, underwater setting, realistic textures. 12. Close-up, princess with a jellyfish crown, featuring highly detailed tentacles and bioluminescent patterns, underwater scene, 8k resolution. 13. Macro shot: princess, ornate jellyfish headdress with intricate skeletal details, neon lighting, underwater environment, sharp focus. 14. Princess portrait: elaborate jellyfish crown with glowing neon accents, detailed textures, underwater background, cinematic lighting. 15. Highly detailed image of a princess wearing a jellyfish crown, showcasing the intricate textures and neon glow, underwater environment. Category: Alternative Interpretations: 16. A regal figure, partially obscured by a mask of intricately woven light and shadow, with bioluminescent accents, underwater, 8k resolution. 17. A captivating woman, her face partially hidden by a fantastical headpiece reminiscent of deepsea life, bathed in ethereal light. 18. A mysterious and elegant woman with a headpiece combining natural and technological elements, emitting a soft, otherworldly glow. 19. An enigmatic figure in an underwater setting, adorned with a crown of ethereal light and intricate, bioluminescent patterns. 20. A vision of elegance and mystery: a woman with a unique headpiece inspired by marine life, emitting a gentle, glowing light. Category: Stylistic Variations: 21. Art Deco style portrait: a princess wearing a geometric jellyfish crown, vibrant neon colors, underwater setting, elegant composition. 22. Surrealist portrait: a princess with a melting jellyfish crown, neon lighting, dreamlike underwater scene, unexpected elements. 23. Impressionistic underwater portrait: a princess with a jellyfish crown, soft focus, pastel neon colors, flowing lines and shapes. 24. A princess with a jellyfish headdress in the style of Alphonse Mucha, elegant lines, soft neon lighting, underwater background. 25. A princess wearing a jellyfish crown, painted in the style of Gustav Klimt, with gold and vibrant neon accents, underwater scene. User User selects prompt number 12: "Close-up, princess with a jellyfish crown, featuring highly detailed tentacles and bioluminescent patterns, underwater scene, 8k resolution" Agent Category: Refined Details & Lighting: 1. Close-up portrait of a princess wearing a bioluminescent jellyfish crown, intricate details on tentacles, soft underwater lighting, 8k resolution, hyperrealistic. 2. Princess with a jellyfish crown, detailed tentacles with iridescent scales, soft blue and pink lighting, underwater scene, serene expression. 3. Underwater portrait: princess, jellyfish crown with glowing accents, subtle shimmering particles, cinematic lighting, highly detailed textures. 4. Close-up, princess, jellyfish crown, vibrant coral reef background, realistic rendering, sharp focus, detailed textures, soft lighting. 5. Princess, jellyfish crown, realistic jellyfish anatomy, submerged in crystal clear water, natural lighting, ethereal atmosphere. Category: Stylistic Variations: 6. Art Nouveau princess with a jellyfish crown, flowing lines, pastel colors, underwater setting, elegant composition. 7. Princess, jellyfish crown, Art Deco style, geometric patterns, neon lighting, underwater Art Deco architecture. 8. Surrealist portrait: princess, melting jellyfish crown, dreamlike underwater environment, vibrantcolors, distorted perspective. 9. Impressionist princess, jellyfish crown, soft brushstrokes, underwater scene, blurred background,pastel hues. 10. Princess, jellyfish crown, in the style of Gustav Klimt, gold leaf accents, underwater scene,shimmering effect. Category: Environmental Changes: 11. Princess, jellyfish crown, deep ocean trench, bioluminescent creatures, dark mysterious background, 8k resolution. 12. Princess, jellyfish crown, glowing coral reef, diverse marine life, vibrant colors, shallow water, sunlight filtering through. 13. Princess, jellyfish crown, ancient sunken city, ruins, mystical underwater atmosphere, detailed textures, cinematic lighting. 14. Princess, jellyfish crown, mysterious underwater cave, glowing crystals, ethereal light, sense of wonder, 8k resolution. 15. Underwater princess, jellyfish crown, surrounded by schools of luminous fish, vibrant coral, crystal-clear water, cinematic lighting. Category: Alternative Headwear Interpretations: 16. Close-up portrait of a princess with a headdress made of intricately woven sea anemones and jellyfish, underwater, ethereal lighting. 17. Princess, ornate headdress inspired by jellyfish, intricate details, flowing fabric, underwater, cinematic lighting, 8k resolution. 18. Portrait of a princess with a crown sculpted from fossilized jellyfish, underwater ruins background, mysterious atmosphere. 19. Princess with a headdress resembling a blooming jellyfish, vibrant colors, detailed textures, underwater, soft lighting. 20. Close-up of a princess wearing a delicate headdress crafted from translucent jellyfish, ethereal glow, underwater scene, serene expression. Category: Completely Different Prompts (Rephrasing): 21. A breathtaking underwater portrait of a regal woman, adorned with a luminous, organic headpiece, evoking a sense of otherworldly beauty. 22. An ethereal underwater scene depicting a queen, her crown formed by bioluminescent sea creatures, radiating an enchanting glow. 23. A captivating image of a majestic figure submerged in an ocean, wearing a radiant headpiece reminiscent of jellyfish, in a hyperrealistic style. 24. A mesmerizing underwater portrait, featuring a serene woman with a crown made of glowing jellyfish, showcasing intricate details and vibrant colors. 25. An elegant woman, her head crowned with a dazzling display of bioluminescent jellyfish, creating a stunning underwater spectacle. User User selects prompt number 22: "An ethereal underwater scene depicting a queen, her crown formed by bioluminescent sea creatures, radiating an enchanting glow."