# private_federated_learning_using_preferenceoptimized_synthetic_data__b539c94d.pdf Private Federated Learning using Preference-Optimized Synthetic Data Charlie Hou * 1 Mei-Yu Wang * 2 Yige Zhu * Daniel Lazar 3 Giulia Fanti 1 In practical settings, differentially private federated learning (DP-FL) is the dominant method for training models from private, on-device client data. Recent work has suggested that DP-FL may be enhanced or outperformed by methods that use DP synthetic data (Wu et al., 2024; Hou et al., 2024). The primary algorithms for generating DP synthetic data for FL applications require careful prompt engineering based on public information and/or iterative private client feedback. Our key insight is that the private client feedback collected by prior DP synthetic data methods (Hou et al., 2024; Xie et al., 2024) can be viewed as an RL reward. Our algorithm, Policy Optimization for Private Data (POPri) harnesses client feedback using policy optimization algorithms such as Direct Preference Optimization (DPO) to fine-tune LLMs to generate high-quality DP synthetic data. To evaluate POPri, we release Large Fed Bench, a new federated text benchmark for uncontaminated LLM evaluations on federated client data. POPri closes the gap in performance between the fullyprivate and non-private settings by up to 58%, compared to 28% for prior synthetic data methods, and 3% for state-of-the-art DP federated learning methods. The code and data are available at https://github.com/meiyuw/POPri. 1. Introduction Many important machine learning (ML) applications feature sensitive datasets that are distributed across client devices (e.g. mobile devices). Such ML models are often hosted on client devices. These on-device models offer privacy, latency, and storage benefits relative to centrally-hosted mod- *Equal contribution 1Department of ECE, Carnegie Mellon University, Pittsburgh, PA 2Pittsburgh Supercomputing Center, Pittsburgh, USA 3Coldrays, Tucson, AZ. Correspondence to: Charlie Hou . Proceedings of the 42 nd International Conference on Machine Learning, Vancouver, Canada. PMLR 267, 2025. Copyright 2025 by the author(s). els. Examples include Google s GBoard (Hard et al., 2019; Xu et al., 2023b; Wu et al., 2024) and Apple s mobile automatic speech recognition system (Paulik et al., 2021). Today, federated learning (FL) is the most widely-used approach in practice for learning on-device models; it trains models locally on user devices and aggregates model updates on a central server (Mc Mahan et al., 2017b). FL protects the privacy of client data in part by adopting differentially private (DP) (Dwork, 2006) optimization techniques, a combination we refer to as DP-FL (Mc Mahan et al., 2017b; Kairouz et al., 2021b; Nguyen et al., 2022; Xu et al., 2023a). With breakthroughs in large language model (LLM) capabilities (Anil et al., 2023; Team et al., 2023; Achiam et al., 2023; Guo et al., 2025) several research teams have used LLMs to better train models on private client data. A common strategy applies standard optimization algorithms (e.g., DP stochastic gradient descent, DP-SGD (Abadi et al., 2016)) to fine-tune models on private client data (Kurakin et al., 2023; Charles et al., 2024). These approaches have an important limitation in the on-device setting: most LLMs today are too large to fit on client devices, let alone train on them (Radford et al., 2019; Touvron et al., 2023). To sidestep the size issue, Wu et al. (2024); Hou et al. (2024) view the problem of learning from distributed, private client data (partially) as a DP synthetic data problem. These approaches use LLM-assisted workflows to generate privacypreserving synthetic data, similar to client data, at the server; then they train the on-device model at the server on the synthetic data. This avoids storing the LLM on client devices. In more detail, Wu et al. (2024) use prior public information about the clients to create LLM-generated synthetic data for pretraining. For example, for their Google GBoard virtual keyboard application, they use prompts like Imagine you are a [GENDER] of age [AGE]. Write some examples of chat messages. to generate synthetic samples. This prompt was designed entirely using prior qualitative information about the data on client devices. However, prior information may not always be available. Moreover, this prompt was not refined based on clients realized data, which could limit the relevance of the resulting synthetic data. Pr E-Text (Hou et al., 2024) instead uses Private Evolution (PE) (Lin et al., 2023; Xie et al., 2024; Lin et al., 2025b) to learn prompts that are relevant to client data. PE itera- Private Federated Learning using Preference-Optimized Synthetic Data Figure 1. Left: Private Evolution (PE)-based techniques. Clients generate low-dimensional statistics which summarize the similarity of the synthetic data to their private samples. These are privately aggregated to refine the synthetic data generation for future iterations. Traditional PE (brown) uses a prompt-based method. POPri (blue) improves a naive fine-tuning method (PE+SFT, purple) by fine-tuning the LLM using policy optimization rather than fine-tuning directly on aggregated client feedback. Right: Next-token prediction accuracy on the bio Rxiv dataset at privacy level ϵ = 1. POPri closes the accuracy gap between the fully-private and non-private settings by 58%, compared to 23% for prior synthetic data methods, and 3% for DP federated learning methods. tively sends synthetic data samples to clients for feedback; each client privately measures the closeness of synthetic samples to their own data, discarding irrelevant samples. It returns this feedback to the central server, which crafts a new prompt based on the most relevant synthetic samples. Finally, an LLM uses the generated synthetic data to finetune a downstream model. This method of utilizing LLMs for on-device learning has some shortcomings: (1) it relies entirely on prompting to teach the LLM to generate relevant synthetic data, which may not be as effective as fine-tuning the weights. (2) It discards irrelevant samples, which may themselves contain valuable information, as shown in reinforcement learning with human feedback (RLHF) (Ouyang et al., 2022). In this paper, we demonstrate how to better utilize LLMs for on-device learning: we propose POPri (Policy Optimization for Private Data), an algorithm that reformulates synthetic data-based approaches for private learning as an LLM policy optimization problem. In POPri, we directly fine-tune an LLM s weights to improve the (DP-noised) similarity scores between generated synthetic data and private client samples. The fine-tuned LLM is used to generate synthetic data, which is used to train a downstream model. Contributions. In summary, our contributions are: (1) We propose POPri, a novel method that casts private learning under the synthetic data framework as an LLM policy optimization problem. Prior work in this space relied on PE, which uses client feedback exclusively to generate new prompts (Hou et al., 2024; Xie et al., 2024). We alter this feedback to instead provide client rewards, and subsequently exploit recent advances in policy optimization (Rafailov et al., 2023). This recasting allows us to more effectively exploit the capabilities of LLMs for on-device learning problems. (2) We create and maintain Large Fed Bench, a new uncontaminated benchmark of federated client data separated by client for the era of LLMs. The datasets in this benchmark consist of: (1) congressional records in English-speaking countries, and (2) abstracts from bio Rxiv, collected starting in April 2023. To our knowledge, this is the first dataset that provides researchers with both (a) over 1,000 clients (congressional records contains 134k clients and bio Rxiv contains 57k as of August 2024), and (b) regular updates, allowing researchers to easily filter data to avoid contaminated evaluations (Magar & Schwartz, 2022; Zhou et al., 2023; Yang et al., 2023; Roberts et al., 2023). (3) We demonstrate the utility of POPri on this new benchmark set of datasets, as well as two central (i.e., the setting where all data is present on the server and no server-client communication is needed) DP benchmarks from prior work (Yu et al., 2023; Xie et al., 2024). Across all datasets and tasks (we consider next token prediction and text classification), POPri achieves the best downstream metrics. For example, Figure 1 shows that on our bio Rxiv dataset at a privacy level of ϵ = 1.0, POPri outperforms PE-based algorithms by 6 full percentage points, and closes the gap between fully private and non-private baselines by over 58%, compared to 23% for PE. It outperforms DP-FL-based methods by even more. Additional experimental details, results, and ablations are provided in Section 5. 2. Problem Statement and Background 2.1. Problem Statement We consider a set S of clients, S = {S1, . . . , Sn}, where Si = {s(i) 1 , . . . , s(i) mi} denotes the private text data of client i [n], and mi denotes the number of text samples held by client i. We consider the partial participation setting, where only a subset of clients can participate in communication with the server at any point in time (Kairouz et al., 2021a; Private Federated Learning using Preference-Optimized Synthetic Data Mc Mahan et al., 2017b), which is consistent with practical private on-device learning deployments. We assume L clients participate in each round t T and denote this set St. We do not assume an a priori upper bound on mi. A central server is given a pre-trained downstream model Φ, which it wants to align with the private client data S. We call the aligned downstream model Φ. In the process of learning Φ, the server may make use of a pre-trained public LLM Ψ. We observe that Ψ and Φ are different models in general; we will assume the server has access to the weights of both Φ and Ψ. The server is subject to two restrictions: (1) client data cannot leave client devices, and (2) the final model Φ must protect user-level differential privacy (DP): User-level (distributed) differential privacy (DP). We say two datasets S and S are neighboring if they differ in at most one client s data. That is, there exists an i [n] such that for all j = i, Sj = S j. A randomized mechanism M is (ϵ, δ)-DP if, for any pair of neighboring datasets S, S that differ by an entire client s data and any possible output set E, it holds that Pr[M(S) E] eϵPr[M(S ) E]+δ. The post-processing property of a DP mechanism ensures that any data-independent transformation applied to its output preserves the same DP guarantees. (Dwork, 2006; Dwork & Roth, 2014). We also evaluate on central DP baselines, so we define central DP below; in that case the final model Φ should protect central DP: Central (example-level) differential privacy (DP). We say two datasets (both fully present on the server) S = {s1, . . . , sm} and S = {s 1, . . . , s m} are neighboring if they differ in at most one example s data. That is, there exists an i [m] such that for all j = i, sj = s j. A randomized mechanism M is (ϵ, δ)-DP if, for any pair of neighboring datasets S, S that differ by one sample and any possible output set E, it holds that Pr[M(S) E] eϵPr[M(S ) E] + δ. Goal. The server seeks an algorithm to optimize the downstream performance (in our paper, this is either next token prediction accuracy or text classification accuracy) of Φ on a test set of private data, subject to an (ϵ, δ)-DP constraint. 2.2. Related Work There are two main approaches for learning on private data. DP optimization-based approaches. In natural language processing (NLP) tasks with privacy constraints, DP optimization algorithms (e.g., DP-SGD (Abadi et al., 2016)) are often used to fine-tune massively pretrained LLMs on private data (Bommasani & Schofield, 2019; Kurakin et al., 2023; Charles et al., 2024). However, in settings where client data cannot leave client devices due to privacy con- cerns, central servers cannot conduct this private fine-tuning. An alternative approach is to train models directly on client devices, using a server to coordinate information exchange between clients; in DP federated learning (DP-FL) (Mc Mahan et al., 2017b; Kairouz et al., 2021a), (small) model weights are iteratively sent to clients for on-device DP optimization. DP-FL has struggled to keep up with the growing size of LLMs; many LLMs cannot be stored or trained on client devices (Collins et al., 2023). Recent work explores how to train LLMs in the DP-FL framework. Proposed approaches include training only subsets of parameters (Charles et al., 2023), as well as memory-efficient zero-order optimization (Zhang et al., 2024; Malladi et al., 2023). However, these methods still require the storage of the entire model on-device, limiting their practicality. Synthetic data-based approaches. An alternative approach to DP optimization involves generating private synthetic data using LLMs, followed by directly fine-tuning downstream models. Synthetic data can be generated on the server side, which bypasses client-side hardware constraints. The post-processing property of DP also implies that DP synthetic data can be used repeatedly without incurring additional privacy loss (Yue et al., 2023a). In the centralized DP setting (where the server is trusted to gather all the data, as opposed to our private on-device setting), prior studies have shown that training downstream models on DP synthetic text achieves performance comparable to privately training on real data (Yue et al., 2023a; Mattern et al., 2022; Xie et al., 2024). In the private on-device setting, Hou et al. (2024) show that fine-tuning a small model on user-level DP synthetic text data on the server side can actually outperform DP-FL, with a significant reduction in communication and computation cost. Similarly, Wu et al. (2024) show that pretraining an FL model on private synthetic data can improve the final outcome of DP-FL. One approach for generating synthetic text data is to finetune an LLM (with DP-SGD) on private data (Kurakin et al., 2023; Yu et al., 2024) and then using the LLM to generate synthetic data. However, client hardware constraints render this approach infeasible on-device. Recent works have relied instead on privacy-aware prompt engineering to generate synthetic data (Wu et al., 2024; Xie et al., 2018; Hou et al., 2024). An important framework by Lin et al. (2023; 2025a) called Private Evolution (PE) is the basis for several competitive DP synthetic text algorithms, including Aug-PE (Xie et al., 2024) and Pr E-Text (Hou et al., 2024). Roughly, these algorithms use the public LLM Ψ to generate synthetic data, score each synthetic data according to its closeness to the client data, and discard synthetic data with low scores. The surviving synthetic data are used as in-context examples for Ψ to generate synthetic data. In concurrent work to ours, Zou et al. extend the PE framework to Private Federated Learning using Preference-Optimized Synthetic Data generate synthetic data from multiple pretrained language models (LMs), and present good and bad responses to the LMs in the next round for in-context learning (Zou et al., 2025). Private Evolution may sacrifice data quality in two ways: First, it uses in-context learning, which is often less effective than fine-tuning (Mosbach et al., 2023). Second, discarding low-score synthetic data may lose useful information (Ouyang et al., 2022). We address both by turning the DP synthetic generation problem into an LLM policy optimization problem. The core idea of POPri (Policy Optimization for Private Data) is a natural reformulation of private on-device learning from synthetic data as an LLM policy optimization problem, which enables the use of powerful LLM alignment methods like DPO (Rafailov et al., 2023). In this section, we detail the POPri design principles and algorithm. POPri s design is based on two related questions. 1. What client feedback should we collect for finetuning? Three natural options arise: (1) DP Data. Clients could directly transmit DP synthetic data samples for fine-tuning, e.g., using a method like DPPrompt (Utpala et al., 2023). DP-Prompt uses an LLM to summarize text at a temperature specified by the desired DP ϵ level. However, DP text cannot be aggregated into a single statistic, which prevents the use of secure aggregation (Bonawitz et al., 2016); this increases the noise needed to reach a given DP guarantee. As such, prior work has shown that DP-Prompt is not competitive with other private on-device learning methods (Hou et al., 2024). We favor aggregation-compatible representations of client data, such as summary statistics or model parameters. (2) DP Model Parameters. A second alternative is to send the parameters of either the LLM Ψ or the downstream model Φ to the client and train on the private samples with DP-SGD (Abadi et al., 2016). These parameters are compatible with secure aggregation (Bonawitz et al., 2016), which makes more efficient use of DP budget. However, Ψ cannot be sent to clients because of client storage constraints. Sending Φ is the DP-FL approach, which is one of our baselines. (3) DP Statistics. Finally, we could collect low-dimensional statistics capturing the quality of synthetic data samples. In PE, the server generates K synthetic data samples (Xie et al., 2024; Hou et al., 2024), and each client computes a histogram counting how often each of the private samples is closest to one of the K samples. This K-dimensional histogram can be made DP by adding (comparatively) little noise, and it is amenable to secure aggregation (Xie et al., 2024; Hou et al., 2024). We view such low-dimensional statistics as the most promising option, as they have lower communication and storage costs, and they make better use of the privacy budget. In a departure from PE, we design the low-dimensional statistics collected by POPri to enable building a preference dataset. We ask the server to generate J samples from each of K prompts; each client then scores the K J samples according to how well they represent the client s data, and the server aggregates the scores for all the synthetic samples. Using these scores, the server can construct a higher scoring response and lower scoring response pair (a preference pair ) for each of the K prompts. The benefit of this new design ties directly to the next question. 2. How should we use client feedback? Given a vector summarizing the quality of synthetic data samples, how should we use it? A few options arise: (1) In-Context Learning. We could use the highest-scoring synthetic samples as in-context examples to prompt the LLM Ψ. This is the PE approach (Hou et al., 2024; Xie et al., 2024). However, in-context learning typically performs worse than finetuning-based approaches (Mosbach et al., 2023), and we find experimentally that POPri outperforms Private Evolution (PE) (Figure 1, Table 1). POPri PE + SFT Evaluation Data Figure 2. 2-PCA visualization of synthetic data from POPri and PE+SFT, and evaluation data. We see that POPri s synthetic data distribution (left) is much closer to the evaluation data distribution (right) than the PE+SFT synthetic data distribution (middle). Naive fine-tuning with SFT on PE-generated synthetic data does not make best use of client feedback. (2) Supervised Fine-Tuning (SFT). One could directly finetune the LLM Ψ on the highest scoring samples using nextword-prediction loss. This is analogous to the SFT baseline evaluated in the RLHF (Ouyang et al., 2022) and DPO (Rafailov et al., 2023) papers, which showed that RLHF and DPO outperform SFT. The reason is that the highest scoring samples while better than the low-scoring samples are not perfect responses to the prompt. The SFT loss trains the LLM to treat high-scoring samples as perfect responses, which is misaligned with the LLM s task. Empirically, we see that this approach (PE+SFT) produces synthetic data that is not representative of the private data (Figure 2) and has poor downstream performance (Table 1). (3) Policy Optimization (PO). Policy optimization-based Private Federated Learning using Preference-Optimized Synthetic Data Algorithm 1 POPri 1: Input: Clients private data {Si}i [n], Number of rounds T, Number of generated samples Nsyn, Noise multiplier σ, LLM Ψ, embedding model Γ, base prompt η, participating clients in each round St, rejected index ℓ, random prompt generator Λ( ), number of clients sampled L 2: Output: Synthetic data Ssyn,T +1 3: 4: All clients i [n] embed private samples, Ei = Γ(Si) 5: Server initializes LLM Ψ1 = Ψ 6: for t 1 ... T do 7: Server: 8: Initialize the response vector R = 9: for k 1 . . . K do 10: Generate prompt ηk = Λ(η), 11: Generate J responses Rkj = Ψt(ηk), j [J] 12: end for 13: Send embeddings Esyn,t = {Γ(Rkj)}k [K],j [J] to all clients in St 14: 15: Client i St: 16: Scoresi,t SIMILARITY(Esyn,t, Ei) 17: Send Scoresi,t + N(0, σ2I/L) to Server 18: 19: Server: 20: Secure aggregate scores: Scorest = 1 i St Scoresi,t 21: Set P[k, j] as the j-th highest score response for prompt ηk, according to Scorest 22: Initialize preference dataset Pt = 23: for k 1 . . . K do 24: Select positive synthetic sample: Pt[k, 1] = Pt[k, 1] 25: Select negative synthetic sample: Pt[k, 2] = Pt[k, ℓ] 26: end for 27: Fine-tune: Ψt+1 DPO(Ψt, {ηk}k [K], Pt) 28: end for 29: Server: 30: Output final synthetic data Ssyn,T +1 from ΨT methods like DPO (Rafailov et al., 2023) instead directly optimize the LLM to produce higher-scoring samples (where the score can be defined by the user of the algorithm). In other words, they are designed to directly make use of the low-dimensional scores we collect from client feedback. Hence, we expect such methods to produce higher quality synthetic data, as evaluated on downstream tasks. 3.1. POPri Algorithm Pseudocode can be found in Algorithm 1. We highlight the algorithmically new steps (that differ from PE) in blue . 1. Synthetic sample generation. We generate K prompts (details in Appendix B.2). A prompt is generated by randomly sampling three samples from Ωand prompting LLa MA-3-8B (Touvron et al., 2023) to generate a fourth sample given the first three samples as examples. The exact prompt is given in Appendix B. For each of the K prompts, we generate J synthetic samples (by running the prompt independently J times). In total, the server generates K J synthetic samples, embeds them using a small sentence embedding model Γ and sends the embeddings to every client in St, i.e., the clients sampled in round t. 2. Scoring synthetic data using DP client feedback. Next, each client in St scores the synthetic samples. Specifically, each client calculates, for each of the K J synthetic samples, its cosine similarity with each of the client s private samples, averaged over the client s samples (Algorithm 4). The use of cosine similarity differs from PE, which uses a nearest neighbors histogram (Lin et al., 2023; Hou et al., 2024; Xie et al., 2024) using cosine similarity is critical to the performance of POPri as we found in our ablations (see Section 5.2). These similarities for every synthetic sample are arranged into a vector. We clip this vector to a norm of 1, which caps the contribution of each client (similar to how gradient updates are clipped per client in DP-FL (Mc Mahan et al., 2017b)). Clipping is done primarily for privacy reasons, as we will elaborate later. Clipping also ensures that the contribution of clients with large amounts of data does not overwhelm the contribution of clients with small amounts of data. We then add N(0, σ2I/L) (where I is the identity matrix of size KJ KJ) noise to the resulting vector to ensure DP (σ2 controls the (ϵ, δ)). Finally, we aggregate scores via secure aggregation (Bonawitz et al., 2016), yielding a DP score for each synthetic sample that reflects its relevance to client data. 3. LLM Policy Optimization. The key insight of our paper is that by generating J synthetic samples from K prompts and scoring all of them using DP client feedback, we can create a preference dataset where for each of the K prompts, we can assemble a good sample and a bad sample . This design choice allows the usage of powerful LLM policy optimization algorithms (we choose DPO (Rafailov et al., 2023)) to finetune the LLM Ψ. In detail, each of the K prompts have J synthetic samples which are ranked according to the scores we gathered. Then for each of the K prompts, we set the highest scoring sample as the chosen sample and the ℓ-th highest scoring sample as the rejected sample . This resulting preference dataset can then be passed, along with the LLM Ψ, into the DPO preference optimization loss (Rafailov et al., 2023): min Ψ E x,yω yr log s τ log(Ψ(yω|x) Ψ(yr|x) ) τ log(Ψref(yω|x) Ψref(yr|x) ) where Ψref a fixed checkpoint for the LLM (we use the public checkpoint of the LLM), τ is a parameter controlling deviation of Ψ from Ψref, x is the prompt, yω is the chosen sample, yr is the rejected sample, Ψ(y|x) is the probability Private Federated Learning using Preference-Optimized Synthetic Data Table 1. Accuracy (%, ) of different algorithms on a variety of tasks and datasets (bio Rxiv, Congress, Pub Med are next-token-prediction accuracy, Open Review is text classification accuracy). The highest accuracy across all methods is in bold. All standard deviation error bars are less than 0.5. Dataset Method Data Type On-device Model ϵ = ϵ = 7 ϵ = 1 ϵ = 0 DP-Fed Avg Original Distil GPT2 41.5 27.9 DP-FTRL Original 29.0 28.2 PE Synthetic 31.0 31.1 PE + SFT Synthetic 28.6 28.6 POPri (ours) Synthetic 34.4 34.8 DP-Fed Avg Original Distil GPT2 35.7 26.9 DP-FTRL Original 29.1 29.0 PE Synthetic 27.3 27.0 PE + SFT Synthetic 27.1 27.1 POPri (ours) Synthetic 30.6 30.4 Pub Med (Yue et al., 2023b) PE Llama-2-7b-chat-hf, Synthetic (2000) BERTsmall 47.6 27.5 PE Opt-6.7b, Synthetic (2000) 27.9 POPri (ours) Synthetic (2000) 29.4 Open Review (Xie et al., 2024) PE Llama-2-7b-chat-hf, Synthetic (2000) ROBERTAbase 50.8 37.0 32.0 PE Opt-6.7b, Synthetic (2000) 32.1 POPri (ours) Synthetic (2000) 40.2 of generating y given x for Ψ, and s is the sigmoid function. The expectation is taken with respect to the empirical distribution (i.e. real samples). The DPO loss will be used to finetune Ψ to generate more samples similar to the chosen sample and fewer like the rejected sample. To reduce GPU memory use, we use Lo RA (Hu et al., 2021) on all the attention matrices and up/down projection matrices with a rank of 4, α = 8. After fine-tuning over the K prompts and preference pairs, we return back to step (2) and generate new synthetic data using the newly fine-tuned Ψ. 4. Synthetic data generation for downstream tasks. Using the final version of Ψ, we generate a large set of synthetic data Ssyn,T +1 which is used to fine-tune Φ into Φ. Φ is then sent to all the client devices, where they can perform inference without communicating information to the server. Privacy guarantees. Because each client s vector is clipped to 1, and the only information revealed to the server (or any other party) is the aggregated vector, the sensitivity of the algorithm is 1. We add N(0, σ2I/L) noise to each client s vector, so the vector given to the server has noise N(0, σ2I), satisfying the Gaussian Mechanism with sensitivity 1. To calculate privacy, we can use a privacy accountant like OPACUS.ACCOUNTANTS.ANALYSIS.RDP (Yousefpour et al., 2021), and input T (the number of rounds we run the algorithm, q (the fraction of clients sampled per round), δ, and set σ to get the desired ϵ value. 4. Large Fed Bench: A Federated Benchmark for LLM Evaluation Today, the most widely-used evaluation datasets for federated learning of text models come from the work of Reddi et al. (2020); they include text from Stack Overflow posts and Shakespeare plays. These datasets pose two evaluation challenges: (1) They pre-tokenize inputs in a non-invertible way, which prevents researchers from using custom tokenizers adopted by several LLMs. (2) The datasets may lead to contaminated evaluations. As state-of-the-art LLMs have been trained on large swaths of the public internet, old public benchmark datasets may be in the training data of many LLMs (Magar & Schwartz, 2022; Zhou et al., 2023; Yang et al., 2023; Roberts et al., 2023). To our knowledge, one work proposes a benchmark dataset for federated LLMs (Ye et al., 2024). The datasets in this paper have at most 747 clients, which may be insufficient for simulating production use cases. Further, they do not explicitly avoid contamination. We release Large Fed Bench, a benchmark comprising two new datasets, Congressional Speeches and bio Rxiv, for experiments over federated client data. These datasets (a) allow researchers to easily avoid contamination, and (b) provide enough distinct clients to simulate production settings. Congressional Speeches ( Congress )1 is a dataset of 134k speeches or debates scraped from congressional or parliamentary transcripts in the US, UK, and Canada. We treat each speech as a separate client, and samples are created as successive 64-token spans within the speech. bio Rxiv2 is a dataset of 57k abstracts, each of which we consider a client dataset of strings, scraped from biology papers. Samples are 64-token spans of the abstract. More details on the datasets are included in Appendix F. A key feature of our datasets is that they are updated every 1https://huggingface.co/datasets/ hazylavender/Congressional Dataset 2https://huggingface.co/datasets/ hazylavender/biorxiv-abstract Private Federated Learning using Preference-Optimized Synthetic Data Figure 3. Next-token prediction accuracy performance of four methods as a function of the number of clients sampled per round out of 10000. We see that across different client participation scenarios, POPri consistently performs the best. 6 months and sorted by date. Hence, researchers can easily select datasets that were generated after their model s knowledge cutoff date. In this paper, we use data from Large Fed Bench published between the dates of April 2023 to August 2024 to avoid contamination with the latest LLM we evaluate our algorithms with, LLa MA-3-8B (AI@Meta, 2024) which has a knowledge cutoff of March 2023. 5. Experiments Datasets and tasks. For next token prediction accuracy, we evaluate POPri on the Large Fed Bench datasets (Congress and bio Rxiv), as well as Pub Med (Yu et al., 2023; Xie et al., 2024) used in the evaluation of Private Evolution (Aug-PE) (Xie et al., 2024). Pub Med contains abstracts of medical papers published between August 1-7, 2023 (details in Appendix D.2.2). For text classification, we evaluate POPri on Open Review consisting of ICLR 2023 reviews published on November 5, 2022 which was used in the evaluation of Private Evolution (Aug-PE) (Xie et al., 2024). Note that Pub Med and Open Review are evaluations in the central DP setting, where the entire dataset is present on the server and no server-client communication is needed. To execute POPri on Pub Med, we use the central DP version of POPri, detailed in Algorithm 2. For Open Review, we employ conditional generation3 (similar to PE (Lin et al., 2023; Xie et al., 2024), where the generation is conditioned on being given a class to generate. The (central) conditional generation version of POPri is detailed in Algorithm 3. Models. Next token prediction tasks: We use LLa MA-38B for the LLM Ψ (Grattafiori et al., 2024), which has a knowledge cutoff date of March 2023 (AI@Meta, 2024). For embedding models (used in measuring semantic distance between text samples), we use all-Mini LM-L6-v2 sentence transformer (Reimers & Gurevych, 2019b). We 3Simply speaking, we score synthetic data generated for each class separately but combine the preference datasets into one to finetune the LLM Ψ. choose Distil GPT2 (Sanh et al., 2019) as the downstream on-device language model for the Large Fed Bench evaluations, which has only 82M parameters, and BERTsmall as the downstream model for the Pub Med evaluation to be consistent with Xie et al. (2024). For synthetic text generation (using the LLM Ψ), we set the maximum sequence length to 64 for the bio Rxiv and Congressional Speeches evaluations and 512 for Pub Med/Open Review. Text classification: We use LLa MA-2-7b-chat-hf for the LLM Ψ (Touvron et al., 2023) to ensure our evaluation was not contaminated, as the knowledge cutoff for LLa MA-27b-chat-hf is September 2022 (before the publish date of the ICLR 2023 reviews). For the embedding model we use the sentence-t5-xl sentence transformer (Reimers & Gurevych, 2019a), and use Ro BERTabase as the downstream model to be consistent with Xie et al. (2024). Metrics. We primarily evaluate each method on accuracy (next-token or text classification) of the final downstream on-device model Φ. In some ablations we also measure the distance of the synthetic dataset to the private dataset using the Fr echet Inception Distance (FID) (Heusel et al., 2017). During training, we evaluate the models on the validation dataset and select the checkpoint that achieves the best validation performance as the model that is evaluated on the test set. Baselines. We compare POPri to several baselines: (1) DP-Fed Avg (Mc Mahan et al., 2017a) (2) DP-FTRL (Kairouz et al., 2021a) (3) Private Evolution (Pr E-Text (Hou et al., 2024) and Aug-PE (Xie et al., 2024)). DP-Fed Avg and DP-FTRL directly privately fine-tune the downstream model Φ on the client data. Private Evolution (Pr E-Text and Aug-PE) generates synthetic data on which the downstream on-device model Φ is finetuned. Note that on the Pub Med and Open Review dataset, we compare to Aug-PE results obtained with models of similar size to the model we use (7B-8B parameters) and which are not potentially Private Federated Learning using Preference-Optimized Synthetic Data contaminated (i.e. model was possibly trained on the benchmark dataset). We also include ϵ = 0 (fully private) and ϵ = (fully non-private) baselines. The ϵ = 0 baseline for the Large Fed Bench evaluations evaluates the public Distil GPT2 checkpoint on the test sets with no further fine-tuning. The ϵ = baseline is the downstream model finetuned directly on the private training set centralized on the server with no noise. The ϵ = 0 baseline for Open Review is the accuracy obtained by predicting everything to be the most populous class. More details about the setup can be found in Appendices C and D.2. Privacy Analysis. All baselines use a privacy guarantee of (ϵ, δ)-DP where δ=3 10 6 and ϵ=1 or ϵ=7 for each of the bio Rxiv and Congressional Speeches datasets. For Pub Med/Open Review, we set δ < 1 Npriv log(Npriv) (Npriv is the number of private samples). We follow the privacy accounting method detailed in Section 3 for POPri. Details for all baselines are in Appendix D.1. 5.1. Main Results Table 1 lists the accuracy (next token prediction and text classification) achieved by baseline methods (DP-Fed Avg, DP-FTRL, Private Evolution) and POPri. In this table, we assume full participation (no client sampling) for fair comparison to baselines, some of which do not have client sampling versions. We find that POPri outperforms all the baseline algorithms. Furthermore, in the ϵ = 1 setting POPri closes the gap between fully private learning (ϵ = 0) and fully non-private learning (ϵ = ) by 40-58% depending on the setting, compared to PE which closes 1-28%. For all methods tested, the measured accuracy values do not depend strongly on ϵ. This has been observed in prior work on DP synthetic data using LLMs (Xie et al., 2024; Hou et al., 2024). POPri outperforms Private Evolution (Aug-PE) even when holding our synthetic sample budget to 2000. Note that synthetic samples are cheap in POPri (we could generate many more) because we have access to the full model, while Xie et al. (2024) only assume access to a model API. Cost analysis case study. In Table 2 we analyze the perround communication and computation costs (and per-client, for download/upload/client runtime costs) of Fed Avg (a representative and cheap method among the DP-FL-based methods), PE, and POPri on the bio Rxiv dataset experiment with 1000 clients sampled per round. For Fed Avg, each round the sampled clients download and upload the downstream model, which in our case is Distil GPT2. This is an 82M (82 million) parameter model leading to a download and upload cost of 82M floats. The client runtime cost comes from local gradient computation, and server runtime is negligible because the server only needs to average model deltas from the clients. For PE, the communication cost comes from each client downloading K = 1800 sentence embeddings of size 384 resulting in a download cost of 700K (700,000) floats, and uploading a histogram of size 1800 resulting in an upload cost of 1800 floats. The client runtime cost comes from calculating a nearest neighbors histogram and the server runtime cost for PE comes mainly from using the LLM Ψ to generate synthetic samples each round. In POPri each client downloads K J = 1800 10 sentence embeddings for a download cost of 7M floats and uploads a vector of size 18,000 for an upload cost of 18,000 floats. The client runtime cost of POPri comes from calculating the cosine similarities, and the server runtime comes from both using Ψ to generate synthetic samples and running DPO. Interpretation. In summary, POPri is much more communication-efficient and client compute-efficient than Fed Avg, while using much more server compute. On the other hand, POPri is generally more communicationand computationally-expensive than PE. At the same time, POPri has the best downstream performance among all three methods, as seen in Table 1. Hence, POPri can be a suitable method when (1) server compute is cheap and powerful, and (2) getting the best synthetic data/downstream model quality is important. 5.2. Ablations Cosine similarity vs. Nearest neighbors histogram. Private Evolution (Lin et al., 2023; Hou et al., 2024; Xie et al., 2024) uses a DP nearest neighbors histogram calculation to score the quality of synthetic samples. The DP nearest neighbors histogram sets the score of a particular synthetic sample to the number of private samples that are closest to that particular synthetic sample (under some text embedding function). In POPri, we instead set the score of a particular synthetic sample to the average cosine similarity between that particular synthetic sample and all private samples (under some text embedding function). We find that cosine similarity works much better than a nearest neighbors histogram (Figure 6), possibly because nearest neighbor histograms produce sparser scores, often assigning zero to all synthetic samples associated with a given prompt. In this setting, the chosen and rejected samples for preference optimization end up being essentially random. In contrast, cosine similarity provides denser scoring that allows the construction of meaningful preference pairs for all prompts. Partial client participation. In each round a fixed number of clients is subsampled uniformly at random for feedback generation. Figure 3 shows the next-token prediction accuracy (%) of four algorithms for different numbers of clients per round. POPri consistently outperforms all of the baselines, regardless of the client sampling rate. Moreover, POPri s accuracy is not sensitive to the client sampling rate. Private Federated Learning using Preference-Optimized Synthetic Data 0 1 2 3 4 5 6 7 8 9 10 11 12 13 Text embedding PCA data (d=2) 0 1 2 3 4 5 6 7 8 9 10 11 12 13 Median distance to the medoid FID=0.38 Round0 FID=0.10 Round6 FID=0.07 Round7 FID=0.15 Round9 FID=0.26 Round12 Evaluation Dataset Figure 4. PCA visualization of POPri synthetic data embeddings over rounds. Right 6 Panels: PCA-2 plots for synthetic data and evaluation data embeddings from the best checkpoint each round for 20 iterations. The orange (round 7) and maroon point clouds represent the round with the lowest FID score and the validation dataset, respectively. Top Left Panel: FID score vs. rounds. Bottom Left Panel: Median distance to the medoid vs rounds. Running POPri for too many rounds appears to cause overfitting. Method Download Upload Client Runtime Server Runtime (floats) (floats) (GPU sec) (GPU sec) Fed Avg 82 million 82 million 4.8 PE 700,000 1,800 0.0027 326.25 POPri 7 million 18,000 0.01 13,547.84 Reduction factor (Fed Avg / POPri) 11.71 4555 480.0 Reduction factor (PE / POPri) 0.100 0.100 0.270 0.024 Table 2. Table setting. Communication and computation cost comparison per round (and per-client for download/upload/client runtime cost) across methods on the bio Rxiv dataset with 1000 clients sampled per round. Download and upload are measured in floats; runtimes are measured in GPU seconds (lower is better). Reduction factor (X / POPri) is the cost of method X divided by the cost of POPri for the given resource; green is a reduction, red is an increase. Server runtime for Fed Avg is left blank as it is negligible compared to other methods. Overall, we view POPri as suitable when server compute is relatively cheap, and improved sample quality is important enough to justify higher on-device communication and computation costs relative to PE (Table 1). Data Distribution Evolution. Synthetic datasets are often generated using a language model distinct from the one being aligned (Guo et al., 2024), making the alignment phase inherently off-policy as the model evolves during training. This is reflected in the synthetic data, where the FID score (relative to a held-out evaluation set) worsens after improving. Figure 4 shows PCA visualizations of synthetic data embeddings across alignment iterations, while the left panels plot the FID score and median distance to the medoid in the PCA space. The data distribution transitions from being initially clustered to (roughly) matching the true data distribution, back to being clustered, likely due to overfitting. Early stopping based on validation metrics can help. How to select rejected samples. Unlike vanilla DPO, we can select the chosen and rejected sample pair from the J samples for each of the K prompts. We consistently choose the highest-scoring sample (rank 1) as the chosen sample, but there are different options for the rejected sample. We found that the middle-ranked sample (e.g., ℓ= 5th-ranked out of J = 10) yields the best results, rather than using the last-ranked sample. If the rejected sample is too dissimilar to client data, then the preference pair is uninformative. However, choosing a sample that is too similar to client data (e.g., rank 2) for the rejected sample could lead to incorrect preference pairs due to DP noise swapping rankings. We use the 5th-ranked sample, and justify it experimentally in Appendix E.3. 6. Conclusion Private on-device learning is important when data is stored on edge devices with hardware, storage, and privacy constraints. We propose POPri, which recasts synthetic databased approaches for private learning as an LLM policy optimization problem. POPri makes several novel design choices in how it gathers and utilizes client feedback to generate DP synthetic data, which is used to finetune a downstream on-device model. POPri outperforms DP-FL and synthetic data baselines on a variety of tasks, including on a large-scale Large Fed Bench, a new federated benchmark we have curated. Private Federated Learning using Preference-Optimized Synthetic Data Impact Statement In this paper, we train models satisfying differential privacy guarantees. When using differential privacy as a tool for protecting user data, it is important to communicate to users what the privacy guarantees mean to be able to obtain informed consent. The algorithms in this paper also use LLMs, which were trained on large scale public text data. While this data was public, explicit consent may not have been given for its use in training the models. The algorithms using LLMs in the paper make no claims about the privacy guarantees of data used in the pretraining of the LLMs. While our work aims to show how synthetic data can be useful for federated learning, it also poses a number of ethical risks, including the generation of biased or harmful content. In particular, our method (and all variants of private evolution) inherits the biases and undesirable aspects of the public LLM. For example, suppose the public LLM only generates text in English, but some clients private data is all in Spanish. In these settings, clients would be forced to vote on synthetic samples, even if potentially none of them are relevant to the client. This may cause the client to contribute data reinforcing a model that is actively not useful (or even harmful) to the client. In contrast, DP-SGD methods do not suffer from this shortcoming, because they do not rely on a public LLM. This problem raises an important point how can we design DP synthetic data algorithms in which clients can stem the biases or failures of the public LLM, based on their own data? This important question is beyond the scope of the current paper. Acknowledgments This work was supported in part by NSF grants CCF2338772 and CNS-2148359, as well as C3.ai, Bosch, Intel, and the Sloan Foundation. This work used Bridges-2 GPU (Brown et al., 2021; Buitrago & Nystrom, 2021) at the Pittsburgh Supercomputing Center through allocation CIS240135 and CIS240937 from the Advanced Cyberinfrastructure Coordination Ecosystem: Services & Support (ACCESS) program, which is supported by National Science Foundation grants 2138259, 2138286, 2138307, 2137603, and 2138296 (Boerner et al., 2023). The authors acknowledge the National Artificial Intelligence Research Resource (NAIRR) Pilot, the AI and Big Data group at the Pittsburgh Supercomputing Center, and NCSA Delta GPU for contributing to this research result. Abadi, M., Chu, A., Goodfellow, I., Mc Mahan, H. B., Mironov, I., Talwar, K., and Zhang, L. Deep learning with differential privacy. In Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security, pp. 308 318. ACM, 2016. Achiam, J., Adler, S., Agarwal, S., Ahmad, L., Akkaya, I., Aleman, F. L., Almeida, D., Altenschmidt, J., Altman, S., Anadkat, S., et al. Gpt-4 technical report. ar Xiv preprint ar Xiv:2303.08774, 2023. AI@Meta. Llama 3 model card. 2024. URL https://github.com/meta-llama/llama3/ blob/main/MODEL_CARD.md. Anil, R., Dai, A. M., Firat, O., Johnson, M., Lepikhin, D., Passos, A., Shakeri, S., Taropa, E., Bailey, P., Chen, Z., et al. Palm 2 technical report. ar Xiv preprint ar Xiv:2305.10403, 2023. Boerner, T. J., Deems, S., Furlani, T. R., Knuth, S. L., and Towns, J. Access: Advancing innovation: Nsf s advanced cyberinfrastructure coordination ecosystem: Services & support. In Practice and Experience in Advanced Research Computing 2023: Computing for the Common Good, pp. 173 176. 2023. Bommasani, R., W. S. and Schofield, X. Towards private synthetic text generation. In Neur IPS 2019 Machine Learning with Guarantees Workshop, 2019. Bonawitz, K. A., Ivanov, V., Kreuter, B., Marcedone, A., Mc Mahan, H. B., Patel, S., Ramage, D., Segal, A., and Seth, K. Practical secure aggregation for federated learning on user-held data. In NIPS Workshop on Private Multi-Party Machine Learning, 2016. URL https://arxiv.org/abs/1611.04482. Brown, S. T., Buitrago, P., Hanna, E., Sanielevici, S., Scibek, R., and Nystrom, N. A. Bridges-2: A platform for rapidly-evolving and data intensive research. In Practice and Experience in Advanced Research Computing 2021: Evolution Across All Dimensions, PEARC 21, New York, NY, USA, 2021. Association for Computing Machinery. ISBN 9781450382922. doi: 10. 1145/3437359.3465593. URL https://doi.org/ 10.1145/3437359.3465593. Buitrago, P. A. and Nystrom, N. A. Neocortex and bridges-2: A high performance ai+hpc ecosystem for science, discovery, and societal good. Communications in computer and information science, 1327, 2021. doi: 10.1007/978-3-030-68035-0 15. URL https: //par.nsf.gov/biblio/10274872. Charles, Z., Mitchell, N., Pillutla, K., Reneer, M., and Garrett, Z. Towards federated foundation models: Scalable dataset pipelines for group-structured learning. ar Xiv preprint ar Xiv:2307.09619, 2023. Private Federated Learning using Preference-Optimized Synthetic Data Charles, Z., Ganesh, A., Mc Kenna, R., Mc Mahan, H. B., Mitchell, N. E., Pillutla, K., and Rush, J. K. Fine-tuning large language models with user-level differential privacy. In ICML 2024 Workshop on Theoretical Foundations of Foundation Models, 2024. Collins, L., Wu, S., Oh, S., and Sim, K. C. Profit: Benchmarking personalization and robustness trade-off in federated prompt tuning. ar Xiv preprint ar Xiv:2310.04627, 2023. Dwork, C. Differential privacy. In International colloquium on automata, languages, and programming, pp. 1 12. Springer, 2006. Dwork, C. and Roth, A. The algorithmic foundations of differential privacy. Found. Trends Theor. Comput. Sci., 9(3 4):211 407, August 2014. ISSN 1551-305X. doi: 10.1561/0400000042. URL https://doi.org/10. 1561/0400000042. Grattafiori, A., Dubey, A., and et al. The llama 3 herd of models, 2024. URL https://arxiv.org/abs/ 2407.21783. Guo, D., Yang, D., Zhang, H., Song, J., Zhang, R., Xu, R., Zhu, Q., Ma, S., Wang, P., Bi, X., et al. Deepseek-r1: Incentivizing reasoning capability in llms via reinforcement learning. ar Xiv preprint ar Xiv:2501.12948, 2025. Guo, S., Zhang, B., Liu, T., Liu, T., Khalman, M., Llinares, F., Rame, A., Mesnard, T., Zhao, Y., Piot, B., Ferret, J., and Blondel, M. Direct Language Model Alignment from Online AI Feedback. ar Xiv e-prints, art. ar Xiv:2402.04792, February 2024. doi: 10.48550/ar Xiv. 2402.04792. Hard, A., Rao, K., Mathews, R., Ramaswamy, S., Beaufays, F., Augenstein, S., Eichner, H., Kiddon, C., and Ramage, D. Federated learning for mobile keyboard prediction, 2019. URL https://arxiv.org/abs/ 1811.03604. Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., and Hochreiter, S. Gans trained by a two time-scale update rule converge to a local nash equilibrium. Neur IPS, 30, 2017. Hou, C., Shrivastava, A., Zhan, H., Conway, R., Le, T., Sagar, A., Fanti, G., and Lazar, D. Pre-text: Training language models on private federated data in the age of llms, 2024. URL https://arxiv.org/abs/ 2406.02958. Hu, E. J., Shen, Y., Wallis, P., Allen-Zhu, Z., Li, Y., Wang, S., Wang, L., and Chen, W. Lora: Low-rank adaptation of large language models. ar Xiv preprint ar Xiv:2106.09685, 2021. Kairouz, P., Mc Mahan, B., Song, S., Thakkar, O., Thakurta, A., and Xu, Z. Practical and private (deep) learning without sampling or shuffling. In ICML, 2021a. Kairouz, P., Mc Mahan, H. B., Avent, B., Bellet, A., Bennis, M., Bhagoji, A. N., Bonawitz, K., Charles, Z., Cormode, G., Cummings, R., et al. Advances and open problems in federated learning. Foundations and Trends in Machine Learning, 14(1 2):1 210, 2021b. Kurakin, A., Ponomareva, N., Syed, U., Mac Dermed, L., and Terzis, A. Harnessing large-language models to generate private synthetic text. ar Xiv preprint ar Xiv:2306.01684, 2023. Lin, Z., Sekar, V., and Fanti, G. On the privacy properties of gan-generated samples. In AISTATS, pp. 1522 1530. PMLR, 2021. Lin, Z., Gopi, S., Kulkarni, J., Nori, H., and Yekhanin, S. Differentially private synthetic data via foundation model apis 1: Images. ar Xiv preprint ar Xiv:2305.15560, 2023. Lin, Z., Baltrusaitis, T., Wang, W., and Yekhanin, S. Differentially private synthetic data via apis 3: Using simulators instead of foundation model. ar Xiv preprint ar Xiv:2502.05505, 2025a. Lin, Z., Baltrusaitis, T., Wang, W., and Yekhanin, S. Differentially private synthetic data via apis 3: Using simulators instead of foundation model, 2025b. URL https: //arxiv.org/abs/2502.05505. Magar, I. and Schwartz, R. Data contamination: From memorization to exploitation, 2022. URL https:// arxiv.org/abs/2203.08242. Malladi, S., Gao, T., Nichani, E., Damian, A., Lee, J. D., Chen, D., and Arora, S. Fine-tuning language models with just forward passes. Advances in Neural Information Processing Systems, 36:53038 53075, 2023. Mattern, J., Jin, Z., Weggenmann, B., Schoelkopf, B., and Sachan, M. Differentially private language models for secure data sharing. In Goldberg, Y., Kozareva, Z., and Zhang, Y. (eds.), Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pp. 4860 4873, Abu Dhabi, United Arab Emirates, December 2022. Association for Computational Linguistics. doi: 10.18653/v1/2022.emnlp-main.323. URL https:// aclanthology.org/2022.emnlp-main.323. Mc Mahan, B., Moore, E., Ramage, D., Hampson, S., and Arcas, B. A. y. Communication-Efficient Learning of Deep Networks from Decentralized Data. In Singh, A. and Zhu, J. (eds.), Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, volume 54 of Proceedings of Machine Learning Private Federated Learning using Preference-Optimized Synthetic Data Research, pp. 1273 1282. PMLR, 20 22 Apr 2017a. URL https://proceedings.mlr.press/v54/ mcmahan17a.html. Mc Mahan, B., Ramage, D., Talwar, K., and Zhang, L. Learning DP recurrent language models. 2017b. Mitchell, E. A note on dpo with noisy preferences & relationship to ipo, November 25 2023. URL https: //ericmitchell.ai/cdpo.pdf. Version 1.1. Mosbach, M., Pimentel, T., Ravfogel, S., Klakow, D., and Elazar, Y. Few-shot fine-tuning vs. in-context learning: A fair comparison and evaluation. ar Xiv preprint ar Xiv:2305.16938, 2023. Nguyen, J., Malik, K., Zhan, H., Yousefpour, A., Rabbat, M., Malek, M., and Huba, D. Federated learning with buffered asynchronous aggregation. In International Conference on Artificial Intelligence and Statistics, pp. 3581 3607. PMLR, 2022. Ouyang, L., Wu, J., Jiang, X., Almeida, D., Wainwright, C., Mishkin, P., Zhang, C., Agarwal, S., Slama, K., Ray, A., et al. Training language models to follow instructions with human feedback. Advances in neural information processing systems, 35:27730 27744, 2022. Paulik, M., Seigel, M., Mason, H., Telaar, D., Kluivers, J., van Dalen, R., Lau, C. W., Carlson, L., Granqvist, F., Vandevelde, C., Agarwal, S., Freudiger, J., Byde, A., Bhowmick, A., Kapoor, G., Beaumont, S., Aine Cahill, Hughes, D., Javidbakht, O., Dong, F., Rishi, R., and Hung, S. Federated evaluation and tuning for on-device personalization: System design applications, 2021. URL https://arxiv.org/abs/2102.08503. Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., Sutskever, I., et al. Language models are unsupervised multitask learners. Open AI blog, 1(8):9, 2019. Rafailov, R., Sharma, A., Mitchell, E., Manning, C. D., Ermon, S., and Finn, C. Direct preference optimization: Your language model is secretly a reward model. In Oh, A., Naumann, T., Globerson, A., Saenko, K., Hardt, M., and Levine, S. (eds.), Advances in Neural Information Processing Systems, volume 36, pp. 53728 53741. Curran Associates, Inc., 2023. Raffel, C., Shazeer, N., Roberts, A., Lee, K., Narang, S., Matena, M., Zhou, Y., Li, W., and Liu, P. J. Exploring the limits of transfer learning with a unified text-to-text transformer. ar Xiv e-prints, 2019. Reddi, S., Charles, Z., Zaheer, M., Garrett, Z., Rush, K., Koneˇcn y, J., Kumar, S., and Mc Mahan, H. B. Adaptive federated optimization. ar Xiv preprint ar Xiv:2003.00295, 2020. Reimers, N. and Gurevych, I. Sentence-bert: Sentence embeddings using siamese bert-networks. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, 11 2019a. URL https://arxiv.org/ abs/1908.10084. Reimers, N. and Gurevych, I. Sentence-BERT: Sentence embeddings using Siamese BERT-networks. In Inui, K., Jiang, J., Ng, V., and Wan, X. (eds.), Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 3982 3992, Hong Kong, China, November 2019b. Association for Computational Linguistics. doi: 10.18653/v1/D19-1410. URL https: //aclanthology.org/D19-1410/. Roberts, M., Thakur, H., Herlihy, C., White, C., and Dooley, S. Data contamination through the lens of time, 2023. URL https://arxiv.org/abs/2310.10628. Sanh, V., Debut, L., Chaumond, J., and Wolf, T. Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter. ar Xiv preprint ar Xiv:1910.01108, 2019. Team, G., Anil, R., Borgeaud, S., Alayrac, J.-B., Yu, J., Soricut, R., Schalkwyk, J., Dai, A. M., Hauth, A., Millican, K., et al. Gemini: a family of highly capable multimodal models. ar Xiv preprint ar Xiv:2312.11805, 2023. Touvron, H., Lavril, T., Izacard, G., Martinet, X., Lachaux, M.-A., Lacroix, T., Rozi ere, B., Goyal, N., Hambro, E., Azhar, F., et al. Llama: Open and efficient foundation language models. ar Xiv preprint ar Xiv:2302.13971, 2023. Turc, I., Chang, M.-W., Lee, K., and Toutanova, K. Well-read students learn better: On the importance of pre-training compact models, 2019. URL https:// arxiv.org/abs/1908.08962. Utpala, S., Hooker, S., and Chen, P.-Y. Locally differentially private document generation using zero shot prompting. In Bouamor, H., Pino, J., and Bali, K. (eds.), Findings of the Association for Computational Linguistics: EMNLP 2023, pp. 8442 8457, Singapore, December 2023. Association for Computational Linguistics. doi: 10.18653/v1/2023.findings-emnlp. 566. URL https://aclanthology.org/2023. findings-emnlp.566/. Wu, S., Xu, Z., Zhang, Y., Zhang, Y., and Ramage, D. Prompt public large language models to synthesize data for private on-device applications. ar Xiv preprint ar Xiv:2404.04360, 2024. Private Federated Learning using Preference-Optimized Synthetic Data Xie, C., Lin, Z., Backurs, A., Gopi, S., Yu, D., Inan, H. A., Nori, H., Jiang, H., Zhang, H., Lee, Y. T., Li, B., and Yekhanin, S. Differentially private synthetic data via foundation model APIs 2: Text. In Forty-first International Conference on Machine Learning, 2024. URL https: //openreview.net/forum?id=LWD7upg1ob. Xie, L., Lin, K., Wang, S., Wang, F., and Zhou, J. Differentially private generative adversarial network. ar Xiv preprint ar Xiv:1802.06739, 2018. Xu, Z., Collins, M., Wang, Y., Panait, L., Oh, S., Augenstein, S., Liu, T., Schroff, F., and Mc Mahan, H. B. Learning to generate image embeddings with user-level differential privacy. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 7969 7980, 2023a. Xu, Z., Zhang, Y., Andrew, G., Choquette-Choo, C. A., Kairouz, P., Mc Mahan, H. B., Rosenstock, J., and Zhang, Y. Federated learning of gboard language models with differential privacy, 2023b. URL https://arxiv. org/abs/2305.18465. Yang, S., Chiang, W.-L., Zheng, L., Gonzalez, J. E., and Stoica, I. Rethinking benchmark and contamination for language models with rephrased samples, 2023. URL https://arxiv.org/abs/2311.04850. Ye, R., Ge, R., Zhu, X., Chai, J., Du, Y., Liu, Y., Wang, Y., and Chen, S. Fedllm-bench: Realistic benchmarks for federated learning of large language models. ar Xiv preprint ar Xiv:2406.04845, 2024. Yousefpour, A., Shilov, I., Sablayrolles, A., Testuggine, D., Prasad, K., Malek, M., Nguyen, J., Ghosh, S., Bharadwaj, A., Zhao, J., Cormode, G., and Mironov, I. Opacus: User-friendly differential privacy library in pytorch. In Neur IPS 2021 Workshop Privacy in Machine Learning, 2021. URL https://openreview.net/forum? id=Eop KEYBo I-. Yu, D., Backurs, A., Gopi, S., Inan, H., Kulkarni, J., Lin, Z., Xie, C., Zhang, H., and Zhang, W. Training private and efficient language models with synthetic data from LLMs. In Socially Responsible Language Modelling Research, 2023. URL https://openreview.net/forum? id=FKwt Kzgl Fb. Yu, D., Kairouz, P., Oh, S., and Xu, Z. Privacy-preserving instructions for aligning large language models, 2024. URL https://arxiv.org/abs/2402.13659. Yue, X., Inan, H., Li, X., Kumar, G., Mc Anallen, J., Shajari, H., Sun, H., Levitan, D., and Sim, R. Synthetic text generation with differential privacy: A simple and practical recipe. In Rogers, A., Boyd-Graber, J., and Okazaki, N. (eds.), Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 1321 1342, Toronto, Canada, July 2023a. Association for Computational Linguistics. doi: 10.18653/v1/2023.acl-long.74. URL https:// aclanthology.org/2023.acl-long.74. Yue, X., Inan, H. A., Li, X., Kumar, G., Mc Anallen, J., Shajari, H., Sun, H., Levitan, D., and Sim, R. Synthetic text generation with differential privacy: A simple and practical recipe, 2023b. URL https://arxiv.org/ abs/2210.14348. Zhang, L., Li, B., Thekumparampil, K. K., Oh, S., and He, N. Dpzero: Private fine-tuning of language models without backpropagation. In Forty-first International Conference on Machine Learning, 2024. Zhou, K., Zhu, Y., Chen, Z., Chen, W., Zhao, W. X., Chen, X., Lin, Y., Wen, J.-R., and Han, J. Don t make your llm an evaluation benchmark cheater, 2023. URL https: //arxiv.org/abs/2311.01964. Zou, T., Liu, Y., Li, P., Xiong, Y., Zhang, J., Liu, J., Ye, X., Ouyang, Y., and Zhang, Y.-Q. Contrastive private data synthesis via weighted multi-plm fusion. ar Xiv preprint ar Xiv:2502.00245, 2025. Private Federated Learning using Preference-Optimized Synthetic Data A. Algorithmic Details Here we provide the pseudocode for the centralized version of POPri and the central + conditional generation version of POPri. Algorithm 2 POPri (central DP, unconditional) 1: Input: Number of iterations T, Noise multiplier σ, LLM Ψ, embedding model Γ, base prompt η, random prompt generator Λ( ), rejected index ℓ, private dataset S, K number of prompts, J number of responses per prompt 2: Output: LLM for generating synthetic data ΨT +1 3: 4: Embed all private samples E = Γ(S) 5: Initialize LLM Ψ1 = Ψ 6: for t 1 ... T do 7: Initialize the response vector R = 8: for k 1 . . . K do 9: Generate prompt ηk = Λ(η), 10: Generate J responses Rkj = Ψt(ηk), j [J] 11: end for 12: Calculate embeddings Esyn,t = {Γ(Rkj)}k [K],j [J] 13: Scorest CENTRALSCORE(Esyn,t, E) + N(0, σ2I) 14: Set P[k, j] as the j-th highest score response for prompt ηk, according to Scorest 15: Initialize preference dataset Pt = 16: for k 1 . . . K do 17: Select positive synthetic sample: Pt[k, 1] = Pt[k, 1] 18: Select negative synthetic sample: Pt[k, 2] = Pt[k, ℓ] 19: end for 20: Fine-tune: Ψt+1 DPO(Ψt, {ηk}k [K], Pt) 21: end for 22: Output ΨT +1 Algorithm 3 POPri (central DP, conditional) 1: Input: Number of iterations T, Noise multiplier σ, LLM Ψ, embedding model Γ, base prompt η, conditional (class-specified) random prompt generator Λ( , ), rejected index ℓ, private dataset S, K number of prompts, J number of responses per prompt, number of classes B 2: Output: LLM for generating synthetic data ΨT +1 3: Embed private samples for each class i = 1...B, Ei = Γ({s}F (s)=i,s S) where F(s) is the class index of sample s 4: Initialize LLM Ψ1 = Ψ 5: for t 1 ... T do 6: Initialize B response vectors R = { , ... } = {R(1), ..., R(B)} 7: for b ... B do 8: for k 1 . . . K do 9: Generate prompt ηk = Λ(η, b), 10: Generate J responses R(b) kj = Ψt(ηk), j [J] 11: end for 12: Calculate embeddings E(b) syn,t = {Γ(R(b) kj )}k [K],j [J] 13: Scorest CENTRALSCORE(E(b) syn,t, Eb) + N(0, σ2I) 14: Set P[k, j] as the j-th highest score response for prompt ηk, according to Scorest 15: Initialize preference dataset P(b) t = 16: for k 1 . . . K do 17: Select positive synthetic sample: P(b) t [k, 1] = Pt[k, 1] 18: Select negative synthetic sample: P(b) t [k, 2] = Pt[k, ℓ] 19: Set prompt P(b) t [k, 3] = ηk 20: end for 21: end for 22: Fine-tune: Pt = SB b=1 P (b) t , Ψt+1 DPO(Ψt, Pt) 23: end for 24: Output ΨT +1 Below is the similarity scoring function we use for the federated setting. Private Federated Learning using Preference-Optimized Synthetic Data Algorithm 4 SIMILARITY 1: Input: Set of embeddings of private client data Ei = {emb(s(i) 1 ), . . . , emb(s(i) mi)} for i St, embeddings of synthetic data Esyn, total synthetic samples M = K J Scores 0M 2: Scores[j] = 1 mi P epri,ej epri ej for ej Esyn 3: return Scores/max 1, Scores 2 Below is the similarity scoring function we use for the central DP setting. Algorithm 5 CENTRALSCORE 1: Input: Embeddings of private data E, embeddings of synthetic data Esyn Scores 0M 2: Scores[j] = (1/|E|) P epri E epri,ej epri ej for ej Esyn 3: return Scores B. Implementation Details of POPri B.1. Model and Hyperparameters We choose LLa MA-3-8B as the data generator in POPri and we fine-tune it iteratively during the course of the algorithm. To fine-tune the LLa MA-3-8B model, we use Lo RA fine-tuning with rank 4, α = 8, applied to all the projection matrices in LLa MA-3-8B. We adapt the Adam W optimizer with a cosine learning rate scheduler with the learning rate ranging from 3 10 7 to 8 10 7. In the Congress and bio Rxiv evaluations, the sample set Ωis a subset of the c4 dataset (Raffel et al., 2019), which is a large scale dataset from 2019, which we use for fair comparison with Private Evolution (Pr E-Text), though we do not know their exact initial sample set because they did not release it. For the Pub Med evaluation, the sample set Ω is a set of 2000 samples generated using the Pub Med generation prompt in Table 16 of the Aug-PE paper, generated by LLa MA-3-8B-Instruct (which has a knowledge cutoff of March 2023), for comparison with Aug-PE (Xie et al., 2024). For each iteration, we fine-tune the models for 2 epochs and select the best checkpoint with the lowest FID score relative to the validation dataset. This checkpoint is used for synthetic data generation and as the starting point for the next iteration. The batch size is set to 24. In each round we generate 18000 synthetic data samples for the clients to evaluate. This is accomplished with 1800 prompts, each generating 10 samples for clients to rank. We select the 1st and 5th ranked sample for a given prompt for the selected and rejected data samples in the DPO preference dataset. We describe the experiments regarding which rank to use for constructing the preference dataset in detail in Appendix Section E.3. To test the scaling relation with the number of clients per round and the total number of clients participating in the training, we set up the parameters and privacy budget shown in Table 3. The all-Mini LM-L6-v2 sentence transformer model is used as the embedding model in POPri. We note that we adopt sentence-t5-base sentence transformer for Pub Med during the step of fine-tuning BERTsmall, which follows the setting in AUG-PE. We ensure POPri follows privacy guarantee of (ϵ, δ)-DP = (1, 3 10 6) or (7, 3 10 6) for both the bio Rxiv and the Congressional Speeches datasets and run with 20 iterations for DP-Fed Avg, DP-FTRL, Pr E-Text for comparison. For AUG-PE, we set (ϵ, δ)-DP = (1, 2.72 10 6) or (4, 2.72 10 6). Pub Med experiments are run with 10 iterations. In terms of models for downstream tasks: For Bio Rxiv & Congressional Speeches, we fine-tuned the pre-trained Distill GPT2 for next-token prediction. We set the max sequence length as 64, number of generated synthetic data as 1,000,000, the batch size as 160, the learning rate as 2e 4, and the number of epochs as 80. For Pub Med, to compare with (Yue et al., 2023b), we follow their procedure to leverage pre-trained BERTsmall (Turc et al., 2019). We set the max sequence length as 512, number of generated synthetic data as 2000, batch size as 32, learning rate as 3e-4, the weight decay as 0.01, and the number of epochs as 10. To compare with (Xie et al., 2024), we set up the (ϵ, δ)-DP value and hypterparameter according to their choice. For example, they set δ = 1 Npriv log(Npriv) following (Yue et al., 2023b). To achieve δ = {1,4}, we use noise multiplier σ = {13.7, 3.87} for 10 iterations under DP Private Federated Learning using Preference-Optimized Synthetic Data List of 6 diverse original text samples: Original Text Sample 1 The observations showed that the object is four million times more massive than the sun and is the size of one astronomical unit (AU), a span equal to Earth's distance from the sun. Sgr A* has a mass density at least a trillion times greater than any known cosmic object. Original Text Sample 2 In response to the general question, they need to study self-protection away from their marital baggage. They need to learn about home security, mobile security, the nature of crime, de-escalation, the law, escape tactics, awareness, and on and on. When it Original Text Sample 3 Under the Patriot Act of 2001, the government significantly expanded its authority in regards to electronic surveillance (Henderson, 2002). One of the chief complaints is that the government can investigate anything that is considered significant. The problem here is that there is Original Text Sample 4 The life history advance program shall be funded from any of the following: monies provided by the general fund; amounts in the presidential family partnership fund; or monies provided by the revolving fund. Original Text Sample 5 As you meet with employers this summer, get in touch with the team.... Figure 5. The synthetic data generation prompt for POPri. The black text marks the input prompt, and the brown text after Original Text Sample 4 is generated. The generated text between Original Text Sample 4 and Original Text Sample 5 is collected and used as a synthetic sample. on all Pub Med data. Note that our noise multiplier values are slightly different than (Xie et al., 2024) due to different methods for calculating differential privacy. B.2. Prompt Design To compare with other data generator methods, we adopt the prompts used in the baseline models against which we compare. We generate the synthetic data using an approach similar to that in Pr E-Text (Hou et al., 2024). Figure 5 shows an example of the prompt we use for prompting LLa MA-3-BB for generating synthetic data. For bio Rxiv/Congress, we randomly take text samples from the c4 (Raffel et al., 2019) dataset as our examples in the prompt. For Pub Med, while running POPri, we still adopt the prompt shown in Figure 5 but reduce the number of examples to two in order to accommodate longer sequence lengths, randomly sampling generated abstracts from LLa MA-3-8B. For Open Review, we prompt the model directly to generate paper reviews (similarly to (Xie et al., 2024)). C. Implementation Details of Baseline Models In this section we provide implementation details for the baseline algorithms. We use two DP-FL baselines: DP-Fed Avg and DP-FTRL. For the PE baseline, we implement Pr E-Text (Hou et al., 2024) for the evaluations on the bio Rxiv and Congressional Speeches datasets. For the PE baselines on the Pub Med dataset we directly compare against the Aug-PE results from Xie et al. (2024). C.1. DP-Fed Avg We employ the Fed Avg federated optimization algorithm (Mc Mahan et al., 2017b) to fully fine-tune Distil GPT2, avoiding linear probing due to its poor performance in DP language models (Lin et al., 2021). Our training configuration includes a batch size of 2, a sequence length of 64, and 20 rounds for Table 1 and 50 rounds for Figure 3, and either full or partial client participation. For differential privacy (DP), we utilize secure aggregation (Bonawitz et al., 2016) and introduce Gaussian noise (Mc Mahan et al., 2017b). We evaluate the model using next-token prediction accuracy across various numbers of training epochs on the clients. We tune the learning rate within the range [0.001, 0.1, 0.1] and the clipping threshold between [0.01, 0.1, 1.0], selecting the model with the best performance on the evaluation set for reporting. The noise is scaled to ensure a privacy guarantee of (ϵ, δ)-DP where δ = 3 10 6 and ϵ = {1,7}, representing two distinct privacy regimes. The noise multipliers are σ = {19.3, 3.35} when considering all the data, and the settings for partial participation experiments Private Federated Learning using Preference-Optimized Synthetic Data Table 3. Experiment privacy budget settings. Total # of # of clients σ1 a, ϵ = 7 σ1 a, ϵ = 1 σ2 b, ϵ = 7 σ2 b, ϵ = 1 clients per round 10000 1000 3.4 19.5 10000 5000 15.5 30.8 10000 10000 30.6 30.8 72000 72000 3.35 19.3 3.35 19.5 133000 133000 3.35 19.3 3.35 19.5 a For DP-Fed Avg, Pr E-Text, POPri. b For DP-FTRL are shown in Table 3. C.2. DP-FTRL We also use the DP variant of Follow-The-Regularized-Leader (DP-FTRL) algorithm (Kairouz et al., 2021a) to fully fine-tune Distil GPT2. The hyperparameter settings are similar to DP-Fed Avg other than the noise multipliers. The noise multipliers are σ = {19.5, 3.35} when considering all the data, and the settings for partial participation experiments are shown in Table 3. C.3. Pr E-Text We follow similar settings as Hou et al. (2024) with some modifications. The privacy budget is similar to DP-Fed Avg and POPri, with a privacy guarantee of (ϵ, δ)-DP where δ = 3 10 6 and ϵ = {1,7} with σ = {19.3, 3.35} for full participation and partial participation in Table 3. We set the thresholds H = 0.1626, T = 20, and Nsyn = 1024. We adopt the all-Mini LML6-v2 sentence transformer model for text embedding generation. D. Experimental Details D.1. Privacy Accounting The precise privacy settings we use and their corresponding ϵ values, as calculated by their corresponding privacy budget computation methods, are reported in Table 3. DP-Fed Avg (Mc Mahan et al., 2017b) and Private Evolution (Pr E-Text) (Hou et al., 2024) both use the Gaussian mechanism, and thus use similar computations. In both cases, we use the privacy accountant of the Opacus library (Yousefpour et al., 2021). For DP-Fed Avg, we calculate privacy by inputting the number of rounds, the client sampling ratio, setting the noise multiplier to be the product of σ and the clipping threshold, choosing a δ 1/|S|, and setting σ for the desired ϵ. Private Evolution (Pr E-Text) (Hou et al., 2024) also uses the Gaussian mechanism, so we use the same accounting except the noise multiplier is the product of σ and the maximum number of samples per client. For DP-FTRL, we follow the privacy accounting methods from their implementation. For Private Evolution (Aug-PE) (Xie et al., 2024), we report their reported ϵ directly. D.2. Evaluation Details for Different Datasets D.2.1. LARGEFEDBENCH EVALUATION For the bio Rxiv and Congressional Speeches datasets, we use the Pr E-Text version of Private Evolution because the Pr E-Text evaluation focused on datasets with samples with max sequence length of 64. D.2.2. PUBMED AND OPENREVIEW EVALUATION For Pub Med and Open Review, our Private Evolution baseline compares to Aug-PE, which has already been evaluated on Pub Med and Open Review (Xie et al., 2024). Note that Pub Med and Open Review was used by Xie et al. (2024) to evaluate central DP algorithms. In the central DP setting, there are no clients; all private data is held at the server and the goal is to release a model with DP guarantees. The notion of neighboring dataset in central DP is a centrally held dataset that is the same except for a single data sample. To compare our algorithm directly with results reported for Private Evolution (Aug-PE) (Xie et al., 2024), we replicate the central DP setting for this dataset by having one Pub Med abstract per client and Private Federated Learning using Preference-Optimized Synthetic Data sampling all clients every iteration (or round , in our case). E. Ablation Studies E.1. Cosine similarity vs Nearest neighbors histogram In this section we perform an ablation justifying the choice of cosine similarity as a scoring function over the nearest neighbor histogram employed by Private Evolution. We find that using cosine similarity works much better than nearest neighbors histogram for our use case, because nearest neighbors histogram is too sparse to ensure the construction of meaningful preference pairs for POPri. Figure 6. Left: FID scores of POPri using NN histogram scoring vs. POPri using cosine similarity. Right: After the client feedback stage, we measure the percentage of the time the non-noised and non-clipped score (nearest neighbor histogram scoring or cosine similarity scoring) of the chosen sample is higher than the rejected sample. For cosine similarity, this recovery rate is much higher (nearly 100%) than in nearest neighbors histogram. Interpretation. Nearest neighbors histogram is much sparser than cosine similarity, often assigning zero to all synthetic samples associated with a given prompt in POPri. This leads to preference pairs often being completely noisy. Cosine similarity provides denser scoring that allows the construction of meaningful preference pairs for all prompts. Figure 7. In this experiment, we investigate whether the use of label noise-resistant alignment methods could allow the use of higher-ranked rejected samples. To do this, we used the third-ranked sample as the rejected sample, and evaluated different settings for conservative DPO (c DPO) (Mitchell, 2023). We used the bio Rxiv dataset experiment setting, set eps=7, learning rate = 8e-7. We find that by tuning the level of conservative-ness we may be able to improve slightly on vanilla DPO. E.2. Alignment methods We also experiment with a noise-resistant (or robust) DPO method, conservative DPO (c DPO) (Mitchell, 2023), to see if by using it we can select a higher ranked rejected sample (recall we use the 5th ranked, and higher ranked samples would introduce more noise into the preference pairs). In Figure 7 we find that it can help slightly when choosing a higher rejected sample ranking. Private Federated Learning using Preference-Optimized Synthetic Data E.3. Rejected sample selection We construct the DPO preference data via client feedback by generating ten samples from the same prompt and then picking the selected and the rejected samples. The samples with the highest scores among the ten examples are picked as the selected sample in the DPO preference dataset. We experiment on which rank should be utilized as the rejected sample in the DPO preference dataset. In Fig 8 we further explore the effects by examining the rejected and selected sample FID scores as a function of round. In the left panel where the selected sample FID values are shown, their magnitude and trends behave similarly before they reach the best results (marked by colored dashed vertical lines). For the rejected sample FID shown in the right panel, the 5th rank rejected samples yield the lowest FID score and therefore smaller gap between the preference sample pairs. However, we also find that higher rank does not always yield better results. This may result from the boundary between the rejected and selected samples becoming undistinguishable for rank < 5th due to DP noise. We therefore select 5th rank samples as our rejected DPO preference samples. Figure 8. Ablation study for selecting rejected samples in the preference data. Here we generate 10 samples for each prompt and select Nth ranked data as the rejected sample, where N is 5, 7, or 10. The vertical lines indicate the round at which the best next-word-prediction accuracy was achieved for each choice of rank. Note that the model that produces the lowest overall FID (not the lowest selected sample FID or the lowest rejected sample FID) is the best synthetic data generation model, since on the final round all generated samples are utilized to form the synthetic dataset. We hypothesize that round 7 corresponds to the highest accuracy for the rank 5 model because after that point, the selected sample FID is higher than the rejected sample FID, which would mean the preference dataset has become mis-aligned with the objective of generating good synthetic data. bio Rxiv Congressional Speeches Figure 9. The distribution of how many tokens are in each client s dataset for the bio Rxiv and Congressional Speeches datasets. Private Federated Learning using Preference-Optimized Synthetic Data Table 4. Dataset details. Dataset # Train Samples # Validation Samples # Test Samples Max Sequence Length Average # of samples per client bio Rxiv 72000 2000 1584 64 6.6 2.6 Congressional Speeches 133000 4200 1547 64 5.0 16.3 Pub Med 75316 14423 4453 512 1 Figure 10. A t-SNE clustering of the Congressional Speeches dataset. US data is colored in purple, UK data is colored in orange, and Canada data is colored in green. We find that the three datasets form distinct clusters and also distinct sub-clusters. F. Datasets bio Rxiv. This dataset consists of abstracts from bio Rxiv papers with appropriate copyright permission from April 2023 to August 2024. This was done by using the bio Rxiv public API to retrieve the abstracts of the paper with permitted licenses (i.e. CC BY NC ND , CC BY ND , CC BY NC , CC BY , CC0 ). This dataset consists of 72k abstracts (clients), each of which we split into chunks of 64 tokens to form samples. Congressional Speeches. This dataset consists of speeches from US, UK and Canada congressional/parliamentary transcripts from April 2023 to August 2024. All speeches are published under a permissive license which allows for third-party use (as detailed in the dataset cards). There are 134k speeches (clients) in total, and 1930 unique speakers. We collected this dataset by using public APIs to retrieve data from each country s official congressional/parliamentary library website. Then we sanitized the data by removing (1) boilerplate procedural language, (2) sentences with more than 30% of the characters not being letters, and (3) some written notation that does not correspond to spoken words. We split each speech into chunks of 64 tokens each. We believe that this dataset is a major contribution because spoken language may be more resistant to contamination (especially for the UK and Canada parliamentary debates). Because they are more conversational and have a large degree of improvisation (many debates are off-the-cuff), they are less likely to be generated by LLMs. Because Congressional Speeches contains a diverse collection of speeches across speakers and also countries, the dataset forms many distinct clusters, reflecting the diversity of the dataset (Figure 10). We will update the dataset periodically with the latest data to allow future researchers to test their algorithms or ideas against an uncontaminated dataset.