# measuring_diversity_in_synthetic_datasets__381cdb25.pdf Measuring Diversity in Synthetic Datasets Yuchang Zhu 1 Huizhe Zhang 1 Bingzhe Wu 2 Jintang Li 3 4 Zibin Zheng 3 4 Peilin Zhao 5 6 Liang Chen * 1 Yatao Bian * 6 7 Large language models (LLMs) are widely adopted to generate synthetic datasets for various natural language processing (NLP) tasks, such as text classification and summarization. However, accurately measuring the diversity of these synthetic datasets an aspect crucial for robust model performance remains a significant challenge. In this paper, we introduce DCScore, a novel method for measuring synthetic dataset diversity from a classification perspective. Specifically, DCScore formulates diversity evaluation as a sample classification task, leveraging mutual relationships among samples. We further provide theoretical verification of the diversity-related axioms satisfied by DCScore, highlighting its role as a principled diversity evaluation method. Experimental results on synthetic datasets reveal that DCScore enjoys a stronger correlation with multiple diversity pseudo-truths of evaluated datasets, underscoring its effectiveness. Moreover, both empirical and theoretical evidence demonstrate that DCScore substantially reduces computational costs compared to existing methods. Code is available at: https://github.com/ bluewhalelab/dcscore. Work done during an internship at Tencent AI Lab *Corresponding author 1School of Computer Science and Engineering, Sun Yat-sen University, Guangzhou, China 2School of Artificial Intelligence, Shenzhen University, Shenzhen, China 3School of Software Engineering, Sun Yat-sen University, Zhuhai, China 4Zhuhai Key Laboratory of Trusted Large Language Models, Zhuhai, China 5School of Artificial Intelligence, Shanghai Jiao Tong University, Shanghai, China 6Tencent AI Lab, Shenzhen, China 7Department of Computer Science, National University of Singapore, Singapore. Correspondence to: Yatao Bian , Liang Chen . Proceedings of the 42 nd International Conference on Machine Learning, Vancouver, Canada. PMLR 267, 2025. Copyright 2025 by the author(s). 1. Introduction Large language models (LLMs) have shown exceptional performance across a range of fields, such as chatbots (Achiam et al., 2023), computer programming (Gu, 2023), and reasoning (Yuan et al., 2024). Inspired by their remarkable capacities, some research (Ye et al., 2022; Abdullin et al., 2024; Ding et al., 2024) employs LLMs as dataset generators to mitigate the shortage of training data. Although generated data facilitates model optimization, recent studies (Yu et al., 2024; Lee et al., 2023) suggest that a lack of diversity within the dataset measured by the variation between samples (Long et al., 2024) may lead to performance degradation in some scenarios. Although previous studies (Yu et al., 2024; Wang et al., 2022) leverage well-designed generation strategies to create highly diverse synthetic datasets, they consistently neglect to evaluate the diversity of these datasets. Additionally, a principled diversity evaluation metric serves not only to guide LLM generators in creating more diverse data but also extends its utility to data selection (Cao et al., 2023), quantifying augmentation performance (Yang et al., 2024a), and assessing mode collapse (Dan Friedman & Dieng, 2023). Thus, a diversity evaluation method for synthetic datasets is becoming increasingly important. Since the diversity evaluation of synthetic datasets remains under-explored, a natural solution is to directly employ diversity evaluation methods from related fields, such as natural language processing (NLP) (Khurana et al., 2023) and machine learning (ML) (Jordan & Mitchell, 2015). Specifically, efforts to measure diversity within these domains can be summarized into three categories: N-gram-based method (Zhu et al., 2018; Mishra et al., 2020), Referencebased method (Heusel et al., 2017; C ıfka et al., 2018), and Transformation-based method (Du & Black, 2019; Zhang et al., 2024). The n-gram-based method evaluates diversity through n-gram statistics, e.g., distinct-n (Li et al., 2015), focusing on textual form rather than semantic content. To align the diversity criteria with human judgment, the referencebased method has emerged as a promising alternative. This approach employs a reference distribution or data as an approximation of human judgment and calculates the similarity between the evaluated data and the reference data (Holtzman et al., 2019). However, collecting reference data can be both time-consuming and may introduce potential biases. Measuring Diversity in Synthetic Datasets Drawing inspiration from deep representation learning (Butepage et al., 2017; Zhang et al., 2021), the transformation-based method evaluates diversity by first mapping the data into the representation space and then performing diversity summarization (Tevet & Berant, 2020). For data mapping, various embedding functions can be employed to facilitate the transformation, with a popular approach being the use of sentence transformers such as Sentence-Bert (Reimers, 2019) and Sim CSE (Gao et al., 2021). Owing to the versatility of embedding functions, transformation-based methods can simultaneously consider extensive aspects, such as semantics and style, to encode data representations, thereby providing a more comprehensive evaluation of diversity. However, this type of method is hindered by suboptimal computational efficiency caused by the high-complexity diversity summarization, such as eigenvalue computation (Dan Friedman & Dieng, 2023). In a nutshell, existing diversity evaluation methods in NLP and ML suffer from inherent limitations in the synthetic dataset evaluation scenario. To effectively evaluate the diversity of synthetic datasets, the following challenges must be tackled: (1) Holistic Analysis. The diversity evaluation of synthetic datasets is a holistic analysis task, necessitating consideration of the impact of each sample on the final evaluation results. (2) Axiomatic Requirements. Prior research has suggested several axioms that diversity metrics should ideally satisfy. To ensure reasonable evaluation, a wellestablished diversity evaluation should exhibit properties corresponding to these axioms. (3) Lower Computational Costs. As a growing volume of synthetic datasets, a diversity evaluation method with lower computational costs is highly desirable to ensure data quality and model performance. To sum up, the essence of diversity is associated with the identification of differences between samples, and the ability to distinguish these differences is a key element in the classification process (Quine, 1969). Motivated by this observation, we propose a synthetic dataset diversity evaluation method from a classification perspective, namely, DCScore. Notably, DCScore tackles the three aforementioned challenges. Firstly, DCScore treats the evaluation of each sample in the synthetic dataset as an independent classification task, providing a holistic analysis. Secondly, theoretical verification demonstrates that DCScore satisfies four axioms outlined in (Leinster & Cobbold, 2012), including effective number, identical samples, symmetry, and monotonicity. Lastly, both empirical and theoretical evidence suggest that DCScore effectively evaluates the diversity of synthetic datasets, while demonstrating lower computational costs compared to existing methods. Our contributions can be summarized as follows: We propose DCScore, a classification-based diversity evaluation method for synthetic datasets. The core idea behind DCScore is to treat diversity evaluation as a sample classification task, enabling the capture of mutual relationships among samples. We theoretically validate that DCScore adheres to several intuitive axioms suggested by (Leinster & Cobbold, 2012), demonstrating its superiority. Additionally, we theoretically validate the lower complexity of DCScore under general kernels. Extensive experiments show that DCScore exhibits a stronger correlation with multiple diversity pseudotruths compared to baseline metrics. We also perform a computational cost experiment to confirm the lower computational cost of DCScore. 2. Related Work We give a brief literature review of diversity evaluation methods. Limited by space, further related works on LLM dataset generators and application of diversity evaluation methods can be found in Appendix A. 2.1. Diversity Evaluation Methods With the development of LLMs as dataset generators, the diversity evaluation of synthetic datasets has become a challenging task and remains under-explored in recent evaluation studies (Liang et al., 2022; tatsu lab, 2023; Ribeiro et al., 2020). The most comparable diversity evaluation research can be traced back to studies in NLP and ML, which can be summarized into the n-gram-based method (Mishra et al., 2020), reference-based method (Heusel et al., 2017), and transformation-based method (Lai et al., 2020). N-gram-based Methods. The n-gram-based method is the most popular lexical diversity evaluation method, leveraging n-grams to capture differences in sentence form (Yu et al., 2024). Commonly used n-gram-based diversity metrics include distinct n-grams (distinct-n) (Song et al., 2024), self BLEU (Shu et al., 2019), and ROUGE-L (Wang et al., 2022; Padmakumar & He, 2023). However, this type of method only captures differences in text form, thereby overlooking differences in other aspects such as semantics and style. Reference-based Methods. Diversity evaluation is a subjective task, leading to a reliance on human judgment. Consequently, the reference-based method evaluates diversity by comparing the distribution of the evaluated data to that of a reference dataset (Heusel et al., 2017). MAUVE (Pillutla et al., 2021) exemplifies this idea by employing a divergencebased metric to capture correlations with human judgment. Regarding the natural language inference (NLI) training set as the reference dataset, (Stasaski & Hearst, 2022) first trains an NLI model to infer the relationship between pairs of generated texts and then calculates diversity based on Measuring Diversity in Synthetic Datasets these inference results. Due to the challenges in collecting reference datasets, recent studies (Kynk a anniemi et al., 2019; Le Bronnec et al., 2024) propose evaluating diversity through precision and recall. (Naeem et al., 2020) introduces density and coverage as solutions to the susceptibility of precision and recall to outliers. Despite these advancements, the reference-based method remains significantly constrained because of the need for reference datasets. Transformation-based Methods. The transformationbased (Lee et al., 2023) method leverages well-designed models to generate representations of the evaluated data. Then, the diversity of these representations is summarized using techniques such as clustering (Du & Black, 2019) and eigenvalue computation (Dan Friedman & Dieng, 2023) of Vendi Score (Dan Friedman & Dieng, 2023). In line with Vendi Score (Dan Friedman & Dieng, 2023), RKE (Jalali et al., 2023) introduces an information-theoretic method for evaluating diversity in multimodal distributions, while FKEA (Ospanov et al., 2024) enhances the computational efficiency of both Vendi Score and RKE. Owing to the superior performance of representation learning, this type of method considers various aspects of the evaluated data, including semantics, form, and style, offering greater flexibility compared to the other two methods. However, its dependence on high-complexity summarization techniques, such as eigenvalue computation, limits its scalability in evaluating the diversity of synthetic datasets. In summary, existing methods primarily focus on NLP and ML fields and are challenging to apply directly to synthetic dataset diversity evaluation. Different from the above-mentioned studies, our work is dedicated to the holistic diversity evaluation of synthetic datasets. Additionally, to ensure flexible evaluation, our work aims to evaluate diversity-sensitive components that impact the performance of trained models in terms of diversity. 3. Preliminaries 3.1. LLM as a Dataset Generator Since the exceptional performance of LLMs, previous works (Dai et al., 2023; Yoo et al., 2021) employ LLMs as a dataset generator or for data augmentation purposes. LLMs significantly reduce the cost of label annotation and data collection (Tan et al., 2024), and in several tasks, even outperform human annotators (Gilardi et al., 2023). While some studies attempt to use LLMs to generate datasets from scratch, it is a challenging task for LLMs. In most cases, a pre-trained LLM, denoted as M, takes the data Dsup to be augmented and the generation task T as input, and outputs the augmented dataset D. Formally, this process can be formulated as follows: D M(T, Dsup) (1) Table 1. Categories of Dsup. In the column of Examples , texts belonging to Dsup are highlighted in gray , while texts associated with T are marked using an underline. Generations Dsup Examples {xi} {yi} {xi} Question: It s a boring movie. The sentiment of the movie review is Answer: Negative. {yi} {xi} {yi} Question: The movie review in positive sentiment is Answer: Good film! {Tseed} {Ti} {Tseed} Question: Following are examples of movie review and their sentiment labels. Generate samples according to these examples. Example 1: A boring movie. (Negative); Example 2: Oh, wonderful movie! (Positive). Answer: A meaningful movie. (Positive) where T can be texts describing the generation task, such as annotation. Dsup, which comprises a small number of seed samples or unlabeled inputs, serves as supplementary materials to facilitate data augmentation. For example, we want LLMs to perform an annotation task for sentiment labeling, such as determining whether the sentiment is positive or negative. If we assume Dsup to be It s a boring movie. , the description of T could be The sentiment of the movie review is . D = {Ti}n i=1 = {(xi, yi)}n i=1 is the generated dataset with n samples, where Ti = (xi, yi), xi, and yi are the input-output sample, the input text, and the output text, respectively. Let TD denote the downstream task of D, when TD is the question-answering task, xi and yi represent the question and answer, respectively. It is worth noting that not all components in D are generated by M, which is related to the category of Dsup. As shown in Table 1, Dsup can be divided into three categories, namely input text, output text, and seed samples. In Table 1, and Tseed represent the direction of generation and seed samples, respectively. For example, xi yi signifies that, given input text xi denoted as Dsup, M processes Dsup and T, generating the output text yi. 3.2. Problem Formulation The diversity evaluation of the dataset is a sample richness evaluation problem. Based on the generation scenarios presented in Table 1, we find that in certain downstream tasks, the diversity of some components in the synthetic dataset does not influence the performance of the trained models. Conversely, the diversity of other components significantly impacts model performance. We refer to these components, whose diversity influences performance, as diversity-sensitive components, denoted as Ti. For example, the input text xi in sentiment classification tasks is the diversity-sensitive component. Conversely, the output text yi, which represents the sentiment label of the sample and is typically a numerical label (e.g., 0 or 1), does not influence model performance in terms of diversity. Thus, the output text cannot be considered as the diversity-sensitive Measuring Diversity in Synthetic Datasets component. It should be underscored that diversity-sensitive components vary across downstream tasks. The diversity evaluation of synthetic datasets can be transformed into the diversity evaluation of diversity-sensitive components. Given a synthetic dataset D = {Ti}n i=1, we define { Ti}n i=1 as a collection of diversity-sensitive components. The problem of diversity evaluation of D can be defined as follows: Diversity Score Eval({ Ti}n i=1) (2) where Eval( ) is the diversity evaluation function, which takes { Ti}n i=1 as inputs and outputs the diversity score. 4. Present Work In this section, we first introduce our proposed method DCScore from a classification perspective. Then, we present the properties of DCScore followed by theoretical proofs. Finally, we provide a detailed complexity analysis of DCScore and the transformation-based counterpart. 4.1. DCScore: Measuring Diversity from a Classification Perspective Due to the intrinsic nature of measuring sample differences in diversity evaluation, it is natural to evaluate diversity as a classification task. Consequently, we propose DCScore, which formulates diversity evaluation as a sample classification task. Specifically, the difference between samples can be measured through an n-classification task, where evaluating n sample datasets involves n n-classification tasks, with each sample corresponding to a distinct category. As shown in Figure 1, DCScore consists of three stages: text representation, pairwise similarity, and diversity summarization. According to the problem formulation in section 3.2, DCScore outputs the diversity of synthetic datasets by evaluating diversity-sensitive components. Let D = {Ti}n i=1 = {(xi, yi)}n i=1 denote a synthetic dataset comprising n input-output samples, and { Ti}n i=1 represents the diversity-sensitive components. DCScore adheres to the paradigm of the transformation-based method to evaluate the diversity of D. Specifically, given Ti, DCScore first applies an embedding function Φ to extract the sample representation hi = Φ( Ti). For all samples in D, we obtain the sample representation matrix H Rn d across all samples, where d denotes the dimension of sample representations. Subsequently, DCScore utilizes a kernel function Kernel( ) to calculate a kernel matrix K, where K Rn n and entry K[i, j] represents similarity between Ti and Tj. From a classification perspective, K[i, j] can be considered as the logit of Ti being classified into category cj. where cj corresponds to Tj. Thus, a sample that is highly distinguishable from others (i.e., correctly classified into its own category with high probability) contributes more to the overall diversity score. Formally, the aforementioned process can be formulated as follows: H = Φ({ Ti}n i=1), K = Kernel(H), (3) where Kernel( ) calculates pairwise similarity, with viable options including inner product and RBF kernel. For Φ, a more expressive embedding function can be employed, such as one trained using a well-designed framework like Sentence-Bert (Reimers, 2019). Based on K, DCScore leverages a classification function with K, denoted as f K( ), to compute the classification probability matrix P Rn n. Here, a natural option for f K( ) is the Softmax function. For Ti, the probability that Ti is classified as category cj can be formulated as follows: P(c = cj| Ti) = P[i, j] = f K(K[i, j]) = exp (K[i, j]/τ) P j exp (K[i, j]/τ), (4) where τ is a temperature hyperparameter to control the classification resolution. A smaller τ amplifies sample similarity differences, implying a higher classification resolution, while a larger value yields the opposite effect. According to Eq. (4), if the evaluated dataset exhibits high sample richness, indicating greater diversity, each sample is likely to be classified into its own category. If the diversity is low, all samples will be randomly classified. Based on P, DCScore calculates diversity of D as the trace of P, which can be formulated as follows: Definition 4.1 (DCScore). Let D = {Ti}n i=1 denote the synthetic dataset with n samples, and let { Ti}n i=1 represent a set of diversity-sensitive components within {Ti}n i=1. Denote Pi as the classification probability vector of Ti. By conducting the classification task for all Ti and obtaining the probability matrix P = [P1, P2, ..., Pn], DCScore for D is defined as the trace of P: DCScore(D) = tr(P) = i=1 P[i, i]. (5) Notably, the process described above is just one implementation of DCScore, with other potential implementations to be explored in future work. 4.2. Properties of DCScore We provide theoretical proof that DCScore satisfies several axioms (Leinster & Cobbold, 2012) defined for a principled diversity metric. DCScore meets four axioms: effective number, identical samples, symmetry, and monotonicity axioms. These axioms ensure a reasonable and robust diversity evaluation. The matched axioms of DCScore are outlined below, while their proofs are detailed in Appendix B. Measuring Diversity in Synthetic Datasets Probability Matrix 𝑷 𝑓𝑲 Embedder Diversity-sensitive 𝐷𝐶𝑆𝑐𝑜𝑟𝑒𝒟 Text Representation Pairwise Similarity Diversity Summarization Figure 1. The overview of DCScore. DCScore consists of text representation, pairwise similarity, and diversity summarization stages. Effective number: Diversity should be defined as the effective number of samples in a dataset, ranging from 1 to n. DCScore meets this axiom, as evidenced by its behavior: DCScore equals 1 when all samples in D are identical and equals n when all samples are distinct. Identical samples: Given two identical datasets D1 and D2, the diversity of D generated by merging these two datasets remains unchanged. The values of DCScore are the same across D1, D2, and D , i.e., DCScore(D1) = DCScore(D2) = DCScore(D ). (6) Symmetry: Diversity remains constant regardless of the order of the samples, exhibiting permutation invariance. Let π( ) denote the permutation function for the sample order, DCScore remains unchanged for any sample permutation of D, i.e., DCScore(D) = DCScore(π(D)). (7) Monotonicity: The diversity of a dataset decreases as the sample similarity increases. Given two datasets D1 and D2, and a new sample Tn+1, where the samples in D1 and D2 are entirely different, and DCScore(D1) = DCScore(D2) = n. If Tn+1 is more similar to the samples in D2 than to those in D1 and is added to both datasets, then for the merged datasets D 1 and D 2, DCScore satisfies the following equation. DCScore(D 1) > DCScore(D 2). (8) 4.3. Complexity Analysis We provide a time complexity analysis of DCScore in Table 2. We compare DCScore with Vendi Score due to their similarity, finding that DCScore has lower computational complexity with non-linear kernels. Denoting Okernel as the complexity associated with general kernels (i.e., kernels other than linear kernels), we analyze the complexity in the pairwise similarity and summarization Table 2. Complexity analysis of DCScore and Vendi Score. Okernel represents the complexity of the kernel function. General Kernels Inner Product Pairwise Similarity Vendi Score O(n2 Okernel) O(d2n) DCScore O(n2d) Summarization Vendi Score O(n3) O(d3) DCScore O(n2) O(n2) Total Vendi Score O(n2 Okernel + n3) O(d2n + d3) = O(d2n) DCScore O(n2 Okernel + n2) O(n2d + n2) stages. In the pairwise similarity stage, the computation of pairwise similarities results in a complexity of O(n2) for DCScore. When combined with the complexity Okernel of the general kernel computation, DCScore exhibits a total complexity of O(n2 Okernel). In the summarization stage, DCScore has a complexity of O(n2) due to the softmax operation. Consequently, the overall complexity of DCScore for general kernels is O(n2 Okernel + n2). In contrast, Vendi Score has a total complexity of O(n2 Okernel + n3), where the pairwise similarity stage is identical to that of DCScore, while the summarization stage incurs a complexity of O(n3) due to the eigenvalue computation. Thus, for general kernels, DCScore demonstrates lower complexity than Vendi Score. However, when the inner product is employed as the kernel function and n d, Vendi Score can significantly reduce the complexity by replacing the pairwise similarity XXT with XT X, where X Rn d. This results in complexities of O(d2n) for the pairwise similarity stage and O(d3) for the summarization stage. In this regard, DCScore has a complexity of O(n2d + n2), which is slightly worse than that of Vendi Score. We can leverage efficient techniques, such as those proposed in (Shim et al., 2017), (Milakov & Gimelshein, 2018), and (Wen et al., 2023), to reduce the computational cost of DCScore. Compared to Vendi Score, DCScore maintains lower complexity in most cases, as empirically validated in Section 5.3 and Appendix D.1. While Vendi Score has lower complexity with the inner product kernel, experiments in Appendix D.1 indicate that its computation time is similar to DCScore. Ensuring low Measuring Diversity in Synthetic Datasets Evaluated Dataset 𝒟 Diversity Evaluator Correlation Diversity Score Diversity Pseudo-truth Figure 2. Experimental settings of correlation evaluation. computational complexity across multiple kernels is more advantageous than achieving it with a single kernel. 5. Experiments We conduct experiments to verify the effectiveness of DCScore by examining correlation, computational cost, hyperparameter sensitivity, and further probing. Limited by space, we provide additional experiments in Appendix D. 5.1. Experimental Settings To verify the effectiveness of DCScore, we conduct a series of correlation experiments following the setup in (Tevet & Berant, 2020). As shown in Figure 2, the core idea of our experimental evaluation is to correlate the diversity measurement results of DCScore with diversity pseudo-truths, such as the Softmax temperature τg of dataset generation, human judgment, and LLMs evaluation. Specifically, we evaluate l generated datasets to obtain l diversity scores and then calculate the correlation with diversity pseudo-truths. To calculate the correlation between measured diversity scores and diversity pseudo-truths, we employ Spearman s ρ (Spearman, 1961), a measure of rank correlation ranging from -1 to 1, with higher absolute values indicating stronger correlations. Limited by space, we present detailed experimental settings in Appendix C. Datasets. We utilize two categories of datasets in our experiments: self-generated datasets and publicly available generated datasets. Self-generated datasets are generated through two data generation strategies (Li et al., 2023): zeroshot and few-shot settings. We generate datasets for two natural language processing tasks: text classification and story completion. Additionally, we utilize three publicly available existing datasets, including SST2 (Socher et al., 2013), Yelp (Zhang et al., 2015), and AG News (Zhang et al., 2015), and their Attr Prompt-augmented version (Yu et al., 2024). Detailed information about these datasets can be found in Appendix C.1. Generation Models. To generate datasets through zeroshot and few-shot settings, we employ two commonly used LLMs as our dataset generators, including Llama2-13B (13B) and Llama2-70B (70B) (Touvron et al., 2023). Table 3. Correlation (Spearman s ρ) between τg and diversity evaluation methods on datasets generated by different settings (Zeroshot or Few-shot). Spearman s ρ varies between -1 and +1, with 0 implying no correlation. Best results are indicated in bold. Zero-shot setting Few-shot setting Text classification Story completion Text classification Story completion 13B 70B 13B 70B 13B 70B 13B 70B Distinct-n 0.9909 0.9870 0.9766 0.9701 0.9857 0.9766 0.9779 0.9935 K-means Inertia -0.1143 0.9688 0.9454 0.8727 0.7104 0.7273 0.9662 0.9662 Vendi Score 0.9961 0.9818 0.9870 0.9922 0.9909 0.9857 0.9857 0.9961 DCScore 0.9961 0.9779 0.9844 0.9792 0.9909 0.9883 0.9857 0.9974 Baseline Methods. We compare DCScore with three baseline methods detailed in Appendix E, i.e., Distinct-n (Li et al., 2015), K-means Inertia (Du & Black, 2019), and Vendi Score (Dan Friedman & Dieng, 2023). 5.2. Correlation Evaluation We investigate the correlation between the diversity evaluation of DCScore and diversity pseudo-truths, such as τg, human judgment, and LLMs evaluation. We compare DCScore with all baselines on self-generated datasets. 5.2.1. CORRELATION WITH τg Evaluation on self-generated datasets. Previous works (Caccia et al., 2018; Tevet & Berant, 2020; Chung et al., 2023) have demonstrated a positive correlation between τg and the diversity of generated texts, making τg as a reasonable diversity pseudo-truth. LLMs with lower τg generate less diverse content, whereas higher τg values yield more diverse content. Thus, we evaluate the performance of DCScore on self-generated datasets with varying τg, ranging from 0.2 to 1.2 at 0.05 intervals. We present more information about self-generated datasets in Appendix C.1.1. Table 3 displays the correlation results of all methods. All methods accurately capture the true diversity of generated datasets, as demonstrated by high Spearman s ρ values. DCScore performs on par with Vendi Score while providing better scalability for larger synthetic datasets, as discussed in Section 5.3. DCScore outperforms all baseline methods under the few-shot setting across all datasets, highlighting its effectiveness. K-means Inertia exhibits the weakest correlation on the text classification dataset generated by the 13B model under the zero-shot setting, potentially due to its sensitivity to the number of cluster centroids. Overall, DCScore outperforms all baselines in most cases, and its evaluation results exhibit a strong correlation with the diversity pseudo-truth according to (Akoglu, 2018). Visualization. We further provide a visualization of the diversity evaluation results for DCScore. For each generated dataset, we prompt LLMs to produce 10 distinct answers corresponding to a single prompt, forming an evaluation batch. We then evaluate diversity using the batch evaluation protocol outlined in Appendix C.2.3. Measuring Diversity in Synthetic Datasets 0.25 0.50 0.75 1.00 1.25 Temperature ( g) Text Classification Few-shot_13B Zero-shot_13B Few-shot_70B Zero-shot_70B 0.25 0.50 0.75 1.00 1.25 Temperature ( g) Story Completion Few-shot_13B Zero-shot_13B Few-shot_70B Zero-shot_70B Figure 3. Diversity evaluation of DCScore on datasets generated using varying τg. DCScore shows a strong correlation with τg, indicating its effectiveness in evaluating the dataset diversity. Based on the above-mentioned settings, a completely diverse dataset may yield a diversity score of 10. As shown in Figure 3, DCScore exhibits a strong positive correlation with τg. In most cases, when τg > 0.75, DCScore scores a generated dataset with a diversity value of approximately 10. For text classification, the 13B generation model under the few-shot setting demonstrates a distinct diversity change pattern compared to others. This phenomenon stems from the inherent limitations of the 13B generation model, which struggles to effectively comprehend and execute more intricate or multi-faceted instructions. As a result, generated datasets only exhibit marginal improvements in diversity. 5.2.2. CORRELATION WITH HUMAN JUDGMENT Diversity evaluation is a subjective task, and an ideal method should align well with human judgment. Thus, we investigate the correlation between DCScore and human judgment. We enlist three human evaluators to perform pairwise diversity comparisons among datasets with varying τg values and report the diversity ranking by averaging the win rate across evaluators. We conduct all evaluations five times to report average results. Limited by space, we present details of human evaluation in Appendix C.2.5. Table 4 presents pairwise correlation between human judgment, τg, and DCScore. Table 4 indicates a strong correlation between human judgment and τg, supporting the use of human judgment as a diversity pseudo-truth. Based on this observation, DCScore performs better in two settings: Story-Few (story completion data generated under the few-shot setting) and Text-Zero (text classification data generated under the zero-shot setting). This is confirmed by higher human-DCScore correlation in these two settings. For Story-Zero and Text-Few settings, we observe more identical content in the initial portions of the diversitysensitive components within an evaluation batch. In these cases, human evaluators tend to disregard the identical content and base their judgments on the latter sections. However, DCScore is affected by the identical content, resulting in a lower pairwise correlation. Despite this, the correlation remains strong, as demonstrated by previous studies (Akoglu, 2018). Table 4. Pairwise correlation (Spearman s ρ) between human, temperature (τg), and DCScore. DCScore indicates a strong correlation with human judgment. Story-Few Story-Zero Text-Few Text-Zero Human-DCScore 0.9040 0.04 0.7870 0.10 0.7915 0.16 0.8798 0.10 τg-DCScore 0.9086 0.07 0.7829 0.16 0.8400 0.16 0.8971 0.07 τg-Human 0.9276 0.02 0.9194 0.06 0.9770 0.02 0.9255 0.08 Table 5. Pairwise correlation (Spearman s ρ) between GPT-4, temperature (τg), and DCScore. DCScore indicates a strong correlation with GPT-4 evaluation results. Story-Few Story-Zero Text-Few Text-Zero GPT-4-DCScore 0.6057 0.30 0.9010 0.04 0.6131 0.18 0.9052 0.09 τg-DCScore 0.6757 0.30 0.8782 0.08 0.5714 0.27 0.9336 0.06 τg-GPT-4 0.9086 0.07 0.7829 0.16 0.8400 0.16 0.8971 0.07 5.2.3. CORRELATION WITH LLM EVALUATOR To further verify the effectiveness of DCScore, we investigate the evaluation correlation between DCScore and LLMs. Following the setting in Section 5.2.2, we employ GPT-4 to conduct pairwise comparisons between two generated datasets with different τg. These generated datasets are identical to those used in Section 5.2.2. Based on the pairwise comparison results, we obtain the diversity ranking outcomes. Regarding GPT-4 evaluation results as the diversity pseudo-truth, we report the pairwise evaluation correlation between DCScore, GPT-4, and τg in Table 5. We observe that DCScore exhibits strong correlations with GPT-4 and τg in zero-shot settings. By comparing the results of τg-DCScore and τg-GPT-4 , we find that DCScore outperforms the GPT-4 evaluator in terms of correlation with τg in zero-shot settings. Regarding the correlation performance in few-shot settings, we notice lower correlations of all baseline methods compared to zero-shot settings. We guess that this phenomenon is related to the distributions of the generated datasets. Although DCScore exhibits lower correlations (about 0.6) with GPT-4, this result can still be considered a strong correlation according to (Akoglu, 2018). In summary, DCScore exhibits a strong correlation (Spearman s ρ 0.6), with three diversity pseudo-truths: τg, human judgment, and LLMs evaluation, thereby verifying the effectiveness of DCScore. 5.3. Computational Cost The computational cost is crucial in diversity evaluation methods, especially with the increasing sample sizes of synthetic datasets. For a fair comparison, we only present the computation times of transformation-based methods: DCScore, K-means Inertia, and Vendi Score. We truncate the text length of three datasets (SST2/Yelp/AG News Attr Prompt, the Attr Prompt (Yu et al., 2024) augmented version of SST2/Yelp/AG News datasets) to 50 tokens and Measuring Diversity in Synthetic Datasets 0 2000 4000 Sample Number SST2-Attr Prompt K-means Inertia Vendi Score DCScore 0 2000 4000 Sample Number Yelp-Attr Prompt K-means Inertia Vendi Score DCScore 0 2000 4000 Sample Number AG News-Attr Prompt K-means Inertia Vendi Score DCScore Figure 4. Computation times under different sample sizes. DCScore outperforms all baselines in computational cost. Table 6. Comparison of computation time between DCScore and Vendi Score on SST2. Kernels SST2 Sample num 4k 8k 16k 32k 64k Inner product Vendi Score 4.65 0.28 9.84 0.26 19.02 0.70 37.31 1.88 76.19 1.91 DCScore 4.58 0.29 10.03 0.17 20.42 0.39 42.91 1.59 112.47 2.43 RBF kernel Vendi Score 5.86 0.06 12.41 0.49 32.94 0.40 100.36 1.44 449.14 10.35 DCScore 5.22 0.33 9.94 0.42 21.20 0.75 46.57 1.47 117.06 1.91 Poly kernel Vendi Score 5.73 0.06 12.72 0.41 31.47 0.97 98.31 0.25 453.11 2.53 DCScore 5.09 0.28 10.27 0.12 20.12 1.02 46.25 1.82 123.51 3.40 record the computation times of three methods with varying sample sizes in the range of {100, 500, 1000, 2000, 4000}. As shown in Figure 4, we repeat the experiments five times to report the final results. DCScore and K-means Inertia exhibit nearly identical computation times. However, DCScore significantly outperforms K-means Inertia in correlation with τg, as evidenced in Section 5.2.1. Compared to Vendi Score, DCScore demonstrates a speed advantage of approximately 16%, or more than one second, when processing 4000 samples. Analyzing the time complexity of these two methods, and disregarding the selection of the kernel function, we find that for a dataset with n samples, where n d is not satisfied, the computational complexity of DCScore in diversity summarization is O(n2) due to the softmax computation. In contrast, Vendi Score requires finding the eigenvalues of an n n matrix, resulting in a computational complexity of O(n3). Consequently, DCScore offers significantly lower time complexity than Vendi Score while sacrificing little in diversity evaluation performance. However, as detailed in the complexity analysis shown in Section 4.3, when n d and inner products are used as the kernel function, the total complexity of Vendi Score can be reduced to O(d2n). Thus, we evaluate computational costs on larger datasets, i.e., satisfying n d. As shown in Table 6, DCScore exhibits a notable advantage in computational efficiency when using non-linear kernels, e.g., RBF and Poly kernel. We present additional experimental results and a more in-depth analysis in Appendix D.1. A recent study (Ospanov et al., 2024) employs the random Fourier features framework to decrease the computational cost of entropy-based diversity evaluation methods, such as Vendi Score (Dan Friedman & Dieng, 2023) and RKE (Jalali et al., 2023). We follow the experimental settings of Table 6 and leverage different random seeds for data sampling. Limited by space, we present the experimental results in Text Classification Few-shot_13B Zero-shot_13B Few-shot_70B Zero-shot_70B Story Completion Few-shot_13B Zero-shot_13B Few-shot_70B Zero-shot_70B Figure 5. Hyperparameter sensitivity analysis w.r.t τ on selfgenerated datasets. Appendix D.2. In summary, DCScore demonstrates lower computation time compared to the efficiency-improved versions of Vendi Score and RKE in most cases. 5.4. Hyperparameter Sensitivity According to Eq. (4), the temperature (τ) in the Softmax function is a critical hyperparameter that affects classification resolution. To investigate this, we conduct a hyperparameter sensitivity analysis of DCScore w.r.t. τ on self-generated datasets used in Section 5.2.1. We vary τ within the range of {0.0001, 0.001, 0.1, 0.5, 1, 10}. Figure 5 presents hyperparameter sensitivity results on datasets for text classification and story completion tasks. Overall, lower τ values result in lower Spearman s ρ, even indicating a negative correlation, while higher τ values do the opposite. From Eq. (4), a higher τ reduces pairwise similarity differences, leading to a more uniform distribution of classification probabilities for each sample. This phenomenon can be regarded as a lower classification resolution, i.e., the classification function f K has poorer discrimination power. Furthermore, the correlation result of the 13B generation model under the few-shot setting for the text classification task remains stable despite variations in τ. This phenomenon has the same explanation as in Figure 3. 5.5. Further Probe 5.5.1. DOWNSTREAM TASK TRAINING To investigate the correlation between DCScore and downstream task training, we train text classification models using self-generated datasets under zero-shot and few-shot settings. We vary τg of self-generated datasets within the range of {0.2, 0.7, 1.2}. More details of training datasets and hyperparameters are presented in Appendix C.1.1 and C.2.2, respectively. As shown in Table 7, models trained on more diverse datasets achieve better accuracy, likely due to their improved generalization capabilities (Gontijo-Lopes et al., 2020). Notably, the increased training data diversity makes model fitting more difficult, necessitating additional epochs to achieve optimal accuracy. Additionally, the diversity evaluated by DCScore has a similar trend to the accuracy Measuring Diversity in Synthetic Datasets Table 7. Downstream task training performance and diversity evaluation on self-generated datasets with τg = {0.2, 0.7, 1.2}. Accuracy DCScore τg=0.2 τg=0.7 τg=1.2 τg=0.2 τg=0.7 τg=1.2 Zero-shot 89.10 89.70 90.37 481.76 1745.42 2082.42 Few-shot 70.07 73.19 73.41 1376.43 1958.16 2047.90 2 4 6 8 10 Label Num. Normalized Diversity Inception V3 Vendi Score DCScore 2 4 6 8 10 Label Num. Normalized Diversity Vendi Score DCScore Figure 6. Diversity evaluation on colored MNIST using two different embedding functions (Inception V3 and Dino V2). A higher label number signifies greater dataset diversity. performance in the zero-shot setting, further demonstrating the effectiveness of DCScore. Detailed experimental results and further analysis are provided in Appendix D.3. 5.5.2. DIVERSITY EVALUATION ON VISUAL MODALITY Similar to text generation, there is increasing attention on image generation. Thus, we further verify the effectiveness of DCScore on the image dataset evaluation. Specifically, following the setting of a recent study (Ospanov et al., 2024), we leverage the colored MNIST (Deng, 2012) as the evaluated dataset and use the label number as the diversity ground truth. Here, a larger label number indicates a higher diversity of the evaluated dataset. We conduct a comparison between DCScore and Vendi Score, using inner product as the kernel function, while adopting Inception V3 and Dino V2 as the embedding functions. For DCScore, τ is set to 1. Figure 6 illustrates the diversity evaluation results on the image dataset, with diversity results normalized by dividing by the highest result of each method. We observe that both DCScore and Vendi Score exhibit a positive correlation with the number of labels. When using the Inception V3 model as the embedding function, DCScore demonstrates higher correlations with the number of labels compared to Vendi Score, as evidenced by a line that is closer to the diagonal. Additionally, DCScore presents more consistent results across different embedding functions compared to Vendi Score, indicating stable performance with respect to the embedding function. 5.5.3. DIVERSITY EVALUATION ON DATASETS WITH DUPLICATE SAMPLES To further investigate the limitations of DCScore, we conduct diversity evaluations on datasets with duplicate sam- Table 8. Comparison of evaluation stability between DCScore and Vendi Score on datasets with duplicate samples. n denotes the number of duplicate samples, and A.P. represents Attr Prompt. Datasets Methods n=0 n=10 n=100 n=1000 n=3000 n=6000 SST2 A.P. DCScore 5937.00 5936.99 5937.10 5936.81 5936.73 5936.99 Vendi Score 11.31 11.31 11.31 11.32 11.29 11.31 Yelp A.P. DCScore 5939.94 5940.01 5939.82 5939.91 5940.21 5939.94 Vendi Score 8.22 8.22 8.22 8.24 8.24 8.22 AG News A.P. DCScore 5998.04 5998.04 5998.04 5998.04 5997.75 5998.04 Vendi Score 40.71 40.70 40.70 40.63 40.69 40.71 ples. Specifically, we randomly select n samples from the evaluated datasets to serve as duplicates and incorporate them into the original datasets. In our experiments, we use SST2-Attr Prompt, Yelp-Attr Prompt, and AG News-Attr Prompt as evaluated datasets, setting n to {0, 10, 100, 1000, 3000, 6000}. Table 8 presents the diversity evaluation values of DCScore and Vendi Score, using inner product as the kernel function. These values can be interpreted as the number of effective samples. As shown in Table 8, we observe that both DCScore and Vendi Score maintain stable evaluation results as the number of duplicate samples increases. This indicates that both methods are robust in terms of the impact of duplicates. Additionally, the results for Vendi Score tend to be lower, suggesting an underestimation of dataset diversity. 6. Conclusion In this work, we investigate the diversity evaluation of synthetic datasets, a topic systematically under-explored in existing research. To this end, we present DCScore, a diversity evaluation method from a classification perspective. DCScore regards the holistic diversity evaluation as the classification task at the sample level, thereby facilitating the capture of mutual relationships between samples. We provide theoretical guarantees demonstrating that DCScore meets the axiom requirements (Leinster & Cobbold, 2012) for a principled diversity evaluation method. Experiments on synthetic datasets reveal that DCScore exhibits better correlations with various diversity pseudo-truths, including τg, human judgment, and LLMs evaluation. Meanwhile, DCScore exhibits significantly lower computational cost compared to transformation-based counterparts. Finally, we hope our work encourages future research to pay more attention to the diversity of synthetic datasets and promotes the wider application of these datasets. Limitation: Although we have verified the effectiveness of DCScore for unimodal data evaluation, DCScore is not yet capable of directly evaluating multimodal data, such as image-text pairs. A key challenge lies in the extraction and fusion of representations from multimodal data. As multimodal LLMs continue to enhance their generative capabilities, evaluating the diversity of multimodal-generated datasets emerges as an important research problem. Measuring Diversity in Synthetic Datasets Acknowledgements The research is supported by the National Key R&D Program of China under grant No. 2022YFF0902500, the Guangdong Basic and Applied Basic Research Foundation, China (No. 2023A1515011050), and Tencent AI Lab RBFR2024004. Impact Statement This paper presents a novel and efficient method for measuring the diversity of synthetic datasets, aiming to advance fields of machine learning and LLMs. The ability to accurately evaluate the diversity of synthetic data has significant implications for various applications, including but not limited to improving the robustness and fairness of ML models, enhancing the quality of synthetic data used in training, and fostering innovation in generative models. From an ethical perspective, our work contributes to the responsible development of AI by providing tools that can help identify and mitigate biases in synthetic data. Ensuring diversity in datasets is crucial for creating equitable AI systems that perform well across different demographic groups and use cases. By promoting diversity, our method can help prevent the perpetuation of existing biases and support the creation of more inclusive technologies. For societal impact, our approach has the potential to impact a wide range of industries, from healthcare to finance, by enabling the generation of more representative and diverse datasets. This can lead to more accurate and fair decisionmaking processes, ultimately benefiting society as a whole. Overall, while our primary goal is to advance fields of machine learning and LLMs, we believe that our work also addresses important ethical considerations and has the potential to contribute positively to society by promoting fairness and inclusivity in AI systems. Abdullin, Y., Molla-Aliod, D., Ofoghi, B., Yearwood, J., and Li, Q. Synthetic dialogue dataset generation using llm agents. ar Xiv preprint ar Xiv:2401.17461, 2024. Achiam, J., Adler, S., Agarwal, S., Ahmad, L., Akkaya, I., Aleman, F. L., Almeida, D., Altenschmidt, J., Altman, S., Anadkat, S., et al. Gpt-4 technical report. ar Xiv preprint ar Xiv:2303.08774, 2023. Akoglu, H. User s guide to correlation coefficients. Turkish journal of emergency medicine, 18(3):91 93, 2018. Butepage, J., Black, M. J., Kragic, D., and Kjellstrom, H. Deep representation learning for human motion prediction and classification. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 6158 6166, 2017. Caccia, M., Caccia, L., Fedus, W., Larochelle, H., Pineau, J., and Charlin, L. Language gans falling short. ar Xiv preprint ar Xiv:1811.02549, 2018. Cao, Y., Kang, Y., Wang, C., and Sun, L. Instruction mining: Instruction data selection for tuning large language models. ar Xiv preprint ar Xiv:2307.06290, 2023. Chen, D., Lee, C., Lu, Y., Rosati, D., and Yu, Z. Mixture of soft prompts for controllable data generation. ar Xiv preprint ar Xiv:2303.01580, 2023. Chung, J. J. Y., Kamar, E., and Amershi, S. Increasing diversity while maintaining accuracy: Text data generation with large language models and human interventions. ar Xiv preprint ar Xiv:2306.04140, 2023. C ıfka, O., Severyn, A., Alfonseca, E., and Filippova, K. Eval all, trust a few, do wrong to none: Comparing sentence generation models. ar Xiv preprint ar Xiv:1804.07972, 2018. Cubuk, E. D., Zoph, B., Shlens, J., and Le, Q. V. Randaugment: Practical data augmentation with no separate search. ar Xiv preprint ar Xiv:1909.13719, 2(4):7, 2019. Dai, H., Liu, Z., Liao, W., Huang, X., Cao, Y., Wu, Z., Zhao, L., Xu, S., Liu, W., Liu, N., et al. Auggpt: Leveraging chatgpt for text data augmentation. ar Xiv preprint ar Xiv:2302.13007, 2023. Dan Friedman, D. and Dieng, A. B. The vendi score: A diversity evaluation metric for machine learning. Transactions on machine learning research, 2023. Deng, L. The mnist database of handwritten digit images for machine learning research [best of the web]. IEEE signal processing magazine, 29(6):141 142, 2012. Dieng, A. B., Ruiz, F. J., Blei, D. M., and Titsias, M. K. Prescribed generative adversarial networks. ar Xiv preprint ar Xiv:1910.04302, 2019. Ding, B., Qin, C., Liu, L., Chia, Y. K., Joty, S., Li, B., and Bing, L. Is gpt-3 a good data annotator? ar Xiv preprint ar Xiv:2212.10450, 2022. Ding, B., Qin, C., Zhao, R., Luo, T., Li, X., Chen, G., Xia, W., Hu, J., Luu, A. T., and Joty, S. Data augmentation using llms: Data perspectives, learning paradigms and challenges. ar Xiv preprint ar Xiv:2403.02990, 2024. dos Santos, V. G., Santos, G. L., Lynn, T., and Benatallah, B. Identifying citizen-related issues from social media using llm-based data augmentation. In International Conference on Advanced Information Systems Engineering, pp. 531 546. Springer, 2024. Measuring Diversity in Synthetic Datasets Du, W. and Black, A. W. Boosting dialog response generation. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 2019. Evuru, C. K. R., Ghosh, S., Kumar, S., Tyagi, U., Manocha, D., et al. Coda: Constrained generation based data augmentation for low-resource nlp. ar Xiv preprint ar Xiv:2404.00415, 2024. Gao, T., Yao, X., and Chen, D. Simcse: Simple contrastive learning of sentence embeddings. ar Xiv preprint ar Xiv:2104.08821, 2021. Gilardi, F., Alizadeh, M., and Kubli, M. Chatgpt outperforms crowd workers for text-annotation tasks. Proceedings of the National Academy of Sciences, 120 (30):e2305016120, 2023. Gontijo-Lopes, R., Smullin, S. J., Cubuk, E. D., and Dyer, E. Affinity and diversity: Quantifying mechanisms of data augmentation. ar Xiv preprint ar Xiv:2002.08973, 2020. Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., and Bengio, Y. Generative adversarial networks. Communications of the ACM, 63(11):139 144, 2020. Gu, Q. Llm-based code generation method for golang compiler testing. In Proceedings of the 31st ACM Joint European Software Engineering Conference and Symposium on the Foundations of Software Engineering, pp. 2201 2203, 2023. Gupta, H., Scaria, K., Anantheswaran, U., Verma, S., Parmar, M., Sawant, S. A., Mishra, S., and Baral, C. Targen: Targeted data generation with large language models. ar Xiv preprint ar Xiv:2310.17876, 2023. Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., and Hochreiter, S. Gans trained by a two time-scale update rule converge to a local nash equilibrium. Advances in neural information processing systems, 30, 2017. Holtzman, A., Buys, J., Du, L., Forbes, M., and Choi, Y. The curious case of neural text degeneration. ar Xiv preprint ar Xiv:1904.09751, 2019. Hu, E. J., Shen, Y., Wallis, P., Allen-Zhu, Z., Li, Y., Wang, S., Wang, L., and Chen, W. Lora: Low-rank adaptation of large language models. ar Xiv preprint ar Xiv:2106.09685, 2021. Huang, S., Zhao, J., Li, Y., and Wang, L. Learning preference model for llms via automatic preference data generation. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pp. 9187 9199, 2023. Huber, M., Luu, A. T., Boutros, F., Kuijper, A., and Damer, N. Bias and diversity in synthetic-based face recognition. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 6215 6226, 2024. Jalali, M., Li, C. T., and Farnia, F. An informationtheoretic evaluation of generative models in learning multi-modal distributions. Advances in Neural Information Processing Systems, 36:9931 9943, 2023. Jordan, M. I. and Mitchell, T. M. Machine learning: Trends, perspectives, and prospects. Science, 349(6245):255 260, 2015. Khurana, D., Koli, A., Khatter, K., and Singh, S. Natural language processing: state of the art, current trends and challenges. Multimedia tools and applications, 82(3): 3713 3744, 2023. Kynk a anniemi, T., Karras, T., Laine, S., Lehtinen, J., and Aila, T. Improved precision and recall metric for assessing generative models. Advances in neural information processing systems, 32, 2019. Lai, Y.-A., Zhu, X., Zhang, Y., and Diab, M. Diversity, density, and homogeneity: Quantitative characteristic metrics for text collections. ar Xiv preprint ar Xiv:2003.08529, 2020. Le Bronnec, F., V erine, A., Negrevergne, B., Chevaleyre, Y., and Allauzen, A. Exploring precision and recall to assess the quality and diversity of llms. In 62nd Annual Meeting of the Association for Computational Linguistics, 2024. Lee, A., Miranda, B., Sundar, S., and Koyejo, S. Beyond scale: the diversity coefficient as a data quality metric demonstrates llms are pre-trained on formally diverse data. ar Xiv preprint ar Xiv:2306.13840, 2023. Lee, S., Kim, H., and Lee, J. Graddiv: Adversarial robustness of randomized neural networks via gradient diversity regularization. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45(2):2645 2651, 2022. Leinster, T. and Cobbold, C. A. Measuring diversity: the importance of species similarity. Ecology, 93(3):477 489, 2012. Leng, X., Chen, Y., Tang, X., and Bian, Y. Rich feature learning via diversification. In Workshop on Spurious Correlation and Shortcut Learning: Foundations and Solutions, 2025. URL https://openreview. net/forum?id=dv23kah60r. Li, J., Galley, M., Brockett, C., Gao, J., and Dolan, B. A diversity-promoting objective function for neural conversation models. ar Xiv preprint ar Xiv:1510.03055, 2015. Measuring Diversity in Synthetic Datasets Li, Y., Ding, K., Wang, J., and Lee, K. Empowering large language models for textual data augmentation. ar Xiv preprint ar Xiv:2404.17642, 2024a. Li, Z., Zhu, H., Lu, Z., and Yin, M. Synthetic data generation with large language models for text classification: Potential and limitations. ar Xiv preprint ar Xiv:2310.07849, 2023. Li, Z., Si, L., Guo, C., Yang, Y., and Cao, Q. Data augmentation for text-based person retrieval using large language models. ar Xiv preprint ar Xiv:2405.11971, 2024b. Liang, P., Bommasani, R., Lee, T., Tsipras, D., Soylu, D., Yasunaga, M., Zhang, Y., Narayanan, D., Wu, Y., Kumar, A., et al. Holistic evaluation of language models. ar Xiv preprint ar Xiv:2211.09110, 2022. Liu, S., Wang, T., Bau, D., Zhu, J.-Y., and Torralba, A. Diverse image generation via self-conditioned gans. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 14286 14295, 2020. Liu, Y. Roberta: A robustly optimized bert pretraining approach. ar Xiv preprint ar Xiv:1907.11692, 2019. Long, L., Wang, R., Xiao, R., Zhao, J., Ding, X., Chen, G., and Wang, H. On llms-driven synthetic data generation, curation, and evaluation: A survey. ar Xiv preprint ar Xiv:2406.15126, 2024. Loshchilov, I. Decoupled weight decay regularization. ar Xiv preprint ar Xiv:1711.05101, 2017. Maas, A. L., Daly, R. E., Pham, P. T., Huang, D., Ng, A. Y., and Potts, C. Learning word vectors for sentiment analysis. In Lin, D., Matsumoto, Y., and Mihalcea, R. (eds.), Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, pp. 142 150, Portland, Oregon, USA, June 2011. Association for Computational Linguistics. URL https://aclanthology.org/ P11-1015. Mahmoudi, G., Behkamkia, B., and Eetemadi, S. Zero-shot stance detection using contextual data generation with llms. ar Xiv preprint ar Xiv:2405.11637, 2024. Milakov, M. and Gimelshein, N. Online normalizer calculation for softmax. ar Xiv preprint ar Xiv:1805.02867, 2018. Mishra, S., Arunkumar, A., Sachdeva, B., Bryan, C., and Baral, C. Dqi: Measuring data quality in nlp. ar Xiv preprint ar Xiv:2005.00816, 2020. Mostafazadeh, N., Chambers, N., He, X., Parikh, D., Batra, D., Vanderwende, L., Kohli, P., and Allen, J. A corpus and evaluation framework for deeper understanding of commonsense stories. ar Xiv preprint ar Xiv:1604.01696, 2016. Naeem, M. F., Oh, S. J., Uh, Y., Choi, Y., and Yoo, J. Reliable fidelity and diversity metrics for generative models. In International conference on machine learning, pp. 7176 7185. PMLR, 2020. Ospanov, A., Zhang, J., Jalali, M., Cao, X., Bogdanov, A., and Farnia, F. Towards a scalable referencefree evaluation of generative models. ar Xiv preprint ar Xiv:2407.02961, 2024. Padmakumar, V. and He, H. Does writing with language models reduce content diversity? ar Xiv preprint ar Xiv:2309.05196, 2023. Park, D. S., Chan, W., Zhang, Y., Chiu, C.-C., Zoph, B., Cubuk, E. D., and Le, Q. V. Specaugment: A simple data augmentation method for automatic speech recognition. ar Xiv preprint ar Xiv:1904.08779, 2019. Pillutla, K., Swayamdipta, S., Zellers, R., Thickstun, J., Welleck, S., Choi, Y., and Harchaoui, Z. Mauve: Measuring the gap between neural text and human text using divergence frontiers. Advances in Neural Information Processing Systems, 34:4816 4828, 2021. princeton nlp. unsup-simcse-bert-base-uncased, 2021. URL https://huggingface.co/princeton-nlp/ unsup-simcse-bert-base-uncased. Quine, W. V. Ontological relativity and other essays, 1969. Reimers, N. Sentence-bert: Sentence embeddings using siamese bert-networks. ar Xiv preprint ar Xiv:1908.10084, 2019. Ribeiro, M. T., Wu, T., Guestrin, C., and Singh, S. Beyond accuracy: Behavioral testing of nlp models with checklist. ar Xiv preprint ar Xiv:2005.04118, 2020. Sahu, G. and Laradji, I. H. Mixsumm: Topic-based data augmentation using llms for low-resource extractive text summarization. ar Xiv preprint ar Xiv:2407.07341, 2024. Samuel, V., Aynaou, H., Chowdhury, A. G., Ramanan, K. V., and Chadha, A. Can llms augment low-resource reading comprehension datasets? opportunities and challenges. ar Xiv preprint ar Xiv:2309.12426, 2023. Seeger, M. Gaussian processes for machine learning. International journal of neural systems, 14(02):69 106, 2004. Shim, K., Lee, M., Choi, I., Boo, Y., and Sung, W. Svdsoftmax: Fast softmax approximation on large vocabulary neural networks. Advances in neural information processing systems, 30, 2017. Measuring Diversity in Synthetic Datasets Shu, R., Nakayama, H., and Cho, K. Generating diverse translations with sentence codes. In Proceedings of the 57th annual meeting of the association for computational linguistics, pp. 1823 1827, 2019. Socher, R., Perelygin, A., Wu, J., Chuang, J., Manning, C. D., Ng, A. Y., and Potts, C. Recursive deep models for semantic compositionality over a sentiment treebank. In Proceedings of the 2013 conference on empirical methods in natural language processing, pp. 1631 1642, 2013. Song, F., Yu, B., Lang, H., Yu, H., Huang, F., Wang, H., and Li, Y. Scaling data diversity for fine-tuning language models in human alignment. ar Xiv preprint ar Xiv:2403.11124, 2024. Spearman, C. The proof and measurement of association between two things. 1961. Stasaski, K. and Hearst, M. A. Semantic diversity in dialogue with natural language inference. ar Xiv preprint ar Xiv:2205.01497, 2022. Tan, Z., Beigi, A., Wang, S., Guo, R., Bhattacharjee, A., Jiang, B., Karami, M., Li, J., Cheng, L., and Liu, H. Large language models for data annotation: A survey. ar Xiv preprint ar Xiv:2402.13446, 2024. tatsu lab. Alpacaeval : An automatic evaluator for instruction-following language models, 2023. URL https://github.com/tatsu-lab/alpaca_ eval?tab=readme-ov-file#evaluators. Tevet, G. and Berant, J. Evaluating the evaluation of diversity in natural language generation. ar Xiv preprint ar Xiv:2004.02990, 2020. Touvron, H., Martin, L., Stone, K., Albert, P., Almahairi, A., Babaei, Y., Bashlykov, N., Batra, S., Bhargava, P., Bhosale, S., et al. Llama 2: Open foundation and fine-tuned chat models. ar Xiv preprint ar Xiv:2307.09288, 2023. Wang, Y., Kordi, Y., Mishra, S., Liu, A., Smith, N. A., Khashabi, D., and Hajishirzi, H. Self-instruct: Aligning language models with self-generated instructions. ar Xiv preprint ar Xiv:2212.10560, 2022. Wen, Y., Liu, W., Feng, Y., Raj, B., Singh, R., Weller, A., Black, M. J., and Sch olkopf, B. Pairwise similarity learning is simple. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 5308 5318, 2023. White, J., Fu, Q., Hays, S., Sandborn, M., Olea, C., Gilbert, H., Elnashar, A., Spencer-Smith, J., and Schmidt, D. C. A prompt pattern catalog to enhance prompt engineering with chatgpt. ar Xiv preprint ar Xiv:2302.11382, 2023. Yang, S., Guo, S., Zhao, J., and Shen, F. Investigating the effectiveness of data augmentation from similarity and diversity: An empirical study. Pattern Recognition, 148: 110204, 2024a. Yang, S., Liu, X., Dong, X., and Fu, B. Mini-da: Improving your model performance through minimal data augmentation using llm. In Proceedings of the Fifth Workshop on Data Science with Human-in-the-Loop (Da SH 2024), pp. 25 30, 2024b. Ye, J., Gao, J., Li, Q., Xu, H., Feng, J., Wu, Z., Yu, T., and Kong, L. Zerogen: Efficient zero-shot learning via dataset generation. ar Xiv preprint ar Xiv:2202.07922, 2022. Ye, J., Xu, N., Wang, Y., Zhou, J., Zhang, Q., Gui, T., and Huang, X. Llm-da: Data augmentation via large language models for few-shot named entity recognition. ar Xiv preprint ar Xiv:2402.14568, 2024. Yoo, K. M., Park, D., Kang, J., Lee, S.-W., and Park, W. Gpt3mix: Leveraging large-scale language models for text augmentation. ar Xiv preprint ar Xiv:2104.08826, 2021. Yu, L., Zhang, W., Wang, J., and Yu, Y. Seqgan: Sequence generative adversarial nets with policy gradient. In Proceedings of the AAAI conference on artificial intelligence, volume 31, 2017. Yu, Y., Zhuang, Y., Zhang, J., Meng, Y., Ratner, A. J., Krishna, R., Shen, J., and Zhang, C. Large language model as attributed training data generator: A tale of diversity and bias. Advances in Neural Information Processing Systems, 36, 2024. Yuan, J., Tang, R., Jiang, X., and Hu, X. Large language models for healthcare data augmentation: An example on patient-trial matching. In AMIA Annual Symposium Proceedings, volume 2023, pp. 1324. American Medical Informatics Association, 2023. Yuan, L., Cui, G., Wang, H., Ding, N., Wang, X., Deng, J., Shan, B., Chen, H., Xie, R., Lin, Y., et al. Advancing llm reasoning generalists with preference trees. ar Xiv preprint ar Xiv:2404.02078, 2024. Zhang, H. mixup: Beyond empirical risk minimization. ar Xiv preprint ar Xiv:1710.09412, 2017. Zhang, T., Peng, B., and Bollegala, D. Improving diversity of commonsense generation by large language models via in-context learning. ar Xiv preprint ar Xiv:2404.16807, 2024. Zhang, X., Zhao, J., and Le Cun, Y. Character-level convolutional networks for text classification. Advances in neural information processing systems, 28, 2015. Measuring Diversity in Synthetic Datasets Zhang, Y., He, R., Liu, Z., Bing, L., and Li, H. Bootstrapped unsupervised sentence representation learning. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 5168 5180, 2021. Zhu, Y., Lu, S., Zheng, L., Guo, J., Zhang, W., Wang, J., and Yu, Y. Texygen: A benchmarking platform for text generation models. In The 41st international ACM SIGIR conference on research & development in information retrieval, pp. 1097 1100, 2018. Measuring Diversity in Synthetic Datasets A. Additional Related Work Limited by space, we provide a literature review of the LLM dataset generator and application of diversity evaluation methods as follows. A.1. LLM Dataset Generator Recent studies (Ding et al., 2022; Chung et al., 2023) leverage LLMs to augment existing datasets or generate a dataset from scratch, demonstrating the effectiveness in improving dataset quality and reducing data collection costs. Generally, efforts to employ LLMs as dataset generators can be categorized into three strategies: Prompt-guided (Li et al., 2023), Dataset-guided (Ye et al., 2022), and Instruct-guided (Samuel et al., 2023). Prompt-guided and Dataset-guided Strategies. The prompt-guided strategy, a prominent data augmentation approach using LLMs, involves designing task-specific prompts to guide LLMs to augment data in a few-shot (Yoo et al., 2021) or zero-shot (Mahmoudi et al., 2024) manner. Due to its simplicity and effectiveness, subsequent works extend this strategy to various scenarios, such as medical (Yuan et al., 2023), person retrieval (Li et al., 2024b), and social media scenario (dos Santos et al., 2024). However, simple prompt engineering has limitations in fully exploiting the capabilities of LLMs, leading to the development of multi-level prompt designs (Ye et al., 2024) and targeted sample augmentation (Yang et al., 2024b). To further harness the potential of LLMs, the dataset-guided strategy employs LLMs to generate a training set and then trains a task-specific model to annotate unlabeled data (Sahu & Laradji, 2024). The dataset-guided strategy aims to approximate the distribution of targeted scenarios, but it is currently only applicable to text classification tasks. Instruct-guided Strategy. Previous studies (White et al., 2023) indicate that the design of prompts significantly impacts the performance of LLMs, spurring research into the instruct-guided strategy. Generally speaking, the instruct-guided strategy leverages LLMs to generate instructions that guide another LLM in dataset generation (Evuru et al., 2024). These instructions typically relate to context (Samuel et al., 2023), criteria (Huang et al., 2023), and tasks (Wang et al., 2022). To further improve the quality of instructions, efforts have been concentrated on selecting optimal instructions (Li et al., 2024a), integrating soft instructions (Chen et al., 2023), and implementing self-correction mechanisms (Gupta et al., 2023). In a nutshell, LLMs are employed to generate or augment datasets through prompt engineering and multi-step strategies, which encompass various application scenarios and downstream tasks. Meanwhile, the diversity of synthetic datasets emerges as a critical factor in measuring data quality. In our work, we focus on the diversity evaluation of synthetic datasets derived from any dataset generation strategies. A.2. Application of Diversity Evaluation Methods Extending beyond diversity quantification, these methods demonstrate wider utility, such as quantifying augmentation performance and evaluating mode collapse. Thus, we present a literature review of other applications of the diversity evaluation. Quantifying Augmentation Performance. As data augmentation becomes an essential component in the training of deep neural networks (Zhang, 2017; Park et al., 2019), researchers gradually explore a better quantification of the quality of data augmentation. Some studies (Cubuk et al., 2019) suggest that the effectiveness of data augmentation arises from the increased diversity of the data. Inspired by this observation, a series of studies have introduced diversity evaluation metrics into the performance assessment of data augmentation strategies. Specifically, they consider diversity as one aspect of evaluating the quality of augmented data, thereby determining the effectiveness of data augmentation. For instance, (Gontijo-Lopes et al., 2020) utilizes the fundamental idea that models find it more challenging to fit more diverse data, comparing metrics such as training loss and training time before and after augmentation to assess diversity. Similarly, (Yang et al., 2024a) evaluates diversity by examining the eigenvalues and eigenvectors of the similarity matrix of samples before and after augmentation. Evaluating Mode Collapse. Generative adversarial networks (GANs) (Goodfellow et al., 2020) suffer from a well-known phenomenon called mode collapse, which can result in a lack of diversity in the generated samples (Dieng et al., 2019). Consequently, existing studies assess mode collapse by evaluating the diversity of the generated samples. For instance, a common approach is to train an MNIST classifier and then count the number of unique classes predicted for the generated samples. Following this paradigm, Vendi Score (Dan Friedman & Dieng, 2023) compares the generation diversity of Pres GAN (Dieng et al., 2019) and Self-conditioned GAN (Liu et al., 2020). Additionally, some studies (Yu et al., 2017; Zhu et al., 2018; Caccia et al., 2018) employ different metrics to evaluate the diversity of generated samples from GANs. Measuring Diversity in Synthetic Datasets Other Applications. In addition to the aforementioned applications, diversity evaluation metrics have valuable applications in various areas, including sample selection for datasets (Cao et al., 2023), enhancing adversarial robustness (Lee et al., 2022) and out of distribution robustness (Leng et al., 2025), and eliminating biases within datasets (Huber et al., 2024). B. Proof of Properties of DCScore We theoretically confirm that DCScore satisfies several intuitive axioms pointed out by previous studies (Leinster & Cobbold, 2012), thereby demonstrating its role as a principled diversity evaluation method. Effective number (Restated): Diversity should be defined as the effective number of samples in a dataset, ranging from 1 to n. DCScore meets this axiom, as evidenced by its behavior: DCScore equals 1 when all samples in D are identical and equals n when all samples are distinct. Proof. For DCScore, if all samples in a dataset are the same, the probability of any given sample being classified into all categories is the same, i.e., for all i, j = {1, 2, ..., n}, P[i, i] = P[i, j] = 1 n. Then, we have DCScore = Pn i=1 1 n = 1. If all samples in the dataset are distinct, for all i, j = {1, 2, ..., n}, P[i, i] = 1. In other words, the classification function confidently predicts that Ti belongs to the i-th category. Then, we have DCScore tending to n. Identical samples (Restated): Given two identical datasets D1 and D2, the diversity of the synthetic dataset D generated by merging these two datasets remains unchanged. The values of DCScore are the same across D1, D2, and D , i.e., DCScore(D1) = DCScore(D2) = DCScore(D ). (9) Proof. Assuming that D1 and D2 are completely identical, and the samples within each dataset are entirely different, i.e., DCScore(D1) = DCScore(D2) = n. Let P = [P1, ..., Pn, ..., P2n] denote the probability matrix of the merged dataset D = D1 D2 = {Ti}2n i=1. For 1 i n, Ti = Tn+i, where Ti D1, Tn+i D2. Consequently, for each diversity-sensitive component Ti in D , P[i, i] = P[i, n + i] = 1 2. Finally, DCScore(D ) = P2n i=1 1 2 = n. However, the assumption that all samples in the dataset are completely different may be too stringent. We further provide a proof with a more relaxed assumption. Suppose that D1 and D2 are completely identical, with K1 and K2 denoting the kernel matrices for D1 and D2, respectively. In this case, we have K1 = K2 as follows: k1 1,1 k1 1,2 k1 1,n k1 2,1 k1 2,2 k1 2,n ... ... ... ... k1 n,1 k1 n,2 k1 n,n k2 1,1 k2 1,2 k2 1,n k2 2,1 k2 2,2 k2 2,n ... ... ... ... k2 n,1 k2 n,2 k2 n,n According to Eq. (4), for the i-th diversity-sensitive component in D1 and D2, the probability of being classified as category ci can be computed as follows: P1[i, i] = P2[i, i] = k1 i,i P j k1 i,j = k2 i,i P j k2 i,j . (11) For a merged dataset D = D1 D2 = {Ti}2n i=1, when 1 i n, we have Ti = Tn+i, where Ti D1, Tn+i D2. Since the newly added data samples do not affect the pairwise similarity, the kernel matrix K for D can be formulated as follows: k1 1,1 k1 1,n k2 1,1 k2 1,n ... ... ... ... ... ... k1 n,1 k1 n,n k2 n,1 k2 n,n k2 1,1 k2 1,n k1 1,1 k1 1,n ... ... ... ... ... ... k2 n,1 k2 n,n k1 n,1 k1 n,n Measuring Diversity in Synthetic Datasets Analogous to Eq. (11), for 1 i n, the probability of the i-th diversity-sensitive component in D being classified as category ci can be computed as follows: P [i, i] = k1 i,i P j k1 i,j + P = k1 i,i 2 P j k1 i,j = k1 i,i 2 P 2P1[i, i] = 1 For n + 1 i 2n, we obtain the same result as depicted in Eq. (13). Consequently, the diversity of D can be computed as follows: DCScore(D ) = i P1[i, i] + 1 i P1[i, i] = = DCScore(D1) = DCScore(D2). Symmetry (Restated): Diversity remains constant regardless of the order of the samples, exhibiting permutation invariance. Let π( ) denote the permutation function for the sample order, DCScore remains unchanged for any sample permutation of D, i.e., DCScore(D) = DCScore(π(D)). (15) Proof. According to Eq. (4), the order of samples does not affect the classification task. Thus, the diagonal elements of P remain unchanged, indicating the symmetry property of DCScore. Monotonicity (Restated): The diversity of a dataset decreases as the sample similarity increases. Given two datasets D1 and D2, and a new sample Tn+1, where the samples in D1 and D2 are entirely different, and DCScore(D1) = DCScore(D2) = n. If Tn+1 is more similar to the samples in D2 than to those in D1 and is added to both datasets, then for the merged datasets D 1 and D 2, DCScore satisfies the following equation. DCScore(D 1) > DCScore(D 2). (16) Proof. For D 1 = {T 1 1 , T 1 2 , ..., T 1 n , Tn+1} and D 2 = {T 2 1 , T 2 2 , ..., T 2 n , Tn+1}, we have S(T 1 i , Tn+1) < S(T 2 j , Tn+1) for any i, j = {1, 2, ..., n}. Here, S( , ) is the similarity function. In this regard, the classification function f( ) exhibits lower confidence when classifying dataset D 2, resulting in a lower probability that the i-th sample is classified into the i-th class, thereby leading to PD 1[i, i] > PD 2[i, i]. Then, the following formula is satisfied: PD 1[i, i] > PD 2[i, i] DCScore(D 1) > DCScore(D 2), (17) where PD 1, PD 2 are the probability matrix of D 1, D 2, respectively. C. Experimental Settings C.1. Datasets Two types of generated datasets, including self-generated datasets and publicly available generated datasets, are employed in our experiments. We provide detailed information on these datasets below. Measuring Diversity in Synthetic Datasets C.1.1. SELF-GENERATED DATASETS In our experiments, we utilize three different self-generated datasets. We employ two commonly used LLMs as our dataset generator, including Llama2-13B (13B) and Llama2-70B (70B) (Touvron et al., 2023). To prompt LLMs to generate datasets, we design two prompts corresponding to Zero-shot and Few-shot generation settings, respectively. Additionally, self-generated datasets involve two natural language processing tasks: text classification and story completion. We set the maximum number of generated tokens to 100 and 30 for text classification and story completion tasks, respectively. The detailed generation information is offered as follows. Generation Settings. We use three different generation settings across different experiments. The detailed information is shown as follows. Datasets on Section 5.2.1, Section 5.4, Appendix D.4, and Appendix D.5. We generate 21 sub-datasets corresponding to different τg by varying τg from 0.2 to 1.2 with 0.05 intervals. For each sub-dataset, we employ LLMs (13B or 70B) to generate sets of 10 responses per context. Specifically, each sub-dataset consists of 100 samples. Datasets on Section 5.2.2 and Section 5.2.3. We employ the 70B model to generate 6 sub-datasets corresponding to different τg by varying τg from 0.2 to 1.2 with 0.2 intervals. Each sub-dataset includes 5 samples corresponding to a context. To repeat experiments five times, we use five different contexts to prompt the 70B model to generate 5 sub-datasets for each τg. Datasets on Appendix 5.5.1 and Appendix D.3. In zero-shot or few-shot settings, we utilize the 70B model to generate three sub-datasets for the text classification task, corresponding to τg = {0.2, 0.7, 1.2}, respectively. Unlike other settings that provide only one example, in this experiment, we adopt a few-shot setting where four examples and their corresponding labels are given, including two positive examples and two negative examples. Each sub-dataset contains 3,000 samples, and a context is employed to prompt the 70B model to generate five samples. To train text classification models on each sub-dataset, we randomly split 2,100 samples to the training set for each sub-dataset and gather the remaining 900 samples into the testing set across all three sub-datasets. Consequently, we construct a test set comprising 1,800 samples. Prompt Settings. Following the setting of (Li et al., 2023), we design different prompts for zero-shot and few-shot settings, respectively. Here, we provide a seed example for the few-shot setting, whereas the zero-shot setting does not receive any. For the text classification task under the zero-shot setting, we require LLMs to generate movie reviews with Sci-fi/Action/Drama/Comedy/Romance topics. Each movie review contains a single sentiment of either positive or negative, which is regarded as the text label. For the story completion task, we require LLMs to complete the story according to the given context. The detailed prompt setting is provided in Table 9. In Table 9, {style} will be replaced with one topic within {Sci-fi, Action, Drama, Comedy, Romance} and {pos or neg} will be replaced with one label within {Positive, Negative}. {num of words} will be replaced with 50 . {story q} will be replaced by the first three sentences of each sample in the ROC Stories dataset. C.1.2. PUBLICLY AVAILABLE GENERATED DATASETS We use SST2 (Socher et al., 2013), Yelp (Zhang et al., 2015), and AG News (Zhang et al., 2015), and their augmented version based on Attr Prompt (Yu et al., 2024). For three original datasets, we randomly sample data from training sets and apply this to the computational cost analysis in Section 5.3, Appendix D.1, and Appendix D.2. For three augmented datasets, each dataset has 6000 samples. We sample different sub-datasets based on these three datasets, applied to Section 5.3 and Section 5.5.3, respectively. The details are as follows. Datasets on Section 5.3. We remove samples with text token lengths less than 50 in the three datasets and then truncate each sample to a length of 50 tokens. Based on the above, we set up sub-datasets with randomly selected samples of 100, 500, 1000, 2000, and 4000. C.2. Implementation Details For three transformation-based methods, including DCScore, Vendi Score, and K-means Inertia, we employ unsup-simcsebert-base-uncased (princeton nlp, 2021) as the weight of the embedding function. For all language models used to generate Measuring Diversity in Synthetic Datasets Table 9. Prompt settings for zero-shot and few-shot settings. Contents that need to be replaced are highlighted in gray . NLG Tasks Zero-shot Few-shot Text Classification Now you are a movie critic. You are given a movie genre/style and a length requirement. You must come up with a movie that corresponds to the genre/style and write a review that meets the length requirement. Write a film review for a {style} movie to express {pos or neg} feedback. Each review should have {num of words} words. Be sure to express your personal insights and feelings. Please be creative and write unique movie reviews. Now you are a movie critic. You are given a movie genre/style and a length requirement. You must come up with a movie that corresponds to the genre/style and write a review that meets the length requirement. Write a film review according to the given example. Make sure your review expresses the same sentiment (positive or negative) as the example. Each review should have {num of words} words. Be sure to express your personal insights and feelings. Please be creative and write unique movie reviews. The following is the example: #An example from IMDB (Maas et al., 2011)# Story Completion Question: {story q} Answer: Complete the story according to the given example. Example: #An example from ROC Stories (Mostafazadeh et al., 2016)# Question: {story q} Answer: the dataset, we set the top-p and top-k parameters to 1 and -1, respectively. Additionally, we limit the maximum number of newly generated tokens to 100 for the text classification task and 30 for the story completion task. All experiments are conducted on 8 NVIDIA Tesla V100 GPU with 32GB of memory. For self-generated datasets, we only evaluate the diversity of generated components. C.2.1. HYPERPARAMETER SETTINGS OF DIVERSITY EVALUATION For DCScore, Vendi Score, and K-means Inertia, we fix the batch size of generating sample representation at 128 across all experiments. Given the varying hyperparameters for each diversity evaluation method, we provide the detailed settings for each method below: DCScore. We employ the inner product as Kernel( ), and Softmax as f K( ). Except for hyperparameter sensitivity experiments, we set τ in Eq. (4) to 1 for all other experiments. Distinct-n. We use 5-grams to calculate distinct-n. K-means Inertia. We set the number of clusters to 10 for all experiments. Vendi Score. We employ the inner product as Kernel( ). C.2.2. HYPERPARAMETER SETTINGS OF DOWNSTREAM TASK TRAINING To train text classification models, we employ Ro BERTa (Liu, 2019) as the encoder and utilize the representations from the last layer of the encoder as the classifier s input. We employ Lo RA (Hu et al., 2021) to finetune the encoder and the Measuring Diversity in Synthetic Datasets Task: Use the following criteria to compare two sets of sentences and judge which set of sentences better meets the criteria. Each set of sentences contains 5 sentences, each starting with # . Evaluation Criteria: Diversity includes richness of vocabulary, variation in sentence structure, and diversity of topics, as detailed below: - Vocabulary richness: Whether the vocabulary used in the sentences is diverse and includes uncommon words - Sentence Structure Variation: Whether there is variation in sentence structure and different grammar structures are used. - Topic Diversity: Whether the sentences cover different topics or concepts. Please compare the following two sets of sentences and analyze them based on the above criteria: Sentence set 1: xxx Sentence set 2: xxx Analyze each sentence and based on the above criteria, give your judgement: Which sentence better meets the criteria? Please directly answer with "Sentence set 1 wins" or "Sentence set 2 wins". If it is impossible to judge, then answer with "No winner". Task Definition Diversity Definition Sent. Set 1 Sent. Set 2 Prompt for GPT-4 Evaluation = + + + Prompt Demo (The background color of the text signifies the category to which it belongs, including the task definition, diversity definition, general prompt, and sentence sets.) Figure 7. Prompt settings for GPT-4 evaluations. classifier on self-generated datasets. Specifically, we fix the Lo RA scaling factor to 32 and the rank of the update matrices to 8. We use Adam W (Loshchilov, 2017) with an initial learning rate of 5e-5 and linear learning rate decay as our optimizer. Additionally, we set the batch size per GPU as 32 and epochs as 120. For the number of training GPUs, we employ 8 GPUs for zero-shot settings and 4 GPUs for few-shot settings. Therefore, the different training steps for zero-shot and few-shot settings are shown in Figure 8. C.2.3. EVALUATION PROTOCOL In our experiments, we employ diversity evaluation methods to score the diversity of sub-datasets using two evaluation protocols: overall evaluation and batch evaluation. While K-means Inertia uses the overall evaluation protocol, all other methods utilize the batch evaluation protocol. The detailed settings for the two evaluation protocols are as follows: Batch evaluation. Due to a context or prompt associated with several samples in a sub-dataset, the batch evaluation protocol requires that evaluation methods treat samples generated from the same context as a single batch. The evaluation results are then averaged across all batches of the entire sub-dataset. Overall evaluation. We consider each sample in a sub-dataset as independent, meaning each sample is generated by a distinct context or prompt. Based on this assumption, the overall evaluation protocol requires evaluation methods to directly measure the diversity of the entire sub-dataset. C.2.4. PROMPT SETTINGS OF LLM EVALUATION In Section 5.2.3, we use GPT-4 to perform pairwise diversity comparisons. To guide GPT-4 in making these comparisons, we employ a well-designed prompt, as illustrated in Figure 7. The prompt for GPT-4 evaluations includes the task definition, diversity definition, general prompt, and sentence sets to be compared. C.2.5. SETTINGS OF HUMAN EVALUATION According to Appendix C.1.1, datasets in Section 5.2.2 are generated at six temperatures, with five results per context (prompt) in each sub-dataset. During the evaluation, evaluators are asked to select the more diverse sub-dataset from pairs of sub-datasets. Across six temperatures, this results in 15 comparisons, and with three evaluators, a total of 45 judgments are made. Sub-datasets are ranked by the frequency of being chosen as more diverse. This process is repeated five times with different contexts to derive the final human diversity ranking. Measuring Diversity in Synthetic Datasets Table 10. Comparison of computation time between DCScore and Vendi Score on Yelp. Kernels Yelp Sample num 4k 8k 16k 32k 64k Inner product Vendi Score 57.96 0.35 114.64 1.63 227.76 7.04 451.49 19.73 912.60 25.69 DCScore 57.95 0.31 115.35 1.16 232.49 1.34 448.98 23.94 961.29 2.86 RBF kernel Vendi Score 59.31 0.06 118.15 0.91 242.06 7.60 527.99 2.89 1272.93 21.15 DCScore 58.49 0.14 116.29 0.92 232.94 3.09 471.18 7.80 953.62 17.21 Poly kernel Vendi Score 59.48 0.05 118.94 0.95 234.08 11.72 522.82 3.04 1313.55 12.64 DCScore 58.73 0.08 117.02 0.90 227.72 9.51 462.45 13.91 988.53 1.10 Table 11. Comparison of computation time between DCScore and Vendi Score on AG News. Kernels AG News Sample num 4k 8k 16k 32k 64k Inner product Vendi Score 14.56 1.16 30.20 1.15 63.70 1.39 127.25 1.13 254.20 11.71 DCScore 14.61 1.15 30.61 1.77 63.57 2.68 129.70 4.17 284.76 12.30 RBF kernel Vendi Score 16.69 1.54 33.69 1.47 80.09 2.34 185.79 6.44 617.06 12.51 DCScore 16.01 1.53 31.06 0.96 69.15 1.32 129.36 5.56 297.29 3.67 Poly kernel Vendi Score 17.60 0.62 36.16 1.27 79.34 1.57 190.96 2.75 632.69 10.14 DCScore 16.88 0.59 33.78 1.28 68.18 1.66 138.18 3.82 303.06 11.40 D. Additional Experiments D.1. Computational Cost for Larger Datasets In the case where the inner product is used as the kernel function and n d, Vendi Score can significantly reduce computational complexity. To ensure a fair comparison, we compare the computation times of Vendi Score and DCScore on larger datasets, as well as under different kernel functions. Specifically, we employ SST2, Yelp, and AG News as the evaluated datasets. We randomly select 4k, 8k, 16k, 32k, 64k samples and record the computation times for both methods across these different sample sizes. We repeat the experiments 5 times to report the mean and standard deviation. As shown in Tables 6, 10, and 11, DCScore has a shorter computation time than Vendi Score in most cases, with Vendi Score only exhibiting a computational advantage when the inner product is used and n d. Furthermore, as the sample size increases, the efficiency advantage of DCScore becomes more pronounced. When using a polynomial kernel, on the SST2 dataset, DCScore requires only one-third of the computation time of Vendi Score when the sample size reaches 64k. In contrast, although Vendi Score has a computational advantage in the case of the inner product, the difference compared to DCScore is not significant. The experimental results are consistent with our complexity analysis presented in Section 4.3. Overall, DCScore outperforms Vendi Score in terms of computation time across most kernel functions. Vendi Score exhibits a computational time advantage only when the inner product is used as the kernel function, which will limit its applicability. As shown in Chapter 4 of (Seeger, 2004), it is essential to employ different kernel functions to accommodate a wider range of scenarios. D.2. Comparison of Computational Cost between DCScore and Efficiency-improved Methods A recent work (Ospanov et al., 2024) employs the random Fourier features framework to reduce the computational cost of Vendi Score (Dan Friedman & Dieng, 2023) and RKE (Jalali et al., 2023). To further investigate the computational efficiency of DCScore, we compare it with efficiency-improved versions of two entropy-based diversity evaluation methods, namely FKEA-Vendi and FKEA-RKE. Here, we set α to 1 for FKEA-Vendi and 2 for FKEA-RKE. Notably, we follow the Measuring Diversity in Synthetic Datasets Table 12. Comparison of computation time between DCScore and two efficiency-improved methods (FKEA-Vendi and FKEA-RKE). Datasets Sample num 4k 8k 16k 32k 64k SST2 FKEA-Vendi 18.43 0.14 36.97 0.03 74.03 0.20 147.78 0.30 295.14 0.41 FKEA-RKE 18.46 0.08 37.08 0.12 73.98 0.09 147.99 0.09 297.46 2.02 DCScore 3.26 0.08 6.75 0.08 13.94 0.06 31.32 0.34 90.36 1.24 Yelp FKEA-Vendi 26.28 0.13 52.69 0.24 106.59 2.25 212.52 2.22 422.51 1.19 FKEA-RKE 26.19 0.13 52.68 0.24 105.27 0.22 212.38 1.87 422.94 1.55 DCScore 37.53 0.08 75.19 0.10 151.59 0.39 306.80 0.33 641.60 0.60 AG News FKEA-Vendi 20.64 0.05 41.66 0.06 83.02 0.11 165.92 0.25 338.16 14.26 FKEA-RKE 20.68 0.04 41.68 0.03 83.04 0.14 165.85 0.23 331.58 0.29 DCScore 10.29 0.49 21.12 0.73 43.84 1.02 91.64 0.86 213.30 2.15 0 250 500 750 1000 Step Zero-shot Setting 0 500 1000 1500 2000 Step Few-shot Setting Figure 8. Loss curves of the downstream task training. experimental settings described in Appendix D.1 and leverage different random seeds for data sampling. Table 12 illustrates the computation time comparison across the SST2, Yelp, and AG News datasets. DCScore demonstrates lower computation time compared to FKEA-Vendi and FKEA-RKE on the SST2 and AG News datasets. However, the results are reversed for the Yelp dataset. Additionally, we observe that FKEA-Vendi does not improve computation time compared to Vendi Score, possibly because FKEA focuses solely on reducing the complexity of the summarization (eigenvalue computation) phase, whereas our computation time encompasses the entire calculation process. D.3. Correlation with Downstream Task Training To investigate the correlation between DCScore and downstream task training, we train text classification models using selfgenerated datasets under zero-shot and few-shot settings. We vary the generation temperature τg of self-generated datasets within the range of {0.2, 0.7, 1.2}. More details of training datasets and hyperparameters are presented in Appendix C.1.1 and C.2.2, respectively. Figure 8 shows the loss curves of these trained classification models. In the zero-shot setting, we observe increasing optimal loss values as τg varied from 0.2 to 1.2, indicating that the model is more easily fitted to datasets with limited diversity. However, as shown in Table 7, models trained on more diverse datasets achieve better accuracy, which can be attributed to their enhanced generalization capabilities. From Table 7, the diversity evaluated by DCScore has a similar trend to the accuracy performance in the zero-shot setting, further demonstrating the effectiveness of DCScore in diversity evaluation. In the few-shot setting, we observe a trend in optimal loss variation similar to that in the zero-shot setting, as shown in Figure 8. However, Figure 9 reveals that at epoch 120, models trained on datasets generated with τg = 0.7 outperform those Measuring Diversity in Synthetic Datasets 0 2000 4000 6000 Step Loss Curves Epoch 360 Epoch 240 Epoch 120 360 240 120 Epoch 73.41 73.19 Model Accuracy Figure 9. Loss curves and accuracy of models trained on generated dataset with τg = 1.2. Table 13. Correlation (Spearman s ρ) results of DCScore with various embedding functions. Spearman s ρ varies between -1 and +1, with 0 implying no correlation. Best results are indicated in bold. Embedding models Zero-shot setting Few-shot setting Text classification Story completion Text classification Story completion 13B 70B 13B 70B 13B 70B 13B 70B Sim CSE (unsup-simcse-bert-base-uncased) 0.9961 0.9779 0.9844 0.9792 0.9909 0.9883 0.9857 0.9974 Sim CSE (sup-simcse-roberta-large) 0.9909 0.9753 0.9883 0.9883 0.9792 0.9935 0.9779 0.9623 Sentence BERT (all-mpnet-base-v2) 0.9896 0.9740 0.9870 0.9909 0.9766 0.9870 0.9857 0.9870 BGE (bge-large-en-v1.5) 0.9909 0.9896 0.9922 0.9948 0.9857 0.9922 0.9870 0.9922 using τg = 1.2. This phenomenon can be attributed to the higher diversity of datasets generated at a higher τg, resulting in increased fitting difficulty. Under the current settings, the number of training epochs for the dataset generated at a temperature of 1.2 is insufficient, preventing the trained model from achieving optimal performance. To validate this hypothesis, we increase the number of epochs to 240 and 360 and train models on the dataset generated at a temperature of 1.2. The final training loss and accuracy of these models are shown in Figure 9. We observe that as the number of epochs increases, the model s loss gradually decreases, and its performance improves progressively. Ultimately, the model s accuracy outperforms that of models trained on datasets generated at temperatures of 0.2 and 0.7. Moreover, from Table 7, models trained on datasets from the zero-shot setting outperform those trained on datasets from the few-shot setting. However, this discrepancy arises from the different test sets used in the two settings, making direct performance comparisons inappropriate. D.4. Impact of Embedding Functions Φ The paradigm of the transformation-based method enables DCScore to utilize various embedding functions tailored to different scenarios. Consequently, we investigate the impact of embedding functions on self-generated datasets used in Section 5.2.1. As shown in Table 13, we compare the correlation of the diversity evaluation results of DCScore across 4 different embedding functions with diversity pseudo-truths, where the model names in parentheses within the embedding function refer to those available on Hugging Face. Our findings indicate that DCScore exhibits strong correlations with diversity pseudo-truths across various embedding functions. Notably, DCScore utilizing the BGE embedding function achieves the best results in half of the cases. Additionally, the minimum correlation in Table 13 exceeds 0.96, which is classified as a strong correlation according to (Akoglu, 2018). This result also supports the following two conclusions: (1) the embedding function used effectively captures the differences among samples from multiple perspectives, and (2) DCScore is sufficiently adaptable to different embedding functions while maintaining stable performance. Measuring Diversity in Synthetic Datasets Table 14. Correlation (Spearman s ρ) results of DCScore with various kernel functions. Spearman s ρ varies between -1 and +1, with 0 implying no correlation. Best results are indicated in bold. Embedding models Zero-shot setting Few-shot setting Text classification Story completion Text classification Story completion 13B 70B 13B 70B 13B 70B 13B 70B Inner product 0.9961 0.9779 0.9844 0.9792 0.9909 0.9883 0.9857 0.9974 laplacian kernel 0.9935 0.9831 0.9883 0.9727 0.9597 0.9649 0.9701 0.9922 RBF kernel 0.9935 0.9818 0.9896 0.9753 0.9740 0.9727 0.9792 0.9922 polynomial kernel 0.9870 0.9584 0.9714 0.9506 0.9182 0.9182 0.9857 0.9896 D.5. Impact of Kernel Functions Similar to Appendix D.4, we investigate the impact of different kernel functions on the performance of DCScore. Specifically, this experimental setup is identical to that in Appendix D.4. As shown in Table 14, we find that DCScore demonstrates stable performance across various kernel functions. However, the influence of the kernel function is slightly more pronounced than that of the embedding function, as indicated by the greater fluctuations in correlation among the different kernel functions. Furthermore, we observe that DCScore achieves optimal performance in the case of the inner product. Overall, DCScore consistently maintains strong diversity evaluation performance across different kernel functions. E. Baseline Methods We present the detailed modeling of baseline methods into DCScore as follows: Distinct-n. Distinct-n (Li et al., 2015) is a prevalent diversity metric depending on n-grams, where n signifies the number of successive items. Distinct-n calculates the proportion of unique n-grams to all n-grams. The n-grams operation falls under the text representation stage, while the step of obtaining a unique set of n-grams corresponds to the pairwise similarity stage. Typically, a high form similarity among samples in the evaluated dataset results in a smaller unique n-gram set. Finally, ratio calculations belong to the diversity summarization stage. Text Representation: n-grams(Concat(D)), Pairwise Similarity: Unique(n-grams(Concat(D))), Diversity Summarization: Distinct-n(D) = | Unique(n-grams(Concat(D)))| | n-grams(Concat(D))| , where n-grams, Unique, and Concat represent the n-grams, de-overlap process, and concatenate operation, respectively. K-means inertia. K-means Inertia (Du & Black, 2019), a transformation-based method, performs clustering in sample representation and then calculates inertia as diversity outcomes. Here, inertia refers to the square summation of the distance between samples and cluster centroids. Text Representation: H = Φ({ Ti}n i=1), Pairwise Similarity: C = K-means(H), Diversity Summarization: Inertia(D) = X ck C,hj Hck (hj ck)2, (19) where H is the representation of all samples and hi is the representation of the i-th sample, C denotes the cluster centroid set, and ck C represents the k-th cluster centroid. The sample representation associated with the k-th cluster centroid is expressed as hj Hck, while Hck denotes the sample representations corresponding to the k-th cluster centroid. Vendi Score. Vendi Score (Dan Friedman & Dieng, 2023) is a recently proposed diversity metric that falls under the category of the transformation-based method. Based on sample representations, Vendi Score utilizes a kernel function to calculate Measuring Diversity in Synthetic Datasets a similarity matrix K. Subsequently, Vendi Score summarizes diversity as the exponential of the Shannon entropy of the eigenvalues of K/n. Text Representation: H = Φ({ Ti}n i=1), Pairwise Similarity: K = Kernel(H), Diversity Summarization: VS(D) = exp ( i=1 λi log λi), where Kernel( ) is the kernel function, such as the inner product, λi is the i-th eigenvalue of K/n.