# merge3_efficient_evolutionary_merging_on_consumergrade_gpus__6550ec94.pdf MERGE3: Efficient Evolutionary Merging on Consumer-grade GPUs Tommaso Mencattini * 1 Adrian Robert Minut * 2 Donato Crisostomi 2 Andrea Santilli 2 Emanuele Rodol a 2 Evolutionary model merging enables the creation of high-performing multi-task models but remains computationally prohibitive for consumer hardware. We introduce MERGE3, an efficient framework that makes evolutionary merging of Large Language Models (LLMs) feasible on a single GPU by reducing fitness computation costs 50 while retaining a large fraction of the original performance. MERGE3 achieves this by Extracting a reduced dataset for evaluation, Estimating model abilities using Item Response Theory (IRT), and Evolving optimal merges via IRT-based performance estimators. Our method enables state-of-the-art multilingual and crosslingual merging, transferring knowledge across languages with significantly lower computational overhead. We provide theoretical guarantees and an open-source library, democratizing highquality model merging. github.com/tommasomncttn/merge3 1. Introduction Model merging has become a powerful and accessible approach for developing new state-of-the-art models without the need for cluster-grade computing typically required for large model training (Yang et al., 2024a). Its key advantage lies in performing the merging process post-hoc directly in the parameters of endpoint models that is, pre-existing models (either base or fine-tuned) that serve as the components of the merging process eliminating the need for training and significantly reducing the demand for expensive computational resources. *Equal contribution 1School of Computer and Communication Science, Ecole Polytechnique F ed erale de Lausanne, Lausanne, Switzerland 2Department of Computer Science, Sapienza University of Rome, Rome, Italy. Correspondence to: Tommaso Mencattini . Proceedings of the 42 nd International Conference on Machine Learning, Vancouver, Canada. PMLR 267, 2025. Copyright 2025 by the author(s). 0 1 2 3 4 TFLOPs 106 MERGE3-20 MERGE3-50 MERGE3-100 Evo LLM-JP-7B 24h NVIDIA 4090 TIES-DARE Task Arithmetic 50x reduction Figure 1. Accuracy on Japanese GSM8K over fitness evaluation FLOPs. MERGE3 is competitive with a model evolved on the full dataset by only using a consumer-grade GPU and 2% of the data (point size reflects data amount). This approach has significantly broadened access to the field, with ML practitioners producing competitive models out of existing ones on standard consumer GPUs1(Ilharco et al., 2022). However, although computationally inexpensive, most of the existing approaches are quite rudimentary, require ad-hoc choices, and are usually based on ungrounded trial-and-error strategies for selecting the merge coefficients, which ultimately limits their downstream performance (Yadav et al., 2023; Yu et al., 2024). On the other hand, recent work has shown that evolutionary merging can produce models of unprecedented quality by automating the hyperparameter search for merging coefficients (Akiba et al., 2025). While this technique can incorporate any standard merging method, such models are absent from public leaderboards likely due to a mismatch between the high computational demands of evolutionary merging and single-GPU setups typical of merging practitioners. Indeed this computational cost is significantly high: computing the fitness function requires generating and evaluating answers for each dataset element, for each candidate in every evolutionary step. As shown in Figure 1, the fitness computation alone in the 1,000-trial evolutionary merge from Akiba et al. (2025) requires approximately 4 106 TFLOPs, with the full algorithm demanding largely over a month of 1At the time of writing, around 30% of models on the Hugging Face Open LLM leaderboard are merged models. Efficient Evolutionary Merging on Consumer-grade GPUs continuous computation if run on a single NVIDIA 4090 ( C.3.3) with 24 GB of VRAM. Requiring repeated and costly runs of large language models, the fitness evaluation is the primary bottleneck. This makes evolutionary merging out of reach on consumer hardware, potentially excluding the very users it was meant to empower. In this paper, we address this challenge by introducing MERGE3, an evolutionary merging framework that runs on a single consumer GPU with competitive results (see fig. 1). Unlike the competing approach, MERGE3 operates with just 0.077 106 TFLOPs, namely a 50-fold reduction. This drastic decrease in computational cost makes it feasible on consumer hardware, freeing up FLOPs for further optimization or additional tasks. Our approach starts by Extracting a reduced subset of the fitness evaluation dataset, significantly alleviating the computational bottleneck of fitness computation (fig. 2). However, this reduction risks losing accuracy if the subset lacks diversity. To address this, we apply Item Response Theory (IRT) (Lord et al., 1968) a well-established statistical framework to bridge the gap between reduced-dataset evaluations and full-dataset performance. Specifically, we first Estimate the latent abilities of the endpoint models using IRT, ensuring the merged models accurately reflect their components strengths. Then, we Evolve the endpoint models with IRT-based performance estimators designed for model merging, assuming the merged model s ability is a combination of those of the endpoint models. This approach significantly improves the efficiency and accuracy of fitness estimation, integrating merging-specific insights into performance estimation theory while maintaining high accuracy with reduced datasets. Experimental results show that MERGE3 effectively transfers mathematical skills by merging a strong math model with three language-specific models, achieving 10 20% higher accuracy than standard merging baselines in each language. Building on this, we evolve a single multilingual model by merging Italian, English, German, and Dutch models, outperforming individually fine-tuned models by up to 19% on ARC (Clark et al., 2018), a widely used benchmark for reasoning. Furthermore, MERGE3 achieves competitive accuracy on Japanese GSM8K (Cobbe et al., 2021), matching models evolved on full datasets while maintaining high efficiency, demonstrating that our evolutionary strategy preserves performance while drastically reducing computational costs. To summarize, our contributions are fourfold: We introduce a novel, efficient evolutionary model merging framework leveraging Item Response Theory, making merging feasible on consumer hardware. We demonstrate its effectiveness in transferring skills Figure 2. MERGE3 for math + Japanese merging (GSM8K). The method Extracts a reduced evolutionary dataset, Estimates ability parameters (γ) via Item Response Theory (IRT) based on their response correctness, and Evolves the endpoint models through iterative merging. Leveraging an IRT-based performance estimator, it approximates full-dataset fitness with reduced data, cutting fitness estimation costs while preserving full-dataset accuracy making evolutionary merging feasible on consumer GPUs. across languages and synthesizing state-of-the-art multilingual models without standard training. We advance the theoretical foundations of performance estimation in model merging and provide formal guarantees for our proposed estimators. We release a modular library for evolutionary merging on consumer GPUs, alongside a suite of state-of-theart models for several low-resource languages. 2. Related Work Model Merging has emerged as an efficient alternative to ensembling by integrating existing models without any additional training. One set of methods identifies neuron permutations that align the models into a shared optimization basin, allowing them to be merged through straightforward averaging (Ainsworth et al., 2022; Jordan et al., 2023; Stoica et al.; Pe na et al., 2023; Crisostomi et al., 2025). Closer to our work, multi-task model merging focuses on the case where a single pre-trained model is finetuned for different tasks (Ilharco et al., 2022; Yadav et al., 2023; Yu et al., 2024; Matena & Raffel; Wortsman et al., 2022; Davari & Belilovsky, 2025; Wang et al., 2024; Zhou et al., 2024; Gargiulo et al., 2025). In this direction, several works address task interference by pruning or selectively combining parameters e.g., TIES-merging (Yadav et al., 2023), Model Breadcrumbs (Davari & Belilovsky, 2025), and DARE Merging (Yu et al., 2024) or by opti- Efficient Evolutionary Merging on Consumer-grade GPUs mizing merge coefficients (Yang et al.), introducing taskspecific modules (Yang et al., 2024b), and disentangling weights (Ortiz-Jimenez et al., 2024). Evolutionary Algorithms. Evolutionary Algorithms are black-box optimization algorithms operating on a population of potential solutions by evolving them through generations with operators such as selection, mutation, recombination, and crossover (B ack & Schwefel, 1993; P etrowski & Ben-Hamida, 2017; Dasgupta & Michalewicz, 1997). Recent applications include neural architecture search (Real et al., 2019) and hyperparameter tuning (Vincent & Jidesh, 2023), where evolutionary methods efficiently navigate large design spaces without manual intervention. The fitness function is crucial, as it evaluates the quality of each solution, guiding the selection process by favoring higher-scoring (fitter) solutions for reproduction (Eiben & Smith, 2015). Closest to our work, Akiba et al. (2025) propose to apply evolutionary algorithms to optimize model merging recipes, eliminating the need for trial-and-error in combining parameters. In this context, the most obvious candidate for a fitness function is simply the performance of the resulting model over a held-out validation set. Item Response Theory. Item Response Theory (IRT) (Cai et al., 2016; Van der Linden, 2018; Brzezi nska, 2020; Lord et al., 1968) is a paradigm to design, analyze, and score responses to tests such as SAT or GRE (An & Yung, 2014; Kingston & Dorans, 1982; Petersen et al., 1982). Based on the relationship between individuals performances on a test item and the test takers levels of performance on the corresponding required ability, IRT has recently spread from psychometrics to natural language processing. In this direction, Lalor et al. (2016) leverage IRT s latent dimensions to evaluate language models, while Vania et al. (2021) use it to analyze benchmark saturation in NLP evaluations. More relevant to our work, Zhuang et al. (2023) and Polo et al. (2024) employ IRT-driven adaptive testing to alleviate the computational burden of large-scale evaluations for large language models (LLMs). Although their focus is on LLM evaluation, which shares similarities with the efficient evaluation of fitness functions in model merging, our work builds on these approaches to design IRT-based estimators specifically tailored for model merging. Unlike prior applications of IRT, which are limited to LLM evaluations, our approach adapts the framework to address the unique challenges of evolutionary model merging, enabling efficient and accurate fitness estimation. Our method MERGE3 speeds up evolutionary model merging by reducing the computational cost of fitness eval- uation. It achieves this by shrinking the fitness evaluation dataset and using IRT-based performance estimators to maintain full-dataset accuracy from subset evaluations. Figure 2 shows an overview of our method, while we present below the pseudo-code for the end-to-end MERGE3 algorithm. Algorithm 1 The full MERGE3 algorithm. Require: Dataset D, models {M1, M2, . . . , Mn}, iterations T Ensure: Pareto-optimal merged models 1: D RANDOMSAMPLE D, k # Sample k items from D 2: {γ1, . . . , γn} ESTIMATEABILITIES {M1, . . . , Mn}, D 3: P Generate Initial Population{M1, . . . , Mn} 4: for t 1 to T do 5: for all M P do 6: λ FITLAMBDA M, {γ1, . . . , γn}, D 7: preds GETPREDICTIONS M, D, λ 8: corr GETCORRECTNESS preds, D 9: F(M) ESTIMATEFITNESS corr, λ 10: end for 11: P SELECTPARENTS P, f # Select based on fitness 12: P APPLYMUTATION P 13: P APPLYCROSSOVER P # Generate offspring 14: end for 15: return PARETOFRONT P 3.1. Extract & Estimate Evaluating the fitness function involves generating and assessing answers for each data sample, repeated across all models in the population at every evolutionary step. Given the computational demands of evolutionary algorithms and LLMs, this process is highly intensive. To mitigate this, we reduce the dataset D to a smaller subset D D with | D| |D|. After exploring various subsampling strategies, we found uniform random sampling as effective as more complex methods (see appendix C.1) and adopted it for simplicity. Since dataset reduction is not our main focus, we leave further optimizations for future work. Reducing the dataset speeds up evaluation but does not guarantee identical results particularly when the subset is significantly smaller, as in our case. To bridge this gap, we build an IRT-based estimator that adjusts for this discrepancy, effectively estimating performance to reflect fulldataset results (Lord et al., 1968; Polo et al., 2024). IRT model. We first define an estimator to assess each endpoint model s inherent abilities, derived from the latents of a Bayesian network. This ensures that merging preserves individual model strengths. In the Evolve step ( 3.2), the estimated latent abilities are fed to a performance estimator to compute the final fitness. To estimate LLM abilities, we build on Polo et al. (2024), who applied IRT to evaluate LLM performance; however, Efficient Evolutionary Merging on Consumer-grade GPUs while they used IRT for benchmarking, we extend it to estimate inherent abilities relevant for model merging, and explicitly use them to guide merging in the Evolve step. In IRT, latent variables (γ) represent a model s underlying abilities, while manifest variables (Y ) indicate response correctness. The framework models the probability of a correct response based on model abilities and item characteristics (e.g., difficulty). IRT defines this probability as: P(Yim = 1 | γm, αi, βi) = 1 1 + exp( α i γm + βi) (1) Here, γm Rd represents model m s latent abilities, αi Rd defines the ability dimensions needed to answer example i, and βi denotes its difficulty. A model is more likely to answer correctly when its abilities (γm) align with the example s required traits (αi) and less likely when the difficulty (βi) is higher. Yim is a binary variable indicating whether model m correctly predicts example i (1 if correct, 0 otherwise). Crucially, this approach estimates a model s likelihood of answering correctly without directly analyzing the example s content, relying solely on the estimated IRT parameters (γm, αi, βi). Fitting. We use variational inference to efficiently estimate both example-specific (αi, βi) and model-specific (γm) parameters within a hierarchical Bayesian model (Lalor & Rodriguez, 2023), initialized as detailed in appendix B.1. Following Polo et al. (2024), we estimate αi and βi using correctness data (Yim) from publicly available model evaluations, namely the Open LLM leaderboard. To estimate γm, each endpoint model generates answers for the full evaluation dataset, which are then used to assess correctness (Yi) (see Figure 2). This procedure is repeated for each model m, producing the corresponding γm (γ1 and γ2 in the Figure). To summarize, unlike previous work, where IRT latent abilities remain hidden variables, we explicitly derive γm as an ability estimator to quantify each model s strengths. Additionally, rather than estimating γm from a subset, we compute it using the full evaluation dataset, providing a more comprehensive measure of model ability, which we now leverage to enhance the merging process. 3.2. Evolve: Performance Estimator The performance estimator, a key part of the Evolve step, efficiently approximates the fitness function, which measures the merged model s accuracy. Since fitness evaluation runs repeatedly during evolution (once per model per iteration), reducing its computational cost is crucial. In- stead of evaluating the full dataset, the estimator predicts performance using only the endpoint models abilities and the reduced dataset from previous steps, significantly accelerating the process. We introduce two novel performance estimators for merging: merged performance Item Response Theory estimator (MP-IRT) and generalized merged performance Item Response Theory estimator (GMP-IRT). Since model merging linearly combines weights, we assume that the latent abilities of the merged model (e.g., problem-solving or linguistic capabilities) are also a linear combination of the endpoints abilities. This makes our approach far more efficient, estimating only the interpolation coefficients (λi) instead of recomputing the full ability vector γ of the merged model from scratch (as done in P-IRT and GP-IRT (Polo et al., 2024)). Assumption 1 (Linear Combination of Latent Abilities). Let {m0, m1, . . . , mn} be endpoint models with latent ability vectors γi. If a new model m is formed as a linear combination of their parameters, its ability vector γ m can be expressed as: j=1 λj γj = [γ1, . . . , γn] λ (2) where λ = (λ1, . . . , λn) are the interpolation coefficients. This assumption allows us to compute the multidimensional IRT model (Eq. 1) for model merging as a linear combination of the individual models abilities: pi m = P(Yi m = 1 | λ1γ1 + λ2γ2, αi, βi) = 1 1 + exp α i λ1γ1 + λ2γ2 + βi (3) Since the endpoint models latent abilities γj were preestimated over the full dataset D in the Estimate step, we only need the subset D to estimate the interpolation coefficients λj via MLE. Performance Estimators. To estimate the accuracy of the merged model m using only the reduced dataset D and pi m, we define the merged performance-IRT (MP-IRT) estimator as: ˆZmp-IRT m = ˆτ | D| i D Yi m + 1 ˆτ |D \ D| i D\ D ˆpi m (4) where ˆτ = | D| |D| downweights smaller subsets that may be noisier. In practice, we are considering the observed correctness for the data points we have access to, while ˆpi m predictions are used for the rest, enabling accurate performance estimation across all examples despite evaluating only a subset, where ˆpi m = Efficient Evolutionary Merging on Consumer-grade GPUs 10 20 30 50 100 Number of Samples Absolute Error 10 20 30 50 100 Number of Samples 10 20 30 50 100 Number of Samples Truthful QA 10 20 30 50 100 Number of Samples p-IRT gp-IRT mp-IRT (ours) gmp-IRT (ours) Random Figure 3. Performance Estimators: Absolute error of various estimators as a function of sample size (lower is better). Our MP-IRT and GMP-IRT estimators consistently achieve lower error across various sample sizes and datasets. Additional results available in Figure 13. P Yi m = 1 ˆλ1ˆγ1 + ˆλ2ˆγ2, ˆαi, ˆβi is the distribution defined by plugging into eq. (3) the parameter found via MLE. Although designed for model merging, ˆZmp-IRT m inherits certain limitations of P-IRT (Polo et al., 2024), such as non-uniform weighting and imperfect IRT fits. To mitigate these, we define a generalized estimator that interpolates between ˆZmp-IRT m and the observed correctness on D: ˆZgmp-IRT m = c X i D wi ˆYi m + (1 c) ˆZmp-IRT m (5) where c is a heuristic scalar and wi are uniform per-sample weights. We discuss in appendix C.2.3 the optimal choice for c. Although model merging can sometimes degrade performance due to weight interference suggesting nonlinear ability interactions our assumption is empirically supported as we are interested only in evolved models that show a positive performance gain. As validated in our experiments ( 4.1), our custom estimators, designed around this assumption, outperform standard IRT estimators. 3.3. Evolve: Evolutionary Search The final step of our algorithm frames model merging as a multi-objective optimization problem. Each merging objective F( m, Di) represents the performance of the merged model m on task i. In practice, we select a multi-objective evolutionary algorithm (e.g., NSGA-II (Deb et al., 2002)) and a merging strategy (e.g., TIES (Yadav et al., 2023)), aiming to optimize the corresponding Pareto front, formally defined as: PF D(Θ) = θi Θ : θj Θ s.t. θj θi Truthful QA Winogrande Truthful QA Winogrande γ{p,gp} IRT γ{mp,gmp} IRT (ours) Figure 4. Ability Estimator: Cosine similarity between estimated and true abilities for different tasks (higher is better). Our estimated abilities γ{mp,gmp} IRT better approximate true abilities. where denotes Pareto-dominance. A model m Paretodominates m if: F F D : F(m; D) F(m ; D) F F D : F(m; D) < F(m ; D) This means m is strictly better in at least one metric and no worse in all others. Models on the Pareto front are thus not dominated by any other model. In our setting, to reduce computational costs, we approximate optimization using F D instead of F D, where D D is obtained by the extraction step. Performance on D is then estimated using the performance estimator. 4. Experiments In this section, we evaluate MERGE3, demonstrating its effectiveness in evolutionary model merging on consumer- Efficient Evolutionary Merging on Consumer-grade GPUs 0.06 0.07 0.08 (a) To Romanian. 0.11 0.13 0.14 0.19 0.20 0.24 (b) To German. 0.02 0.02 0.03 (c) To Dutch. Figure 5. Cross-lingual skill transfer: merging math models (dark blue) with language-specific models (red) effectively transfers mathematical skills across languages (green - our method) compared to baselines (white). Accuracy on GSM8K for each target language. Shi SA-Gamma-7B Arithmo2-7B Task Arithmetic Evo LLM-JP-7B Figure 6. Accuracy of merged models for Japanese GSM8K. grade GPUs. We first validate the proposed ability and performance estimators, assessing their accuracy in approximating full-dataset evaluations. Next, we examine crosslingual transfer, where MERGE3 enables efficient merging of multilingual models, improving mathematical reasoning across languages. Thereafter, we evaluate its ability to synthesize multilingual models, surpassing individual finetuned baselines while remaining computationally efficient. Finally, we analyze the performance of MERGE3 on different GPUs. All the merging experiments were performed with our custom-made library Mergenetic (see Appendix A) on a RTX 4090 GPU featuring 24 GB of VRAM, while employing a batch size of 8, 4-bit quantization, and models comprising 7 billion parameters (see Appendix B). 4.1. Validating Estimators In this section, we empirically validate our mergedperformance estimators by comparing them against standard P-IRT and GP-IRT estimators (Polo et al., 2024) across five benchmark datasets: GSM8K (Cobbe et al., 2021), Winogrande (Sakaguchi et al., 2021), Truthful QA (Lin et al., 2022), Hellaswag (Zellers et al., 2019), and ARC (Clark et al., 2018). Due to space limitations, additional results are provided in Appendix C. Ability Estimators. To validate our ability estimators we compare their inferred latent ability vectors to the reference ground-truth vectors Γ. Specifically, we measure the cosine similarity and the Euclidean distance from the ground-truth Γ both for γ{mp,gmp} IRT, estimated with our merged-performance IRT approaches, and γ{p,gp} IRT, estimated with the P-IRT and GP-IRT estimators (Polo et al., 2024). Here, Γm is computed by fitting the IRT model (as in section 3.1) to each merged model m using its entire set of responses on the full dataset D. Incorporating all available data, Γm serves as our best proxy for the model s true ability. Conversely, both γ{mp,gmp} IRT m and γ{p,gp} IRT m are estimated using only a smaller subset D D of size n. Figure 4 shows the results of this comparison for n = 10 and n = 20, while the results for n = 15, 30, 50, 100 are reported in Appendix C.2 along with the same experiment over different languages. Across all five benchmark tasks our proposed ability estimator γ{mp,gmp} IRT m consistently yields ability vectors with higher cosine similarity to Γ than γ{p,gp} IRT m . This trend is evident across both subset sizes, highlighting the robustness of our approach even with limited data. The superior performance of γ{mp,gmp} IRT m empirically validates Assumption 1, confirming that an IRTbased ability estimator designed around this assumption provides more accurate ability estimates than a generalpurpose alternative. Performance Estimators. To assess the accuracy of our proposed performance estimators, we measure their absolute estimation error across different sample sizes. Specifically, we evaluate the performance estimates of six merged models using random sampling, P-IRT, GP-IRT (Polo et al., 2024), MP-IRT, and GMP-IRT across various subset sizes. The resulting absolute errors shown in Figure 3 are reported for ARC, GSM8K, Truthful QA, and an aggregate average across all five benchmarks. As shown in the figure, our proposed estimators, MPIRT and GMP-IRT, consistently achieve lower absolute error compared to GP-IRT and P-IRT. While all IRT-based Efficient Evolutionary Merging on Consumer-grade GPUs methods outperform random sampling, the incorporation of merged-performance IRT significantly enhances estimation accuracy. Notably, both MP-IRT and GMP-IRT maintain low empirical error and reduced variance even when operating with very small subsets (| D| 1.5% of the full dataset). This highlights the robustness of our approach in low-data regimes. Since lower empirical error often correlates with reduced expected error (as formalized in Section 5), we adopt MPIRT and GMP-IRT as our primary estimators for evolving merged language models in subsequent experiments. 4.2. Cross-Lingual Transfer of Mathematical Skills To assess the transfer of mathematical reasoning from English to other languages, we merge an English mathspecialized model with a Mistral-7B (Jiang et al., 2023) fine-tuned on each target language, then evaluate on the corresponding GSM8K translations (Cobbe et al., 2021). Appendix B.2 provides details on the specific models used for merging. Following Akiba et al. (2025), we label an answer correct only if it is both accurate and written in the target language. We benchmark our approach against three commonly used merging baselines Task Arithmetic (Ilharco et al., 2022), TIES (Yadav et al., 2023) and DARE (Yu et al., 2024). Following standard practice in the merging community, we apply either TIES and DARE jointly or SLERP (Shoemake, 1985). As shown in fig. 5, merging a language-specific fine-tuning with a math-specialized model consistently surpasses both endpoint models by 10 20% in accuracy on the translated GSM8K. In contrast, standard baselines often yield sub-optimal merges, performing worse than the endpoints themselves. This highlights the importance of optimized merging coefficients and motivates our evolutionary framework. Next, we evaluate our method for transferring math skills from English to Japanese and compare it to Evo Merge (Akiba et al., 2025), which serves as an upper bound by computing fitness on the full dataset. As illustrated in figure 6, our approach confirms the significant gains seen for the other languages, greatly surpassing both the performance of the endpoint models and that of the merging baselines. While the accuracy is lower than that of the model obtained by computing the fitness on the full dataset as done by Akiba et al. (2025), figure 1 shows that our approximation yields a method that is 50 more efficient, effectively making evolutionary merging feasible on a single consumer GPU. Mistral-7B-v0.1 Figure 7. Accuracy of the base model (Mistral-7B), the Italian Endpoint (IT), Self-Merge and MERGE3 models on the Italiantranslated version of GSM8k. 4.3. Ablation Study: Self-Merging In this section, we present an ablation study to test whether the observed improvements in the merged models arise from genuine cross-lingual knowledge transfer or merely from fitting to the prompt template. To structure this analysis, we formalize our inquiry through two hypotheses: Null Hypothesis (H0). The improvements seen in the merged models are due to the model fitting itself on the prompt template, rather than any cross-lingual knowledge exchange. Alternative Hypothesis (H1). The improvements arise from actual cross-lingual knowledge transfer and are not merely the result of fitting the prompt template. To evaluate these hypotheses, we propose a self-merging procedure. Concretely, we take the linguistic model and merge it with itself using the standard MERGE3 methodology outlined in algorithm 1. Under H0, if the improvements are solely due to the prompt template, merging the model with itself should lead to performance gains (i.e., the merged model would still fit the template). Conversely, under H1, if cross-lingual knowledge transfer is responsible for the enhanced performance, self-merging should not yield improvements. In fact, additional noise could even degrade performance relative to the baseline. We conducted this self-merging experiment on the Italian model using the GSM8K dataset. The results, shown in fig. 7, reveal that performance actually decreases when the model is merged with itself. This observation strongly supports the alternative hypothesis (H1): the performance gains in cross-lingual merges indeed stem from genuine knowledge transfer, rather than mere adaptation to a prompt template. 4.4. Evolving a Multilingual model We next combine individually fine-tuned models for {IT, EN, DE, NL} into a single multilingual model. Ap- Efficient Evolutionary Merging on Consumer-grade GPUs Table 1. Evolving a multilingual model. For each language, we report the accuracy on the corresponding translated ARC of both the language-specific model and the evolved multilingual model. Model Accuracy ( ) Italian English German Dutch Finetuned 0.61 0.75 0.61 0.50 MERGE3 0.69 ( 8%) 0.79 ( 4%) 0.72 ( 11%) 0.69 ( 19%) pendix B.2 provides details on the specific models used for each language. As shown in table 1, the resulting merged model surpasses each language-specific variant by up to 19% in accuracy on the ARC-Challenge dataset (Clark et al., 2018). Even more notably, it outperforms all its constituent endpoints, demonstrating a clear positive transfer of knowledge across languages. Beyond the clear accuracy boosts in each language, a few key insights stand out. First, the largest improvement occurs for Dutch (from 50% to 69%), suggesting that merging particularly benefits languages where the baseline performance is lower. Second, even English, which starts from the highest baseline, still gains by 4%, indicating that positive transfer is not limited to low-resource or weaker endpoints. Finally, the fact that the merged model outperforms all individual fine-tunings (rather than landing between them) points to a genuine cross-lingual synergy, wherein knowledge from each language-specific model collectively strengthens the multilingual result. These conclusions are further strengthened by the ablation study in section 4.3, where we assess whether the observed improvements in the merged models arise from genuine cross-lingual knowledge transfer. 5. Theoretical Analysis In this section, we provide theoretical guarantees for our performance estimator, demonstrating that its estimated accuracy is a reliable approximation of full-dataset accuracy. We provide formal guarantees for its performance, analyze its stability under dataset reduction, and explain why it remains a robust proxy for the true fitness of the merged models. This analysis not only solidifies the estimator s theoretical foundation but also offers practical insights into its behavior in finite-data and asymptotic regimes. The section is structured as follows: first ( 5.1), we derive a correlation between the accuracy of the performance estimator and the quality of the minimum found by solving an optimization problem using that performance estimator as objective function; second ( 5.2), we study the asymptotic properties of the performance estimator as the dataset size approaches infinity, formalizing it as an unbiased estimator; and finally ( 5.3), we demonstrate that our performance estimator behaves in expectation within a ϵ-bound of the accuracy on the true optimum dataset. The proofs for all the theorems and propositions presented below are outlined in appendix D. 5.1. Part I: ϵ-Stable Estimators and ϵ-Optimality Preservation We first consider a performance metric F(θ; D) for θ Θ Rn, where D is a dataset. If we choose a smaller subset D D to approximate this metric, denoted F(θ; D), we wish to control the loss in optimality incurred by replacing F(θ; D) with F(θ; D). Definition 1 (ϵ-Stability.). Given two datasets D and D, we say F( ; D) is ϵ-stable with respect to F( ; D) if, for all θ Θ, F(θ; D) F(θ; D) ϵ Under this condition, minimizing F( ; D) yields an objective value within ϵ of minimizing F( ; D). Formally: Theorem 2 (ϵ-Optimality Preservation). Let D be a dataset, let D D be a subset, and let F( ; D) be ϵ-stable with respect to F( ; D), with a fixed ϵ > 0. Define θ = argmin θ Θ F(θ; D) and ˆθ = argmin θ Θ F(θ; D) Then F(θ ; D) F(ˆθ; D) ϵ Thus, ϵ-stability ensures that any global minimizer on D achieves an objective value on D no worse than ϵ from the true global optimum. Nevertheless, uniformly bounding F(θ; D) F(θ; D) for all θ may be too strong in practice. For this reason, we introduce: Definition 3 (ϵ-Stability in expectation). Given two datasets D and D, we say F( ; D) is ϵ-stable in expectation with respect to F( ; D) if E D F(θ; D) F(θ; D) ϵ where the expectation is over the (random) choice of D Under this relaxed notion, we still obtain a similar control on the expected suboptimality gap: Theorem 4 (Expected ϵ-Stability of the Minimum). Suppose F( ; D) is ϵ-stable in expectation with respect to F( ; D). Let m := min θ Θ F(θ; D) and bm( D) := min θ Θ F(θ; D) Then m E D bm( D) ϵ Hence, even if stability only holds on average, the expected gap between the global optimum on D and the optimum on D remains at most ϵ. Efficient Evolutionary Merging on Consumer-grade GPUs 5.2. Part II: Theoretical Guarantees for MP-IRT We now apply these ideas to our proposed MP-IRT estimator (cf. 3.1). We first show that MP-IRT is asymptotically unbiased, and then combine this fact with Theorem 4 to argue that MP-IRT-based minimizers remain close to those that minimize the full-dataset performance measure. Asymptotic unbiasedness. The following proposition establishes that, as D grows, ˆZmp-IRT converges in probability to the true performance Z. Its proof relies on classical limit arguments for unbiased estimators. Proposition 5 (Asymptotic unbiasedness of MP-IRT). Assume: (i) ˆλ λ in probability as |ˆI| , (ii) for each i I, the true values αi, βi, θ1, θ2 are known, with supi I αi 2 c for a fixed c, (iii) linear inheritance of abilities (cf. Assumption 1) holds. Then, for all j, l, E ˆZjl Yi0l, . . . , Yikl E Zjl Yi0l, . . . , Yikl 0 in probability as |ˆI| . Thus, for sufficiently large subsets D, the discrepancy between ˆZ m and Z m can be made arbitrarily small with high probability. 5.3. Part III: performance preservation via MP-IRT We now conclude that MP-IRT preserves near-optimality when we train on a suitably large D D. Since Proposition 5 asserts that ˆZ approximates Z well for large | D|, it follows (under mild conditions) that MP-IRT remains ϵstable in expectation. Hence, Theorem 4 shows that minimizing ˆZ on D yields, on average, a solution within ϵ of the full-dataset optimum. Theorem 6 (Asymptotic performance preservation of MP-IRT). Let D D be a random subset used to compute ˆZmp-IRT. Suppose that, as | D| , ˆZmp-IRT converges in probability to Z (the true performance on D), and that ˆZmp-IRT is ϵ-stable in expectation for sufficiently large | D|. Then the expected global optimum of ˆZmp-IRT on D differs from that of Z on D by at most ϵ. As | D| , ϵ 0. Finite-sample analysis via the Law of Large Numbers. In practice, we rarely have | D| . Instead, one can appeal to expected ϵ-stability (Theorem 4) and then estimate the corresponding expectation empirically. For instance, one may draw multiple subsets D1, . . . , DS at random from D and compute F(θ; D) F(θ; Ds) as an empirical approximation to E D |F(θ; D) F(θ; D)| . By the Law of Large Numbers, if this empirical Table 2. Comparison of Evolve methods by number of trials, estimated total time on a single NVIDIA 4090, sample size used for Fitness computation, and final accuracy on GSM8K. The number of trials is the result of population size iterations, parameters of the Genetic Algorithms of each method, and represents the total number of merged models evaluated during the entire Evolve run. Method Nmodels Estimated total time Sample size Accuracy Evo LLM-JP-7B 1000 62 days 1000 0.49 MERGE3 100 175 21h 100 0.42 MERGE3 50 175 12h 20m 50 0.38 MERGE3 30 175 10h 30m 30 0.38 MERGE3 20 175 10h 15m 20 0.34 average remains small (say, ϵ), then the true expectation is also small. Consequently, Theorem 4 implies that the optimal solution on each Ds is within ϵ of the global optimum on D, on average. Conclusion. In summary, MP-IRT inherits asymptotic consistency from p-IRT while requiring only a subset D D. By showing it is ϵ-stable (in expectation) for large | D|, we conclude that optimizing on D yields (on average) a solution close to the true optimum on D. In finite-sample regimes, multiple random draws of D can be used to empirically verify that the discrepancy remains small, thereby justifying the practical use of MP-IRT on moderately sized subsets. 6. Technical Details We summarize the GPU timing results for MERGE3 in table 2, comparing evaluation and merge times across different hardware setups. These findings highlight the practical feasibility of our approach even on older GPUs. For additional experimental details refer to appendix C.3.3. 7. Conclusions We introduced MERGE3, an evolutionary merging framework that makes high-quality model merging feasible on a single consumer GPU. By combining a subset-based approach with IRT-driven performance estimation, MERGE3 reduces merging costs by up to fifty-fold compared to prior methods without sacrificing the quality of the merged model. Our experiments demonstrate successful cross-lingual transfer in mathematics (e.g., from English to Japanese), as well as the synthesis of new multilingual models that outperform each of their language-specific endpoints. Overall, MERGE3 expands the practical reach of evolutionary merging, allowing everyday practitioners to benefit from advanced multi-task and multilingual model compositions at a fraction of the usual computational cost. Efficient Evolutionary Merging on Consumer-grade GPUs Impact Statement The introduction of MERGE3 provides an efficient and accessible method for evolutionary model merging on consumer-grade GPUs. By combining dataset reduction techniques and Item Response Theory (IRT)-based performance estimations, MERGE3 significantly lowers computational requirements while maintaining competitive performance. This enables researchers and developers to synthesize high-quality multilingual and cross-lingual models without requiring cluster-scale hardware. The open-source release of MERGE3 aims to make evolutionary model merging widely accessible, fostering further innovation in resource-constrained environments. With applications in multilingual NLP and low-resource language modeling, MERGE3 addresses practical challenges in the field, offering an efficient solution for creating state-of-theart models on standard hardware. ACKNOWLEDGMENTS We acknowledge ISCRA for awarding this project access to the LEONARDO supercomputer, owned by the Euro HPC Joint Undertaking, hosted by CINECA (Italy). We also acknowledge support from Sapienza University of Rome, under project BEAT (Better d Eep le Arning securi Ty) and through the Seed of ERC grant MINT.AI . Ainsworth, S., Hayase, J., and Srinivasa, S. Git Re Basin: Merging models modulo permutation symmetries. In The Eleventh International Conference on Learning Representations, 2022. Akiba, T., Shing, M., Tang, Y., Sun, Q., and Ha, D. Evolutionary optimization of model merging recipes. Nature Machine Intelligence, Jan 2025. ISSN 2522-5839. doi: 10.1038/s42256-024-00975-8. URL https://doi. org/10.1038/s42256-024-00975-8. Alves, D. M., Pombal, J., Guerreiro, N. M., Martins, P. H., Alves, J., Farajian, A., Peters, B., Rei, R., Fernandes, P., Agrawal, S., Colombo, P., de Souza, J. G. C., and Martins, A. Tower: An open multilingual large language model for translation-related tasks. In First Conference on Language Modeling, 2024. URL https: //openreview.net/forum?id=EHPns3h Vkj. An, X. and Yung, Y.-F. Item response theory: What it is and how you can use the irt procedure to apply it. SAS Institute Inc, 10(4):364 2014, 2014. Blank, J. and Deb, K. Pymoo: Multi-objective optimization in python. IEEE Access, 8:89497 89509, 2020. ISSN 2169-3536. doi: 10.1109/access.2020. 2990567. URL http://dx.doi.org/10.1109/ ACCESS.2020.2990567. Brzezi nska, J. Item response theory models in the measurement theory. Communications in Statistics-Simulation and Computation, 49(12):3299 3313, 2020. B ack, T. and Schwefel, H.-P. An overview of evolutionary algorithms for parameter optimization. Evolutionary Computation, 1(1):1 23, 1993. doi: 10.1162/evco.1993. 1.1.1. Cai, L., Choi, K., Hansen, M., and Harrell, L. Item response theory. Annual Review of Statistics and Its Application, 3:297 321, 2016. Clark, P., Cowhey, I., Etzioni, O., Khot, T., Sabharwal, A., Schoenick, C., and Tafjord, O. Think you have solved question answering? try arc, the AI2 reasoning challenge. Co RR, abs/1803.05457, 2018. URL http://arxiv.org/abs/1803.05457. Cobbe, K., Kosaraju, V., Bavarian, M., Chen, M., Jun, H., Kaiser, L., Plappert, M., Tworek, J., Hilton, J., Nakano, R., et al. Training verifiers to solve math word problems. ar Xiv preprint ar Xiv:2110.14168, 2021. Crisostomi, D., Fumero, M., Baieri, D., Bernard, F., and Rodol a, E. c2m3: Cycle-consistent multi-model merging. In Advances in Neural Information Processing Systems, volume 37, 2025. Dasgupta, D. and Michalewicz, Z. Evolutionary algorithms an overview. Evolutionary algorithms in engineering applications, pp. 3 28, 1997. Davari, M. and Belilovsky, E. Model breadcrumbs: Scaling multi-task model merging with sparse masks. In European Conference on Computer Vision, pp. 270 287. Springer, 2025. Deb, K., Pratap, A., Agarwal, S., and Meyarivan, T. A fast and elitist multiobjective genetic algorithm: Nsga-ii. IEEE Transactions on Evolutionary Computation, 6(2): 182 197, 2002. doi: 10.1109/4235.996017. Deb, K., Sindhya, K., and Okabe, T. Self-adaptive simulated binary crossover for real-parameter optimization. In Proceedings of the 9th Annual Conference on Genetic and Evolutionary Computation, GECCO 07, pp. 1187 1194, New York, NY, USA, 2007. Association for Computing Machinery. ISBN 9781595936974. doi: 10. 1145/1276958.1277190. URL https://doi.org/ 10.1145/1276958.1277190. Eiben, A. and Smith, J. Introduction to Evolutionary Computing. Springer, 2015. Efficient Evolutionary Merging on Consumer-grade GPUs Gao, L., Tow, J., Abbasi, B., Biderman, S., Black, S., Di Pofi, A., Foster, C., Golding, L., Hsu, J., Le Noac h, A., Li, H., Mc Donell, K., Muennighoff, N., Ociepa, C., Phang, J., Reynolds, L., Schoelkopf, H., Skowron, A., Sutawika, L., Tang, E., Thite, A., Wang, B., Wang, K., and Zou, A. A framework for few-shot language model evaluation, 07 2024. URL https://zenodo.org/ records/12608602. Gargiulo, A. A., Crisostomi, D., Bucarelli, M. S., Scardapane, S., Silvestri, F., and Rodol a, E. Task singular vectors: Reducing task interference in model merging. In Proc. CVPR, 2025. Goddard, C., Siriwardhana, S., Ehghaghi, M., Meyers, L., Karpukhin, V., Benedict, B., Mc Quade, M., and Solawetz, J. Arcee s Merge Kit: A toolkit for merging large language models. In Dernoncourt, F., Preot iuc-Pietro, D., and Shimorina, A. (eds.), Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing: Industry Track, pp. 477 485, Miami, Florida, US, November 2024. Association for Computational Linguistics. doi: 10.18653/v1/2024.emnlp-industry. 36. URL https://aclanthology.org/2024. emnlp-industry.36/. Ilharco, G., Ribeiro, M. T., Wortsman, M., Gururangan, S., Schmidt, L., Hajishirzi, H., and Farhadi, A. Editing models with task arithmetic. The Eleventh International Conference on Learning Representations, 2022. Jiang, A., Sablayrolles, A., Mensch, A., Bamford, C., Chaplot, D., de las Casas, D., Bressand, F., Lengyel, G., Lample, G., Saulnier, L., Lavaud, L., Lachaux, M.-A., Stock, P., Le Scao, T., Lavril, T., Wang, T., Lacroix, T., and El Sayed, W. Mistral 7b. ar Xiv preprint ar Xiv:2310.06825, 2023. URL https://arxiv. org/abs/2310.06825. Jordan, K., Sedghi, H., Saukh, O., Entezari, R., and Neyshabur, B. REPAIR: REnormalizing permuted activations for interpolation repair. In The Eleventh International Conference on Learning Representations, January 2023. Joulin, A., Grave, E., Bojanowski, P., and Mikolov, T. Bag of tricks for efficient text classification. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume 2, Short Papers, pp. 427 431. Association for Computational Linguistics, April 2017. Kingston, N. M. and Dorans, N. J. The feasibility of using item response theory as a psychometric model for the gre aptitude test. ETS Research Report Series, 1982(1): i 148, 1982. Lalor, J. P. and Rodriguez, P. py-irt: A scalable item response theory library for python. INFORMS Journal on Computing, 35(1):5 13, 2023. Lalor, J. P., Wu, H., and Yu, H. Building an evaluation scale using item response theory. In Proceedings of the Conference on Empirical Methods in Natural Language Processing. Conference on Empirical Methods in Natural Language Processing, volume 2016, pp. 648. NIH Public Access, 2016. Lin, S., Hilton, J., and Evans, O. Truthful QA: Measuring how models mimic human falsehoods. In Muresan, S., Nakov, P., and Villavicencio, A. (eds.), Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 3214 3252, Dublin, Ireland, May 2022. Association for Computational Linguistics. doi: 10.18653/v1/2022.acl-long. 229. URL https://aclanthology.org/2022. acl-long.229/. Lord, F., Novick, M., and Birnbaum, A. Statistical theories of mental test scores. 1968. Matena, M. and Raffel, C. Merging models with fisherweighted averaging. Minut, A. R., Mencattini, T., Santilli, A., Crisostomi, D., and Rodol a, E. Mergenetic: a simple evolutionary model merging library, 2025. URL https://arxiv.org/ abs/2505.11427. Ortiz-Jimenez, G., Favero, A., and Frossard, P. Task arithmetic in the tangent space: Improved editing of pretrained models. Advances in Neural Information Processing Systems, 36, 2024. Pe na, F. A. G., Medeiros, H. R., Dubail, T., Aminbeidokhti, M., Granger, E., and Pedersoli, M. Re-basin via implicit sinkhorn differentiation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 20237 20246, 2023. Petersen, N. S. et al. Using item response theory to equate scholastic aptitude test scores. 1982. P etrowski, A. and Ben-Hamida, S. Evolutionary algorithms. John Wiley & Sons, 2017. Polo, F. M., Weber, L., Choshen, L., Sun, Y., Xu, G., and Yurochkin, M. tinybenchmarks: evaluating llms with fewer examples. In Forty-first International Conference on Machine Learning, 2024. Real, E., Aggarwal, A., Huang, Y., and Le, Q. V. Regularized evolution for image classifier architecture search. In Proceedings of the aaai conference on artificial intelligence, volume 33, pp. 4780 4789, 2019. Efficient Evolutionary Merging on Consumer-grade GPUs Sakaguchi, K., Bras, R. L., Bhagavatula, C., and Choi, Y. Winogrande: An adversarial winograd schema challenge at scale. Communications of the ACM, 64(9):99 106, 2021. Shoemake, K. Animating rotation with quaternion curves. SIGGRAPH Comput. Graph., 19(3):245 254, July 1985. ISSN 0097-8930. doi: 10.1145/325165.325242. URL https://doi.org/10.1145/325165.325242. Stoica, G., Bolya, D., Bjorner, J. B., Ramesh, P., Hearn, T., and Hoffman, J. Zipit! merging models from different tasks without training. In The Twelfth International Conference on Learning Representations. Thellmann, K., Stadler, B., Fromm, M., Buschhoff, J. S., Jude, A., Barth, F., Leveling, J., Flores-Herr, N., K ohler, J., J akel, R., and Ali, M. Towards cross-lingual llm evaluation for european languages, 2024. Van der Linden, W. J. Handbook of item response theory: Three volume set. CRC Press, 2018. Vania, C., Htut, P. M., Huang, W., Mungra, D., Pang, R. Y., Phang, J., Liu, H., Cho, K., and Bowman, S. R. Comparing test sets with item response theory. In Zong, C., Xia, F., Li, W., and Navigli, R. (eds.), Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 1141 1158, Online, August 2021. Association for Computational Linguistics. doi: 10.18653/v1/2021.acl-long.92. URL https: //aclanthology.org/2021.acl-long.92/. Vincent, A. M. and Jidesh, P. An improved hyperparameter optimization framework for automl systems using evolutionary algorithms. Scientific Reports, 13(1):4737, 2023. Wang, K., Dimitriadis, N., Ortiz-Jimenez, G., Fleuret, F., and Frossard, P. Localizing task information for improved model merging and compression. In Salakhutdinov, R., Kolter, Z., Heller, K., Weller, A., Oliver, N., Scarlett, J., and Berkenkamp, F. (eds.), Proceedings of the 41st International Conference on Machine Learning, volume 235 of Proceedings of Machine Learning Research, pp. 50268 50287. PMLR, 21 27 Jul 2024. URL https://proceedings.mlr. press/v235/wang24k.html. White, C., Zela, A., Ru, R., Liu, Y., and Hutter, F. How powerful are performance predictors in neural architecture search? Advances in Neural Information Processing Systems, 34, 2021. Wortsman, M., Ilharco, G., Gadre, S. Y., Roelofs, R., Gontijo-Lopes, R., Morcos, A. S., Namkoong, H., Farhadi, A., Carmon, Y., Kornblith, S., and Schmidt, L. Model soups: averaging weights of multiple finetuned models improves accuracy without increasing inference time. In Chaudhuri, K., Jegelka, S., Song, L., Szepesvari, C., Niu, G., and Sabato, S. (eds.), Proceedings of the 39th International Conference on Machine Learning, volume 162 of Proceedings of Machine Learning Research, pp. 23965 23998. PMLR, 17 23 Jul 2022. URL https://proceedings.mlr. press/v162/wortsman22a.html. Yadav, P., Tam, D., Choshen, L., Raffel, C. A., and Bansal, M. Ties-merging: Resolving interference when merging models. In Oh, A., Naumann, T., Globerson, A., Saenko, K., Hardt, M., and Levine, S. (eds.), Advances in Neural Information Processing Systems, volume 36, pp. 7093 7115. Curran Associates, Inc., 2023. Yang, E., Wang, Z., Shen, L., Liu, S., Guo, G., Wang, X., and Tao, D. Adamerging: Adaptive model merging for multi-task learning. In The Twelfth International Conference on Learning Representations. Yang, E., Shen, L., Guo, G., Wang, X., Cao, X., Zhang, J., and Tao, D. Model merging in llms, mllms, and beyond: Methods, theories, applications and opportunities. ar Xiv preprint ar Xiv:2408.07666, 2024a. Yang, E., Shen, L., Wang, Z., Guo, G., Chen, X., Wang, X., and Tao, D. Representation surgery for multi-task model merging. In Salakhutdinov, R., Kolter, Z., Heller, K., Weller, A., Oliver, N., Scarlett, J., and Berkenkamp, F. (eds.), Proceedings of the 41st International Conference on Machine Learning, volume 235 of Proceedings of Machine Learning Research, pp. 56332 56356. PMLR, 21 27 Jul 2024b. URL https://proceedings. mlr.press/v235/yang24t.html. Yu, L., Yu, B., Yu, H., Huang, F., and Li, Y. Language models are super mario: Absorbing abilities from homologous models as a free lunch. In Salakhutdinov, R., Kolter, Z., Heller, K., Weller, A., Oliver, N., Scarlett, J., and Berkenkamp, F. (eds.), Proceedings of the 41st International Conference on Machine Learning, volume 235 of Proceedings of Machine Learning Research, pp. 57755 57775. PMLR, 21 27 Jul 2024. URL https://proceedings.mlr. press/v235/yu24p.html. Zellers, R., Holtzman, A., Bisk, Y., Farhadi, A., and Choi, Y. Hella Swag: Can a machine really finish your sentence? In Korhonen, A., Traum, D., and M arquez, L. (eds.), Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp. 4791 4800, Florence, Italy, July 2019. Association for Computational Linguistics. doi: 10.18653/v1/P19-1472. URL https://aclanthology.org/P19-1472/. Efficient Evolutionary Merging on Consumer-grade GPUs Zhou, L., Solombrino, D., Crisostomi, D., Bucarelli, M. S., Silvestri, F., and Rodol a, E. Atm: Improving model merging by alternating tuning and merging. ar Xiv preprint ar Xiv:2411.03055, 2024. Zhuang, Y., Liu, Q., Ning, Y., Huang, W., Lv, R., Huang, Z., Zhao, G., Zhang, Z., Mao, Q., Wang, S., et al. Efficiently measuring the cognitive ability of llms: An adaptive testing perspective. ar Xiv preprint ar Xiv:2306.10512, 2023. Table 3. Overview of supported merging methods in Mergenetic. Method Multi-Model Uses Base Model Linear (Model Soups) Yes No SLERP No Yes Task Arithmetic Yes Yes TIES Yes Yes DARE (TIES) Yes Yes DARE (Task Arithmetic) Yes Yes A. Mergenetic Each experiment was run using a library developed specifically for this paper, which will be released as open-source software, called Mergenetic (Minut et al., 2025). This library allows for defining a merging problem as either a single-objective or multi-objective optimization problem, where one only needs to specify the merging method,a fitness function, and select the hyperparameters for a chosen evolutionary algorithm. The implementation relies on Mergekit (Goddard et al., 2024) for merging the models, Pymoo (Blank & Deb, 2020) for optimizing the objective function through evolutionary algorithms, and Lm-Evaluation-Harness (Gao et al., 2024) for implementing some of the fitness functions. In table 3 we outline the supported merging methods, while in table 4 we outline the currently available evolutionary algorithms. We believe this library is a significant contribution as it facilitates evolutionary model merging and aligns well with the paper s approach, which aims to reduce computational burden. It can be a valuable tool for the community and for users interested in cross-lingual transfer or creating multilingual models for target low-resource languages. B. Additional Details This section provides additional implementation and experimental details that were not included in the main paper. B.1. IRT Fitting Details As previously stated, we used the implementation from Polo et al. (2024) and adopted their configuration settings. Specifically, we used γm N(µγ1d, 1/uγId), αi N(µα1d, 1/uαId), and βi N(µβ, 1/uβ). Following Polo et al. (2024), we also applied (hyper)priors to the prior parameters using the software for fitting hierarchical Bayesian models (Lalor & Rodriguez, 2023): µγ N(0, 10), uγ Γ(1, 1), µα N(0, 10), uα Γ(1, 1), µβ N(0, 10), and uβ Γ(1, 1). For both the model and example-specific parameters γm, αi, and βi, we take Efficient Evolutionary Merging on Consumer-grade GPUs Table 4. Overview of supported Pymoo s evolutionary algorithms in Mergenetic. Algorithm Class Objective(s) Constraints Genetic Algorithm GA single x Differential Evolution DE single x Biased Random Key GA BRKGA single x Nelder Mead Nelder Mead single x Pattern Search Pattern Search single x CMAES CMAES single Evolutionary Strategy ES single SRES SRES single x ISRES ISRES single x NSGA-II NSGA2 multi x R-NSGA-II RNSGA2 multi x NSGA-III NSGA3 many x U-NSGA-III UNSGA3 many x R-NSGA-III RNSGA3 many x MOEAD MOEAD many AGE-MOEA AGEMOEA many C-TAEA CTAEA many x SMS-EMOA SMS-EMOA many x RVEA RVEA many x their point estimates as the means of their respective variational distributions. The γ model dimensionality is set to 15 following the parameter choice suggested by Polo et al. (2024). B.2. Experimental Details Models. One key assumption of model merging is that the endpoint models lie within the same basin (Ilharco et al., 2022). This means that merging arbitrary models is not feasible; rather, all models involved must be fine-tuned versions of the same base model. To satisfy this requirement, we selected several fine-tuned models from the Hugging Face Hub that originated from the same base model. Specifically, we focused on models fine-tuned starting from Mistral-7B (Jiang et al., 2023), following common best practices in the community (Akiba et al., 2025). Table 5 lists all the models used in our experiments, along with their corresponding names on the Hugging Face Hub. A total of 27 models were considered for our experiments. Number of models. The initialization step, shared across all evolutionary algorithms, involves sampling merging configurations (e.g., interpolation coefficients) and applying them to merge the endpoint models. Consequently, MERGE3 requires the same number of models as any standard merging approach. In the rest of the section, we provide further details for reproducing the experiments in section 4.2 and section 4.4 of the main paper. Table 5. Mistral-based models with shortened column headers and names. Role can be either E, M or B, referring to endpoint, merge or base model respectively. Spec refers instead to specialization, with mth, ger, ita, jpn, dut and gen referring to Math, German, Italian, Japanese, Dutch and General respectively. We finally have the author and model ID as per the Huggingface. Role Spec Author Model E mth upaya07 Arithmo2-Mistral-7B E mth,jpn Sakana AI Evo LLM-JP-v1-7B E mth GAIR Abel-7B-002 E mth meta-math Meta Math-Mistral-7B B gen mistralai Mistral-7B-v0.1 E ger jphme em german mistral v01 E ger Leo LM leo-mistral-hessianai-7b E ita Deep Mount00 Mistral-Ita-7b E jpn augmxnt shisa-gamma-7b-v1 E dut Bram Vanroy GEITje-7B-ultra E ro Open LLM-Ro Ro Mistral-7b-Instruct M gen chlee10 T3Q-Merge-Mistral7B E gen liminerity M7-7b E gen yam-peleg Experiment26-7B M gen Practice LLM SOLAR-tail-10.7B-Merge-v1.0 E gen upstage SOLAR-10.7B-v1.0 E gen Yhyu13 LMCocktail-10.7B-v1 M gen Fuse AI Fuse Chat-7B-Slerp M gen Fuse AI Fuse Chat-7B-TA E gen Fuse AI Open Chat-3.5-7B-Mixtral E gen Fuse AI Open Chat-3.5-7B-Solar M gen jan-hq supermario-slerp-v3 E gen jan-hq supermario-slerp-v2 E gen jan-hq supermario-v2 M gen superlazycoder Neural Pipe-7B-slerp E gen Open Pipe mistral-ft-optimized-1218 E gen mlabonne Neural Hermes-2.5-Mistral-7B B.2.1. CROSS-LINGUAL TRANSFER In the cross-lingual transfer evolutionary merging, we evolved four merged models with mathematical capabilities in different languages: Japanese, Romanian, German, and Dutch. In each of these experiments, we deployed an ad-hoc genetic algorithm for single-objective optimization. We employed the Simulated Binary Crossover (Deb et al., 2007) operator to generate offspring solutions by combining parent solutions. To maintain diversity and explore the search space, we applied Polynomial Mutation (Deb et al., 2007), which introduces small perturbations to offspring solutions and enhances the algorithm s ability to escape local optima. This combination of SBX and PM effectively balances exploration and exploitation, facilitating efficient convergence toward optimal solutions. Furthermore, guided by empirical tests, we decided to deploy SLERP to evolve solutions for the Romanian and Dutch problems, while we used a combination of TIES and DARE for the Japanese and the German ones. We deployed four different sizes of the fitness datasets for Japanese, namely 20, 30, 50, and 100, in order to obtain a more detailed analysis of the method for comparison with the work of (Akiba et al., 2025). On the other hand, we kept the fitness dataset size fixed to 20 for all other aforemen- Efficient Evolutionary Merging on Consumer-grade GPUs tioned experiments. The fitness dataset was extracted from the test set of GSM8K, and we used the remaining, nonoverlapping samples as test set for evaluating the model. To get the language-specific versions of GSM8K, we used Unbabel/Tower Instruct-7B-v0.2 (Alves et al., 2024) to translate the datasets. In each experiment, the population size was fixed to 25 and the number of iterations to 7. To check the correctness of the solution, following Akiba et al. (2025), we used a regex to extract the last numerical value returned in the model s answer and compare it with the ground truth. The solution is also checked to be in the correct language with the language identifier from fast Text (Joulin et al., 2017). The mathematical models used in combination with TIES and DARE were Abel-7B-002 and Arithmo2-Mistral-7B, whereas we used Meta Math-Mistral-7B in combination with SLERP. Moreover, we employed the following language-specialized models: shisa-gamma-7b-v1, em german mistral v01, GEITje-7B-ultra, and Ro Mistral-7b-Instruct. More information about these models can be found in table 5. Lastly, we evaluated Evo LLM-JP-v1-7B (Akiba et al., 2025) under the same conditions as MERGE3 to assess its accuracy, following the prompting structure outlined by Akiba et al. (2025). B.2.2. MULTI-LINGUAL TRANSFER In this experiment, we tackle the ARC dataset in multiple languages (Italian, Dutch, German, and English)2 (Thellmann et al., 2024) using a multi-objective evolutionary merging procedure based this time on NSGA-II (Deb et al., 2002). We configure the population size to 25 and the number of evolutionary iterations to 7. We deployed a combination of TIES and DARE as merging strategy. As in previous settings, both the fitness function and the test metrics operate by extracting the final model-generated choice via a regex, but this time they look for an instance from the set {A, B, C, D} rather than a number. On top of this, we employed a dataset composed by 20 datapoints for each language from the relative translation of ARC to compute the fitness, and we extracted the test set as for the previous experiments. Furthermore, unlike the single-objective approach described earlier, here we explicitly optimize multiple objectives simultaneously. This time, the employed models are Mistral-Ita-7b, GEITje-7B-ultra, leo-mistral-hessianai-7b, and the base model Mistral-7B-v0.1. 2We used the dataset on the Hugging Face Hub from open GPT-X/arcx Table 6. Notation used in the paper. Notation Description D Full dataset. D Reduced subset of the dataset. Di Subdataset for task i. γm Latent abilities of model m. Γm True latent abilities of model m. γ{p,gp} IRT m Latent abilities of model m via P-IRT ability estimator. γ{mp,gmp} IRT m Latent abilities of model m via MP-IRT ability estimator. αi, βi IRT parameters related to dataset item i. λ Interpolation coefficients for latent abilities. ˆλ, ˆγ, ˆα, ˆβ MLE of the aforementioned parameters. pi,m IRT model for datapoint i and model m. ˆpi,m IRT model for datapoint i and model m parametrized by MLE estimators of α, β, γ, λ. m Merged language model. Yi,m Sample-level correctness of model m for example i. ˆZ MP-IRT Merged performance estimator MP-IRT. ˆZ GMP-IRT Generalized merged performance estimator GMP-IRT. F(m) Fitness value of a model m. θ Parameters being optimized in evolutionary search. P FD Pareto front defined by function s set F and data D θ Global optimum on D. ˆθ Global optimum on D. N Number of samples in the dataset. B.2.3. ABILITY AND PERFORMANCE ESTIMATOR In these experiments (reported in section 4.1 and section 4.1) we used the test set of the standard version of GSM8K, Hella Swag, ARC, Winogrande, and Truthful QA. Furthermore, we used 6 different models to test the different performance of the ability and performance estimator: SOLAR-tail-10.7B-Merge-v1.0, Fuse Chat-7B-Slerp, Neural Pipe-7B-slerp, T3Q-Merge-Mistral7B, Fuse Chat-7B-TA, and supermario-slerp-v3. These models were chosen as already available on the Open LLM Leaderboard. C. Additional Experiments We report here additional experiments and analyses. C.1. Extract Step In the extract step outlined in section 3.1, random sampling has been proposed as the main method to subsample the dataset D D. While we explored various dataset subsampling strategies, we ultimately opted for uniform random sampling, as our experiments showed that more complex approaches offered no significant advantage over this simpler method. In this section we report some of the experiments behind this decision and the two alternative methods tried in the extraction step: IRT Clustering (IRT), introduced by Polo et al. (2024), and a custom Representation Clustering (RC) method. Efficient Evolutionary Merging on Consumer-grade GPUs C.1.1. IRT CLUSTERING Given a dataset D and the parameter of a fitted IRT model α and β, one can define a low-dimensional embedding of each datapoint i D by Ei = [αi βi]. Therefore, IRTclustering obtains a representative subset by first obtaining a clustering over this embedding space through K-Means, and then choosing the points closest to the centroids as representative samples. C.1.2. REPRESENTATION CLUSTERING Let {mj}M j=1 be the set of endpoint models, and let D = {xi}N i=1 be our full dataset. For each sample xi, we first encode it into a high-dimensional vector by concatenating model-specific embeddings. Concretely, we compute: t=1 Ei,j,t Rd, where Ei,j,t is the embedding of the t-th token of sample xi under model mj, and Ti is the number of tokens in xi. We form the concatenated representation: Ei = [Ei,1 Ei,2 Ei,M] RM d. Since Ei can be very high-dimensional, we apply Principal Component Analysis (PCA) to project Ei onto a lowerdimensional space: Ei = PCAk(Ei) Rk, k M d. Next, we apply k-means clustering to the reduced embeddings { Ei}N i=1: min {ck}K k=1 i=1 min 1 k K Ei ck 2, where ck is the centroid of the k-th cluster. This partitions the dataset into K clusters, each capturing a distinct region of the representation space. From each cluster k, we select the representative sample xi k whose embedding Ei k is closest to the centroid ck: i k = arg min xi Ck Ei ck , where Ck is the set of samples assigned to cluster k. To approximate the full-dataset metrics from the selected subset D = {xi k}K k=1, we assign a weight to each representative sample. Since the size of the cluster Ck indicates how prevalent that region of representation space is, we define wi k = |Ck| |D| . These weights ensure that the contribution of each representative sample to the overall metric reflects the true proportion of samples that it represents in the original dataset. By evaluating a new model m only on D and using {wi k} to calculate a weighted average, we approximate m s performance on the full dataset D at a fraction of the computational cost. A schematic overview of the full process is outlined in algorithm 2. Algorithm 2 Representation Clustering Extractor Require: Dataset D, Endpoint Models m1, ..., mn, Desired subset size K Ensure: Subset of size K with weights wi 1: for i in D do 2: Ei [] 3: for m in {m1, . . . , mn} do 4: Eim embed i with model m 5: Ei Ei|Eim 6: end for 7: end for 8: {Ei}i D PCA({Ei}i D) 9: Apply k-means clustering to {Ei}i D, obtaining K centroids {ck}K k=1 10: For each cluster k, select the closest example i k = arg mini D Ei ck 2 11: Let Ck = {i D | arg minc {ck}K k=1 Ei c 2 = ck} be the set of examples in cluster k 12: Assign weights wi k = |Ck| |D| for k = 1, ..., K 13: return i k, wi k K k=1 C.1.3. EXPERIMENTS To compare the performance of the Sample Extractors, we followed a procedure similar to that described in section 4.1, computing the absolute estimation error for each extractor. For random sampling, the accuracy estimator was obtained via uniform averaging, whereas for IRT and RC it was obtained via weighted averaging. We evaluated the estimator in two different settings: (1) merging a math model with a language-tuned model (similar to the crosslingual setting of section 4.2) for several languages (Italian, German, Romanian, Dutch) and testing the extractor on the corresponding translations of GSM8K (see fig. 8), and (2) merging several math models and testing the extractor on the English version of GSM8K (see fig. 9). Focusing on fig. 8, we see that performance variability is somewhat higher (larger error bars) due to different language-specific datasets. Even so, Random sampling never falls behind IRT or RC, especially for small sample sizes. By the time the subset size reaches 50 or more examples, all three methods converge to comparable accuracyerror levels, underscoring the robustness of Random sampling. Instead, in fig. 9, the trends are broadly similar for RC and Random sampling, while slightly worse for IRT. Again, as the dataset sample size grows, overall error drops and the gap among methods narrows. Efficient Evolutionary Merging on Consumer-grade GPUs 10 15 20 30 50 100 Number of Examples Accuracy Error Sampling Method Random IRT Representation Clustering Figure 8. Extractors across Languages: Absolute error of the estimated accuracy of Sample Extractors, averaged across merges of language-specific and English Math finetunings of Mistral-7B-v0.1, evaluated on translations of GSM8K and presented as a function of the number of dataset samples. 10 15 20 30 50 100 Number of Examples Accuracy Error Sampling Method Random IRT Latent Clustering Figure 9. Extractors across Merges: Absolute error of the estimated accuracy of Sample Extractors, averaged across merges of English Math models based on Mistral-7B-v0.1, evaluated on GSM8K and presented as a function of the dataset sample size. To sum up, the Random sampler can sometimes lag slightly behind the more sophisticated IRT and RC. Nevertheless, neither of these methods has a clear advantage over the others. Given its simplicity and negligible overhead, the Random strategy stands out as a highly practical choice for dataset subsampling especially when the marginal improvements of more complex methods do not clearly justify their added complexity. C.2. Estimation step C.2.1. ADDITIONAL EXPERIMENT FOR ABILITY ESTIMATOR We report in fig. 10 the Euclidean distance between the estimated and ground-truth ability vectors across different sample sizes. The results are consistent with the case n = 10, 20 seen in fig. 4, with our estimated ability vec- italian romanian italian romanian italian romanian italian romanian 10 20 30 40 50 60 italian romanian italian romanian γ{p,gp} IRT γ{mp,gmp} IRT (ours) Figure 10. Ability Estimator over languages: Euclidean distance (lower is better) between estimated and true abilities for different languages. italian romanian italian romanian 0.4 0.2 0.0 0.2 0.4 0.6 0.8 italian romanian italian romanian 0.2 0.0 0.2 0.4 0.6 0.8 italian romanian italian romanian spanish 0.2 γ{p,gp} IRT γ{mp,gmp} IRT (ours) Figure 11. Ability Estimator over languages: Cosine similarity (higher is better) between estimated and true abilities for different languages. tor being significantly closer to the ground-truth one compared to the ability vector estimated by p IRT and gp-IRT. Similarly, we report the corresponding cosine similarity in fig. 11, confirming much higher similarity in our case. C.2.2. ADDITIONAL EXPERIMENT FOR PERFORMANCE ESTIMATOR We report in fig. 13 the evaluation of performance estimators across Winogrande and Hellaswag, extending the results in fig. 3. Efficient Evolutionary Merging on Consumer-grade GPUs Truthful QA Winogrande Truthful QA Winogrande Truthful QA Winogrande γ{p,gp} IRT γ{mp,gmp} IRT (ours) Figure 12. Ability Estimator over tasks: Cosine similarity (higher is better) between estimated and true abilities for different tasks. 10 20 30 50 100 Number of Samples Absolute Error 10 20 30 50 100 Number of Samples p-IRT gp-IRT mp-IRT (ours) gmp-IRT (ours) Random Figure 13. Performance Estimators over Winogrande and Hellaswag. Absolute error of various estimators as a function of sample size (lower is better). gmp-IRT consistently achieves lower error. C.2.3. HYPERPARAMETER ANALYSIS FOR PERFORMANCE ESTIMATOR We now analyse the optimal choice of the scalar c required in GMP-IRT and GP-IRT (see eq. (5)). In the experiments reported in the main paper (section 4.1) and above (appendix C.2.2), we used as a heuristic c = 1 2. Despite its empirical effectiveness, this uniform interpolation may not be optimal across all model pairs and data regimes. Therefore, we introduce a grid-search-based strategy to estimate an improved value of c and empirically validate its potential to reduce estimation error. Methodology. We propose a two-step approach to selecting the optimal interpolation coefficient c for use in the GMP-IRT estimator. The procedure is as follows: 1. Optimizing c for Endpoint Models For a given dataset D 3, we find a c [0, 1] by minimizing the absolute estimation error of GP-IRT on each individual model. This yields a set of optimal coefficients {c1, . . . , c M}, one for each endpoint model M. 2. Averaging c for the Merged Model. To obtain a suitable interpolation coefficient for GMP-IRT, we compute the average of the optimal values across the end- 3Such a dataset is available because it relies solely on the correctness of the endpoint models answers, rather than on the unknown answers of the merged model. point models, c = 1 M PM m=1 cm, and use this as the parameter for the merged model s GMP-IRT. Results Discussion We evaluate the absolute estimation error across five benchmarks, comparing the adaptive averaging strategy for c (denoted by ) to the baseline fixed heuristic c = 1 2. Results for GMP-IRT and GP-IRT , along with the baseline models GMP-IRT, GP-IRT, MPIRT, and P-IRT, are reported in table 7. Table 7. Absolute estimation error across datasets. The symbol indicates the adaptive c strategy described above. Lower is better. Dataset GMP-IRT GMP-IRT GP-IRT GP-IRT MP-IRT ARC 0.035 0.040 0.046 0.049 0.048 Winogrande 0.018 0.031 0.032 0.037 0.036 GSM8K 0.057 0.057 0.074 0.064 0.062 Hella Swag 0.046 0.056 0.077 0.071 0.047 Truthful QA 0.040 0.045 0.062 0.055 0.044 Across most datasets, the adaptive coefficient strategy yields consistent improvements in both GMP-IRT and GPIRT, with the largest gains observed on Winogrande and Hella Swag. Notably, GMP-IRT with a tuned c performs on par or better than any other method across all benchmarks. C.2.4. SPEARMAN CORRELATION Following White et al. (2021), we report the Spearman correlation between the ground-truth ranking and the ranking induced by each estimator s predictions. We compare the best merging-specific estimator, GMP-IRT, against the strongest vanilla baseline, GP-IRT. The results, averaged across dataset sizes and types, are shown in fig. 14. Notably, GMP-IRT achieves the highest correlation in each setting, further underscoring the benefits of using estimators specifically designed for the model merging context. C.3. Evolve Step C.3.1. ADDITIONAL EXPERIMENT FOR MULTILINGUAL EVOLUTION: COMPARISON WITH IN-CONTEXT LEARNING A natural question when evaluating merging-based methods is how their performance compares to inference-time adaptation strategies such as In-Context Learning (ICL). In particular, given access to a small validation set, one might ask whether directly providing these examples as input context at evaluation time can match the performance achieved through evolutionary merging. To investigate this, we evaluate a 20-shot ICL setup in the multilingual transfer setting introduced in section 4.4. Prompts are constructed using 20 validation examples and prepended to the input at inference time. We apply this setup to two merging baselines, TIES-DARE and Task Efficient Evolutionary Merging on Consumer-grade GPUs n = 10 n = 20 n = 30 n = 50 n = 100 0.0 Spearman ρ ( better) Across Sizes ARC GSM8K Hella Swag Truthful Winogrande 0.0 Spearman ρ ( better) Across Tasks GMP-IRT GP-IRT Figure 14. Spearman rank correlations using the appendix C.2.3 setup. Higher is better. GMP-IRT shows substantially higher correlation than GP-IRT. Arithmetic, and compare the results against our method using GMP-IRT with a fitness dataset of size 20. Across all Table 8. Accuracy on multilingual ARC using few-shot ICL (20shot) versus MERGE3 (GMP-IRT-20). Method DE IT NL EN TIES-DARE Few-shot (20) 0.227 0.226 0.227 0.226 Task Arithmetic Few-shot (20) 0.427 0.406 0.491 0.566 MERGE3 (GMP-IRT-20) 0.720 0.690 0.690 0.790 languages, MERGE3 significantly outperforms the ICLaugmented baselines. While ICL can provide moderate improvements over the original models, it increases inferencetime memory usage and latency due to the expanded context. In contrast, the merged models produced by our method operate without additional overhead and are immediately deployable as standalone networks. C.3.2. ADDITIONAL EXPERIMENT FOR CROSS-LINGUAL EVOLUTION: ANALYZING NEGATIVE TRANSFER While our main analysis focused on cross-lingual transfer of MERGE3 (section 4.2), we did not explicitly examine the phenomenon of negative transfer; that is, cases where merging degrades performance on specific inputs. In this subsection, we formally define negative transfer in the context of multiple-choice questions (MCQs) and introduce a framework for measuring its prevalence in merged multilingual models. We then present an analysis of negative transfer in GSM8K. Methodology We consider a multiple-choice question (MCQ) evaluation setup, where knowledge is operationalized as a model s ability to correctly answer a question. Let m1, m2, . . . , m K denote the set of K endpoint models, and let m represent the merged model resulting from their combination. Following our earlier notation (see table 6), correctness for a given sample i is defined as a binary variable: Yi,mj {0, 1}: indicates whether endpoint model mj answers sample i correctly, Yi, m {0, 1}: indicates whether the merged model answers sample i correctly. We define negative transfer on example i as occurring when at least one of the base models answers correctly, but the merged model fails: j {1, . . . , K} such that Yi,mj = 1 and Yi, m = 0. To track this, we introduce a binary indicator variable ni for each input: ( 1, if j : Yi,mj = 1 and Yi, m = 0, 0, otherwise. Finally, we compute the Negative Transfer Rate (NTR) as the proportion of examples exhibiting negative transfer among those for which at least one base model answered correctly: NTR = PN i=1 ni PN i=1 1 j {1, . . . , K} : Yi,mj = 1 . This metric provides a task-level perspective on the potential degradation introduced by merging and complements the aggregate performance measures reported in the main paper. Results Discussion We compute the NTR using the same experimental setting described in section 4.2. As shown in fig. 15, MERGE3 consistently yields substantially lower negative transfer than SLERP, TIES, and Task Arithmetic across all languages. This indicates that MERGE3 not only improves average accuracy but also preserves correct knowledge from its component models, thereby maintaining per-example competence during merging. Efficient Evolutionary Merging on Consumer-grade GPUs Dutch Romanian German 0.0 Error ( better) MERGE3 SLERP TIES TA Figure 15. Negative Transfer Rate across languages. Lower is better. MERGE3 shows substantially less negative transfer than standard baselines. C.3.3. ADDITIONAL EXPERIMENT: TIME & FLOPS REQUIREMENTS EVOLUTIONARY MERGING Hardware Setting. To compare the efficiency of different model evaluation strategies, we measured the time required to evolve merged LLM models using a single NVIDIA 4090 with 24GB of VRAM, and report the Throughput R in table 9. We also benchmark evaluation and merging times across three GPU models (3090, 4090, V100) to illustrate practical runtimes for MERGE3 on both modern and older hardware. We report the results in table 10 Results Discussion. Over a 12-hour period, we were able to evaluate 8 models on 1000 samples of GSM8K with a single NVIDIA 4090, allowing us to estimate that evaluating 1000 models would take approximately 62 days under similar conditions. In contrast, MERGE3 enabled the evaluation of a larger number of merged models in significantly less time by using a reduced dataset. These results suggest that researchers and practitioners could leverage consumer-grade GPUs for efficient LLM merging and evaluation, making rapid experimentation of model merging methods more accessible. We report in table 2 the estimated total time of the Evolve runs, which we calculated using the following formula: T(Nmodels) = Nmodels RDataset Size Lastly, table 10 shows that MERGE3 maintains practical runtimes across a range of GPUs. While the 4090 offers the fastest evaluation, older hardware like the V100 still supports feasible experimentation, highlighting the framework s accessibility and the generalizability of results across different consumer GPUs. FLOPs Calculation. We provide a Jupyter Notebook that describes the FLOPs calculations for our experiments in the supplementary material, based on the calc-flops li- Table 9. Throughput (R) in models per hour for different sample sizes per fitness evaluation on GSM8K. These estimates are based on 12-hour Evolve runs on a single NVIDIA 4090 with 24GB of VRAM. Sample size 1000 100 50 30 20 Throughput (Models/Hour) 0.67 8.33 14.17 16.67 17.08 Table 10. Evaluation and merge time across different GPU models using Mistral-7B on 10 examples (4-bit, SLERP). GPU Model Eval Time (s) Merge Time (s) NVIDIA 3090 24GB 65 135 NVIDIA 4090 24GB 45 160 NVIDIA V100 32GB 80 220 brary4. This script has been used to estimate the FLOPs for the experiment in Figure 1. D. Mathematical proofs We outline in table 6 a scheme of the notation used throughout the paper. D.1. Proof of Theorem 2 Proof. Let m := F(θ ; D) and ˆm := F(ˆθ; D). We must show that | m ˆm| ϵ. 1. By ϵ-stability, for all θ Θ: F(θ; D) F(θ; D) ϵ. In particular, for θ = θ , F(θ ; D) F(θ ; D) ϵ. Hence F(θ ; D) F(θ ; D) ϵ and F(θ ; D) F(θ ; D) + ϵ. 2. Since ˆθ is the minimizer of F( ; D), we have F(ˆθ; D) F(θ ; D). Because θ is the minimizer of F( ; D), F(ˆθ; D) F(θ ; D). 4https://github.com/Mr Yx J/ calculate-flops.pytorch. Efficient Evolutionary Merging on Consumer-grade GPUs 3. To bound ˆm m, we can add and subtract F(θ ; D) to have ˆm m = F(ˆθ; D) F(θ ; D) + F(θ ; D) F(θ ; D) . The first term is 0 (since ˆθ is a minimizer on D), and the second term is ϵ. Hence ˆm m 0 + ϵ = ϵ. 4. Analogously, to bound m ˆm, we can rewrite m ˆm = F(θ ; D) F(ˆθ; D) + F(ˆθ; D) F(ˆθ; D) . The first term is 0 (since θ is a minimizer on D), and the second term is ϵ. Thus, m ˆm 0 + ϵ = ϵ. 5. Combining these inequalities: ϵ ˆm m ϵ = |m ˆm| ϵ. Hence F(θ ; D) F(ˆθ; D) ϵ, completing the proof. D.2. Proof of Theorem 4 Proof. By hypothesis, for every θ Θ, E D h F(θ; D) F(θ; D) i ϵ. Using Jensen s inequality for the absolute value, E D F(θ; D) F(θ; D) E D h F(θ; D) F(θ; D) i ϵ. ϵ E D F(θ; D) F(θ; D) ϵ for each fixed θ. It thus follows that E D F(θ; D) F(θ; D) + ϵ and E D F(θ; D) F(θ; D) ϵ. Consequently, min θ Θ E D F(θ; D) min θ Θ F(θ; D) + ϵ = m + ϵ, where m := minθ Θ F(θ; D). Meanwhile, by a minversus-expectation (Jensen-type) inequality, E D min θ Θ F(θ; D) min θ Θ E D F(θ; D) . E D bm( D) = E D h min θ Θ F(θ; D) i min θ Θ E D F(θ; D) m ϵ. Combining these two bounds results in m ϵ E D bm( D) m + ϵ and, therefore, m E D bm( D) ϵ. D.3. Proof of Proposition 5 Proof. We must show that E ˆZjl Yi0l, . . . , Yikl E Zjl Yi0l, . . . , Yikl 0 in probability as |ˆI| . Under the assumptions of the proposition (including linear inheritance of abilities, ˆλ λ in probability, and bounded αi ), we may bound this difference as follows: E ˆZjl Yi0l, . . . , Yikl E Zjl Yi0l, . . . , Yikl σ (ˆλ1θl1 + ˆλ2θl2) αi βi σ θ lmαi βi . Since σ is 1/4-Lipschitz on R, we have (ˆλ1θl1 + ˆλ2θl2) θlm αi i ˆIj αi 2 (ˆλ1θl1 + ˆλ2θl2) θlm 2. Since supi Ij αi 2 c, it follows that c (ˆλ1 λ1) θl1 + (ˆλ2 λ2) θl2 2 0 in probability as |ˆI| . (The last step uses ˆλ λ in probability, with θl1, θl2 fixed in Rd.) Hence ˆZjl converges in probability to Zjl, completing the proof. Efficient Evolutionary Merging on Consumer-grade GPUs D.4. Proof of Theorem 2 Proof. By Proposition 5, ˆZmp-IRT becomes arbitrarily close (in probability) to Z as | D| . Under standard regularity conditions, this implies Z(θ; D) ˆZmp-IRT(θ; D) ϵ in expectation, for all sufficiently large | D|, hence ˆZmp-IRT is ϵ-stable in expectation. Applying Theorem 4 completes the argument.