# perturbationrestrained_sequential_model_editing__7aefb3a8.pdf Published as a conference paper at ICLR 2025 PERTURBATION-RESTRAINED SEQUENTIAL MODEL EDITING Jun-Yu Ma1,2, Hong Wang1, Hao-Xiang Xu1,2, Zhen-Hua Ling1,2, Jia-Chen Gu3 1University of Science and Technology of China 2National Engineering Research Center of Speech and Language Information Processing 3University of California, Los Angeles {mjy1999,wanghong1700,nh2001620}@mail.ustc.edu.cn, zhling@ustc.edu.cn, gujc@ucla.edu Model editing is an emerging field that focuses on updating the knowledge embedded within large language models (LLMs) without extensive retraining. However, current model editing methods significantly compromise the general abilities of LLMs as the number of edits increases, and this trade-off poses a substantial challenge to the continual learning of LLMs. In this paper, we first theoretically analyze that the factor affecting the general abilities in sequential model editing lies in the condition number of the edited matrix. The condition number of a matrix represents its numerical sensitivity, and therefore can be used to indicate the extent to which the original knowledge associations stored in LLMs are perturbed after editing. Subsequently, statistical findings demonstrate that the value of this factor becomes larger as the number of edits increases, thereby exacerbating the deterioration of general abilities. To this end, a framework termed Perturbation Restraint on Upper bou Nd for Editing (PRUNE) is proposed, which applies the condition number restraints in sequential editing. These restraints can lower the upper bound on perturbation to edited models, thus preserving the general abilities. Systematically, we conduct experiments employing three editing methods on three LLMs across four downstream tasks. The results show that PRUNE can preserve general abilities while maintaining the editing performance effectively in sequential model editing. The code are available at https://github.com/mjy1111/PRUNE. 1 INTRODUCTION Despite the remarkable capabilities of large language models (LLMs), they encounter challenges such as false or outdated knowledge, and the risk of producing toxic content (Zhang et al., 2023; Peng et al., 2023; Ji et al., 2023; Huang et al., 2023). Given the high cost of retraining LLMs to address these issues, there has been a surge in focus on model editing (Dai et al., 2022; Meng et al., 2022; Mitchell et al., 2022a;b; Meng et al., 2023; Zhang et al., 2024; Hu et al., 2024; Ma et al., 2024; Jiang et al., 2025; Fang et al., 2025), which aims at updating the knowledge of LLMs cost-effectively. Existing editing methods can be roughly classified into either parameter-modifying methods (Mitchell et al., 2022a; Meng et al., 2023) that directly modify a small subset of model parameters, or parameterpreserving methods (Mitchell et al., 2022b; Yu et al., 2024) that integrate additional modules without altering the model parameters. In this paper, we study the parameter-modifying editing methods. Sequential model editing involves making successive edits to the same model over time to continuously update knowledge, as illustrated in Figure 1(a). Recent studies (Gu et al., 2024; Gupta et al., 2024b; Lin et al., 2024; Gupta et al., 2024a) indicate that parameter-modifying editing methods significantly compromise the general abilities of LLMs as the number of edits increases, such as summarization, question answering, and natural language inference. However, these studies neither provide a theoretical analysis of the bottleneck of the general abilities of the edited models, nor propose a solution to preserve these abilities in sequential editing. These affect the scalability of model editing and pose a substantial challenge to the continual learning of LLMs. Corresponding author. Published as a conference paper at ICLR 2025 𝑓𝑓𝑊𝑊1 𝑓𝑓𝑊𝑊2 𝑓𝑓𝑊𝑊𝑛𝑛 Question Answering Reasoning Summarization Sentiment Analysis Performance before editing Performance after regular editing Performance after restrained editing Edit Edit Edit Edit Prune 𝑓𝑓𝑊𝑊 𝑓𝑓𝑊𝑊𝑛𝑛 𝑓𝑓𝑊𝑊𝑛𝑛 (a) Sequential model editing (b) Condition number of matrices (c) General downstream task performance Efficacy Generalization Locality (d) Editing performance Figure 1: (a) Illustration of sequential model editing. (b) The condition number of edited matrix rapidly increases as the number of edits increases. (c) Comparison of general downstream task performance before editing, after regular editing, and after restrained editing by PRUNE. (d) Comparison of editing performance after regular editing and after restrained editing by PRUNE. f W , f Wn and f W n denote the models that are unedited, regularly edited n times, and restrainedly edited by PRUNE respectively. W is denoted as a matrix to be edited. In light of the above issues, we first theoretically analyze through matrix perturbation theory (Vaccaro, 1994; Wedin, 1972) to elucidate a crucial factor affecting the general abilities during sequential editing: the condition number (Smith, 1967; Dedieu, 1997; Sun, 2000) of the edited matrix. The condition number of a matrix represents its numerical sensitivity and therefore can be used to indicate the extent to which the original knowledge associations stored in LLMs are perturbed after editing. As shown in Figure 1(b), statistical findings demonstrate that the condition number of the edited matrix substantially increases as the number of edits increases, thereby exacerbating the perturbation of original knowledge and the deterioration of general abilities. Therefore, we assume that the bottleneck of the general abilities during sequential editing lies in the escalating value of the condition number. Towards continual and scalable model editing, we propose Perturbation Restraint on Upper bou Nd for Editing (PRUNE) based on the above analysis, which applies the condition number restraints in sequential editing to preserve general abilities and maintain new editing knowledge simultaneously. Specifically, the condition number of the edited matrix is restrained by reducing the large singular values (Albano et al., 1988; Wall et al., 2003) of the edit update matrix. Consequently, the upper bound on perturbation to the edited matrix is lowered, thus reducing the perturbation to the original knowledge associations and preserving the general abilities of the edited model, as shown in Figure 1(c). Additionally, we observe that these larger singular values often encapsulate redundant editing overfitting information, so regularizing them will not affect the newly editing knowledge, as shown in Figure 1(d). In this way, the new editing knowledge is embedded into LLMs without affecting their original general abilities. Overall, the proposed editing framework requires only minimal computing resources, and is adaptable to be coupled with multiple existing editing methods. To validate the effectiveness of the proposed PRUNE, our study comprehensively evaluates the edited LLMs for both general abilities and editing performance in sequential editing scenarios. Extensive research involves three popular editing methods, including MEND (Mitchell et al., 2022a), ROME (Meng et al., 2022), and MEMIT (Meng et al., 2023), which are analyzed based on three LLMs including GPT-2 XL (1.5B) (Radford et al., 2019), LLa MA-2 (7B) (Touvron et al., 2023), and LLa MA-3 (8B). Four downstream tasks including reasoning (Cobbe et al., 2021), summarization (Gliwa et al., 2019), open-domain QA (Kwiatkowski et al., 2019), and natural language inference (Dagan et al., 2005) are employed to demonstrate the impact of editing on the general abilities of LLMs. Experimental results demonstrate that the proposed PRUNE can preserve considerable general abilities and maintain almost all editing performance in sequential editing. In essence, our research offers three contributions: (1) This study theoretically analyzes that the escalating value of the condition number of the edited matrix is the bottleneck of sequential model editing. (2) The PRUNE based on the analysis is proposed to preserve the general abilities of the edited model while retaining the editing knowledge. (3) Experimental results including both editing and downstream task performance across three editing methods on three LLMs demonstrate the effectiveness of PRUNE. Published as a conference paper at ICLR 2025 2 RELATED WORK Model Editing Methods From the perspective of whether the model parameters are modified, existing editing methods can be divided into parameter-modifying (Mitchell et al., 2022a; Meng et al., 2022; 2023; Dai et al., 2022) and parameter-preserving methods (Mitchell et al., 2022b; Hartvigsen et al., 2023; Yu et al., 2024). This paper focuses on the former. Previous works have investigated the role of MLP layers in Transformer, showing that MLP layers store knowledge, which can be located in specific neurons and edited (Geva et al., 2021; Da et al., 2021; Geva et al., 2022). KE (Cao et al., 2021) and MEND (Mitchell et al., 2022a) train a hypernetwork to get gradient changes to update model parameters (Mitchell et al., 2022a). Besides, Meng et al. (2022) and Meng et al. (2023) used Locate-Then-Edit strategy, which first located multi-layer perceptron (MLP) storing factual knowledge, and then edited such knowledge by injecting new key-value pair in the MLP module. Parameter-preserving methods do not modify model weights but store the editing facts with an external memory. For example, Mitchell et al. (2022b) stored edits in a base model and learned to reason over them to adjust its predictions as needed. Model Editing Evaluation Some works investigate the paradigm for model editing evaluation (Zhong et al., 2023; Cohen et al., 2024; Ma et al., 2023; Li et al.; Hase et al., 2023; Wu et al., 2023; Gandikota et al., 2023; Ma et al., 2024). Cohen et al. (2024) introduced the ripple effects of model editing, suggesting that editing a particular fact implies that many other facts need to be updated. Ma et al. (2023) constructed a new benchmark to assess the edited model bidirectionally. Besides, Li et al. explored two significant areas of concern: Knowledge Conflict and Knowledge Distortion. These early studies mainly evaluate edited models per edit rather than sequentially, and they focus narrowly on basic factual triples. Recently, some works assess the impact of editing methods on the general abilities of LLMs in sequential editing scenarios. These studies (Gu et al., 2024; Gupta et al., 2024b; Lin et al., 2024; Yang et al., 2024; Gupta et al., 2024a;c) have conducted comprehensive experiments, showing the parameter-modifying methods significantly degrade the model performance on downstream tasks. Matrix Perturbation Theory It plays a crucial role in the field of artificial intelligence (AI) by providing a systematic framework to understand the impact of small changes or perturbations in various AI algorithms and models. Some studies (Harder et al., 2020; Qin et al., 2022; Singh et al., 2024) delve into the interpretability of LLMs, revealing how minor alterations in input features or model parameters influence the model s predictions. This understanding helps uncover significant feature connections within the model architecture. Moreover, it has been instrumental in assessing and enhancing the robustness of models (Chen et al., 2023; 2024). Furthermore, Bird et al. (2020) and Dettmers et al. (2023) have employed it for sensitivity analysis to identify critical factors affecting algorithm performance. It also contributes to the development of efficient optimization techniques (Li et al., 2020; Jiang et al., 2024), improving convergence rates and stability of optimization algorithms. Compared with previous works (Meng et al., 2022; 2023; Yao et al., 2023; Gu et al., 2024; Gupta et al., 2024b; Lin et al., 2024) that are the most relevant, a main difference should be highlighted. They neither theoretically investigate the reasons for general ability degradation, nor propose effective methods to maintain these abilities during sequential editing. In contrast, our study makes the first attempt to theoretically explore the bottleneck of general abilities in sequential editing and proposes the PRUNE framework to preserve these abilities for continual model editing. 3 ANALYSIS ON BOTTLENECK OF SEQUENTIAL MODEL EDITING 3.1 PRELIMINARY Model Editing This task involves modifying the memorized knowledge contained in LMs. Various kinds of complex learned beliefs such as logical, spatial, or numerical knowledge are expected to be edited. In this paper, following previous work (Meng et al., 2022; Zhong et al., 2023; Meng et al., 2023; Zhang et al., 2024), we study editing factual knowledge in the form of (subject s, relation r, object o), e.g., (s = United States, r = President of, o = Donald Trump). An LM is expected to recall a memory representing o given a natural language prompt p(s, r) such as The President of the United States is . Editing a fact is to incorporate a new knowledge triple (s, r, o ) in place of the current one (s, r, o). An edit is represented as e = (s, r, o, o ) for brevity. Given a set of editing facts Published as a conference paper at ICLR 2025 E = {e1, e2, . . .} and an original model fθ0, sequential model editing operationalizes each edit after the last edit1, i.e., K(fθn 1, en) = fθn, where fθn denotes the model after n edits. Singular Value Decomposition SVD (Albano et al., 1988) is a fundamental and effective matrix factorization technique for analyzing matrix structures. Formally, an SVD of a matrix W Rp q is given by W = UΣV T, where U = [u1, u2, ..., up] Rp p, V = [v1, v2, ..., vq] Rq q, and Σ Rp q. ui and vi are the column vectors of U and V , and constitute an orthonormal basis of Rp and Rq respectively. Σ is a diagonal matrix whose diagonal entries are given by the singular values of W in descending order. Additionally, the SVD of W could also be formulated as: W = Pmin{p,q} i=1 σiuiv T i , where σi is singular value, and σ1 σ2 ... σmin{p,q} 0. In the scenario of this paper, W is a full-rank matrix, so σmin{p,q} > 0. 3.2 MATRIX PERTURBATION THEORY ANALYSIS Previous works (Geva et al., 2021; Meng et al., 2022; Gupta et al., 2023; Wang et al., 2024) have analyzed and located that the MLP modules in Transformer (Vaswani et al., 2017) store various kinds of knowledge (Pearl, 2001; Vig et al., 2020). The MLP module of the l-th Transformer layer consists of two projection layers, where the first and second layers are denoted as W l fc and W l proj respectively. W l proj is considered as a linear associative memory which stores knowledge in the form of key-value pairs (ki, vi), and is usually regarded as the editing area (Meng et al., 2022; 2023). In this paper, W l proj is denoted as W for brevity. W is assumed to store many key-value pairs P = {(ki, vi) | i = 1, 2, ...} which satisfies Wki = vi, where ki Rq and vi Rp. Assuming | E |= N in sequential model editing, an edit update matrix Wj is calculated for the edit ej and added to W, which can be formulated as: WN = W + PN j=1 Wj with Wj calculated from fθj 1. Problem Modeling To explore the reasons for the general ability degradation of edited models, we begin by noting that most of the key-value pairs of P correspond to facts unrelated to editing. For the sake of analysis, only the matrix W of a single layer is assumed to be modified. We intuitively hypothesize that for the facts that are irrelevant to the editing fact, the cumulative modifications applied during sequential model editing may lead to significant mismatches in the associations between the original key-value pairs P. Specifically, consider a key-value pair (ki, vi) P. After applying an edit ej that generates Wj and adding it to W, if the extracted value vi remains unchanged, the corresponding key ki needs to be adjusted with an adjustment denoted as kj i . Mathematically, this can be represented as2 WN(ki + PN j=1 kj i ) = vi after N edits. However, during the editing process, it s challenging to guarantee such adjustments completely, leading to inaccuracies in the knowledge extracted from the edited model. To delve deeper, let s analyze how the key ki changes (i.e., PN j=1 kj i ) when its corresponding value vi remains unchanged after N edits. Perturbation Analysis of Single Edit According to matrix perturbation theory (Luo & Tseng, 1994; Vaccaro, 1994; Wedin, 1972), the edit update matrix W from an edit can be regarded as a perturbation3 for W, so we first analyze the situation where W Rp q is appended with a perturbation W. Define W is the generalized inverse (Stewart & Sun, 1990) of W, represents 2-norm, and W = W + W. Theorem 3.1 Consider Wk = v, there exists k such that k = k + k satisfies W k = v. Let k = W v and k = W v, and W is an acute perturbation of W. Then: η 1g(v) + E21 where E11, E12, and E21 are directly related to W. Ψ2(F) is a monotonically increasing function of F and g(v) is a function about v. ˆκ = W W 1 11 , where W11 is square and related to the reduced form of W. Each term on the right-hand side involves ˆκ, which means that the upper 1This paper studies editing a single fact at a time and leaves the exploration of batch editing as future work. 2As Wj Rp q, and we observed p < q in LLMs, so there will be kj i that satisfies this formula. 3We obtained some Wj and found Wj W , which satisfies the definition of perturbation. Published as a conference paper at ICLR 2025 W W20 W50 W100 W150 W200 0 Condition Number Condition Number Singular Value Max Singular Value Min Singular Value W W10 W20 W30 W40 W50 0 Condition Number Condition Number Singular Value Max Singular Value Min Singular Value W W20 W50 W100 W150 W180 0 Condition Number Condition Number Singular Value Max Singular Value Min Singular Value Figure 2: The condition number, maximum singular value and minimum singular value of the edited matrix in sequential editing. Three editing methods including ROME, MEND, and MEMIT are used to edit LLa MA-2 (7B) on the COUNTERFACT (Meng et al., 2022) dataset. For editing methods that modify the parameters of multiple MLP layers, one of them is randomly selected for illustration. W and Wn denote the unedited and edited matrices respectively. bound on the perturbation of the vector k is constrained by ˆκ. Readers can refer to Appendix A.3 for the details and proof of this theorem. However, calculating W 1 11 involves the reduced form of W, which incurs unnecessary additional overhead. Therefore, we consider the following theorem and give an alternative estimation. Theorem 3.2 Let κ = W W , and suppose that γ 1 κ E11 W > 0. Then: According to Theorem 3.2, W 1 11 W 1 11 γ = W γ , so ˆκ κ γ . Here κ = W W = σmax σmin is the condition number of W, where σmax and σmin are the maximum and minimum singular values of W, respectively. Combining Theorem 3.1, we know that the larger κ is, the greater the upper bound on the perturbation of the vector k. Readers can refer to Appendix A for the full theoretical analysis. 3.3 TREND OF THE CONDITION NUMBER DURING SEQUENTIAL EDITING As mentioned above, we have analyzed that the condition number of the edited matrix can be used to indicate the upper bound on the perturbation of the key-value pair associations by a single edit. In order to explore the impact of sequential model editing on these associations, the change trend of the condition number of the edited matrix during sequential editing is illustrated in Figure 2. Surprisingly, we observed that regardless of the editing methods employed, the condition number of the edited matrix exhibited a rapid increase as the number of edits increased, particularly after a large number of edits. According to Theorem 3.1, the adjustment norm kn i 2 corresponding to the n-th edit tends to increase as the number of edits n increases. Therefore, we can draw two conclusions: (1) As more edits are performed, the upper bound of the perturbation caused by a new single edit to the key-value pair associations increases. (2) During the sequential model editing process, the cumulative perturbation of these edits will become larger and larger. These factors further disrupt the stored original knowledge and exacerbate the deterioration of general abilities. As the second conclusion is easy to understand, here is an example for the first point. From the first subfigure of Figure 2, we can observe that the condition number of the W200 matrix after the 200th edit is significantly higher than that of the unedited matrix W. Therefore, the perturbation of the model caused by the 201st edit is likely to be much greater than the perturbation of the model caused by the 1st edit. 4 PRUNE: PERTURBATION RESTRAINT ON UPPER BOUND FOR EDITING Motivation According to the analysis in Section 3, the bottleneck of the general abilities during sequential editing lies in the escalating value of the condition number. Assuming a set of edits {ei} and their corresponding edit update matrices { Wi}, the information contained in these edit update matrices coordinates with each other to a certain extent since the parametric knowledge of LLMs is distributional rather than independent. This editing overfitting is reflected in SVD, where the largest singular value of the edited matrix WN becomes significantly large after the addition of these edit Published as a conference paper at ICLR 2025 update matrices. To illustrate this, consider an extreme example: suppose we make N edits, where each edit changes the answer to the question Who is the president of the United States? to Biden . Each edit update matrix is denoted as W1, and its maximum singular value is δmax. Then the sum of the N edit update matrices is N W1, and its maximum singular value is Nδmax, which is amplified by N times. Therefore, our goal is to reduce the editing overfitting in edited matrix WN as much as possible while also retaining valuable editing information. In this section, a framework termed Perturbation Restraint on Upper bou Nd for Editing (PRUNE) is proposed, which applies the condition number restraints to preserve general abilities and maintain new editing knowledge. Table 1: The maximum singular values of PN j=1 Wj with three edting methods. Other settings are the same as those illustrated in Figure 2. Edits (N) ROME MEMIT MEND 10 7.25 7.46 14.08 50 11.38 15.63 75.53 100 15.62 23.39 127.89 200 57.61 935 191.04 Principle Given an edited matrix with N edits, WN = W + PN j=1 Wj, as shown in Figure 2, its maximum singular value is constantly increasing, while the minimum singular value is basically unchanged as the number of edits N increases. This directly leads to the increasing condition number of the edited matrix. Therefore, our motivation is to restrain the large singular value of the edited matrix to lower the upper bound on the perturbation. If we directly perform SVD operation on WN and reduce its singular values, the original W will be inevitably destroyed. Consequently, an analysis of the singular values of PN j=1 Wj is conducted, and the results in Table 1 present that its maximum singular value becomes very large when N is large. Since the singular values of W are relatively small, we can assume that the large maximum singular value of PN j=1 Wj is the main reason why the maximum singular value of WN is large, our method therefore aims to restrain the large singular values of PN j=1 Wj. Design Firstly, SVD is operated on the original W and PN j=1 Wj respectively as: i=1 σiuiv T i , i=1 ˆσiˆuiˆv T i . (3) This paper considers W to be the main part, and any singular value in PN j=1 Wj should be ensured not to obviously exceed the maximum singular value of W. Subsequently, if any singular value ˆσi of PN j=1 Wj is greater than the maximum singular value of W, it will be restrained with a function F, otherwise it remains unchanged, which could be formulated as: σi = F(ˆσi), if ˆσi > max{σi}, ˆσi, if ˆσi max{σi}. (4) F(ˆσi) = logα(ˆσi) logα(max{σi}) + max{σi}. (5) In the main paper, we use the log function in F to restrain ˆσi. Here α is a hyperparameter to control the degree of restraints, readers can refer to Appendix B.3 for its details for experiments. Besides, we also provide the definition and results of linear function in Appendix C.3. Finally, we obtain the restrained edited matrix W N to replace WN: σiˆuiˆv T i . (6) In this way, the condition number of the edited matrix is reduced (see Appendix C.4) and the upper bound on perturbation is significantly restrained. It is worth noting that the PRUNE proposed here is only used once in Section 5, but can actually be used multiple times in the editing process to better maintain the general ability. We provide a comparison of the two strategies in Appendix C.6. 5 EXPERIMENTS In this section, both the downstream task performance and editing performance of three editing methods on three LLMs were evaluated in sequential model editing. The proposed PRUNE was plug-and-play which can be coupled with these editing methods. Published as a conference paper at ICLR 2025 5.1 BASE LLMS AND EDITING METHODS Experiments were conducted on three LLMs including GPT-2 XL (1.5B) (Radford et al., 2019), LLa MA-2 (7B) (Touvron et al., 2023) and LLa MA-3 (8B)4. Three popular editing methods were selected as the baselines including MEND (Mitchell et al., 2022a), ROME (Meng et al., 2022), and MEMIT (Meng et al., 2023). Appendix B.1 shows the details of these editing methods. 5.2 EDITING DATASETS AND EVALUATION METRICS To make a more comprehensive evaluation, we used two types of knowledge for editing: factual knowledge and conceptual knowledge. (1) For factual knowledge, two popular model editing datasets Zero-Shot Relation Extraction (ZSRE) (Levy et al., 2017) and COUNTERFACT (Meng et al., 2022) were adopted in our experiments. These two datasets are QA datasets. A key distinction between COUNTERFACT and ZSRE datasets is that ZSRE contains true facts, while COUNTERFACT contains counterfactual examples where the new target has a lower probability when compared to the original answer (Gupta et al., 2024b). (2) For conceptual knowledge, the Concept Edit dataset (Wang et al., 2024) was adopted. Due to the limitations of computing resources and pages, most of the experiments in this paper were conducted on factual datasets, with the results presented in Sections 5.4 and 5.5. Meanwhile, Section 5.6 provided some results on conceptual datasets. Readers can refer to Appendix B.2 for examples of each dataset. To assess the editing performance of editing methods, following previous works (Cao et al., 2021; Mitchell et al., 2022a; Meng et al., 2022; 2023; Ma et al., 2024), three fundamental metrics were employed: efficacy, generalization and locality. Given an original model fθ0, an edited model fθn with n times sequential editing. Define 1 as the indicator function. Each edit ei = (si, ri, oi, o i ) has an editing prompt pi, paraphrase prompts PG i , and locality prompts PL i . Efficacy validates whether the edited models could recall the editing fact under editing prompt pi. The assessment is based on Efficacy Score (ES) representing as: Ei[1[ argmaxo Pfθn(o | pi) = o i ] ]. Generalization verifies whether the edited models could recall the editing fact under the paraphrase prompts PG i via Generalization Score (GS): Ei [Ep PG i [1[ argmaxo Pfθn(o | p) = o i ] ]. Locality verifies whether the output of the edited models for inputs out of editing scope remains unchanged under the locality prompts PL i via Locality Score (LS): Ei [Epl PL i [1[ argmaxo Pfθn(o | pl) = ol] ] ], where ol was the original answer of pl. Different from previous studies that assess the edited models after each individual edit (Gupta et al., 2024b; Yao et al., 2023), this paper evaluated whether the final edited models after completing all edits can still recall all preceding edits, which is more challenging and common in real-world. 5.3 DOWNSTREAM TASKS, DATASETS AND METRICS To explore the side effects of sequential model editing on the general abilities of LLMs, four representative tasks with corresponding datasets were adopted for assessment following previous work (Gu et al., 2024; Gupta et al., 2024b; Lin et al., 2024; Zhang et al., 2024), including: Reasoning on the GSM8K (Cobbe et al., 2021), and the results were measured by solve rate. Summarization on the SAMSum (Gliwa et al., 2019), and the results were measured by the average of ROUGE-1, ROUGE-2 and ROUGE-L following Lin (2004). Open-domain QA on the Natural Question (Kwiatkowski et al., 2019), and the results were measured by exact match (EM) with the reference answer after minor normalization as in Chen et al. (2017) and Lee et al. (2019). Natural language inference (NLI) on the RTE (Dagan et al., 2005), and the results were measured by accuracy of two-way classification. For each dataset, some examples were randomly sampled for evaluation. Details of prompts for each task were shown in Appendix B.4. 4https://llama.meta.com/llama3/ Published as a conference paper at ICLR 2025 Figure 3: The downstream task performance (%) of models edited by three editing methods with LLa MA-2 (7B) on the ZSRE dataset. The dashed lines refer to the results of the unrestrained editing methods. The solid lines refer to the results of the editing methods coupled with the proposed PRUNE framework. Statistical significance tests were performed to demonstrate that the improvement in PRUNE compared to baseline was statistically significant (t-test with p-value <0.05). 5.4 GENERAL ABILITIES RESULTS ON FACTUAL KNOWLEDGE Figure 3 illustrates the downstream task performance of editing methods with LLa MA-2 (7B) on the ZSRE dataset. Due to page limitation, results of other LLMs and factual datasets were put in Appendix C.1. These results were analyzed from the following perspectives. Current editing methods significantly compromised general abilities. As depicted by the dashed lines of Figure 3, both the ROME and MEMIT methods initially maintained relatively stable performance in downstream tasks when the number of edits was small ( 50). However, as the number of edits surpassed 100, a noticeable decline in performance was observed across all tasks for both methods. Additionally, the MEND method exhibited significant performance degradation after just 20 sequential edits, indicating its inadequacy as a sequential model editing method. Furthermore, when comparing LLMs of different sizes, a general trend emerged: larger models suffered more pronounced compromises in their general abilities when subjected to the same number of edits. For instance, with 300 edits, MEMIT s performance on GPT2-XL remained largely unchanged, whereas it dwindled to nearly 0 on LLa MA-2 and LLa MA-3. The performance decline was gradual initially but accelerated with increasing edit count. This trend aligned with the fluctuation observed in the size of the condition number, as depicted in Figure 2. When the number of edits was small, the condition number was small, and each new edit introduced relatively minor perturbations to the model. However, as the number of edits increased, the condition number underwent a substantial increase. Consequently, each subsequent edit exerted a significant perturbation on the model, leading to a pronounced impairment of its general abilities. These results substantiated the analysis presented in Section 3.3. The proposed PRUNE can preserve considerable general abilities. As shown by the solid lines of Figure 3, when MEMIT was coupled with PRUNE and subjected to 100 edits, its downstream tasks performance remained close to that of the unedited model. However, for the unrestrained MEMIT, downstream task performance had plummeted to nearly 0 by this point. This consistent trend was also observed with ROME and MEND. Nevertheless, for models edited using the unrestrained MEND method, performance degradation was stark after just 10 edits. Even with the addition of PRUNE, preservation could only be extended up to 20 edits. This suggests that while PRUNE effectively preserves general abilities, it does have an upper limit determined by the unrestrained editing method. 5.5 EDITING PERFORMANCE RESULTS ON FACTUAL KNOWLEDGE Figure 4 shows three metrics used for measuring the editing performance with LLa MA-2 (7B) on the ZSRE dataset. Other results were put in Appendix C.2. Three conclusions can be drawn. Previous editing facts were forgotten as the number of edits increased. As shown by the dashed lines of Figure 4, the decline in efficacy and generalization suggests that in sequential editing scenarios, post-edited models gradually forget knowledge acquired from previous edits after a few iterations. Published as a conference paper at ICLR 2025 Figure 4: The editing performance (%) of editing methods with LLa MA-2 (7B) on the ZSRE dataset. The dashed lines refer to the results of the unrestrained editing methods. The solid lines refer to the results of the editing methods coupled with the proposed PRUNE. Statistical significance tests were performed to demonstrate that the improvement in PRUNE compared to baseline was statistically significant (t-test with p-value <0.05). Comparing these editing methods, we also observed a notable drop in efficacy and generalization after hundreds of edits with ROME and MEMIT, whereas these values decreased significantly after only 15 edits with MEND. This indicates that in sequential editing scenarios, the MEND method struggled to successfully integrate new knowledge into LLMs after several edits. Unrelated facts were perturbed as the number of edits increased. The locality metric served as an indicator of perturbation for unrelated facts. It became evident that for each editing method, the locality decreased significantly. Additionally, an observation emerged: when the locality of the edited model was low, the performance of downstream tasks was also low. This observation underscores that perturbations of irrelevant knowledge compromise the general abilities of the edited model. PRUNE can effectively maintain the editing performance. This is shown by the solid lines of Figure 4 and could be analyzed from two aspects. On the one hand, when the number of edits was small, the editing performance of each editing method coupled with PRUNE was about the same as the unrestrained method. On the other hand, it significantly mitigated the forgetting of editing facts and the perturbation of irrelevant facts when the number of edits was large during the sequential editing. Specifically, when the number of edits reached 100, the editing performance of MEMIT was very low. But when coupled with PRUNE, its performance remained relatively stable. These observations further validate our motivation in Section 4, demonstrating that the information in the edit update matrices is coordinated, and that performing too many edits can easily result in overfitting. Therefore, applying a certain degree of restraint to edit perturbations can help preserve the model s general abilities while maintaining the editing knowledge. 5.6 EDITING WITH CONCEPTUAL KNOWLEDGE Section 5.4 and 5.5 analyzed the results on factual knowledge. This section conducted some experiments with ROME on conceptual knowledge using the Concept Edit dataset (Wang et al., 2024) to make a more comprehensive evaluation. For editing performance, in addition to the three basic metrics, this dataset also designed a new metric Instance Change to measure whether the instances under the concept changed accordingly when the definition of the concept was changed. As shown in Table 6, the performance trends of editing and downstream tasks were similar to those observed with the factual datasets. But there are several key differences: (1) When the number of edits was the same, the editing performance of conceptual knowledge was lower than that of factual knowledge. (2) Both editing performance and general abilities deteriorated more quickly than factual knowledge. For example, even if the number of edits was 100, the editing performance and downstream task performance of ROME were very low, while it was still relatively high when editing factual knowledge. (3) The low Instance Change indicated that when the definition of a concept was altered, the instances contained in the original concept were still recognized by the model as belonging to that concept. This shows that this editing method primarily modifies the definition without successfully altering the relationship between concepts and instances, which is not reasonable. These findings indicate that conceptual knowledge is more abstract and more difficult to edit than factual knowledge, highlighting the need to explore editing methods for different types of knowledge. Published as a conference paper at ICLR 2025 Table 2: Evaluation results (%) of LLa MA-2 (7B) edited by ROME on the Concept Edit dataset. Mode General Abilities Editing Performance Method Edits Reasoning Summa Open-QA NLI Efficacy General Locality Instance 20 75.13 11 6.50 24.7 49.15 52.58 35.68 25 50 20.67 4.90 1.50 0.7 55.42 49.45 19.94 12 100 12.29 4.7 0.77 0 28.25 30.18 5.68 10 200 0 4.62 0 0 10.14 8.65 5.31 -8.99 20 89.38 14.34 23.37 63.54 75.66 58.35 71.7 25 50 85.15 14.06 25.29 50.52 56.51 45.55 73.16 8 100 90.78 13.75 21.46 53.17 46.22 42.26 64.06 20 200 72.9 10.55 22.22 46.15 35.82 34.95 46.65 32 5.7 ANALYSIS ON THE FORGETTING OF EDITING FACTS 20 10 0 10 20 30 LLa MA-2 (7B)-ROME VCurrent VEditing VPrune Figure 5: 2-dimensional PCA visualization of first 100 values. The model was edited by ROME with LLa MA-2. Section 3 conducted analysis to elucidate the reasons behind the degradation in general abilities with an increasing number of edits. Subsequent experiments quantitatively demonstrated the effectiveness of PRUNE. Here, we delve into qualitative analysis to explain why editing facts are forgotten and how PRUNE can mitigate this forgetting. Initially, given a set of editing facts E = {e1, e2, . . .}, where | E |= 200. ROME was employed for analysis, and the original matrix was defined as W. During sequential editing, ROME computed key-value pairs (ke j, ve j) of the last subject token to generate Wj for each edit ej to incorporate new facts, satisfying the equation: Wj ke j = ve j. However, when evaluating editing performance, the edited model obtained from the last edit was utilized, thus computing values5: W200 ke j = ˆve j. After adopting PRUNE to ROME, this equation became W 200 ke j = ve j. We hypothesized that if ˆve j was similar to ve j, the editing fact ej could be maintained. Denote VCurrent = {ve j}, VEditing = {ˆve j}, and VP rune = {ve j}. Specifically, these corresponding values of the first 100 edits were used, as they are more prone to be forgotten than the last 100. Principal Component Analysis (PCA) (Gewers et al., 2022) was employed to visualize these values. The first two principal components of each value were calculated and illustrated, as they can represent most of its features (Zheng et al.). As shown in Figure 5, on the one hand, the discrepancy between the principal components of VCurrent and VEditing was markedly large. This indicates that after 200 edits to the model, the values corresponding to the first 100 facts stored in the edited matrix are severely corrupted, leading to significant forgetfulness. On the other hand, after adopting PRUNE, the discrepancy between the principal components of VCurrent and VP rune was small. This demonstrates that PRUNE effectively maintains the values and mitigates the forgetting of editing facts. 6 CONCLUSION AND LIMITATION In this paper, a theoretical analysis is firstly conducted to elucidate that the bottleneck of the general abilities during sequential editing lies in the escalating value of the condition number. Subsequently, a plug-and-play framework called PRUNE is proposed to apply restraints to preserve general abilities and maintain new editing knowledge simultaneously. Comprehensive experiments on various editing methods and LLMs demonstrate the effectiveness of this method. We aspire that our analysis and method will catalyze future research on continual model editing. Limitation Firstly, this paper focuses on editing a single fact at a time in sequential model editing, but some works study updating hundreds of facts simultaneously in batch editing. Therefore, investigating batch-sequential editing could enhance the scalability of model editing. Secondly, it is necessary to explore the performance of larger-size models and more editing methods on more downstream tasks. 5Since ROME only modifies one matrix, the ke j remains the same across these edited models. Published as a conference paper at ICLR 2025 ACKNOWLEDGMENTS We would like to express gratitude to the anonymous reviewers for kind comments. This work is funded by the National Science and Technology Major Project (No. 2023ZD0121103). Alfonso M Albano, J Muench, C Schwartz, AI Mees, and PE Rapp. Singular-value decomposition and the grassberger-procaccia algorithm. Physical review A, 38(6):3017, 1988. Sarah Bird, Miro Dudík, Richard Edgar, Brandon Horn, Roman Lutz, Vanessa Milan, Mehrnoosh Sameki, Hanna Wallach, and Kathleen Walker. Fairlearn: A toolkit for assessing and improving fairness in ai. Microsoft, Tech. Rep. MSR-TR-2020-32, 2020. Nicola De Cao, Wilker Aziz, and Ivan Titov. Editing factual knowledge in language models. In Marie Francine Moens, Xuanjing Huang, Lucia Specia, and Scott Wen-tau Yih (eds.), Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, EMNLP 2021, Virtual Event / Punta Cana, Dominican Republic, 7-11 November, 2021, pp. 6491 6506. Association for Computational Linguistics, 2021. doi: 10.18653/v1/2021.emnlp-main.522. URL https: //doi.org/10.18653/v1/2021.emnlp-main.522. Danqi Chen, Adam Fisch, Jason Weston, and Antoine Bordes. Reading wikipedia to answer opendomain questions. In Regina Barzilay and Min-Yen Kan (eds.), Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, ACL 2017, Vancouver, Canada, July 30 - August 4, Volume 1: Long Papers, pp. 1870 1879. Association for Computational Linguistics, 2017. doi: 10.18653/V1/P17-1171. URL https://doi.org/10.18653/v1/P17-1171. Shuo Chen, Jindong Gu, Zhen Han, Yunpu Ma, Philip H. S. Torr, and Volker Tresp. Benchmarking robustness of adaptation methods on pre-trained vision-language models. In Alice Oh, Tristan Naumann, Amir Globerson, Kate Saenko, Moritz Hardt, and Sergey Levine (eds.), Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, Neur IPS 2023, New Orleans, LA, USA, December 10 - 16, 2023, 2023. URL http://papers.nips.cc/paper_files/paper/2023/ hash/a2a544e43acb8b954dc5846ff0d77ad5-Abstract-Datasets_and_ Benchmarks.html. Zhuotong Chen, Zihu Wang, Yifan Yang, Qianxiao Li, and Zheng Zhang. PID control-based selfhealing to improve the robustness of large language models. Co RR, abs/2404.00828, 2024. doi: 10. 48550/ARXIV.2404.00828. URL https://doi.org/10.48550/ar Xiv.2404.00828. Karl Cobbe, Vineet Kosaraju, Mohammad Bavarian, Mark Chen, Heewoo Jun, Lukasz Kaiser, Matthias Plappert, Jerry Tworek, Jacob Hilton, Reiichiro Nakano, Christopher Hesse, and John Schulman. Training verifiers to solve math word problems. Co RR, abs/2110.14168, 2021. URL https://arxiv.org/abs/2110.14168. Roi Cohen, Eden Biran, Ori Yoran, Amir Globerson, and Mor Geva. Evaluating the ripple effects of knowledge editing in language models. Transactions of the Association for Computational Linguistics, 11:283 298, 2024. Jeff Da, Ronan Le Bras, Ximing Lu, Yejin Choi, and Antoine Bosselut. Analyzing commonsense emergence in few-shot knowledge models. In Danqi Chen, Jonathan Berant, Andrew Mc Callum, and Sameer Singh (eds.), 3rd Conference on Automated Knowledge Base Construction, AKBC 2021, Virtual, October 4-8, 2021, 2021. doi: 10.24432/C5NK5J. URL https://doi.org/ 10.24432/C5NK5J. Ido Dagan, Oren Glickman, and Bernardo Magnini. The PASCAL recognising textual entailment challenge. In Joaquin Quiñonero Candela, Ido Dagan, Bernardo Magnini, and Florence d AlchéBuc (eds.), Machine Learning Challenges, Evaluating Predictive Uncertainty, Visual Object Classification and Recognizing Textual Entailment, First PASCAL Machine Learning Challenges Workshop, MLCW 2005, Southampton, UK, April 11-13, 2005, Revised Selected Papers, volume 3944 of Lecture Notes in Computer Science, pp. 177 190. Springer, 2005. doi: 10.1007/11736790\ _9. URL https://doi.org/10.1007/11736790_9. Published as a conference paper at ICLR 2025 Damai Dai, Li Dong, Yaru Hao, Zhifang Sui, Baobao Chang, and Furu Wei. Knowledge neurons in pretrained transformers. In Smaranda Muresan, Preslav Nakov, and Aline Villavicencio (eds.), Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2022, Dublin, Ireland, May 22-27, 2022, pp. 8493 8502. Association for Computational Linguistics, 2022. doi: 10.18653/v1/2022.acl-long.581. URL https://doi. org/10.18653/v1/2022.acl-long.581. Jean-Pierre Dedieu. Condition operators, condition numbers, and condition number theorem for the generalized eigenvalue problem. Linear algebra and its applications, 263:1 24, 1997. Tim Dettmers, Ruslan Svirschevski, Vage Egiazarian, Denis Kuznedelev, Elias Frantar, Saleh Ashkboos, Alexander Borzunov, Torsten Hoefler, and Dan Alistarh. Spqr: A sparse-quantized representation for near-lossless LLM weight compression. Co RR, abs/2306.03078, 2023. doi: 10. 48550/ARXIV.2306.03078. URL https://doi.org/10.48550/ar Xiv.2306.03078. Junfeng Fang, Houcheng Jiang, Kun Wang, Yunshan Ma, Shi Jie, Xiang Wang, Xiangnan He, and Tat-Seng Chua. Alphaedit: Null-space constrained knowledge editing for language models. The Thirteenth International Conference on Learning Representations, ICLR, 2025. URL https: //openreview.net/pdf?id=Hv Sytvg3Jh. Rohit Gandikota, Joanna Materzynska, Jaden Fiotto-Kaufman, and David Bau. Erasing concepts from diffusion models. Co RR, abs/2303.07345, 2023. doi: 10.48550/ARXIV.2303.07345. URL https://doi.org/10.48550/ar Xiv.2303.07345. Mor Geva, Roei Schuster, Jonathan Berant, and Omer Levy. Transformer feed-forward layers are keyvalue memories. In Marie-Francine Moens, Xuanjing Huang, Lucia Specia, and Scott Wen-tau Yih (eds.), Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, EMNLP 2021, Virtual Event / Punta Cana, Dominican Republic, 7-11 November, 2021, pp. 5484 5495. Association for Computational Linguistics, 2021. doi: 10.18653/v1/2021.emnlp-main.446. URL https://doi.org/10.18653/v1/2021.emnlp-main.446. Mor Geva, Avi Caciularu, Kevin Ro Wang, and Yoav Goldberg. Transformer feed-forward layers build predictions by promoting concepts in the vocabulary space. In Yoav Goldberg, Zornitsa Kozareva, and Yue Zhang (eds.), Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022, Abu Dhabi, United Arab Emirates, December 7-11, 2022, pp. 30 45. Association for Computational Linguistics, 2022. doi: 10.18653/v1/2022.emnlp-main.3. URL https://doi.org/10.18653/v1/2022.emnlp-main.3. Felipe L. Gewers, Gustavo R. Ferreira, Henrique Ferraz de Arruda, Filipi Nascimento Silva, Cesar H. Comin, Diego R. Amancio, and Luciano da Fontoura Costa. Principal component analysis: A natural approach to data exploration. ACM Comput. Surv., 54(4):70:1 70:34, 2022. doi: 10.1145/3447755. URL https://doi.org/10.1145/3447755. Bogdan Gliwa, Iwona Mochol, Maciej Biesek, and Aleksander Wawer. SAMSum corpus: A humanannotated dialogue dataset for abstractive summarization. In Lu Wang, Jackie Chi Kit Cheung, Giuseppe Carenini, and Fei Liu (eds.), Proceedings of the 2nd Workshop on New Frontiers in Summarization, pp. 70 79, Hong Kong, China, November 2019. Association for Computational Linguistics. doi: 10.18653/v1/D19-5409. URL https://aclanthology.org/D19-5409. Jia-Chen Gu, Hao-Xiang Xu, Jun-Yu Ma, Pan Lu, Zhen-Hua Ling, Kai-Wei Chang, and Nanyun Peng. Model editing harms general abilities of large language models: Regularization to the rescue. In Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, pp. 16801 16819, 2024. Akshat Gupta, Sidharth Baskaran, and Gopala Anumanchipalli. Rebuilding ROME : Resolving model collapse during sequential model editing. In Yaser Al-Onaizan, Mohit Bansal, and Yun-Nung Chen (eds.), Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, EMNLP 2024, Miami, FL, USA, November 12-16, 2024, pp. 21738 21744. Association for Computational Linguistics, 2024a. URL https://aclanthology.org/ 2024.emnlp-main.1210. Published as a conference paper at ICLR 2025 Akshat Gupta, Anurag Rao, and Gopala Anumanchipalli. Model editing at scale leads to gradual and catastrophic forgetting. Co RR, abs/2401.07453, 2024b. doi: 10.48550/ARXIV.2401.07453. URL https://doi.org/10.48550/ar Xiv.2401.07453. Akshat Gupta, Dev Sajnani, and Gopala Anumanchipalli. A unified framework for model editing. In Yaser Al-Onaizan, Mohit Bansal, and Yun-Nung Chen (eds.), Findings of the Association for Computational Linguistics: EMNLP 2024, Miami, Florida, USA, November 12-16, 2024, pp. 15403 15418. Association for Computational Linguistics, 2024c. URL https://aclanthology.org/2024.findings-emnlp.903. Anshita Gupta, Debanjan Mondal, Akshay Krishna Sheshadri, Wenlong Zhao, Xiang Lorraine Li, Sarah Wiegreffe, and Niket Tandon. Editing commonsense knowledge in GPT. Co RR, abs/2305.14956, 2023. doi: 10.48550/ARXIV.2305.14956. URL https://doi.org/10. 48550/ar Xiv.2305.14956. Frederik Harder, Matthias Bauer, and Mijung Park. Interpretable and differentially private predictions. In The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, The Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, The Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, February 7-12, 2020, pp. 4083 4090. AAAI Press, 2020. doi: 10.1609/AAAI.V34I04.5827. URL https://doi.org/10.1609/aaai.v34i04.5827. Tom Hartvigsen, Swami Sankaranarayanan, Hamid Palangi, Yoon Kim, and Marzyeh Ghassemi. Aging with GRACE: lifelong model editing with discrete key-value adaptors. In Alice Oh, Tristan Naumann, Amir Globerson, Kate Saenko, Moritz Hardt, and Sergey Levine (eds.), Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, Neur IPS 2023, New Orleans, LA, USA, December 10 - 16, 2023, 2023. URL http://papers.nips.cc/paper_files/paper/2023/hash/ 95b6e2ff961580e03c0a662a63a71812-Abstract-Conference.html. Peter Hase, Mohit Bansal, Been Kim, and Asma Ghandeharioun. Does localization inform editing? surprising differences in causality-based localization vs. knowledge editing in language models. Co RR, abs/2301.04213, 2023. doi: 10.48550/ARXIV.2301.04213. URL https://doi.org/ 10.48550/ar Xiv.2301.04213. Chenhui Hu, Pengfei Cao, Yubo Chen, Kang Liu, and Jun Zhao. Wilke: Wise-layer knowledge editor for lifelong knowledge editing. Co RR, abs/2402.10987, 2024. doi: 10.48550/ARXIV.2402.10987. URL https://doi.org/10.48550/ar Xiv.2402.10987. Lei Huang, Weijiang Yu, Weitao Ma, Weihong Zhong, Zhangyin Feng, Haotian Wang, Qianglong Chen, Weihua Peng, Xiaocheng Feng, Bing Qin, and Ting Liu. A survey on hallucination in large language models: Principles, taxonomy, challenges, and open questions. Co RR, abs/2311.05232, 2023. doi: 10.48550/ARXIV.2311.05232. URL https://doi.org/10.48550/ar Xiv. 2311.05232. Ziwei Ji, Nayeon Lee, Rita Frieske, Tiezheng Yu, Dan Su, Yan Xu, Etsuko Ishii, Yejin Bang, Andrea Madotto, and Pascale Fung. Survey of hallucination in natural language generation. ACM Comput. Surv., 55(12):248:1 248:38, 2023. doi: 10.1145/3571730. URL https://doi.org/10. 1145/3571730. Houcheng Jiang, Junfeng Fang, Ningyu Zhang, Guojun Ma, Mingyang Wan, Xiang Wang, Xiangnan He, and Tat-seng Chua. Anyedit: Edit any knowledge encoded in language models. ar Xiv preprint ar Xiv:2502.05628, 2025. Shuoran Jiang, Qingcai Chen, Youcheng Pan, Yang Xiang, Yukang Lin, Xiangping Wu, Chuanyi Liu, and Xiaobao Song. Zo-adamu optimizer: Adapting perturbation by the momentum and uncertainty in zeroth-order optimization. In Michael J. Wooldridge, Jennifer G. Dy, and Sriraam Natarajan (eds.), Thirty-Eighth AAAI Conference on Artificial Intelligence, AAAI 2024, Thirty Sixth Conference on Innovative Applications of Artificial Intelligence, IAAI 2024, Fourteenth Symposium on Educational Advances in Artificial Intelligence, EAAI 2014, February 20-27, 2024, Vancouver, Canada, pp. 18363 18371. AAAI Press, 2024. doi: 10.1609/AAAI.V38I16.29796. URL https://doi.org/10.1609/aaai.v38i16.29796. Published as a conference paper at ICLR 2025 Tom Kwiatkowski, Jennimaria Palomaki, Olivia Redfield, Michael Collins, Ankur P. Parikh, Chris Alberti, Danielle Epstein, Illia Polosukhin, Jacob Devlin, Kenton Lee, Kristina Toutanova, Llion Jones, Matthew Kelcey, Ming-Wei Chang, Andrew M. Dai, Jakob Uszkoreit, Quoc Le, and Slav Petrov. Natural questions: a benchmark for question answering research. Trans. Assoc. Comput. Linguistics, 7:452 466, 2019. doi: 10.1162/TACL\_A\_00276. URL https://doi.org/10. 1162/tacl_a_00276. Kenton Lee, Ming-Wei Chang, and Kristina Toutanova. Latent retrieval for weakly supervised open domain question answering. In Anna Korhonen, David R. Traum, and Lluís Màrquez (eds.), Proceedings of the 57th Conference of the Association for Computational Linguistics, ACL 2019, Florence, Italy, July 28August 2, 2019, Volume 1: Long Papers, pp. 6086 6096. Association for Computational Linguistics, 2019. doi: 10.18653/V1/P19-1612. URL https: //doi.org/10.18653/v1/p19-1612. Omer Levy, Minjoon Seo, Eunsol Choi, and Luke Zettlemoyer. Zero-shot relation extraction via reading comprehension. In Roger Levy and Lucia Specia (eds.), Proceedings of the 21st Conference on Computational Natural Language Learning (Co NLL 2017), Vancouver, Canada, August 3-4, 2017, pp. 333 342. Association for Computational Linguistics, 2017. doi: 10.18653/v1/K17-1034. URL https://doi.org/10.18653/v1/K17-1034. Hui-Jia Li, Lin Wang, Yan Zhang, and Matjaž Perc. Optimization of identifiability for efficient community detection. New Journal of Physics, 22(6):063035, 2020. Zhoubo Li, Ningyu Zhang, Yunzhi Yao, Mengru Wang, Xi Chen, and Huajun Chen. Unveiling the pitfalls of knowledge editing for large language models. In The Twelfth International Conference on Learning Representations. Chin-Yew Lin. ROUGE: A package for automatic evaluation of summaries. In Text Summarization Branches Out, pp. 74 81, Barcelona, Spain, July 2004. Association for Computational Linguistics. URL https://aclanthology.org/W04-1013. Zihao Lin, Mohammad Beigi, Hongxuan Li, Yufan Zhou, Yuxiang Zhang, Qifan Wang, Wenpeng Yin, and Lifu Huang. Navigating the dual facets: A comprehensive evaluation of sequential memory editing in large language models. Co RR, abs/2402.11122, 2024. doi: 10.48550/ARXIV.2402.11122. URL https://doi.org/10.48550/ar Xiv.2402.11122. Zhi-Quan Luo and Paul Tseng. Perturbation analysis of a condition number for linear systems. SIAM Journal on Matrix Analysis and Applications, 15(2):636 660, 1994. Jun-Yu Ma, Jia-Chen Gu, Zhen-Hua Ling, Quan Liu, and Cong Liu. Untying the reversal curse via bidirectional language model editing. Co RR, abs/2310.10322, 2023. doi: 10.48550/ARXIV.2310. 10322. URL https://doi.org/10.48550/ar Xiv.2310.10322. Jun-Yu Ma, Zhen-Hua Ling, Ningyu Zhang, and Jia-Chen Gu. Neighboring perturbations of knowledge editing on large language models. In Forty-first International Conference on Machine Learning, 2024. Kevin Meng, David Bau, Alex Andonian, and Yonatan Belinkov. Locating and editing factual associations in GPT. In Neur IPS, 2022. URL https://arxiv.org/abs/2202.05262. Kevin Meng, Arnab Sen Sharma, Alex J. Andonian, Yonatan Belinkov, and David Bau. Massediting memory in a transformer. In The Eleventh International Conference on Learning Representations, ICLR 2023, Kigali, Rwanda, May 1-5, 2023. Open Review.net, 2023. URL https://openreview.net/pdf?id=Mkbc AHIYgy S. Eric Mitchell, Charles Lin, Antoine Bosselut, Chelsea Finn, and Christopher D. Manning. Fast model editing at scale. In The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25-29, 2022. Open Review.net, 2022a. URL https://openreview.net/ forum?id=0Dc Zxe Wf OPt. Eric Mitchell, Charles Lin, Antoine Bosselut, Christopher D. Manning, and Chelsea Finn. Memorybased model editing at scale. In Kamalika Chaudhuri, Stefanie Jegelka, Le Song, Csaba Szepesvári, Gang Niu, and Sivan Sabato (eds.), International Conference on Machine Learning, ICML 2022, Published as a conference paper at ICLR 2025 17-23 July 2022, Baltimore, Maryland, USA, volume 162 of Proceedings of Machine Learning Research, pp. 15817 15831. PMLR, 2022b. URL https://proceedings.mlr.press/ v162/mitchell22a.html. Judea Pearl. Direct and indirect effects. In Jack S. Breese and Daphne Koller (eds.), UAI 01: Proceedings of the 17th Conference in Uncertainty in Artificial Intelligence, University of Washington, Seattle, Washington, USA, August 2-5, 2001, pp. 411 420. Morgan Kaufmann, 2001. URL https://dslpitt.org/uai/display Article Details.jsp?mmnu= 1&smnu=2&article_id=126&proceeding_id=17. Baolin Peng, Michel Galley, Pengcheng He, Hao Cheng, Yujia Xie, Yu Hu, Qiuyuan Huang, Lars Liden, Zhou Yu, Weizhu Chen, and Jianfeng Gao. Check your facts and try again: Improving large language models with external knowledge and automated feedback. Co RR, abs/2302.12813, 2023. doi: 10.48550/ar Xiv.2302.12813. URL https://doi.org/10.48550/ar Xiv. 2302.12813. Bin Qin, Fu-Lai Chung, and Shitong Wang. KAT: A knowledge adversarial training method for zeroorder takagi-sugeno-kang fuzzy classifiers. IEEE Trans. Cybern., 52(7):6857 6871, 2022. doi: 10. 1109/TCYB.2020.3034792. URL https://doi.org/10.1109/TCYB.2020.3034792. Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever, et al. Language models are unsupervised multitask learners. Open AI blog, 1(8):9, 2019. Chandan Singh, Jeevana Priya Inala, Michel Galley, Rich Caruana, and Jianfeng Gao. Rethinking interpretability in the era of large language models. Co RR, abs/2402.01761, 2024. doi: 10.48550/ ARXIV.2402.01761. URL https://doi.org/10.48550/ar Xiv.2402.01761. Russell A Smith. The condition numbers of the matrix eigenvalue problem. Numerische Mathematik, 10:232 240, 1967. Gilbert W Stewart and Ji-guang Sun. Matrix perturbation theory. (No Title), 1990. Ji-guang Sun. Condition number and backward error for the generalized singular value decomposition. SIAM Journal on Matrix Analysis and Applications, 22(2):323 341, 2000. Hugo Touvron, Louis Martin, Kevin Stone, Peter Albert, Amjad Almahairi, Yasmine Babaei, Nikolay Bashlykov, Soumya Batra, Prajjwal Bhargava, Shruti Bhosale, Dan Bikel, Lukas Blecher, Cristian Canton-Ferrer, Moya Chen, Guillem Cucurull, David Esiobu, Jude Fernandes, Jeremy Fu, Wenyin Fu, Brian Fuller, Cynthia Gao, Vedanuj Goswami, Naman Goyal, Anthony Hartshorn, Saghar Hosseini, Rui Hou, Hakan Inan, Marcin Kardas, Viktor Kerkez, Madian Khabsa, Isabel Kloumann, Artem Korenev, Punit Singh Koura, Marie-Anne Lachaux, Thibaut Lavril, Jenya Lee, Diana Liskovich, Yinghai Lu, et al. Llama 2: Open foundation and finetuned chat models. Co RR, abs/2307.09288, 2023. doi: 10.48550/ar Xiv.2307.09288. URL https://doi.org/10.48550/ar Xiv.2307.09288. Richard J Vaccaro. A second-order perturbation expansion for the svd. SIAM Journal on Matrix Analysis and Applications, 15(2):661 671, 1994. Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. Attention is all you need. In Isabelle Guyon, Ulrike von Luxburg, Samy Bengio, Hanna M. Wallach, Rob Fergus, S. V. N. Vishwanathan, and Roman Garnett (eds.), Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4-9, 2017, Long Beach, CA, USA, pp. 5998 6008, 2017. URL https://proceedings.neurips.cc/paper/2017/hash/ 3f5ee243547dee91fbd053c1c4a845aa-Abstract.html. Jesse Vig, Sebastian Gehrmann, Yonatan Belinkov, Sharon Qian, Daniel Nevo, Yaron Singer, and Stuart M. Shieber. Investigating gender bias in language models using causal mediation analysis. In Hugo Larochelle, Marc Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin (eds.), Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, Neur IPS 2020, December 6-12, 2020, virtual, 2020. Published as a conference paper at ICLR 2025 Michael E Wall, Andreas Rechtsteiner, and Luis M Rocha. Singular value decomposition and principal component analysis. In A practical approach to microarray data analysis, pp. 91 109. Springer, 2003. Peng Wang, Ningyu Zhang, Xin Xie, Yunzhi Yao, Bozhong Tian, Mengru Wang, Zekun Xi, Siyuan Cheng, Kangwei Liu, Guozhou Zheng, and Huajun Chen. Easyedit: An easy-to-use knowledge editing framework for large language models. Co RR, abs/2308.07269, 2023. doi: 10.48550/ar Xiv. 2308.07269. URL https://doi.org/10.48550/ar Xiv.2308.07269. Xiaohan Wang, Shengyu Mao, Shumin Deng, Yunzhi Yao, Yue Shen, Lei Liang, Jinjie Gu, Huajun Chen, and Ningyu Zhang. Editing conceptual knowledge for large language models. In Yaser Al-Onaizan, Mohit Bansal, and Yun-Nung Chen (eds.), Findings of the Association for Computational Linguistics: EMNLP 2024, Miami, Florida, USA, November 12-16, 2024, pp. 706 724. Association for Computational Linguistics, 2024. URL https://aclanthology. org/2024.findings-emnlp.40. Per-Åke Wedin. Perturbation bounds in connection with singular value decomposition. BIT Numerical Mathematics, 12:99 111, 1972. Suhang Wu, Minlong Peng, Yue Chen, Jinsong Su, and Mingming Sun. Eva-kellm: A new benchmark for evaluating knowledge editing of llms. Co RR, abs/2308.09954, 2023. doi: 10.48550/ARXIV. 2308.09954. URL https://doi.org/10.48550/ar Xiv.2308.09954. Wanli Yang, Fei Sun, Xinyu Ma, Xun Liu, Dawei Yin, and Xueqi Cheng. The butterfly effect of model editing: Few edits can trigger large language models collapse. In Lun-Wei Ku, Andre Martins, and Vivek Srikumar (eds.), Findings of the Association for Computational Linguistics, ACL 2024, Bangkok, Thailand and virtual meeting, August 11-16, 2024, pp. 5419 5437. Association for Computational Linguistics, 2024. doi: 10.18653/V1/2024.FINDINGS-ACL.322. URL https: //doi.org/10.18653/v1/2024.findings-acl.322. Yunzhi Yao, Peng Wang, Bozhong Tian, Siyuan Cheng, Zhoubo Li, Shumin Deng, Huajun Chen, and Ningyu Zhang. Editing large language models: Problems, methods, and opportunities. In Houda Bouamor, Juan Pino, and Kalika Bali (eds.), Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, EMNLP 2023, Singapore, December 610, 2023, pp. 10222 10240. Association for Computational Linguistics, 2023. URL https: //aclanthology.org/2023.emnlp-main.632. Lang Yu, Qin Chen, Jie Zhou, and Liang He. MELO: enhancing model editing with neuronindexed dynamic lora. In Michael J. Wooldridge, Jennifer G. Dy, and Sriraam Natarajan (eds.), Thirty-Eighth AAAI Conference on Artificial Intelligence, AAAI 2024, Thirty-Sixth Conference on Innovative Applications of Artificial Intelligence, IAAI 2024, Fourteenth Symposium on Educational Advances in Artificial Intelligence, EAAI 2014, February 20-27, 2024, Vancouver, Canada, pp. 19449 19457. AAAI Press, 2024. doi: 10.1609/AAAI.V38I17.29916. URL https://doi. org/10.1609/aaai.v38i17.29916. Ningyu Zhang, Yunzhi Yao, Bozhong Tian, Peng Wang, Shumin Deng, Mengru Wang, Zekun Xi, Shengyu Mao, Jintian Zhang, Yuansheng Ni, Siyuan Cheng, Ziwen Xu, Xin Xu, Jia-Chen Gu, Yong Jiang, Pengjun Xie, Fei Huang, Lei Liang, Zhiqiang Zhang, Xiaowei Zhu, Jun Zhou, and Huajun Chen. A comprehensive study of knowledge editing for large language models. Co RR, abs/2401.01286, 2024. doi: 10.48550/ARXIV.2401.01286. URL https://doi.org/10. 48550/ar Xiv.2401.01286. Yue Zhang, Yafu Li, Leyang Cui, Deng Cai, Lemao Liu, Tingchen Fu, Xinting Huang, Enbo Zhao, Yu Zhang, Yulong Chen, Longyue Wang, Anh Tuan Luu, Wei Bi, Freda Shi, and Shuming Shi. Siren s song in the AI ocean: A survey on hallucination in large language models. Co RR, abs/2309.01219, 2023. doi: 10.48550/ar Xiv.2309.01219. URL https: //doi.org/10.48550/ar Xiv.2309.01219. Chujie Zheng, Fan Yin, Hao Zhou, Fandong Meng, Jie Zhou, Kai-Wei Chang, Minlie Huang, and Nanyun Peng. On prompt-driven safeguarding for large language models. In ICLR 2024 Workshop on Secure and Trustworthy Large Language Models. Published as a conference paper at ICLR 2025 Zexuan Zhong, Zhengxuan Wu, Christopher D. Manning, Christopher Potts, and Danqi Chen. Mquake: Assessing knowledge editing in language models via multi-hop questions. In Houda Bouamor, Juan Pino, and Kalika Bali (eds.), Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, EMNLP 2023, Singapore, December 610, 2023, pp. 15686 15702. Association for Computational Linguistics, 2023. URL https: //aclanthology.org/2023.emnlp-main.971. Published as a conference paper at ICLR 2025 A THEORETICAL ANALYSIS BASED ON PERTURBATION THEORY Here, we provide a detailed analysis and proof of Section 3.2. We begin by introducing some definitions and then present several preliminary lemmas and theorems. These lemmas and theorems are finally used to prove Theorem 3, which is most relevant to our problem discussed in Section 3.2. A.1 DEFINITION We discuss the problem Ax = b, where A is a perturbation of A given by A = A + E. We assume b remains unchanged and x represents the corresponding change, satisfying A x = b. Here A Cm n, b Cm. It is noteworthy that in the following derivation, AH denotes the conjugate transpose of A, A represents the generalized inverse of A, and represents 2-norm (Stewart & Sun, 1990). To simplify the problem, we apply a rotation. Specifically, let V = (V1 V2) be a unitary matrix with R(V1) = R(AH), and let U = (U1 U2) be a unitary matrix with R(U1) = R(A), where R refers to the rank. Then U HAV = U H 1 AV1 U H 1 AV2 U H 2 AV1 U H 2 AV2 = A11 0 0 0 where A11 is square and nonsingular. If we set U HEV = U H 1 EV1 U H 1 EV2 U H 2 EV1 U H 2 EV2 = E11 E12 E21 E22 U H AV = A11 + E11 E12 E21 E22 = A11 E12 E21 E22 We will call these transformed, partitioned matrices the reduced form of the problem. Many statements about the original problem have revealing analogues in the reduced form. In this form, x is replaced by V Hx and b is replaced by U Hb. If x and b are partitioned in the forms , b = b1 b2 where x1, b1 Cr, then x1 = A 1 11 b1 (11) and x2 = 0. (12) Moreover, the norm of the residual vector r = b Ax (13) is given by r = b2 . (14) Here, we define the symbol η: and for any F Ck r (k r) the symbol Ψ(F), for the spectral norm: Ψ2(F) = F (1 + F 2)1/2 . (16) Published as a conference paper at ICLR 2025 A.2 PRELIMINARY LEMMAS & THEOREMS After introducing some definitions, we give some preliminary lemmas and theorems, which are used to prove Theorem 3. Lemma 1 Let κ(A) = A A 1 be the condition number of A. If A is nonsingular, then If in addition E A κ(A) < 1, then A is perforce nonsingular and Moreover A 1 A 1 Lemma 2 In the reduced form the matrices A and A are acute if and only if A11 is nonsingular and E22 = E21 A 1 11 E12. (20) In this case, if we set F21 = E21 A 1 11 and F12 = A 1 11 E12, then A = I F21 A11 (I F12) A = (I F12) A 1 11 Lemma 3 The matrix I F = Ψ2(F). (23) Theorem 1 Let A be an acute perturbation of A, and let ˆκ = A A 1 11 . (24) Published as a conference paper at ICLR 2025 , I12 = (I 0) , (26) J21 = I F21 , J12 = (I F12) . (27) A = J 12A 1 11 I 21, hence A A = (J 12 I 12)A 1 11 I 21 + J 12A 1 11 (J 21 I 21) + J 12( A 1 11 A 1 11 )J 21. (28) From Lemma 1 we have the following bound: J 12( A 1 11 A 1 11 )J 21 A 1 11 ˆκ E11 (J 12 I 12)A 1 11 I 21 A 1 11 J 12 I 12 = A 1 11 Ψ2(F12) (30) = A 1 11 Ψ2( A 1 11 E12) (31) and likewise J 12A 1 11 (J 21 I 21) A 1 11 Ψ2 Theorem 2 In Theorem 1, let κ = A A , (34) and suppose that A E11 < 1, (35) A > 0. (36) Proof. From the equation A = J 12 A 1 11 J 21, we have A J 12 A 1 11 J 21 A 1 11 . (39) By Lemma 1, A 1 11 A 1 11 γ = A which establishes equation 37. Also ˆκ κ γ , and the inequality equation 38 follows from equation 25. Published as a conference paper at ICLR 2025 A.3 CORE THEOREM Finally, we give the core theorem used in main paper. Some symbols and definitions have been claimed in Appendix A.1 and A.2. Theorem 3 Let x = A b and x = A b, where A = A + E, and E is an acute perturbation of A. Then x x Proof. By Lemma 2, write x x = J 12( A 1 11 A 1 11 )b1 + (J 12 I 12)A 1 11 b1 + J 12 A 1 11 (J 21 I 21)b. (42) J 12( A 1 11 A 1 11 )b1 ˆκ E11 (J 12 I 12)A 1 11 b1 Ψ2 Now J 12 A 1 11 (J 21 I 21)b = J 12 A 1 11 ((I + F H 21F21) 1 I)b1 + J 12 A 1 11 (I + F H 21F21) 1F H 21b2. (45) To bound the first term in equation 45, note that (I + F H 21F21) 1 I = (I + F H 21F21) 1F H 21F21. Hence J12 A 1 11 ((I + F H 21F21) I)b1 A 1 11 (I + F H 21F21) 1 F H 21 F21b1 A 1 11 E21 A 1 11 b1 A 1 11 E21 2 x For the second term in equation 45 we have J 21 A 1 11 (I + F H 21F21) 1F21b2 A 1 11 2 E21 b2 = A 1 11 2 E21 b2 η 1ˆκ2 E21 b2 A b1 x . (47) The bound equation 41 follows on combining equation 42 equation 47. Readers can refer to this work (Stewart & Sun, 1990) for more details of perturbation analysis. Returning to our problem, consider Wk = v, where (k, v) P. Let W = W + W, where W is the corresponding perturbation matrix. Assuming v remains constant, there exists k such that k = k + k satisfies W k = v. And we have k = W v and k = W v. Applying Theorem 3, we obtain k where E11, E12, E21, and W are directly related, and each term on the right-hand side involves ˆκ. This means that the relative perturbation of the vector k is constrained by ˆκ. According to Theorem 2, ˆκ κ γ , where κ = W W is the condition number of W. This indicates that κ is a robust indicator of the impact of W on the vector k. Published as a conference paper at ICLR 2025 B EXPERIMENTAL SETUP B.1 BASELINE EDITING METHODS Three popular model editing methods were selected as baselines including: MEND (Mitchell et al., 2022a)6: it learned a hypernetwork to produce weight updates by decomposing the fine-tuning gradients into rank-1 form. ROME (Meng et al., 2022)7: it first localized the factual knowledge at a specific layer in the transformer MLP modules, and then updated the knowledge by directly writing new key-value pairs in the MLP module. MEMIT (Meng et al., 2023)8: it extended ROME to edit a large set of facts and updated a set of MLP layers to update knowledge. The ability of these methods were assessed based on Easy Edit9 (Wang et al., 2023), an easy-to-use knowledge editing framework which integrates the released codes and hyperparameters from previous methods. B.2 EDITING DATASETS AND EVALUATION METRICS Table 3 shows the examples of two factual datasets (ZSRE) (Levy et al., 2017) and COUNTERFACT (Meng et al., 2022). Figure 6 shows an example of Concept Edit dataset, which is cited from Wang et al. (2024). More details can refer to the original paper of these datasets. Table 3: The editing datasets of both ZSRE and COUNTERFACT. Datasets Editing prompt ZSRE Which was the record label for New Faces, New Sounds? COUNTERFACT In America, the official language is Figure 6: An example of Concept Edit dataset Besides, following previous works (Meng et al., 2022; Mitchell et al., 2022a; Meng et al., 2023), the editing performance metrics for the ZSRE and COUNTERFACT datasets are efficacy, generalization and locality, but there are some computational differences. In the main paper, the metrics of editing performance are used for the ZSRE dataset. For the COUNTERFACT dataset, here are the details: Efficacy validates whether the edited models could recall the editing fact under editing prompt pi. The assessment is based on Efficacy Score (ES) representing as: Ei[1[ Pfθn(o i | pi) > Pfθn(oi | pi)] ], where 1 is the indicator function. 6https://github.com/eric-mitchell/mend 7https://github.com/kmeng01/rome 8https://github.com/kmeng01/memit 9https://github.com/zjunlp/Easy Edit Published as a conference paper at ICLR 2025 Generalization verifies whether the edited models could recall the editing fact under the paraphrase prompts PG i via Generalization Score (GS): Ei [ Ep PG i [1[ Pfθn(o i | p) > Pfθn(oi | p)] ]. Locality verifies whether the output of the edited models for inputs out of editing scope remains unchanged under the locality prompts PL i via Locality Score (LS): Ei [ Epl PL i [1[ Pfθn(ol | pl) > Pfθn(o i | pl)] ] ], where ol was the original answer of pl. B.3 HYPERPARAMETERS OF PRUNE When conducting experiments, for different editing methods, LLMs and editing datasets, the hyperparameter α in function F of PRUNE is different. Table 4 shows the details of this hyperparameter. e is the base of the natural logarithm. Table 4: The hyperparameters α for PRUNE. Datasets Models ROME MEMIT MEND COUNTERFACT GPT-2 XL 1.2 1.2 1.2 LLa MA-2 1.2 e 1.2 LLa MA-3 1.5 e - ZSRE LLa MA-2 1.2 e e B.4 TASK PROMPTS The prompts for each downstream task were illustrated in Table 5. Table 5: The prompts to LLMs for evaluating their zero-shot performance on these general tasks. Reasoning: Q: {QUESTION} A: Let s think step by step. {HINT} Therefore, the answer (arabic numerals) is: NLI: {SENTENCE1} entails the {SENTENCE2}. True or False? answer: Open-domain QA: Refer to the passage below and answer the following question. Passage: {DOCUMENT} Question: {QUESTION} Summarization: {DIALOGUE} TL;DR: B.5 EXPERIMENTS COMPUTE RESOURCES We used NVIDIA A800 80GB GPU for experiments. For LLa MA-2 (7B) and LLa MA-3 (8B), it occupies about 40+GB memory and costs about 3 hours for each editing method to run 200 edits and then to test downstream tasks . For GPT-2 XL (1.5B), it needs 10+GB and costs about 1.5 hours for each editing method to run 200 edits and then to test downstream tasks. C EXPERIMENTAL RESULTS C.1 RESULTS OF GENERAL ABILITIES Figure 7, 8 and 9 show the downstream task performance of edited models with GPT-2 XL, LLa MA-2 (7B) and LLa MA-3 (8B) on COUNTERFACT dataset. Due to limitations of computing resources, experiments were conducted using only LLa MA-2 (7B) on the ZSRE dataset. We will supplement experiments with other LLMs in the future. Published as a conference paper at ICLR 2025 Figure 7: The downstream task performance (%) of models edited by three editing methods with GPT-2 XL on the COUNTERFACT dataset. Figure 8: The downstream task performance (%) of models edited by three editing methods with LLa MA-2 (7B) on the COUNTERFACT dataset. Figure 9: The downstream task performance (%) of models edited by two editing methods with LLa MA-3 (8B) on the COUNTERFACT dataset. Since the code framework Easy Edit used in this paper does not currently support MEND editing on LLa MA-3, there are no results of MEND here. Published as a conference paper at ICLR 2025 C.2 RESULTS OF EDITING PERFORMANCE Figure 10, 11 and 12 shows the editing performance of edited models with GPT-2 XL, LLa MA-2 (7B) and LLa MA-3 (8B) on COUNTERFACT dataset. Figure 10: The editing performance (%) of three editing methods with GPT-2 XL on COUNTERFACT. Figure 11: The editing performance (%) of three editing methods with LLa MA-2 (7B) on the COUNTERFACT dataset. Figure 12: The editing performance (%) of three editing methods with LLa MA-3 (8B) on COUNTERFACT dataset. C.3 RESULTS OF ANOTHER FUNCTION FOR PRUNE In the main paper, log function is used in F in PRUNE to restrain ˆσi. Here we use the linear function, which could be represented as: F(ˆσi) = 1 β ˆσi + β 1 β max{σi}. Here β > 1 was a hyperparameter and was set as 2 in this section. Figure 13 and 14 respectively show some downstream task performance and editing performance with linear function on COUNTERFACT dataset. Compared with Figure 7 and 10, we observed that although the linear function in PRUNE played a role in preserving general abilities and maintaining editing performance, its effectiveness was noticeably inferior to that of the log function when the number of edits was large. Published as a conference paper at ICLR 2025 Figure 13: The downstream task performance (%) of models edited by three editing methods with GPT-2 XL on the COUNTERFACT dataset. Here the linear function was used in PRUNE. Figure 14: The editing performance (%) of editing methods with GPT-2 XL on the COUNTERFACT dataset. Here the linear function was used in PRUNE. C.4 CONDITION NUMBER WITH PRUNE Figure 15 shows after coupling with PRUNE, the condition number of MEMIT is significantly restrained. W W20 W50 W100 W150 W180 0 Condition Number Condition Number Condition Number-PRUNE Figure 15: The condition number of MEMIT with LLa MA-2 (7B) on the COUNTERFACT dataset. -PRUNE refers to the condition number of MEMIT coupled with the proposed PRUNE. C.5 THE CORRELATION BETWEEN CONDITION NUMBER AND GENERAL ABILITIES Figure 16 simultaneously shows the condition number and general abilities of three editing methods without PRUNE in the sequential editing process. From these experiments, we observed that a dramatic increase in the condition number is often accompanied by a rapid decline in general abilities. Published as a conference paper at ICLR 2025 Figure 16: The condition number and downstream task performance of three editing methods with LLa MA-2 (7B) on the COUNTERFACT dataset. Since MEMIT and MEND have multiple parameters to be edited, we randomly selected one of them to calculate the condition number. C.6 ENHANCED PRUNE We find applying multiple operation of PRUNE for longer-term performances will perform better than only once. The results of applying PRUNE once have been shown in our paper. Here, we make a comparison. For convenience, we list some results of using ROME to edit GPT 2-XL on the Counter Fact dataset. Table 6: comparison of applying once operation of PRUNE and applying multiple multiple operation. Mode General Abilities Editing Performance Method Edits Reasoning Summa Open-QA NLI Efficacy General Locality ROME+PRUNE (once) 100 30.12 12.12 22.63 49.83 99 92 74.10 500 27.26 10.78 17.86 45.34 56.80 46.05 72.32 ROME+PRUNE (multiple) 100 30.17 11.38 22.99 48.83 100 98 76.1 500 32.68 11.52 22.99 51.17 84.6 78.7 75.1 D BROADER IMPACTS This work offers significant advancements in the field of model editing for LLMs. By addressing the challenge of preserving general abilities while performing sequential edits, PRUNE facilitates continual learning and adaptability in LLMs. This can lead to several positive impacts, such as: Enhanced Adaptability. It enables LLMs to update their knowledge base quickly and accurately without extensive retraining. This adaptability is crucial in dynamic environments where up-todate information is vital, such as real-time translation services, personalized learning systems, and interactive virtual assistants. Resource Efficiency. By mitigating the need for full retraining, PRUNE significantly reduces computational resources and energy consumption. This aligns with sustainable AI and makes it more feasible to deploy LLMs in resource-constrained settings. Improved Performance in Specialized Tasks. PRUNE s ability to perform targeted edits without compromising overall model performance can enhance LLMs effectiveness in specialized domains, such as medical diagnostics, legal analysis, and technical support, where precise and updated knowledge is essential. While this work offers many benefits, there are potential negative societal impacts that must be considered: Misuse for Malicious Purposes. The capability to edit LLMs efficiently could be exploited to inject harmful or biased information into models, thereby spreading disinformation or propaganda. This risk is particularly concerning in applications involving social media and news dissemination where LLMs might generate or amplify misleading content. Fairness. Unintended biases could be introduced during the editing process, potentially exacerbating existing biases in LLMs. This could lead to unfair treatment or misrepresentation of specific Published as a conference paper at ICLR 2025 groups, especially if the editing is not conducted with proper oversight and consideration of ethical implications. Privacy Concerns. The ability to update models quickly might also pose privacy risks, as models could be edited to include sensitive or personal information. Ensuring that editing processes do not compromise individual privacy is critical, particularly in applications involving personal data. To mitigate these potential negative impacts, several strategies could be implemented: Gated Release and Monitoring. Limiting access to the framework through gated releases and monitoring its usage can help prevent misuse. Bias and Fairness Audits. Conducting regular audits to assess and address biases in the model editing process can help ensure that edits do not unfairly impact any specific group. Developing guidelines for ethical editing practices is also essential. Privacy Protection Measures. Establishing clear protocols for handling sensitive data during the editing process can help protect privacy. Anonymization and encryption techniques should be employed to safeguard personal information.