# hessianfree_online_certified_unlearning__563d56b2.pdf Published as a conference paper at ICLR 2025 HESSIAN-FREE ONLINE CERTIFIED UNLEARNING Xinbao Qiao1 Meng Zhang1,* Ming Tang2 Ermin Wei3 1Zhejiang University 2Southern University of Science and Technology 3Northwestern University xinbaoqiao@zju.edu.cn Machine unlearning strives to uphold the data owners right to be forgotten by enabling models to selectively forget specific data. Recent advances suggest precomputing and storing statistics extracted from second-order information and implementing unlearning through Newton-style updates. However, the Hessian matrix operations are extremely costly and previous works conduct unlearning for empirical risk minimizer with the convexity assumption, precluding their applicability to high-dimensional over-parameterized models and the nonconvergence condition. In this paper, we propose an efficient Hessian-free unlearning approach. The key idea is to maintain a statistical vector for each training data, computed through affine stochastic recursion of the difference between the retrained and learned models. We prove that our proposed method outperforms the state-of-theart methods in terms of the unlearning and generalization guarantees, the deletion capacity, and the time/storage complexity, under the same regularity conditions. Through the strategy of recollecting statistics for removing data, we develop an online unlearning algorithm that achieves near-instantaneous data removal, as it requires only vector addition. Experiments demonstrate that our proposed scheme surpasses existing results by orders of magnitude in terms of time/storage costs with millisecond-level unlearning execution, while also enhancing test accuracy. 1 INTRODUCTION Recent data protection regulations, e.g. General Data Protection Regulation, aim to safeguard individual privacy. These regulations emphasize the provisions of the Right to be Forgotten, which grants individuals the ability to request the removal of personal data. This regulation mandates that the data controller shall erase personal data without undue delay when data providers request to delete it. Within the context of the right to be forgotten, data owners can request the removal of their personal data s contribution from trained models. This can be achieved through the concept of machine unlearning, which involves erasing data that pertains to the data owner s personal information. A straightforward unlearning algorithm is to retrain a learned model from scratch, which is however impractical as it incurs unaffordable costs. It is especially challenging in time-sensitive applications such as fraud detection or online learning, as delayed updates upon receiving a deletion request would impede the system s ability to respond effectively to subsequent inputs, compromising its performance and reliability. Moreover, attaining a model entirely identical to the one achieved through retraining for an unlearning method represents a notably demanding criterion. Therefore, many approximate methods have been proposed aiming to minimize the impact of forgetting data. Inspired by differential privacy, Guo et al. (2020) introduced a relaxed definition of certified unlearning, which aims to make the output distribution of unlearned algorithm indistinguishable from that of the retraining. Second-order unlearning methods have been recently studied due to their rigorous unlearning certification and the generalization guarantee. Sekhari et al. (2021) initially proposed the Newton update method and introduced the definition of deletion capacity to understand the relationship between generalization performance and the amount of deleted samples. These method and concepts were later extended, e.g. Suriyakumar et al. (2022); Liu et al. (2023); Mahadevan & Mathioudakis (2021); Chien et al. (2022). The key idea of these works is to extract Hessian information of the loss function and achieve computational and memory efficiency by pre-computing and storing these second-order statistics prior to any deletion request. In particular, Sekhari et al. (2021) and Suriyakumar et al. Corresponding Author Published as a conference paper at ICLR 2025 Table 1: Summary of Results. Here, d, n, and m denote the model parameters size, training dataset size, and forgetting dataset size, respectively. The unlearning guarantee is derived under the same regularity conditions as prior second-order unlearning studies, with ρ being a constant less than 1. Methods Computation time Storage Unlearning time Unlearning guarantee Sekhari et al. (2021) O(nd2) O(d2) O(d3 + md2 + md) O(m2/n2) Suriyakumar et al. (2022) O(d3 + nd2) O(d2) O(d2 + md) O(m2/n2) Proposed method O(n2d) O(nd) O(md) O(mρn) (2022) revealed that for models that approximate the minimization of convex and sufficiently smooth empirical minimization objectives, unlearning can be efficiently executed through a Newton style step, facilitated by pre-storing Hessian matrix. These methods and theoretical analyses are subsequently extended to graph learning (Chien et al. (2022)) and minimax learning (Liu et al. (2023)) scenarios. While these works have demonstrated the effectiveness of the second-order method to approximate retrained models with limited updates, they do have a few fundamental challenges: existing secondorder algorithms (i) lack the capability to handle multiple deletion requests in online because they require explicit precomputing Hessian (or its inverse), and (ii) the learned model needs to be empirical risk minimizer, (iii) with assumption of convexity to ensure the invertibility of Hessian. Consequently, these limitations preclude their applicability to high-dimensional models (e.g., over-parameterized neural networks) for computing the Hessian. For non-convex problems (e.g., deep networks), we cannot guarantee the invertibility of the Hessian, and it is non-trivial for a learned model to converge to an empirical risk minimizer. Even for convex problems, the non-stationarity of data (Hazan (2016)) is another common issue in the incremental model update scenarios (e.g., continual learning) where training is sequential but not concurrent, making it challenging for the model to adapt and converge to the optimal solution. Consequently, we are motivated to answer the following key question: (Q) How can efficient second-order unlearning be achieved without incurring the high cost of Hessian operations, and how can effective certified unlearning be implemented under non-convex assumptions for the objectives and non-convergence conditions for the models? In this paper, we take the first step towards tackling (Q) from an innovative Hessian-free perspective. In contrast to existing schemes that extract second-order information exclusively from the optimal learned model, we leverage details from learning models during training procedure. We analyze the difference between the learning model and the retraining model by recollecting the model update trajectory discrepancy. Subsequently, we demonstrate that the discrepancies can be characterized through an affine stochastic recursion. We summarize our main contributions below and in Table 1. Hessian-Free Model Update Method. We propose a method to extract the Hessian information by analyzing each sample s impact on the training trajectory without explicitly computing the Hessian matrix. Our Hessian-free method enables efficient handling of deletion requests in an online manner. Moreover, the proposed method does not require that the learned model is empirical risk minimizer, which enjoys better scalability compared to existing second-order methods. (Section 3). Improved Guarantees and Complexity. Under identical regularity conditions, we analytically show that our proposed method outperforms the state-of-the-art theoretical studies in terms of unlearning guarantee (See Corollary 5), generalization guarantee (See Theorem 6), deletion capacity (See Theorem 7), and computation/storage complexity (See Subsection 4.4). Moreover, the unlearning guarantee can be applied to both convex and non-convex settings. (Section 4). Instantaneous Unlearning Algorithm. Building upon the aforementioned analysis, we develop a Hessian-free certified unlearning algorithm. To our knowledge, our algorithm is near-instantaneous and is currently the most efficient as it requires only a vector addition operation. (Section 4). Extensive Evaluation. We conduct experimental evaluations using a wider range of metrics compared to previous theoretical second-order studies, and release our open source code.1 The experimental results verify our theoretical analysis and demonstrate that our proposed approach surpasses previous certified unlearning works. In particular, our algorithm incurs millisecond-level unlearning runtime to forget per sample with minimal performance degradation. (Sections 5 and 6). 1Our code is available at Hessian Free Certified Unlearning Published as a conference paper at ICLR 2025 1.1 RELATED WORK Machine unlearning can be traced back to Cao & Yang (2015), which defines "the completeness of forgetting", demanding the models obtained through retraining and unlearning exhibit consistency. Exact Unlearning, e.g., Bourtoule et al. (2021); Brophy & Lowd (2021); Schelter et al. (2021); Yan et al. (2022); Chen et al. (2022b;a), typically train the submodels based on different dataset partitions and aggregation strategies (See elaborated exact unlearning background in Appendix A). Approximate Unlearning, e.g. Wu et al. (2020); Nguyen et al. (2020); Golatkar et al. (2020a); Izzo et al. (2021); Mehta et al. (2022); Wu et al. (2022); Tanno et al. (2022); Becker & Liebig (2022); Jagielski et al. (2023); Warnecke et al. (2023); Tarun et al. (2023), seeks to minimize the impact of forgetting data to an acceptable level to tradeoff for computational efficiency, reduced storage costs, and flexibility (See elaborated approximate unlearning background in Appendix A). In particular, Guo et al. (2020) introduces a relaxed definition of (ϵ, δ)-unlearning, in anticipation of ensuring the output distribution of an unlearning algorithm is indistinguishable from that of the retraining. Building upon this framework, diverse methodologies, e.g. Sekhari et al. (2021); Suriyakumar et al. (2022); Gupta et al. (2021); Neel et al. (2021); Liu et al. (2023); Chien et al. (2022), have been devised. In the certified unlearning definition, the Hessian-based approaches (Sekhari et al. (2021); Suriyakumar et al. (2022); Liu et al. (2023)) are aligned with our methodology, which involves pre-computing data statistics that extract second-order information from the Hessian and using these precomputed "recollections" to facilitate data forgetting. Specifically, (1) Newton Step (NS). Sekhari et al. (2021) proposed, through the precomputation and storage of the Hessian offline, to achieve forgetting upon the arrival of a forgetting request using a Newton step. However, NS is limited by its high computational complexity of O(d3), making it impractical for large-scale deep models. Furthermore, it can only perform certified unlearning for empirical minimizers under convexity assumption. (2) Infinitesimal Jackknife (IJ). Suriyakumar et al. (2022) extended the work of NS, attaining strong convexity through a non-smooth regularizer by leveraging the proximal IJ with reduced unlearning time O(d2). However, a few important challenges still remain, i.e., existing second-order methods still require expensive Hessian computations and perform forgetting for empirical risk minimizer with the convexity assumption. In contrast, our method overcomes the limitations of existing second-order methods by achieving near-instantaneous unlearning and performing forgetting for models in any training step. Hence, our approach is applicable to scenarios where models are continuously updated. 2 PROBLEM FORMULATION Learning. Let ℓ(w;z) be loss function for a given parameter w over parameter space W and sample z over instance space Z. Learning can be expressed as a population risk minimization problem: min w F(w) := Ez D[ℓ(w; z)]. (1) Addressing this problem directly is challenging, as the probability distribution D is usually unknown. For a finite dataset S = {zi}n i=1, one often solves the following empirical risk minimization problem: min w FS(w) := 1 i=1 ℓ(w; zi). (2) A standard approach to solve the problem (2) is the stochastic gradient descent (SGD), iterating towards the direction with stepsize ηe,b of the negative gradient (stochastically): we,b+1 we,b ηe,b i Be,b ℓ(we,b; zi) , (3) for the index of epoch update e {0, ..., E+1} and batch update b {0, ..., B+1}, where E is total epochs and B is batches per epoch. Here, Be,b denotes the set that contains the data sampled in the e-th epoch s b-th update, and |Be,b| is the size of Be,b, where |B| is the maximum batchsize. Unlearning. Consider a scenario where a user requests to remove a forgetting dataset U={uj}m j=1 S. To unlearn U, retraining from scratch is considered a naive unlearning method, which aims to minimize FS\U(w). Specifically, for ease of understanding, each sample appears only in one (mini-)batch per epoch, we define the data point uj in U to be sampled in Be,b(uj) during the e-th Published as a conference paper at ICLR 2025 epoch s b(uj)-th update in the learning process, where b(uj) represents the batch update index when uj is sampled. In accordance with the linear scaling rule discussed in Goyal et al. (2017), i.e. in order to effectively leverage batch sizes, one has to adjust the step size in tandem with batch size. In other words, the step size to batch size ratio during retraining is consistent with the learning process. Therefore, the update rule of retraining process during e-th epoch is given by, w U e,b+1 w U e,b ηe,b i Be,b ℓ(w U e,b ; zi). (4) Specifically, for the training data uj sampled into Be,b(uj) in the learning process, the removal of sample uj in the retraining process results in the following update: w U e,b(uj)+1 w U e,b(uj) ηe,b(uj) i Be,b(uj )\{uj} ℓ(w U e,b(uj); zi). (5) To avoid the unaffordable costs incurred by the retraining algorithm, we aim to approximate the retrained model through limited updates on the learned model we,b+1. We thus consider the relaxed definition of model indistinguishability defined by Sekhari et al. (2021). Definition 1 ((ϵ, δ)-certified unlearning). Let S be a training set and Ω: Zn W be an algorithm that trains on S and outputs a model w, T (S) represents the additional data statistics that need to be stored (typically not the entire dataset). Given an output of the learning algorithm Ω W and a set of data deletion requests U, we obtain the results of the unlearning algorithm Ω(Ω(S), T (S)) W and the retraining algorithm Ω(Ω(S\U), ) through a removal mechanism Ω. For 0 < ϵ 1, δ > 0, and W Rd, S Z, we say that the removal mechanism Ωsatisfies (ϵ, δ)-certified unlearning if P( Ω(Ω(S), T (S)) eϵ P( Ω(Ω(S\U), ))) + δ, and P( Ω(Ω(S\U), ) eϵ P( Ω(Ω(S), T (S)) + δ. (6) Although classical Gaussian mechanisms (Dwork & Roth (2014)) assume 0 < ϵ 1, Balle & Wang (2018) addressed ϵ > 1 by using Gaussian cumulative density function instead of a tail bound approximation. For consistency with prior works, we use classical mechanisms in the main text. 3 MAIN INTUITION AND SOLUTION JUSTIFICATION Before detailing our analysis, let us informally motivate our main method. Specifically, we focus on studying the discrepancy between the learning and the retraining model by analyzing the impact of each sample on the update trajectory. Our main observation is that the difference between the retraining and learning models analyzed through SGD recursion can be described by an affine stochastic recursion. Suppose retraining and learning processes share the same initialization, such as using the same random seed or pre-trained model. We now commence the main procedure as follows: Unlearning a single data sample. We start with designing an approximator for a single sample (i.e., U ={u}) on the training (the learning and retraining) updates, considering the difference between the learning model in Equation (3) and the retraining model obtained without the knowledge of u in Equation (4). In particular, we consider the following two cases of model updates in the e-th epoch. Case 1 (u not in Be,b): The difference between retraining and learning updates in the e-th epoch is w u e,b+1 we,b+1 = w u e,b we,b ηe,b ℓ(w u e,b ; zi) ℓ(we,b; zi) . (7) Suppose loss function ℓ G2 Rd , i.e. ℓis twice continuously differentiable. Therefore, from the Taylor expansion of ℓ(w u e,b ; zi) around we,b, we have ℓ(w u e,b ; zi) ℓ(we,b; zi) + 2ℓ(we,b; zi)(w u e,b we,b). Let He,b=1/|Be,b| P i Be,b 2ℓ(we,b; zi) denote the Hessian matrix of the loss function, we can then approximate the recursion of Equation (7) as w u e,b+1 we,b+1 (I ηe,b He,b)(w u e,b we,b). (8) Case 2 (u in Be,b(u)): The difference between retraining and learning updates in the e-th epoch is: w u e,b(u)+1 we,b(u)+1 (I ηe,b He,b)(w u e,b(u) we,b(u))+ ηe,b(u) |Be,b(u)| ℓ(we,b(u); u) | {z } Approximate impact of u in epoch e Published as a conference paper at ICLR 2025 where I denotes identity matrix of appropriate size. Combining the observations in Cases 1 and 2, along with same initialization, we now define the recollection matrix Me,b(u) and the approximator a u E,B, backtracking the overall difference between the retrained model and learned model: e=0 Me,b(u) ℓ(we,b(u); u), where Me,b(u) := ηe,b(u) |Be,b(u)| ˆHE,B 1 e,b(u)+1. (10) We define ˆHE,B 1 e,b(u)+1 = (I ηE,B 1HE,B 1)(I ηE,B 2HE,B 2)...(I ηe,b(u)+1He,b(u)+1), which represents the product of I ηH from E-th epoch s B 1-th update to e-th epoch s b(u)+1-th update.2 This provides an approximation of the impact of sample u on the training trajectory: w u E,B w E,B a u E,B (11) Colloquially, a simple vector a u E,B addition to model w E,B can approximate retrained model w u E,B for forgetting the sample u. We define u E,B =w u E,B w E,B a u E,B as the approximation error. The approximator designed in Equation (11) enables us to exploit the Hessian vector product (HVP) technique (Pearlmutter (1994)) to compute it without explicitly computing Hessian. When computing the product of the Hessian matrix and an arbitrary vector for the neural networks, the HVP technique first obtains gradients during the first backpropagation, multiplying the gradients with the vector, and then performing backpropagation again. This results in a time complexity of O(d) (Dagréou et al. (2024)), which has the same order of magnitude as computing a gradient. We defer the detailed calculation process of Equation (11) using HVP to Appendix B.3. We note that the existing second-order methods require proactively storing a Hessian (or its inverse) (Sekhari et al. (2021); Suriyakumar et al. (2022); Liu et al. (2023)). It is thus unable to deploy HVP in their settings. Note that our proposed analysis does not require to invert Hessian matrix, distinguishing it from existing Newton-like methods (Sekhari et al. (2021); Suriyakumar et al. (2022); Liu et al. (2023)). This eliminates the need for assuming convexity to ensure the invertibility of the Hessian matrix. Moreover, when it comes to forgetting a set of samples, we have, Theorem 1 (Additivity). When m sequential deletion requests arrive, the sum of m approximators is equivalent to performing batch deletion simultaneously. Colloquially, for u1, . . . , um sequence of continuously arriving deletion requests, we demonstrate that a Pm j=1 uj E,B = Pm j=1 a uj E,B. See complete analysis and proofs of Section 3 in Appendix C.1 and Appendix C.2. This elucidates that by storing the impact of updates for each sample, nearly instantaneous data removal can be achieved through vector additions. Note that it is non-trivial to retrain and store a leave-one-out model for each sample, as such a process cannot accommodate sequential deletion requests. If we were to consider all possible combinations of forgetting data, retraining need to pretrain O(2n) models, which yields computation complexity O(2n EBd). The additivity enables our Hessian-free method to handle multiple deletion requests in an online manner. Previous Hessian-based methods, lacking additivity, fail to utilize HVP as the forgetting sample u is unknown before a deletion request arrives. 4 THEORETICAL ANALYSIS AND ALGORITHM DESIGN In this section, we theoretically analyze the ❶Unlearning Guarantee (Theorem 4), ❷Generalization Guarantee (Theorem 6), and ❸Deletion Capacity Guarantee (Theorem 8). Particularly, we demonstrate that under the Assumption 1, our method can outperform existing theoretical certified unlearning methods (Sekhari et al. (2021); Suriyakumar et al. (2022); Liu et al. (2023)) in terms of unlearning guarantee, generalization guarantee, and deletion capacity. In contrast to these works, the analysis of unlearning guarantee is applicable to both convex and non-convex settings, as our proposed method does not involve Hessian inversion. Based on the foregoing analysis, we further develop an almost instantaneous online unlearning algorithm, which is one of the most efficient, as it requires only a vector addition operation to delete each training data. Finally, we conduct storage and computation complexity analysis and show that in most scenarios where over-parameterized models and the ratio of E to B are much smaller than the parameters, it is more efficient than previous works. 2Similar analysis has been conducted in other lines of work to investigate the behavior of SGD dynamics, e.g. Gürbüzbalaban et al. (2021) focused on the properties of the matrix I ηH, which is termed as multiplicative noise, serving as the main source of heavy-tails in SGD dynamics and governing the update behavior of we,b. Published as a conference paper at ICLR 2025 4.1 APPROXIMATION ERROR ANALYSIS To analyze the approximation error U E,B, we introduce the following lemmata. We first show that the geometrically decaying stepsize strategy facilitates us to derive an upper bound for the error term. Assumption 1. For any z and w, loss ℓ(w;z) is λ-strongly convex, L-Lipschitz and M-smooth in w. Lemma 2. For any e, b and z Be,b, there exists G such that G = max ℓ(we,b; z) < . Consider geometrically decaying stepsizes satisfying ηe,b+1= qηe,b, where 0 < q < 1 and η = η0. During the e-th epoch s b-th batch update, we have: w U e,b we,b 2ηG1 qe B+b See proof in Appendix C.3. Then, we further analyze the eigenvalue of the Hessian matrix: Lemma 3. For any e, b and vector v Rd, there exists λ and M such that λI 2ℓ(we,b; zi) MI, and let ρ = max{|1 ηe,bλ|, |1 ηe,b M|} be the spectral radius of I ηe,b He,b, we have (I ηe,b He,b)v ρ v . Combining Equation (11), Lemma 2 and Lemma 3, we have the approximation error bound: Theorem 4 (Unlearning Guarantee). For any subset U ={uj}m j=1 to be forgotten, with 0 E < and q<ρ, the unlearning model w U E,B =w E,B+a U E,B yields an approximation error bounded by U E,B = w U E,B w U E,B O(ηGmρn E/|B|), (13) where the total number of batch updates per epoch B = n/|B| . See proofs and analysis in Appendix C.4. We note that in general, Theorem 4 does not require the objectives to be strongly convex as it does not involve inverting the Hessian. Therefore, our method accommodates most non-convex and non-M-smooth functions. On the other hand, considering that previous theoretical works require the convexity assumption, for further comparing with these works and better analyzing the following generalization results in Subsection 4.2, we show that L2regularized convex problems with random regularization coefficients fall into the scope of Theorem 4. To achieve a tighter bound, we consider the standard Assumption 1, which are similar to those in unlearning works (Suriyakumar et al. (2022); Liu et al. (2023)). We have the following conclusion, Corollary 5. Suppose Assumption 1 is satisfied, and we choose η < 2/(M + λ) to ensure gradient descent convergence and ρ < 1. Then, Theorem 4 yields an approximation error bound by O(mρn). Under same Assumption 1, the derived bound surpasses the state-of-the-art O(m2/n2) in Suriyakumar et al. (2022); Liu et al. (2023), representing an improvement for unlearning guarantee. Moreover, Corollary 5 implies that the unlearning model can closely approximate the retrained model. 4.2 GENERALIZATION PROPERTIES ANALYSIS Despite that our proposed method is applicable to both convex and non-convex settings, we temporarily impose Assumption 1 for the analytical simplicity of generalization guarantee. This assumption will be relaxed for algorithm design, unlearning analyses, and experiments. We further analyze the generalization guarantee and compare the results with previous works. Theorem 6 (Generalization Guarantee). Let Assumption 1 hold. Choose η 2 M+λ such that ρ<1 for 0 E < . Let w be the minimizer of the population risk in Equation (1). Considering noise perturbation to ensure (ϵ, δ)-unlearning, the unlearned model w U E,B satisfies the excess risk bound, F( w U E,B) E[F (w )] = O λ(n m) + ρn E/|B| 2L2 See proof in Appendix C.6. The result of our proposed method yields the order O(d 1 2 ρn). Under these conditions and the fact that n = Ω(m), our proposed method outperforms the performance of the state-of-the-art methods (Suriyakumar et al. (2022); Liu et al. (2023)), which are O(d 1 2 /n2). Moreover, to measure the maximum number of samples to be forgotten while ensuring a good excess risk guarantee, we provide our proposed scheme s deletion capacity, following the analysis in Sekhari et al. (2021). Due to the space limit, we defer the definition and analysis to Appendix B.2. Published as a conference paper at ICLR 2025 Algorithm 1: Hessian-Free Online Unlearning (HF) Algorithm 1 Stage I: Learning model w E,B on dataset S = {zi}n i=1: 2 for e = 0, 1 , E + 1 do 3 for b = 0, 1 , B + 1 do 4 Gradient descent: we,b+1 we,b ηe,b 1 |Be,b| P i Be,b ℓ(we,b; zi) 7 Stage II: Pre-computing and Pre-storing unlearning statistics T (S) = {a uj E,B}n j=1: 8 for j = 1, 2 , n do 9 Recursive HVP based on Algorithm 2: a uj E,B PE e=0 Me,b(u) ℓ(we,b(u); u). 11 Stage III: Unlearning when user requests to forget the subset U={uj}m j=1 on model w E,B: 12 Compute: w U E,B w E,B+a U E,B, σ U E,B 2ln(1.25/δ) ϵ , Delete unlearning statistics: {a uj E,B}m j=1 13 Sample: n N(0, σI), Return: w U E,B = w U E,B + n 4.3 EFFICIENT UNLEARNING ALGORITHM DESIGN Building on the foregoing techniques and analysis of Theorem 4, we propose a memory and computationally efficient unlearning scheme in Algorithm 1. To the best of our knowledge, our algorithm is one of the most efficient method, as it only requires a vector addition by utilizing pre-computed statistics. This is particularly advantageous in time-sensitive scenarios and non-optimal models. We present our approach in Algorithm 1. In Stage I, we perform conventional model training. Upon completing the model training, we proceed to Stage II, where we compute the unlearning statistics vector a u E,B for each sample u. To reduce computational complexity, we leverage the HVP technique as described in Appendix B.3. The computed approximators {a uj E,B}n j=1 are then stored for further use.3 A deletion request initiates Stage III. In this stage, the indistinguishability of the (ϵ, δ)-unlearning, as defined in Definition 1, is achieved through a single vector addition with noise perturbation. Notably, Algorithm 1 does not necessitate accessing any datasets during Stage III. 4.4 COMPARISON TO EXISTING APPROACHES We provide a brief overview of the second-order unlearning algorithms and compare the complexity of our method with these works in terms of pre-computation, storage, and unlearning time. Newton Step (NS). Sekhari et al. (2021) proposed a certified unlearning algorithm by pre-computing Hessian Pn i=1 2ℓ( ˆw;zi). When a request arrives to forget U={uj}m j=1, the update for NS is as follows: w U NS = ˆw + 1 n m i=1 2ℓ( ˆw; zi) X j U 2ℓ( ˆw; uj) -1 X j U ℓ( ˆw; uj). (15) Infinitesimal Jackknife (IJ). Suriyakumar et al. (2022) builded upon the work of NS and facilitated forgetting by pre-computing and storing the inverse Hessian. The update for NS thus is as follows: w U IJ = ˆw + 1 i=1 2ℓ( ˆw; zi) -1 X j U ℓ( ˆw; uj). (16) These methods fail to use HVP, for Hessian P j U 2ℓ( ˆw; uj) in NS and gradient ℓ( ˆw, uj) in both NS and IJ, (i) forgetting sample uj is unknown before deletion request arrives, and (ii) the learned empirical risk minimzer ˆw will be deleted after processing a deletion request, while subsequent unlearned model is also unknown as forgetting sample uj is unpredictable. Therefore, explicit pre-computing is required for Pn i=1 2ℓ( ˆw; zi) in NS and (Pn i=1 2ℓ( ˆw; zi)) 1 in IJ. 3Storing additional data-dependent vectors may incur supplementary privacy risks. A straightforward solution is to have the users retain custody of these vectors. Analogous to Ghazi et al. (2023), upon submitting a deletion request, the user is then required to furnish the corresponding approximators as part of the deletion request. Published as a conference paper at ICLR 2025 Comparison of Precomputation. For a single data, Algorithm 1 entails computing the HVP O(En/|B|) times. To precompute the Hessian-free (HF) approximators a u HF for all n samples, the complexity is O(n2d E/|B|). Typically, the ratio of E to |B| is much smaller than d or n (Smith (2018)), which simplifies the complexity to O(n2d). HF method is highly efficient when dealing with overparameterized models, especially when d n.4 This is because computing the full Hessian for NS requires O(nd2) time, and computing and inverting the Hessian H 1 IJ for IJ requires O(d3+nd2). Comparison of Storage. When we compute the approximator, it requires storing O(En/|B|) models in Stage I, leading to a maximum complexity of O(nd E/|B| + nd), which can simplify to O(2nd). After complete precomputing, Algorithm 1 only needs the storage of O(nd). This is lower than the overhead introduced by storing matrices for NS and IJ in overparameterized models, which is O(d2). Comparison of Unlearning Computation. When an unlearning request with m forgetting data arrives, Algorithm 1 utilizes precomputed and prestored approximators to achieve forgetting, requiring a simple vector addition that takes O(md) time. For NS, it requires computing and inverting different Hessian that depends on the user requesting the deletion. This necessitates substantial computational cost of O(d3+md2+md) each time a deletion request arrives. Although IJ reduces this to O(d2+md) by precomputing the inverse Hessian, it is still not as efficient compared to our method. 5 EXPERIMENTS FROM DEVELOPER S PERSPECTIVE We devised the following experimental components: ❶Verification and ❷Application. Verification experiments are centered on evaluating the disparity between the unlearning model w U E,B and retrained models w U E,B, across both convex and non-convex settings. Application experiments are geared towards evaluating the performance of different unlearning algorithms after perturbing w U E,B with noise n with a specific focus on determining the accuracy, pre-computation/storage, and unlearning time. Considering the space limit, we defer a portion of experiments to the Appendix D. 5.1 VERIFICATION EXPERIMENTS Our principal aim is to validate the differences between the proposed Hessian-free (HF) approximator a U HF and the actual w U E,B w E,B, where a smaller difference indicates the unlearned model is closer to retrained model and also reflects the degree of data forgetting. Therefore, we evaluate the degree of forgetting from two perspectives: distance and correlation. Specifically, we use the L2-norm w U E,B w E,B a U to measure the distance metric. A smaller distance metric indicates the approximators accurately capture the disparity between the retrained and learned models. Additionally, we utilize the Pearson (Wright (1921)) and Spearman (Spearman (1961)) coefficients to assess the correlation between a U and w U E,B w E,B by mapping them from high-dimensional space to scalar loss values, i.e. the approximate loss change ℓ(w E,B+a U; U) ℓ(w E,B; U) and the actual change ℓ(w U E,B; U) ℓ(w E,B; U) on the forgetting dataset U. The correlation scores range from -1 to 1, where the higher score, the more the unlearned model performs like the retrained model. Configurations: We conduct experiments in both convex and non-convex scenarios. Specifically, we trained a multinomial Logistic Regression (LR) with total parameters d = 7850 and a simple convolutional neural network (CNN) with total parameters d = 21840 on MNIST dataset (Deng (2012)) for handwriting digit classification. We apply a cross-entropy loss and the inclusion of an L2-regularization coefficient of 0.5 to ensure that the loss function of LR is strongly convex. For LR, training was performed for 15 epochs with a stepsize of 0.05 and a batch of 32. For CNN, training was carried out for 20 epochs with a stepsize of 0.05 and a batch size of 64. Given these configurations, we separately assess the distance and correlation between approximators a U HF , a U NS , a U IJ at deletion rates in the set {1%, 5%, 10%, 15%, 20%, 25%, 30%}. Following the suggestion in Basu et al. (2021), a damping factor of 0.01 is added to the Hessian to ensure its invertibility when implementing NS and IJ. Besides, we conducted experiments with 7 random seeds to obtain average results. For the sake of clear visualizations, we use the minimum and maximum values as error bars. 4Deep networks for computation vision tasks are typically trained with a number of model parameters that far exceeds the size of the training dataset (Chang et al. (2021)). For general neural language model, the empirical data-to-parameter scaling law is also expected to remain at n d0.74, as suggested in Kaplan et al. (2020). Published as a conference paper at ICLR 2025 (a) Convex Setting (b) Non-Convex Setting Figure 1: Verification experiments I on (a) LR and (b) CNN, respectively. The left of (a)(b) shows the distance to the retrained model, i.e. w U E,B w E,B a U . The right of (a)(b) shows the correlation between the approximate loss change and actual loss change on the forgetting dataset. Table 2: Application experiments I. An online scenario with 20% samples to be forgotten, where each unlearning request involves forgetting a single data point with each execution. A performance gap against Retrain is provided in ( ). We compute the total computational/storage cost for all samples. Model Unlearning Computation Pre-Computaion Storage Test Accuracy (%) Runtime (Sec) Speedup Runtime (Sec) (GB) Unlearned model NS LR 5.09 102 1.34 2.55 103 0.23 86.25 (-1.50) CNN 5.81 103 0.12 2.91 104 1.78 83.50 (-10.25) IJ LR 0.23 102 29.64 2.55 103 0.23 86.25 (-1.50) CNN 1.91 102 3.63 2.91 104 1.78 82.75 (-11.00) HF LR 0.0073 93,376 2.20 102 0.03 87.50 (-0.25) CNN 0.0268 25,821 5.34 102 0.08 91.50 (-2.25) The convex setting. As shown in Figure 1 (a), our approach outperforms previous works. First, by distance evaluation, the approximation error U E,B of the proposed method is lower than that of the previous NS and IJ works. At a deletion rate of 30%, the error is 0.2097, lower than that of NS and IJ, which have errors of 0.246554 and 0.246557, respectively. Second, in assessing the loss change for the forgotten dataset U, our proposed method more accurately captures the actual changes. Particularly when removing 30% of the samples, the proposed method maintains a high correlation, with Pearson and Spearman being 0.96 and 0.95, respectively. These evaluations demonstrate the efficacy of both Hessian-free and Hessian-based methods in accurately approximating the retrained model. Our proposed approach not only achieves a closer approximation to the retrained model but also exhibits lower complexity compared to previous methods. This is substantiated by the results of Application Experiments I, which provide empirical evidence supporting our theoretical findings. The non-convex setting. As illustrated in Figure 1 (b), the proposed HF method demonstrates superior performance, exhibiting less dependency on random variations compared to NS and IJ. When 30% of the samples are removed, our method maintains a lower approximation error of 0.90, surpassing the performance of NS and IJ. Our methods achieve Spearman and Pearson correlation coefficients of 0.81 and 0.74, respectively, accurately capturing actual loss changes, in contrast to NS and IJ which are far from the actual values in non-convex settings. In Appendix D.2, we explain the volatility of the results in the prior works on non-convex setting, and provide more large-scale results, which is not achievable in NS and IJ (Sekhari et al. (2021); Suriyakumar et al. (2022); Liu et al. (2023)) due to out-of-memory to precisely computing the Hessian matrix (or its inverse). 5.2 APPLICATION EXPERIMENTS We compare the cost and performance of our method with the previous certified unlearning methods. Evaluations are conducted from two perspectives: runtime and utility. Specifically, runtime focuses on the time spent pre-computing unlearning statistics and the speedup of the unlearning algorithm compared to the retraining. Moreover, we evaluate the utility by the test accuracy of unlearned model to ensure generalization guarantee. We train LR and CNN with 20% data to be forgotten, where the configuration is identical to the foregoing verification experiments I. We conduct experiments with the unlearning guarantee of ϵ = 1, δ = 10 3 for LR and ϵ = 100, δ = 0.1 for CNN. We provide larger scale results in Appendix D.3, and we have not evaluated NS and IJ due to out-of-memory. Published as a conference paper at ICLR 2025 0 2 4 6 8 10 12 14 60 61 62 63 64 65 66 67 68 Target Model: LR 1 2 3 4 5 102 58 86 87 88 89 90 91 Target Model: CNN Figure 2: Membership Inference Attack I. The first plot shows the MIA-L and error bars show 95% confidence intervals; the second and third plots depict the MIA-U results; and the fourth plot illustrates the privacy-utility tradeoff. Proposed method successfully defends MIA-L and mitigates excessive privacy leakage from MIA-U without sacrificing performance when applied with appropriate noise. Application experiment results. Table 2 demonstrates that the proposed algorithm outperforms other baselines by a significant margin. Strikingly, the proposed algorithm outperforms benchmarks of certified unlearning regarding all metrics with millisecond-level unlearning and minimal performance degradation in both convex and non-convex settings. When models exceed the complexity of LR, precisely estimating the full Hessian becomes infeasible. This conclusion is supported by Table 2 and Mehta et al. (2022). Therefore, we demonstrate that the proposed Hessian-free perspective method exhibits great potentials for over-parameterized models compared to the Hessian-based approaches. 6 EXPERIMENTS FROM ADVERSARY S PERSPECTIVE To further investigate the effectiveness of data removal for auditing, we perform membership inference attack (MIA) experiments from the attacker s perspective, including two state-of-the-art MIA: ❶ MIA against Learning (MIA-L) (Li RA MIA Carlini et al. (2022)) and ❷MIA against Unlearning (MIA-U) (Chen et al. (2021)). Specifically, MIA-L enables an adversary to query a trained ML model to predict whether a specific example was included in the model s training dataset; MIA-U involves querying two versions of the ML model to infer whether a target sample is part of the training set of the learned model but has been removed from the corresponding unlearned model (primarily targeting retrain algorithm). See Appendix D.4 for a detailed introduction and more experimental results. MIA results with 5% data to be forgotten. As shown in the first plot of Figure 2, our proposed method effectively defends against MIA-L even in low-privacy scenarios, as evidenced by effective approximation of the retrained model in the verification experiment. However, MIA-U exploits discrepancies between the outputs of the model s retrained and original versions. The verification experiment also reveals that our unlearned model s loss change correlation coefficient is closely aligned with that of the retrained model, indicating that, like the retrained model, it is similarly vulnerable to such attacks. As depicted in the second and third plots, MIA-U shows high AUC values for the retrained method (light red line) and our method (red line) without perturbation (ϵ = ). Nevertheless, adding appropriate noise to the model can mitigate MIA-U. As demonstrated in the fourth plot, introducing an appropriate noise level (ϵ > 2 for LR and ϵ > 120 for CNN) effectively reduces the excessive privacy leakage without compromising performance. Besides, further increasing the model perturbation to tradeoff privacy-utility does not significantly reduce the effectiveness of such attacks. To better defend MIA-U, we will consider measures like noised SGD in future work. 7 CONCLUSION We proposed a novel Hessian-free certified unlearning method. Our analysis relies on an affine stochastic recursion to characterize the model update discrepancy between the learned and retrained models, which does not depend on inverting Hessian and model optimality. We provide results of unlearning, generalization, deletion capacity guarantees, and time/memory complexity that are superior to state-of-the-art methods. We then develop an algorithm that achieves near-instantaneous unlearning. We show that the proposed method can reduce the precomputation time, storage, and unlearning time to O(n2d), O(nd), O(d), respectively. Experimental results validate that our work benefits certified unlearning with millisecond-level unlearning time and improved test accuracy. Published as a conference paper at ICLR 2025 ACKNOWLEDGMENTS This work was supported in part by NSF grants CMMI-2024774 and ECCS-2030251, and in part by the National Natural Science Foundation of China under Grant 62202427. Krizhevsky Alex. Learning multiple layers of features from tiny images. https://www. cs. toronto. edu/kriz/learning-features-2009-TR. pdf, 2009. 29 Borja Balle and Yu-Xiang Wang. Improving the gaussian mechanism for differential privacy: Analytical calibration and optimal denoising. In Jennifer G. Dy and Andreas Krause (eds.), Proceedings of the 35th International Conference on Machine Learning, ICML 2018, Stockholmsmässan, Stockholm, Sweden, July 10-15, 2018, volume 80 of Proceedings of Machine Learning Research, pp. 403 412. PMLR, 2018. 4 Samyadeep Basu, Xuchen You, and Soheil Feizi. On second-order group influence functions for black-box predictions. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13-18 July 2020, Virtual Event, volume 119 of Proceedings of Machine Learning Research, pp. 715 724. PMLR, 2020. 17, 29, 30 Samyadeep Basu, Phillip Pope, and Soheil Feizi. Influence functions in deep learning are fragile. In 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3-7, 2021. Open Review.net, 2021. 8, 17, 29, 30 Alexander Becker and Thomas Liebig. Certified data removal in sum-product networks. In Peipei Li, Kui Yu, Nitesh V. Chawla, Ronen Feldman, Qing Li, and Xindong Wu (eds.), IEEE International Conference on Knowledge Graph, ICKG 2022, Orlando, FL, USA, November 30 - Dec. 1, 2022, pp. 14 21. IEEE, 2022. 3 Lucas Bourtoule, Varun Chandrasekaran, Christopher A. Choquette-Choo, Hengrui Jia, Adelin Travers, Baiwu Zhang, David Lie, and Nicolas Papernot. Machine unlearning. In 42nd IEEE Symposium on Security and Privacy, SP 2021, San Francisco, CA, USA, 24-27 May 2021, pp. 141 159. IEEE, 2021. 3, 17, 32 Jonathan Brophy and Daniel Lowd. Machine unlearning for random forests. In Marina Meila and Tong Zhang (eds.), Proceedings of the 38th International Conference on Machine Learning, ICML 2021, 18-24 July 2021, Virtual Event, volume 139 of Proceedings of Machine Learning Research, pp. 1092 1104. PMLR, 2021. 3, 17 Yinzhi Cao and Junfeng Yang. Towards making systems forget with machine unlearning. In 2015 IEEE Symposium on Security and Privacy, SP 2015, San Jose, CA, USA, May 17-21, 2015, pp. 463 480. IEEE Computer Society, 2015. 3, 17 Nicholas Carlini, Steve Chien, Milad Nasr, Shuang Song, Andreas Terzis, and Florian Tramèr. Membership inference attacks from first principles. In 43rd IEEE Symposium on Security and Privacy, SP 2022, San Francisco, CA, USA, May 22-26, 2022, pp. 1897 1914. IEEE, 2022. 10, 32 Xiangyu Chang, Yingcong Li, Samet Oymak, and Christos Thrampoulidis. Provable benefits of overparameterization in model compression: From double descent to pruning neural networks. In Thirty-Fifth AAAI Conference on Artificial Intelligence, AAAI 2021, Thirty-Third Conference on Innovative Applications of Artificial Intelligence, IAAI 2021, The Eleventh Symposium on Educational Advances in Artificial Intelligence, EAAI 2021, Virtual Event, February 2-9, 2021, pp. 6974 6983. AAAI Press, 2021. 8 Chong Chen, Fei Sun, Min Zhang, and Bolin Ding. Recommendation unlearning. In Frédérique Laforest, Raphaël Troncy, Elena Simperl, Deepak Agarwal, Aristides Gionis, Ivan Herman, and Lionel Médini (eds.), WWW 22: The ACM Web Conference 2022, Virtual Event, Lyon, France, April 25 - 29, 2022, pp. 2768 2777. ACM, 2022a. 3, 17 Published as a conference paper at ICLR 2025 Min Chen, Zhikun Zhang, Tianhao Wang, Michael Backes, Mathias Humbert, and Yang Zhang. When machine unlearning jeopardizes privacy. In Yongdae Kim, Jong Kim, Giovanni Vigna, and Elaine Shi (eds.), CCS 21: 2021 ACM SIGSAC Conference on Computer and Communications Security, Virtual Event, Republic of Korea, November 15 - 19, 2021, pp. 896 911. ACM, 2021. 10, 32, 33 Min Chen, Zhikun Zhang, Tianhao Wang, Michael Backes, Mathias Humbert, and Yang Zhang. Graph unlearning. In Heng Yin, Angelos Stavrou, Cas Cremers, and Elaine Shi (eds.), Proceedings of the 2022 ACM SIGSAC Conference on Computer and Communications Security, CCS 2022, Los Angeles, CA, USA, November 7-11, 2022, pp. 499 513. ACM, 2022b. 3, 17 Xiangyi Chen, Steven Z Wu, and Mingyi Hong. Understanding gradient clipping in private sgd: A geometric perspective. Advances in Neural Information Processing Systems, 33:13773 13782, 2020. 36 Eli Chien, Chao Pan, and Olgica Milenkovic. Efficient model updates for approximate unlearning of graph-structured data. In The Eleventh International Conference on Learning Representations, 2022. 1, 2, 3, 17, 21 Mathieu Dagréou, Pierre Ablin, Samuel Vaiter, and Thomas Moreau. How to compute hessian-vector products? In The Third Blogpost Track at ICLR 2024, 2024. 5, 19 Li Deng. The mnist database of handwritten digit images for machine learning research. IEEE Signal Processing Magazine, 29(6):141 142, 2012. 8 Cynthia Dwork. Differential privacy. In International colloquium on automata, languages, and programming, pp. 1 12. Springer, 2006. 17 Cynthia Dwork and Aaron Roth. The algorithmic foundations of differential privacy. Found. Trends Theor. Comput. Sci., 9(3-4):211 407, 2014. 4 Badih Ghazi, Pritish Kamath, Ravi Kumar, Pasin Manurangsi, Ayush Sekhari, and Chiyuan Zhang. Ticketed learning-unlearning schemes. In Gergely Neu and Lorenzo Rosasco (eds.), The Thirty Sixth Annual Conference on Learning Theory, COLT 2023, 12-15 July 2023, Bangalore, India, volume 195 of Proceedings of Machine Learning Research, pp. 5110 5139. PMLR, 2023. 7 Justin Gilmer, Behrooz Ghorbani, Ankush Garg, Sneha Kudugunta, Behnam Neyshabur, David Cardoze, George Edward Dahl, Zachary Nado, and Orhan Firat. A loss curvature perspective on training instabilities of deep learning models. In International Conference on Learning Representations, 2022. 18, 34, 35 Antonio Ginart, Melody Guan, Gregory Valiant, and James Y Zou. Making ai forget you: Data deletion in machine learning. Advances in neural information processing systems, 32, 2019. 17 Aditya Golatkar, Alessandro Achille, and Stefano Soatto. Eternal sunshine of the spotless net: Selective forgetting in deep networks. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, June 13-19, 2020, pp. 9301 9309. Computer Vision Foundation / IEEE, 2020a. 3, 17 Aditya Golatkar, Alessandro Achille, and Stefano Soatto. Forgetting outside the box: Scrubbing deep networks of information accessible from input-output observations. In Andrea Vedaldi, Horst Bischof, Thomas Brox, and Jan-Michael Frahm (eds.), Computer Vision - ECCV 2020 - 16th European Conference, Glasgow, UK, August 23-28, 2020, Proceedings, Part XXIX, volume 12374 of Lecture Notes in Computer Science, pp. 383 398. Springer, 2020b. 17 Aditya Golatkar, Alessandro Achille, Avinash Ravichandran, Marzia Polito, and Stefano Soatto. Mixed-privacy forgetting in deep networks. In IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2021, virtual, June 19-25, 2021, pp. 792 801. Computer Vision Foundation / IEEE, 2021. 17 Priya Goyal, Piotr Dollár, Ross B. Girshick, Pieter Noordhuis, Lukasz Wesolowski, Aapo Kyrola, Andrew Tulloch, Yangqing Jia, and Kaiming He. Accurate, large minibatch SGD: training imagenet in 1 hour. Co RR, abs/1706.02677, 2017. 4 Published as a conference paper at ICLR 2025 Chuan Guo, Tom Goldstein, Awni Y. Hannun, and Laurens van der Maaten. Certified data removal from machine learning models. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13-18 July 2020, Virtual Event, volume 119 of Proceedings of Machine Learning Research, pp. 3832 3842. PMLR, 2020. 1, 3, 17 Varun Gupta, Christopher Jung, Seth Neel, Aaron Roth, Saeed Sharifi-Malvajerdi, and Chris Waites. Adaptive machine unlearning. In Marc Aurelio Ranzato, Alina Beygelzimer, Yann N. Dauphin, Percy Liang, and Jennifer Wortman Vaughan (eds.), Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, Neur IPS 2021, December 6-14, 2021, virtual, pp. 16319 16330, 2021. 3 Mert Gürbüzbalaban, Umut Simsekli, and Lingjiong Zhu. The heavy-tail phenomenon in SGD. In Marina Meila and Tong Zhang (eds.), Proceedings of the 38th International Conference on Machine Learning, ICML 2021, 18-24 July 2021, Virtual Event, volume 139 of Proceedings of Machine Learning Research, pp. 3964 3975. PMLR, 2021. 5 Anisa Halimi, Swanand Kadhe, Ambrish Rawat, and Nathalie Baracaldo. Federated unlearning: How to efficiently erase a client in fl? ar Xiv preprint ar Xiv:2207.05521, 2022. 17 Boris Hanin. Which neural net architectures give rise to exploding and vanishing gradients? In S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett (eds.), Advances in Neural Information Processing Systems, volume 31. Curran Associates, Inc., 2018. 36 Satoshi Hara, Atsushi Nitanda, and Takanori Maehara. Data cleansing for models trained with SGD. In Hanna M. Wallach, Hugo Larochelle, Alina Beygelzimer, Florence d Alché-Buc, Emily B. Fox, and Roman Garnett (eds.), Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, Neur IPS 2019, December 8-14, 2019, Vancouver, BC, Canada, pp. 4215 4224, 2019. 17 Elad Hazan. Introduction to online convex optimization. Found. Trends Optim., 2(3-4):157 325, 2016. 2 Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, June 27-30, 2016, pp. 770 778. IEEE Computer Society, 2016. 29, 31 Gary B. Huang, Manu Ramesh, Tamara Berg, and Erik Learned-Miller. Labeled faces in the wild: A database for studying face recognition in unconstrained environments. Technical Report 07-49, University of Massachusetts, Amherst, October 2007. 29 Zachary Izzo, Mary Anne Smart, Kamalika Chaudhuri, and James Zou. Approximate data deletion from machine learning models. In Proceedings of The 24th International Conference on Artificial Intelligence and Statistics, pp. 2008 2016, 2021. 3, 30 Matthew Jagielski, Om Thakkar, Florian Tramèr, Daphne Ippolito, Katherine Lee, Nicholas Carlini, Eric Wallace, Shuang Song, Abhradeep Guha Thakurta, Nicolas Papernot, and Chiyuan Zhang. Measuring forgetting of memorized training examples. In The Eleventh International Conference on Learning Representations, ICLR 2023, Kigali, Rwanda, May 1-5, 2023. Open Review.net, 2023. 3 Jared Kaplan, Sam Mc Candlish, Tom Henighan, Tom B. Brown, Benjamin Chess, Rewon Child, Scott Gray, Alec Radford, Jeffrey Wu, and Dario Amodei. Scaling laws for neural language models. Co RR, abs/2001.08361, 2020. 8 Korbinian Koch and Marcus Soll. No matter how you slice it: Machine unlearning with sisa comes at the expense of minority classes. In 2023 IEEE Conference on Secure and Trustworthy Machine Learning (Sa TML), pp. 622 637. IEEE, 2023. 17 Pang Wei Koh and Percy Liang. Understanding black-box predictions via influence functions. In Doina Precup and Yee Whye Teh (eds.), Proceedings of the 34th International Conference on Machine Learning, ICML 2017, Sydney, NSW, Australia, 6-11 August 2017, volume 70 of Proceedings of Machine Learning Research, pp. 1885 1894. PMLR, 2017. 17, 21 Published as a conference paper at ICLR 2025 Meghdad Kurmanji, Peter Triantafillou, Jamie Hayes, and Eleni Triantafillou. Towards unbounded machine unlearning. Advances in Neural Information Processing Systems, 36, 2024. 17 Xiaoyu Li, Zhenxun Zhuang, and Francesco Orabona. A second look at exponential and cosine step sizes: Simplicity, adaptivity, and performance. In Marina Meila and Tong Zhang (eds.), Proceedings of the 38th International Conference on Machine Learning, ICML 2021, 18-24 July 2021, Virtual Event, volume 139 of Proceedings of Machine Learning Research, pp. 6553 6564. PMLR, 2021. 24 Jiaqi Liu, Jian Lou, Zhan Qin, and Kui Ren. Certified minimax unlearning with generalization rates and deletion capacity. In Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, Neur IPS 2023., 2023. 1, 2, 3, 5, 6, 9, 17, 21, 30 Ziwei Liu, Ping Luo, Xiaogang Wang, and Xiaoou Tang. Deep learning face attributes in the wild. In Proceedings of International Conference on Computer Vision (ICCV), December 2015. 29 Ananth Mahadevan and Michael Mathioudakis. Certifiable machine unlearning for linear models. Co RR, abs/2106.15093, 2021. 1, 21 Vien V Mai and Mikael Johansson. Stability and convergence of stochastic gradient clipping: Beyond lipschitz continuity and smoothness. In International Conference on Machine Learning, pp. 7325 7335. PMLR, 2021. 36 Ronak Mehta, Sourav Pal, Vikas Singh, and Sathya N. Ravi. Deep unlearning via randomized conditionally independent hessians. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2022, New Orleans, LA, USA, June 18-24, 2022, pp. 10412 10421. IEEE, 2022. 3, 10, 17 Seth Neel, Aaron Roth, and Saeed Sharifi-Malvajerdi. Descent-to-delete: Gradient-based methods for machine unlearning. In Vitaly Feldman, Katrina Ligett, and Sivan Sabato (eds.), Algorithmic Learning Theory, 16-19 March 2021, Virtual Conference, Worldwide, volume 132 of Proceedings of Machine Learning Research, pp. 931 962. PMLR, 2021. 3 Quoc Phong Nguyen, Bryan Kian Hsiang Low, and Patrick Jaillet. Variational bayesian unlearning. In Hugo Larochelle, Marc Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin (eds.), Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, Neur IPS 2020, December 6-12, 2020, virtual, 2020. 3 Barak A. Pearlmutter. Fast exact multiplication by the hessian. Neural Comput., 6(1):147 160, 1994. Levent Sagun, Leon Bottou, and Yann Le Cun. Eigenvalues of the hessian in deep learning: Singularity and beyond. ar Xiv preprint ar Xiv:1611.07476, 2016. 36 Levent Sagun, Utku Evci, V. Ugur Güney, Yann N. Dauphin, and Léon Bottou. Empirical analysis of the hessian of over-parametrized neural networks. In 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Workshop Track Proceedings. Open Review.net, 2018. 36 Sebastian Schelter, Stefan Grafberger, and Ted Dunning. Hedgecut: Maintaining randomised trees for low-latency machine unlearning. In Guoliang Li, Zhanhuai Li, Stratos Idreos, and Divesh Srivastava (eds.), SIGMOD 21: International Conference on Management of Data, Virtual Event, China, June 20-25, 2021, pp. 1545 1557. ACM, 2021. 3, 17 Ayush Sekhari, Jayadev Acharya, Gautam Kamath, and Ananda Theertha Suresh. Remember what you want to forget: Algorithms for machine unlearning. In Advances in Neural Information Processing Systems 34, Neur IPS, pp. 18075 18086, 2021. 1, 2, 3, 4, 5, 6, 7, 9, 17, 18, 19, 29, 30 Shai Shalev-Shwartz, Ohad Shamir, Nathan Srebro, and Karthik Sridharan. Stochastic convex optimization. In COLT 2009 - The 22nd Conference on Learning Theory, Montreal, Quebec, Canada, June 18-21, 2009, 2009. 27 Published as a conference paper at ICLR 2025 Leslie N Smith. A disciplined approach to neural network hyper-parameters: Part 1 learning rate, batch size, momentum, and weight decay. ar Xiv preprint ar Xiv:1803.09820, 2018. 8 Charles Spearman. The proof and measurement of association between two things. 1961. 8 Ningxin Su and Baochun Li. Asynchronous federated unlearning. In IEEE INFOCOM 2023 - IEEE Conference on Computer Communications, New York City, NY, USA, May 17-20, 2023, pp. 1 10. IEEE, 2023. 17 Vinith M. Suriyakumar, Ashia C. Wilson, et al. Algorithms that approximate data removal: New results and limitations. In Advances in Neural Information Processing Systems 35, Neur IPS, 2022. 1, 2, 3, 5, 6, 7, 9, 17, 21, 29, 30 Ryutaro Tanno, Melanie F. Pradier, Aditya V. Nori, and Yingzhen Li. Repairing neural networks by leaving the right past behind. In Neur IPS, 2022. 3 Ayush Kumar Tarun, Vikram Singh Chundawat, Murari Mandal, and Mohan S. Kankanhalli. Deep regression unlearning. In Andreas Krause, Emma Brunskill, Kyunghyun Cho, Barbara Engelhardt, Sivan Sabato, and Jonathan Scarlett (eds.), International Conference on Machine Learning, ICML 2023, 23-29 July 2023, Honolulu, Hawaii, USA, volume 202 of Proceedings of Machine Learning Research, pp. 33921 33939. PMLR, 2023. 3, 29 Alexander Warnecke, Lukas Pirch, Christian Wressnegger, and Konrad Rieck. Machine unlearning of features and labels. In 30th Annual Network and Distributed System Security Symposium, NDSS 2023, San Diego, California, USA, February 27 - March 3, 2023. The Internet Society, 2023. 3, 17 Sanford Weisberg and R Dennis Cook. Residuals and influence in regression. 1982. 21 Kaiyue Wen, Tengyu Ma, and Zhiyuan Li. How sharpness-aware minimization minimizes sharpness? In The Eleventh International Conference on Learning Representations, ICLR 2023, Kigali, Rwanda, May 1-5, 2023. Open Review.net, 2023. 18 Sewall Wright. Correlation and causation. Journal of agricultural research, 20(7):557 585, 1921. 8 Ga Wu, Masoud Hashemi, and Christopher Srinivasa. PUMA: performance unchanged model augmentation for training data removal. In Thirty-Sixth AAAI Conference on Artificial Intelligence, AAAI 2022, Thirty-Fourth Conference on Innovative Applications of Artificial Intelligence, IAAI 2022, The Twelveth Symposium on Educational Advances in Artificial Intelligence, EAAI 2022 Virtual Event, February 22 - March 1, 2022, pp. 8675 8682. AAAI Press, 2022. 3, 17 Kun Wu, Jie Shen, Yue Ning, Ting Wang, and Wendy Hui Wang. Certified edge unlearning for graph neural networks. In Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, KDD 23, pp. 2606 2617, 2023. 17 Lei Wu, Chao Ma, and Weinan E. How SGD selects the global minima in over-parameterized learning: A dynamical stability perspective. In Samy Bengio, Hanna M. Wallach, Hugo Larochelle, Kristen Grauman, Nicolò Cesa-Bianchi, and Roman Garnett (eds.), Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, Neur IPS 2018, December 3-8, 2018, Montréal, Canada, pp. 8289 8298, 2018. 34 Yinjun Wu, Edgar Dobriban, and Susan B. Davidson. Deltagrad: Rapid retraining of machine learning models. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13-18 July 2020, Virtual Event, volume 119 of Proceedings of Machine Learning Research, pp. 10355 10366. PMLR, 2020. 3, 30 Haonan Yan, Xiaoguang Li, Ziyao Guo, Hui Li, Fenghua Li, and Xiaodong Lin. ARCANE: an efficient architecture for exact machine unlearning. In Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, IJCAI 2022, Vienna, Austria, 23-29 July 2022, pp. 4006 4013, 2022. 3, 17 Binchi Zhang, Yushun Dong, Tianhao Wang, and Jundong Li. Towards certified unlearning for deep neural networks. ar Xiv preprint ar Xiv:2408.00920, 2024. 17 Jingzhao Zhang, Tianxing He, Suvrit Sra, and Ali Jadbabaie. Why gradient clipping accelerates training: A theoretical justification for adaptivity. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020, 2020. 36 Published as a conference paper at ICLR 2025 CONTENTS OF APPENDIX A An Expanded Version of The Related Work 17 B Discussions 18 B.1 Limitations and Future Works . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18 B.2 Deletion Capacity Guarantee . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18 B.3 Hessian-Vector Products Technique in Stage II of Algorithm 1 . . . . . . . . . . . 19 B.4 Unlearning-Repairing Strategy in Stage III of Algorithm 1 . . . . . . . . . . . . . 20 C Proofs 21 C.1 Detailed Analysis of Equation (11) . . . . . . . . . . . . . . . . . . . . . . . . . . 21 C.2 Detailed Proof of Theorem 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23 C.3 Detailed Proof of Lemma 2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24 C.4 Detailed Proof of Theorem 4 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24 C.5 Additional Definitions of Assumption 1 . . . . . . . . . . . . . . . . . . . . . . . 26 C.6 Detailed Proof of Theorem 6 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27 D Supplementary Experiments 28 D.1 Hardware, Software and Source Code . . . . . . . . . . . . . . . . . . . . . . . . 28 D.2 Additional Verification Experiments . . . . . . . . . . . . . . . . . . . . . . . . . 29 D.3 Additional Application Experiments . . . . . . . . . . . . . . . . . . . . . . . . . 31 D.4 Additional Adversary Experiments . . . . . . . . . . . . . . . . . . . . . . . . . . 32 E Sensitivity Analysis and Ablation Studies 33 E.1 Impact of Step Sizes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34 E.2 Impact of Epochs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35 E.3 Impact of Stepsize Decay . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36 E.4 Impact of Gradient Clipping . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36 Published as a conference paper at ICLR 2025 A AN EXPANDED VERSION OF THE RELATED WORK Machine Unlearning, the terminology can be traced back to the paper of Cao & Yang (2015), and their work defined the notion of "the completeness of forgetting". There are two main categories based on the completeness of forgetting: Exact Unlearning and Approximate Unlearning. Exact Unlearning aims to accelerate the speed of retraining and completely eliminate the influence of forgotten data on learned model. The majority of existing Exact Unlearning methods are derived from the work of Bourtoule et al. (2021), which proposed the notable SISA framework, grounded in ensemble learning techniques. Specifically, the training dataset is partitioned into mutually disjoint shards, and then sub-models are trained on these shards. Upon unlearning, SISA only needs to retrain the sub-model correlated with the relevant data shard and then make the final prediction by assembling the knowledge of each sub-model, which significantly lowers retraining computational costs. Under the SISA framework, other Exact Unlearning methods have also been proposed, e.g., Chen et al. (2022b) extended SISA to graph learning by partitioning the original training graph into disjoint shards, parallelly training a set of shard models, and learning an optimal importance score for each shard model. Rec Eraser (Chen et al. (2022a)) extended SISA to recommendation systems. KNOT (Su & Li (2023)) adopted the SISA for client-level asynchronous federated unlearning during training and introduced a clustered mechanism that divides clients into multiple clusters. Other methods, such as Da RE (Brophy & Lowd (2021)) and Hedge Cut (Schelter et al. (2021)), addressed the issue that SISA cannot be applied to decision trees and proposed exact unlearning solutions; and Ginart et al. (2019) aimed to achieve forgetting in K-means clustering. Furthermore, ARCANE (Yan et al. (2022)) introduced a class-based data partitioning strategy to enhance the performance of SISA. However, the SISA framework adopts a structure that completely separates the data, resembling ensemble learning approaches where predictions or models are aggregated only at the end. This separation restricts the model training process from fully leveraging the entirety of the segmented datasets and also hampers its extensibility, leading to lower accuracy and performance compared to a model trained on the entire unsegmented dataset. In an experimental evaluation conducted by Koch & Soll (2023), it was found that simply down-sampling the data set to achieve a more balanced single shard outperforms SISA in terms of unlearning speed and performance. Approximate Unlearning thus seeks to minimize the impact of forgetting data to an acceptable level to trade off computational efficiency, reduced storage costs, and flexibility. There are many approaches to approximate unlearning, such as Kurmanji et al. (2024) developed a distillation-based method; Halimi et al. (2022) adopted a gradient ascent approach on forgotten data to achieve forgetting. Among these, second-order methods have been widely studied, with most of these methods (e.g., Sekhari et al. (2021); Suriyakumar et al. (2022); Liu et al. (2023); Chien et al. (2022); Wu et al. (2023); Mehta et al. (2022); Warnecke et al. (2023); Wu et al. (2022)) inspired by influence function studies (Koh & Liang (2017); Basu et al. (2021; 2020); Hara et al. (2019)), aiming to extract information from the Hessian of loss function and achieve forgetting through limited model updates. In particular, inspired by differential privacy (Dwork (2006)), Guo et al. (2020) introduced certified unlearning (ϵ, δ), anticipating ensuring that the output distribution of an unlearning algorithm is indistinguishable from that of retraining. Moreover, due to costly Hessian operations, Fisher-based methods (Golatkar et al. (2020a;b; 2021)) have emerged to approximate Hessian. Our paper concentrates on the Hessian-based approach, yet obviates the need for Hessian inversion and explicit calculation of the Hessian. Among these Hessian-based certified unlearning methods, Sekhari et al. (2021); Suriyakumar et al. (2022); Liu et al. (2023) established the theoretical foundation for Hessian-based methods and provided the generalization result in terms of the population risk, as well as derived the deletion capacity guarantee. In the definition of certified unlearning, their works align with our method, which lies in pre-computing statistics of data that extract second-order information from the Hessian and using these recollections to forget data. However, existing second-order methods still require expensive matrix computations and can only perform forgetting for optimal models under convexity assumption. Although Zhang et al. (2024) offered certified unlearning for deep model through adding sufficient large regularization to guarantee local strong convexity, but the optimality problem remains unsolved, particularly due to the non-convexity of modern deep models, making it impractical to find the empirical risk minimizer. Even for convex models, in most non-concurrent training scenarios, such as online learning or continual learning, the model cannot converge to the optimal solution. Published as a conference paper at ICLR 2025 B DISCUSSIONS B.1 LIMITATIONS AND FUTURE WORKS Limitations. Although we have made the above contributions, there are some significant limitations: ①Our proposed algorithm necessitates O(n2d) precomputation and O(nd) storage, which remains costly, particularly in scenarios involving extensive data and larger-scale models. Consequently, conducting experiments concurrently with both large-scale datasets and models is anticipated to significantly extend the time required for precomputing unlearning statistics. ②As demonstrated by the results in Theorem 4, it is important to recognize that the proposed method is highly sensitive to the the stability of the training process or well-conditioned Hessian on non-convex setting. The addition of a large explicit L2-regularizer to the objective or utilizing small step sizes can ensure a meaningful bound. However, we are not aware of any popular non-convex machine learning problem that can satisfy all the conditions of Theorem 4. If the learning process is unstable, unlearning will introduce significant error. For instance, in Appendix E.1, employing an aggressive step size (η > 0.1) can lead to a substantial increase in approximation error. ③While we propose a new method that does not rely on convexity, our analysis builds upon the theoretical framework of previous works for comparison purposes and much of our theoretical results rely on convexity except for our unlearning-related conclusions. Providing a nonconvex theoretical framework is meaningful, and we will explore potential extensions in further research. Future work. Therefore, to solve the problems mentioned above, we briefly propose some solutions. ①Specifically, our proposed framework offers the feasibility to unlearn an arbitrary subset of the entire datasets, allowing a trade-off that further reduces the precomputation and storage complexity, for instance, consider a typical scenario at the user level, where there are k users (k << n). By focusing on the unlearning of a single user s dataset or a subset of users datasets, we can maintain k vectors for k users dataset instead of maintaining n vectors for n all data points. Therefore, we can further reduce precomputation and storage to O(knd) and O(kd), respectively. ②Moreover, based on our analysis in Theorem 4 and Appendix E.1, the main source of the approximation error is ρ, which mainly consists of two components: the step size η and the maximum eigenvalue λ1 of Hessian. Reducing the step size is straightforward method, but we cannot expect every training process to have a sufficiently small step size. Therefore, reducing λ1 during optimization, will not only improve the generalization ability of the learned model but also reduce the approximation error in the unlearning stage. Possible approaches are through stepsize warm-up strategies (Gilmer et al. (2022)) or sharpness-aware minimization (Wen et al. (2023)) to effectively reduce λ1. B.2 DELETION CAPACITY GUARANTEE We analyze the deletion capacity of our method based on the analysis of Sekhari et al. (2021), which formalizes how many samples can be deleted from the model parameterized by the original empirical risk minimizer while maintaining reasonable guarantees on the test loss. We establish the relationship between generalization performance and the amount of deleted samples through deletion capacity, and our analysis indicates that our proposed method outperforms previous works in both generalization and deletion capacity. The lower bound on deletion capacity can be directly derived from the generalization guarantee (Theorem 6). Below, we first formalize the deletion capacity. Definition 2 (Definition of Deletion Capacity (Sekhari et al. (2021))). Let ϵ, δ 0. Let S be a dataset of size n drawn i.i.d. from D. For a pair of learning and unlearning algorithms Ω, Ωthat are (ϵ, δ)-certified unlearning, the deletion capacity mΩ, Ω ϵ,δ (d, n) is defined as the maximum number of samples U to be forgotten, while still ensuring an excess population risk is γ. Specifically, mΩ, Ω ϵ,δ,γ(d, n) := max m | E max U S:|U| m F( Ω(Ω(S); T(S)) F(w ) γ , (17) where the expectation above is with respect to S Dn and output of the algorithms Ωand Ω. Similar to Sekhari et al. (2021), we provide upper bound and lower bound of deletion capacity for our algorithm. We first give the upper bound of deletion capacity. Published as a conference paper at ICLR 2025 Theorem 7 (Upper Bound of Deletion Capacity Sekhari et al. (2021)). Let δ 0.005 and ϵ = 1. There exists a 4-Lipschitz and 1-strongly convex loss function f, and a distribution D such that for any learning algorithm A and removal mechanism M that satisfies (ϵ, δ)-unlearning and has access to all remaining samples S\U, then the deletion capacity is: m A,M ϵ,δ,γ (d, n) cn (18) where c depends on the properties of function f and is strictly less than 1. A comprehensive proof of Theorem 7 is available in Sekhari et al. (2021). Furthermore, based on Theorem 6, we present the lower bound of deletion capacity below: Theorem 8 (Deletion Capacity Guarantee). There exists a learning algorithm Ωand an unlearning algorithm Ωsuch that for any convex (and hence strongly convex), L-Lipschitz, and M-Smoothness loss f and distribution D. Choose step size η 2 M+λ and w U E,B is the empirical risk minimizer of objective function FS\U(w), then we have mΩ, Ω ϵ,δ,γ(d, n) = O (1/ρ)n/d 1 2 . (19) Proof of Theorem 8. Based on Theorem 6, we have F( w U E 1,B) E[F (w )] = O 4L2 λ(n m) + ρn E/|B| 2L2 λ + ρn E/|B| 2LηG Note that ρ < 1. The above upper bound on the excess risk thus implies that we can delete at least ln (1/δ)d 1 2 ρn samples while still ensuring an excess risk guarantee of γ. Accordingly, it yields an order of O (1/ρ)n/d 1 2 . Therefore, we have completed the proof. B.3 HESSIAN-VECTOR PRODUCTS TECHNIQUE IN STAGE II OF ALGORITHM 1 We provide the detailed Hessian-Vector Products (HVP) procedure in Algorithm 2. Notably, when recursively computing the product of the matrix and an arbitrary vector (I ηH)v, we can derive it from v ηHv. Here, Hv involves a HVP through modern automatic differentiation frameworks such as JAX or Py Torch. For more explanation of the HVP, please refer to Dagréou et al. (2024). Algorithm 2: Matrix-Vector Products in Stage II of Algorithm 1 1 Stage II: Pre-computing and Pre-storing statistics T (S) = {a uj E,B}n j=1: 2 initialization {a uj E,B}n j=1 = 0, {v uj e,b }n j=1 = 0; 3 for e = 0, , E 1, E do 4 for b = 0, , B 1, B do 5 for j = 1, 2 , n do 6 Compute: a uj e,b (I ηe,b He,b)a uj e,b + ηe,b |Be,b|(I ηe,b He,b)v uj e,b 7 if gradient of sample uj computed in this update step then 8 v uj e,b ℓ(w uj e,b ; zi) 10 v uj e,b 0 Published as a conference paper at ICLR 2025 B.4 UNLEARNING-REPAIRING STRATEGY IN STAGE III OF ALGORITHM 1 While our approach does not require accessing any datasets in Stage III, we can mitigate the performance degradation caused by excessive deletions and make our algorithm more robust through a simple fine-tuning strategy if we can still access the remaining dataset in Stage III. We only consider this as a preliminary attempt in this section; other experiments do not include this Finetune strategy. This strategy allows our method to be used as a heuristic algorithm in non-privacy scenarios, without the need for theoretical certified guarantees, such as in bias removal. Specifically, when the continuous incoming deletion requests to forget data, we can rapidly implement data removal through vector addition. This involves augmenting the current model with approximators, where errors accumulate during this process and thus cause performance degradation. Suppose, upon the arrival of a new round of deletion requests for a particular model and dataset, the model s accuracy significantly drops due to the removal of a substantial amount of data. As a remedy in Algorithm 3, we can address this performance degradation by fine-tuning the current model for Er epochs on the remaining dataset, referred to as the repairing process. Additionally, by computing the approximators for the remaining dataset during the repairing process, and adding these approximators during the repairing process to the pre-stored approximators during the learning process, we obtain the new set of approximators for the remaining data. In Figure 3, we conduct an experiment on Res Net-18 using the Celeb A dataset to demonstrate the efficacy of the repairing strategy. Figure 3: Unlearning-Repairing Strategy. The blue, red, and green lines respectively represent the learning, unlearning, and repairing process. When a significant amount of data is removed, leading to a loss in model performance, performing fine-tuning on the remaining dataset and recording the approximators during this period can help avoid performance degradation. Algorithm 3: Hessian-Free Unlearning-Repairing Strategy in Stage III of Algorithm 1 1 Stage III: Unlearning when user requests to forget the subset U={uj}m j=1 on model w E,B: 2 Compute: w U E,B w E,B+a U E,B, c U E,B 2ln(1.25/δ) ϵ Delete statistics: {a uj E,B}m j=1 3 Sample: n N(0, σI), Return: w U E,B = w U E,B + n 4 Reparing (Fine-tuning) the model w U E,B on the remaining dataset S\U = {zr}n m r=1 : 5 for e = 0, ..., Er do 6 for b = 0, 1 , B do 7 Compute gradient: g 1 |Be,b| P r Be,b ℓ(we,b; zr), 8 Gradient descent: we,b+1 we,b ηe,bg 11 Pre-computing statistics T (S) = {a ur E,B}n m r=1 of the remaining dataset S\U = {zr}n m r=1 : 12 for r = 1, 2 , n m do 13 Recursive computation by using HVP based on Algorithm 2: 14 a ur E+Er,B PEr e=0 Me,b(ur) ℓ(we,b(ur)); u) + ˆHEr,B 1 0,b(ur)+1 a ur E,B 15 where: Me,b(ur) = ηe,b(ur) |Be,b(ur)| ˆHE,B 1 e,b(ur)+1 Published as a conference paper at ICLR 2025 C.1 DETAILED ANALYSIS OF EQUATION (11) Before proceeding with the analysis of our method, we will first provide the analysis of previous works (Suriyakumar et al. (2022); Liu et al. (2023); Mahadevan & Mathioudakis (2021); Chien et al. (2022); Liu et al. (2023)) and discuss the circumstances that lead to its failure. We first assume that the empirical risk in Equation (2) is twice-differentiable and strictly convex in the parameter space W. The original empirical risk minimizer can be expressed as ˆw := arg min w W FS = arg min w W i=1 ℓ(w; zi). (22) We reweight the sample u in the original empirical risk with a weighting factor ω, ˆw u := arg min w W i=1 ℓ(w; zi) + ωℓ(w; u), where ω = 1 When weighting factor ω = 1 n, Equation (23) is equivalent to the empirical risk minimizer without samle u, i.e. removing a training point u is similar to up-weighting its corresponding weight by ω = 1 n. Following the classical result of Weisberg & Cook (1982), Koh & Liang (2017) aimed to approximate ˆw u by the first-order Taylor expansion around the optimal learned model ˆw, given by, i=1 2ℓ( ˆw; zi) ℓ( ˆw; u). (24) However, we note that these works rely on several key assumptions: the convexity of the objective function to ensure the invertibility of the Hessian matrix, and the optimality of the learned model. These assumptions are often stringent and may not be satisfied in practice, particularly in modern machine learning scenarios where the objective function may be non-convex and the empirical risk minimizer may not be attainable. In addition, in over-parameterized models, the computation of the Hessian matrix also requires a significant memory overhead. In fact, aside from the optimal model, the training process contains more information, and our method aims to utilize the information from the training process to address both issues simultaneously. Here, we provide a more detailed explanation for Equation (11). Specifically, we first consider the impact of sample u during the e-th epoch, and then extend the analysis to all epochs. It should be noted that a necessary setting for our method is that the retraining and learning process have the same initialization, model architecture, and training process. Analysis of Equation (11). For ease of understanding, each sample appears only in one (mini-)batch per epoch. The update rule for the model parameters is typically given by, w E,B+1 w E,B ηE,B i BE,B ℓ(w E,B; zi) . (25) Without loss of generality, for the e-th epoch, we define the sample u to be sampled in the e-th epoch s b(u)-th update. Reweighting the sample u in the each update process (rather than the empirical risk) with a weighting factor ω, Equation (3) and Equation (4) can be represented by w E,B+1 w E,B ηE,B ℓ(w E,B; zi) + 1b=b(u)ω ℓ(w E,B; u) . (26) where 1b=b(u) is an indicator function that equals 1 if the condition b = b(u) is true (that is, sample u is to be involved (sampled) in the b(u)-th update of the e-th epoch), and 0 otherwise. We define that when ω = 0, Equation (26) represents the original learning update, whereas ω = 1 represents the Published as a conference paper at ICLR 2025 update after forgetting training sample u. To obtain the difference between the learning model and retraining model, we calculate the derivative of w E,B+1 w.r.t the weighting factor at ω = 0, given by, w u E,B+1 w E,B+1 = w E,B+1|ω= 1 w E,B+1|ω=0 dw E,B+1 Based on the chain rule, we have, dw E,B+1 ω=0 = dw E,B+1 Let He,b = 1/|Be,b| P i Be,b 2ℓ(we,b; zi) denote the Hessian matrix of the loss function. For Equation (26), we have that dw E,B+1 dw E,B = I ηE,BHE,B. Substituting into Equation (28), ω=0 = (I ηE,BHE,B) dw E,B which is equivalent to Equation (8) in the main text (Case 1), that is, w u e,b+1 we,b+1 (I ηe,b He,b)(w u e,b we,b). (30) Recursively applying the chain rule until 1b=b(u) = 1 during E-th epoch, b=B (I ηE,b HE,b) dw E,b(u)+1 When the indicator function 1b=b(u) = 1, for Equation (26), we obtain, dw E,b(u)+1 ω=0 = dw E,b(u) ω=0 ηE,b(u)HE,b(u) dw E,b(u) |BE,b| ℓ w E,b(u); u , (32) which is equivalent to Equation (9) in the main text (Case 2), that is, w u e,b(u)+1 we,b(u)+1 (I ηe,b(u)He,b(u))(w u e,b(u) we,b(u))+ ηe,b(u) |Be,b(u)| ℓ(we,b(u); u). (33) Substituting Equation (32) into Equation (31), b=B (I ηE,b HE,b) dw E,b(u) b=B (I ηE,b HE,b) ℓ w E,b(u); u . (34) Therefore, the impact of sample u on the model update trajectory during the E-th round is, w u E,B w E,B b=B 1 (I ηE,b HE,b)(w u E,0 w E,0) b=B 1 (I ηE,b HE,b) ηE,b(u) |BE,b(u)| ℓ(w E,b(u); u). Since initially the retraining model w u 0,0 and the learning model w0,0 are the same, the first term of Equation (35) will recursively reduce to the initial model and will equal 0, and as sample u participates in updates once per epoch, the second term is the sum over E epochs.5 Apply the above process recursively to complete the proof and, we obtain the approximator for sample u. e=0 Me,b(u) ℓ(we,b(u); u), where Me,b(u) = ηe,b(u) |Be,b(u)| ˆHE,B 1 e,b(u)+1. (36) We define ˆHE,B 1 e,b(u)+1 =(I ηE,B 1HE,B 1)...(I ηe,b(u)+1He,b(u)+1), which means that the multiplication should begin with recent update and continue backward through previous updates. 5Note that we assume sample u participates in the gradient calculation exactly once per epoch. Therefore, in Equation (36), the number of summations E represents that sample u has participated in E gradient calculations. In general, due to the randomness of sampling, sample u may be sampled multiple times or not at all within a single epoch. Nevertheless, we can derive the results for these situations in the same manner. To facilitate understanding and draw relevant conclusions, all results in this work are based on the simplified assumption. Published as a conference paper at ICLR 2025 C.2 DETAILED PROOF OF THEOREM 1 Now, we demonstrate that in the case of continuous arrival of deletion requests, our algorithm is capable of streaming removing samples. Intuitively, this property can be mainly explained by Taylor s linear expansion. For ease of exposition, we simplify our proof to a scenario with only two samples, u1 and u2. That is, our goal is to prove that when u1 initiates a deletion request and the algorithm is executed, the arrival of a deletion request from u2 after u1 (Streaming Deletion), is fully equivalent to the simultaneous execution of deletion requests from both u1 and u2 (Batch Deletion). Without loss of generality, we assume that b(u1) < b(u2). Based on Equation (11), the approximators for u1 and u2 are denoted as a u1 E,B and a u2 E,B, respectively, as follows: a u1 E,B := e=0 Me,b(u1) ℓ(we,b(u1); u1), where Me,b(u1) = ηe,b(u1) |Be,b(u1)| ˆHE,B 1 e,b(u1)+1, a u2 E,B := e=0 Me,b(u2) ℓ(we,b(u2); u2), where Me,b(u2) = ηe,b(u2) |Be,b(u2)| ˆHE,B 1 e,b(u2)+1 When we simultaneously delete u1 and u2, the difference is, w u2 u1 e,b+1 we,b+1 w u2 u1 e,b we,b ηe,b ℓ(w u2 u1 e,b ; zi) ℓ(we,b; zi) . (38) Consider the Taylor expansion during the E-th epoch, we have, w u2 u1 E,b+1 w E,b+1 (I ηE,b |BE,b|HE,b)(w u2 u1 E,b w E,b). (39) When u1 and u2 are removed respectively, we have the following approximation: w u2 u1 E,b(u2)+1 w E,b(u2)+1 (I ηE,b(u2) |BE,b(u2)|HE,b(u2))(w u2 u1 E,b(u2) w E,b(u2)) + ηE,b(u2) |BE,b(u2)| ℓ(w E,b(u2); u2) | {z } Approximate impact of u2 in e-th epoch w u2 u1 E,b(u1)+1 w E,b(u1)+1 (I ηE,b(u1) |BE,b(u1)|HE,b(u1))(w u2 u1 E,b(u1) w E,b(u1)) + ηE,b(u2) |BE,b(u2)| ℓ(w E,b(u2); u2) | {z } Approximate impact of u2 in e-th epoch + ηE,b(u1) |BE,b(u1)| ℓ(w E,b(u1); u1) | {z } Approximate impact of u1 in e-th epoch (40) Therefore, in the E-th epoch, we have the following affine stochastic recursion, w u2 u1 E,B w E,B b=B 1 (I ηE,b |BE,b|HE,b)(w u2 u1 E,0 w E,0)+ b=B 1 (I ηE,b |BE,b|HE,b) ηE,b(u1) |BE,b(u1)| ℓ(w E,b(u1);u1)+ b=B 1 (I ηE,b |BE,b|HE,b) ηE,b(u2) |BE,b(u2)| ℓ(w E,b(u2);u2). (41) Apply it recursively to complete the proof, and we can simultaneously compute the impact of u1 and u2 on training for all epochs. w u2 u1 E,B w E,B a u E,B := e=0 Me,b(u1) ℓ(we,b(u1); u1) + e=0 Me,b(u2) ℓ(we,b(u2); u2) + . (42) According to Eqs. (37) and (42), we have completed the proof of a u2 u1 E,B = a u1 E,B + a u2 E,B. Published as a conference paper at ICLR 2025 C.3 DETAILED PROOF OF LEMMA 2 Proof of Lemma 2. Now we begin the proof that the remainder term o(w u e,b we,b) in the e-th epoch, b-th round is bounded. Specifically, recalling Equation (4), we have, w u e,b+1 we,b+1 w u e,b we,b ηe,b ℓ(w u e,b ; zi) ℓ(we,b; zi) , w u e,b(u)+1 we,b(u)+1 w u e,b(u) we,b(u) i Be,b(u)\{u} ℓ(w u e,b(u); zi) ℓ(we,b(u); zi) + ηe,b(u) |Be,b(u)| ℓ(we,b(u); u). (43) For all e, b and z Be,b, suppose that exists G such that G=max ℓ(we,b; z) < . w u e,b we,b w u e,b 1 we,b 1 + 2ηe,b 1G. (44) Consider a simple geometrically step size decay strategy ηe,b = qηe,b 1, where 0 < q < 1 is the decay rate, and let η be the initial stepsize.6 Recalling that t = e B + b, then we get w u e,b we,b w u e,b 1 we,b 1 + 2ηGqt 1. (45) Note that the initial retraining model and learning model are identical. Therefore, based on the above recursion, we can obtain the geometric progression as follows, w u e,b we,b 2ηGqt 1 + 2ηGqt 2 + ... + 2ηGq0 Therefore, we have completed the proof of Lemma 2. C.4 DETAILED PROOF OF THEOREM 4 Proof of Theorem 4. Let us recall the approximator of Equation (11) in Section 3, we have, e=0 Me,b(u) ℓ(we,b(u); u), where Me,b(u) = ηe,b(u) |Be,b(u)| ˆHE,B 1 e,b(u)+1. (47) For the E-th epoch B-th update, we have the following Taylor approximation, w u E,B w E,B a u E,B =(I ηE,B 1HE,B 1)(w u E,B 1 w E,B 1) + ηE,B 1o(w u E,B 1 w E,B 1) a u E,B. (48) For convenience, we define w u E,B = w u E,B w E,B, and ˆHE,B 1 = I ηE,B 1HE,B 1. Here, we begin to analyze the error generated by each update, w u E,B a u E,B = ˆHE,B 1 w u E,B 1 + ηE,B 1o( w u E,B 1) a u E,B, (49) 6Li et al. (2021) demonstrated that SGD with exponential stepsizes achieves (almost) optimal convergence rate for smooth non-convex functions, thus not affecting the convergence of learning process in our paper. Published as a conference paper at ICLR 2025 Furthermore, we observe that during the E-th epoch, Equation (49) can be obtained from affine stochastic recursion as follows, w u E,B a u E,B b=0 ˆHE,b w u E,0 a u E 1,B + ηq T B+b(u) b=b(u)+1 ˆHE,b ℓ(w u E,b(u); u) ℓ(w E,b(u); u) | {z } Approximation error term C2 during the E-th epoch + ηq T 1 w u E,B 1 + ηq T 2 ˆHE,B 1 w u E,B 2 + ... + ηq T B b=0 ˆHE,b w u E,0 . | {z } Approximation error term C1 during the E-th epoch (50) Through the above process, we obtain the result of the E-the epoch approximation. Now let s bound the last two terms, i.e., the approximation error term C1 and C2 during the E-th epoch. Recalling Lemma 2, we have that for all t = e B + b, w u e,b = w U e,b we,b 2ηG1 qt Recalling Lemma 3, we have that for all e, b and ρ is the spectral radius of I ηe,b He,b, ˆHe,b w u e,b = (I ηe,b He,b) w u e,b ρ w u e,b ρ 2ηG1 qt Therefore, we can bound the approximation error terms C1 and C2 during the E-th epoch as follows: (1) we first bound the first approximation error term C1 during the E-th epoch. For clarity, We define T as the total number of steps of SGD updates performed when obtaining a E,B, where T = EB +B. q T 1 w u E,B 1 + q T 2 ˆHE,B 1 w u E,B 2 + ... + q T B b=0 ˆHE,b w u E,0 2ηGq T 1 q2(T 1) 1 q + 2ηGq T 2 q2(T 2) 1 q ρ + ... + 2ηGq T B q2(T B) k=T 1 ρT 1 k(qk q2k). (2) we then bound the second approximation error term C2 during the E-th epoch. Suppose the stochastic gradient norm ℓ(we,b; u) is bounded by G, and we thus have the following upper bound, ηq T B+b(u) b=b(u)+1 ˆHE,b( ℓ(w u E,b(u); u) ℓ(w E,b(u); u)) 2Gη |B| q T B+b(u)ρB b(u) 1. Similarly, for each epoch, we have the aforementioned approximation error bound. Moreover, it is noted that when e = 0, the initial models for the retraining process and the learning process are same. Published as a conference paper at ICLR 2025 Therefore, we have the following overall approximate error across all epochs: w u E,B w u E,B η q T 1 w u E,B 1 + q T 2 ˆHE,B 1 w u E,B 2 + ... + q0 E Y b=0 ˆHk,b w u 0,0 + ηq T B+b(u) b=b(u)+1 ˆHE,b( ℓ(w u E,b(u); u) ℓ(w E,b(u); u)) + ... b=b(u)+1 ˆHE,b( ℓ(w u 0,b(u); u) ℓ(w0,b(u); u)) 2η2Gq T 1 q2(T 1) 1 q + 2η2Gq T 2 q2(T 2) 1 q ρ + ... + 2η2C q0 q2 (0) |B| q T B+b(u)ρB b(u) 1 + ... + 2Gη |B| qb(u)ρT b(u) 1 k=0 ρT k 1(qk q2k) + 2ηG e=0 ρT e B b(u) 1qe B+b(u). (55) Suppose q < ρ. Through the polynomial multiplies geometric progression, we obtain, w u E,B w u E,B = 2η2G 1 q ρT 1 T 1 X |B| ρT b(u) 1qb(u) E X ρB q B ρB b(u) 1qb(u), which yields an upper bound on the approximation error. For ease of comparison, we subsequently simplify the aforementioned result, as follows: w u E,B w u E,B 2η2G ρB q B ρB 1 2ηG(ρT q2T ) ηq (ρ q)(ρ q2) + 1 |B|(ρ q) At this point, we have completed the proof of Theorem 4 as follows, w u E,B w u E,B O(2ηGρT ). (58) If Theorem 4 further satisfies Assumption 1, and we choose η 2/(M + λ) to ensure gradient descent convergence as well as ρ < 1, we thus obtain a tight upper bound of O(ρn) for each sample. C.5 ADDITIONAL DEFINITIONS OF ASSUMPTION 1 We provide the following standard definitions of λ-strong convexity, L-Lipschitzness, M-Smoothness. Definition 3 (λ-Strong convexity). The loss function ℓ(w; z) is λ-strongly convex with respect to w, i.e., there exists a constant λ > 0 such that for any z Z and w1, w2 Rd, it holds that ℓ(w1; z) ℓ(w2; z) + ℓ(w2; z) , w1 w2 + λ 2 w1 w2 2 , (59) or equivalently, 2ℓ(w; z) λ for all w Rd. Definition 4 (L-Lipschitzness). The loss function ℓ(w; z) is L-Lipschitz with respect to w, i.e., there exists a constant L > 0 such that for any z Z and w1, w2 Rd, it holds that |ℓ(w1; z) ℓ(w2; z)| L w1 w2 . (60) or equivalently, ℓ(w; z) L for all w Rd. Definition 5 (M-Smoothness). The loss function ℓ(w; z) is M-smooth with respect to w, i.e., there exists a constant M > 0 such that for any z Z and w1, w2 Rd, it holds that ℓ(w1; z) ℓ(w2; z) M w1 w2 , (61) or equivalently, 2ℓ(w; z) M for all w Rd. Published as a conference paper at ICLR 2025 C.6 DETAILED PROOF OF THEOREM 6 Before we commence with our proof, we present some necessary lemmata. Lemma 9 (Shalev-Shwartz et al. (2009)). Let S = {zi}n i=1 where S Dn. Let w denote a minimizer of the population risk in Equation (1) and ˆw U denote a minimizer of the empirical risk without the knowledge of U in Equation (2) which aims to minimize FS\U(w), where |U| = m. Let Assumption 1 hold, we have, E F ˆw U F (w ) 4L2 λ(n m). (62) Proof of Lemma 9. A comprehensive proof of Lemma 9 is available in Shalev-Shwartz et al. (2009). For brevity and to save space, we present our generalization guarantee only for GD. Nevertheless, the results for SGD can be analogously obtained by similar process. Lemma 10. For any z Z, the loss function ℓ(w; z) is λ-strongly convex, L-Lipschitz and Msmooth. Let ˆw U = argminw FS\U(w) without the knowledge of U in problem Equation (2). For all e, b and vector v Rd, the spectral radius of I ηe,b He,b is defined as ρ which is largest absolute eigenvalue of these matrices. We have that after T steps of GD with initial stepsize η 2 λ+M , w U T ˆw U ρT 2M λ , where ρ = max {|1 ηλ| , |1 ηM|} < 1. (63) Proof of Lemma 10. Recalling that, according to the strong convexity and smoothness, λ, M > 0, we have λI 2ℓ(w; z) MI. Therefore, we have (1 ηT λ)I I ηT i BT 2ℓ(w U T ; zi) (1 ηT M)I. (64) Let η 2 λ+M . Therefore, ηT 2 λ+M , which implies that for any T [0, ): |1 ηT λ| < 1, |1 ηT M| < 1. (65) It is seen from the fundamental theorem of calculus that FS\U w U T = FS\U w U T FS\U ˆw U 0 2FS\U w U τ dτ w U T ˆw U , (66) where w U τ := w U T + τ ˆw U w U T . Here, w U τ 0 τ 1 forms a line segment between w U T and ˆw U. Therefore, According to the GD update rule and based on Lemma 3, we have, w U T +1 ˆw U = 0 2FS\U w U τ dτ w U T ˆw U I ηT 2FS\U w U τ w U T ˆw U ρ w U T ˆw U . (67) Apply it recursively, w U T ˆw U ρT w U 0 ˆw U , where ρ = max {|1 ηλ| , |1 ηM|} < 1. (68) By the assumption that λ-strongly convex and L-Lipschitzness, which implies that λ 2 w U ˆw U 2 F(w U) F ˆw U F(w U) F ˆw U 2L2 Therefore, we have completed the proof as follows, w U T ˆw U ρT 2L λ , where ρ = max {|1 ηλ| , |1 ηM|} < 1. (70) Published as a conference paper at ICLR 2025 Building upon the proofs presented in Lemma 9 and Lemma 10, we are now prepared to establish the generalization results stated in Theorem 6. Proof of Theorem 6. Let ˆw U be the empirical risk minimizer of the objective function FS\U(w), where |U| = m. Let T be the total number of update steps. For any z Z the loss function ℓ(w; z) is λ-strongly convex, M-smoothness and L-Lipschitz. For a minimizer w of the population risk in Equation (1) and w U E,B = w E,B + a U E,B + n, we have that E h F( w U E,B) F (w ) i = E h F( w U E,B) F( ˆw U) i + E F ˆw U F (w ) . (71) Based on Lemma 9, we get E F ˆw U F (w ) 4L2 λ(n m). (72) Then we note that the function F(w) satisfies L-Lipschitz, therefore we can obtain, E [F ( w E,B) F (w )] E[L w E,B ˆw U ] + 4L2 λ(n m). (73) For the first term of Equation (73), we have E h w U E,B w U E,B + w U E,B ˆw U i E w U E,B w U E,B | {z } C3 + E w U E,B ˆw U | {z } C4 Therefore, we now bound the first term C3 and the second term C4 of Equation (74) as follows: (1) C4: Based on Lemma 10 and w U E,B = w U T , we have that for η 2 λ+M , E w U E,B ˆw U ρT 2L λ , where ρ = max {|1 ηλ| , |1 ηM|} < 1. (75) (2) C3: Based on Theorem 4, we obtain, E[ w U E,B w U E,B ] = E[ w E,B + a U E,B w U E,B ] + E[ n ] U E,B + where U E,B = 2η2G ρB q B ρB b(u) 1qb(u). (76) And since the loss function ℓ(w; z) satisfies M-smoothness and λ-strongly convex, we thus have that 0 < ρ < 1. Plugging the above bound Equation (75) and Equation (76) into Equation (74), we have E h F( w U E,B) F (w ) i 4L2 λ(n m) + ρT 2L2 2 ln (1.25/δ) ϵ + 1) L U E,B . (77) Therefore, we have completed the proof. Tightness Analysis of Theorem 6. Furthermore, the following settings, where ρ is close to q, might render the bound on the approximation error vacuous and also lead to a high approximation error. We note that the first and third terms of U E,B can be equivalently written as O( ρT q T ρ q + ρT q T ρT 1 PT 1 t=0 ( q ρ)t +ρT B PB e=0( q ρ)e B). These terms increase as ρ approaches q, however are always upper bounded by O(ρT 1 T + ρT B T). In our analysis, T = EB + B is a fixed constant, so this bound remains meaningful, although it may not be tight in some cases. Such as when T , in this scenario ρ is close to 1, the bound therefore is equivalent to O( 2ηG 1 q2 + 2ηG 1 q B ). Fortunately, SGD typically does not need to find a local minima in infinite time, hence ρ is always less than 1. D SUPPLEMENTARY EXPERIMENTS D.1 HARDWARE, SOFTWARE AND SOURCE CODE The experiments were conducted on the NVIDIA Ge Force RTX 4090. The code were implemented in Py Torch 2.0.0 and leverage the CUDA Toolkit version 11.8. Our comprehensive tests were conducted on AMD EPYC 7763 CPU @1.50GHz with 64 cores under Ubuntu20.04.6 LTS. Source Code: Our code is available at Hessian Free Certified Unlearning Published as a conference paper at ICLR 2025 D.2 ADDITIONAL VERIFICATION EXPERIMENTS The verification experiments in this section consist of two parts: (1) We first investigate the reasons behind the randomness of previous works under the non-convex setting. We select forgetting samples under different random seeds, and show that previous works exhibit dependence for forgetting samples. This implies that the forgetting effect in prior studies benefits certain samples while negatively impacting others. (2) Second, we perform experiments on a larger-scale dataset and model with more metrics. Prior studies were unable to achieve this due to the costly Hessian computation. Sensitivity to the forgetting samples. We conduct experiments on a simple CNN and the hyperparameter setup is consistent with the verification experiments I in the main paper. As shown in Figure 4, we provide intuitive insights into the reasons for the decline of prior works in non-convex settings. Specifically, we investigated the L2 norm of the approximate parameter change a U and the norm of the exact parameter change across different selections of the forgetting sample. We observed that the performance of NS and IJ depends on the selection of the forgetting data points, i.e., NS and IJ can approximate actual changes well for some data while generating large approximation errors for others under the non-convex setting, as illustrated in Figure 4. The dependency of forgetting data points results in the decrease in distance and correlation coefficient metrics observed in Figure 1 (b). The observed results for NS and IJ are consistent with previous studies on influence functions (Basu et al. (2021; 2020)), where these methods display random behavior under the non-convex setting, resulting in a decrease in correlation with retraining. In contrast, our approach can effectively approximate the actual norm of parameter changes, regardless of the selection of the forgetting sample. Additional experiments. We further evaluate using larger-scale model Res Net-18 (He et al. (2016)) which features 11M parameters with three datasets: CIFAR-10 (Alex (2009)) for image classification, Celeb A (Liu et al. (2015)) for gender prediction, and LFWPeople (Huang et al. (2007)) for face recognition across 29 different individuals. Since previous second-order works (NS Sekhari et al. (2021)) and (IJ Suriyakumar et al. (2022)) are difficult to compute (Hessian) in this scenario, we instead use the following fast baselines, as described and similarly set up in Tarun et al. (2023): Fine Tune: In the case of finetuning, the learned model is finetuned on the remaining dataset. Neg Grad: In the case of gradient ascent, the learned model is finetuned using the negative of update direction on the forgetting dataset. Configurations: For Fine Tune and Neg Grad, we adjust the hyperparameters and execute the algorithm until the accuracy on the forgetting dataset approaches that of the retrained model. Here are our detailed hyperparameter settings: (1) We conducted evaluation on Res Net-18 trained on CIFAR-10 with 50,000 samples. The learning stage is conducted for 40 epochs with step size of 0.001 and batch size of 256. For Fine Tune, 10 epochs of training are performed with step size of 0.001 and batch size of 256. For Neg Grad, we run gradient ascent for 1 epoch with step size of 5 10 5 and batch size of 2. (2) We conducted evaluation on Res Net-18 trained on LFW with 984 samples, for the classification of 29 facial categories. The learning stage is conducted for 50 epochs with step size of 0.004 and batch size of 40. For Fine Tune, 10 epochs of training are performed with step size of 0.004 and batch (a) Low Deletion Rates (b) High Deletion Rates Figure 4: Verification experiments II. Evaluation shows a comparison between the norm of approximate parameter change a U and norm of exact parameter change w U E,B w E,B across different random seeds. Intuitively, the NS and IJ methods are contingent on the selection of forgetting data. In contrast, our approach consistently approximates the actual values effectively. Published as a conference paper at ICLR 2025 size of 40. For Neg Grad, we run gradient ascent for 1 epoch with step size of 4e 6 and batch size of 1. (3) We conducted evaluation on Res Net-18 trained on Celeb A with 10,000 samples. The learning stage is conducted for 5 epochs with step size of 0.01 and batch size of 64. For Fine Tune, 2 epochs of training are performed with step size of 0.01 and batch size of 64. For Neg Grad, we run gradient ascent for 1 epoch with a step size of 0.0005 and batch size of 1. We threshold the gradients within 10 for the above experiments. Finally, we uniformly unlearned 50 data points for each dataset. We denote the test data as Dt, the remaining data as Dr, and the forgetting data as Df. The following Table 3-5 are our evaluation results: we employ five widely-used metrics to evaluate the performance of unlearning methods: Acc_Dt (accuracy on Dt), error _Dt (error on Dt), error_Dr (error on Dr), error_Df (error on Df), Distance (L2 norm between unlearned and retrained model). Performances of our scheme: We still maintain model indistinguishability from the retrained model on more complex tasks and larger models, where the distance is only 0.06 on CIFAR10 and 0.29 on Celeb A. The results also show that the unlearned model of proposed method performs similarly to the retrained model in terms of the accuracy metric. Moreover, compared to these baselines, our method achieves forgetting at only the millisecond level, demonstrating the potential of our method on larger-scale models. Additionally, we noticed that when training is unstable, our method would lead to larger performance degradation, such as on the LFW dataset. Although our method has similar accuracy on the forgetting dataset, it suffers performance loss on the remaining dataset. This implies that our method is sensitive to the training process and hyperparameters in non-convex scenarios. We summarize the limitations in Appendix B.1, and provide possible schemes to address these problems. Model indistinguishability: For Fine Tune, we can observe that even though the accuracy results are similar to the retrained model, the distance metric is very large (e.g. 2.21 on CIFAR-10 and 2.36 on Celeb A). For Neg Grad, we observe that the accuracy on forgetting dataset Df and remaining dataset Dr would rapidly decrease simultaneously, but even in this case, the distances are is still lower than that of Fine Tune for all three datasets. This demonstrates that a small difference in accuracy does not indicate the model indistinguishability. Therefore, we use the L2 norm and correlation metric as the primary measure of forgetting completeness, which are commonly employed in studies related to influence function (Basu et al. (2021; 2020)). These studies serve as inspiration for prior certified unlearning papers (Sekhari et al. (2021); Suriyakumar et al. (2022); Liu et al. (2023)) that we compare. Measuring forgetting through L2 norm is also used in unlearning research, e.g., Wu et al. (2020); Izzo et al. (2021), but it necessitates retraining a model, which is impractical for real-world applications. Table 3: Verification experiments III. Unlearning results on Resnet-18 trained on CIFAR-10 dataset. Method Acc_Dt (%) Err_Dt (%) Err_Dr (%) Err_Df (%) Distance Runtime (Sec) Fine Tune 79.34 20.66 15.92 22.00 2.21 106.76 Neg Grad 40.61 59.39 59.15 70.00 0.11 0.56 Retrain 79.63 20.37 16.19 24.00 468.65 Ours 79.62 20.38 16.15 22.00 0.06 0.0006 Table 4: Verification experiments IV. Unlearning results on Resnet-18 trained on LFW dataset. Method Acc_Dt (%) Err_Dt (%) Err_Dr (%) Err_Df (%) Distance Runtime (Sec) Fine Tune 76.02 23.98 8.99 4.00 0.52 15.60 Neg Grad 63.01 36.99 15.63 20.00 0.48 0.46 Retrain 80.89 19.11 4.07 26.00 77.43 Ours 71.92 28.08 14.91 22.00 0.33 0.001 Table 5: Verification experiments V. Unlearning results on Resnet-18 trained on Celeb A dataset. Method Acc_Dt (%) Err_Dt (%) Err_Dr (%) Err_Df (%) Distance Runtime (Sec) Fine Tune 95.78 4.22 1.07 4.00 2.36 35.09 Neg Grad 37.95 62.05 58.12 60.0 1.13 0.96 Retrain 95.72 4.28 1.20 4.00 87.72 Ours 95.57 4.43 1.12 2.00 0.29 0.0006 Published as a conference paper at ICLR 2025 D.3 ADDITIONAL APPLICATION EXPERIMENTS Furthermore, we evaluate the performance on different datasets and models in real-world applications. Our evaluations are conducted from two perspectives: runtime and utility. Specifically, run-time focuses on the time spent precomputing unlearning statistics and the speedup of the unlearning algorithm compared to the retraining algorithm. Moreover, we evaluate the utility by the test accuracy of the unlearned model to ensure that generalization performance is not compromised. Finally, we record the area under the curve (AUC) for the MIA-L attack. Configurations: We assess the performance on various datasets and models, as presented in Table 6. The values in parentheses indicate the difference in test accuracy compared to the retrained model. (1) We train a LR and simple CNN on MNIST with 1,000 training data 20% data points to be forgotten, which have setups identical to the aforementioned verification experiments I. We further evaluate on FMNIST with 4,000 training data and 20% data points to be forgotten using CNN and Le Net with a total of 61,706 parameters. The training is conducted for 30 epochs with a stepsize of 0.5 and a batch size of 256. (2) Moreover, we assessed our method on Resnet-18 (He et al. (2016)) trained on the CIFAR10 dataset for image classification with 50,000 data points, Celeb A dataset for gender prediction with 10,000 data points, and LFWPeople dataset for face recognition with 984 data points. we uniformly add noise with a standard deviation of 0.01. The configurations align with those detailed in Table 3-5. In these circumstances, we have not evaluated NS and IJ due to out-of-memory. Application experiment results show that our proposed method is computationally efficient and performs well in realistic environments, outperforming benchmarks of certified unlearning regarding all metrics by a significant margin, as demonstrated in Table 6. It achieves minimal computational and storage overhead while maintaining high performance in both convex and non-convex settings. We further investigate data removal with large-scale datasets using Res Net-18, as shown in Table 6. Our proposed Hessian-free method demonstrates significant potential for over-parameterized models, as previous second-order certified unlearning works are unable to compute the full Hessian matrix in such cases. Notably, for unlearning runtime, by precomputing statistical data before unlearning requests arrive, we only need to execute simple vector additions, achieving millisecond-level efficiency (0.6 ms) with minimal performance degradation in test accuracy. Finally, although our method has lower computation and storage complexity compared to previous works, it still requires O(n2d) precomputation and O(nd) storage which is expensive when facing both larger-scale datasets. We summarize the limitations in Appendix B.1, and provide possible schemes to address these problems. Table 6: Application experiments II. We conduct two types of comparisons: (1) on small-scale models and datasets with previous second-order algorithms, where we compute the total computational/storage cost for all samples; and (2) on large-scale models and datasets with fast unlearning algorithms, where we compute the computational/cost overhead of per-sample forgetting. Method Dataset Model Unlearning Computaion Pre Computaion Storage MIA-L Distance Test Accuracy (%) Runtime (Sec) Speedup Runtime (Sec) (GB) AUC Unlearned model MNIST LR 5.09 102 1.34 2.55 103 0.23 0.51 0.18 86.25 (-1.50) CNN 5.81 103 0.12 2.91 104 1.78 0.51 1.78 83.50 (-10.25) FMNIST CNN 2.31 104 0.54 1.16 105 1.78 0.54 2.41 57.69 (-22.25) Le Net 8.53 104 0.15 4.27 105 14.18 0.54 1.10 62.00 (-18.50) MNIST LR 0.23 102 29.64 2.55 103 0.23 0.52 0.18 86.25 (-1.50) CNN 1.91 102 3.63 2.91 104 1.78 0.51 3.12 82.75 (-11.00) FMNIST CNN 2.95 101 424 1.16 105 1.78 0.51 0.94 75.69 (-4.25) Le Net 3.24 101 402 4.27 105 14.18 0.54 0.57 76.75 (-3.75) MNIST LR 0.0073 93,376 2.20 102 0.03 0.51 0.15 87.50 (-0.25) CNN 0.0268 25,821 5.34 102 0.08 0.51 0.68 91.50 (-2.25) FMNIST CNN 0.127 98,468 3.66 103 0.32 0.51 0.74 77.85 (-2.09) Le Net 0.163 79,828 4.43 103 0.48 0.53 0.53 78.63 (-1.87) Method Dataset Model Unlearning Computaion Pre Computaion Storage MIA-L Distance Test Accuracy (%) Runtime (Sec) Speedup Runtime (Sec) (GB) AUC Unlearned model CIFAR10 106.76 4.39 0.56 2.21 79.34 (-0.29) LFW Resnet18 15.60 4.96 0.53 0.52 76.02 (-4.87) Celeb A 35.09 2.50 0.56 2.36 95.78 (0.06) CIFAR10 0.56 836.88 0.51 0.11 40.61 (-39.03) LFW Resnet18 0.46 168.33 0.46 0.48 63.01 (-17.88) Celeb A 0.96 91.38 0.50 1.13 37.95 (-57.77) CIFAR10 0.0006 781,083 184.53 0.043 0.51 0.06 79.62 (-0.01) LFW Resnet18 0.001 77,430 84.99 0.043 0.56 0.31 71.92 (-8.97) Celeb A 0.0006 146,200 31.90 0.042 0.54 0.29 95.57 (-0.15) Published as a conference paper at ICLR 2025 D.4 ADDITIONAL ADVERSARY EXPERIMENTS Membership Inference Attack against Learning (MIA-L) aims to determine whether a specific example is part of the training set of the learned model. Carlini et al. (2022) introduced the Likelihood Ratio Attack (Li RA), which leverages the model s per-example difficulty scores and employs a well-calibrated Gaussian likelihood estimate to gauge the probability of a sample being part of training data. Finally, the adversary queries the confidence of the target model on a specific example and outputs a parametric likelihood-ratio test. Membership Inference Attack against Unearning (MIA-U) seeks to predict whether a target sample is part of the learned model s training set but not in the corresponding unlearned (retrained) model s training set. Chen et al. (2021) proposed MIA-U to attack Scratch methods, specifically the Retrain and SISA (Bourtoule et al. (2021)) algorithms, where the adversary jointly utilizes the posteriors of the two models as input to the attack model, either by concatenating them or by computing difference. Specifically, it includes five methods to construct features for the attack model: Direct difference (Direct Diff), Sorted difference (Sorted Diff), Direct concatenate (Direct Concat), Sorted concatenate (Sorted Concat), and Euclidean distance (Euc Dist). Configuration: (1) We investigate the Area Under the Curve (AUC) of MIA-U at different deletion rates. (2) We report the Receiver Operating Characteristic (ROC) curve of MIA-U in large-scale experiments as well. (3) Finally, as a complement to the results of the Direct Diff method reported in the main text, we present the ROC of the remaining four feature construction methods in MIA-U. We maintain the other hyperparameters consistent with previous experiments. Different deletion rates. As illustrated in Figure 5, a clear correlation exists between the deletion rate and the effectiveness of MIA-U attacks. Specifically, at a low deletion rate of 1%, the AUC of MIA-U attacks reaches 91% for our proposed method and 89% for the retrain method in the LR model, indicating that unlearning algorithms may increase the risk of privacy breaches. When the deletion rate exceeds 5%, the performance of MIA-U attacks approaches random guessing (AUC = 50%), suggesting that removing a larger amount of data between successive model releases makes it increasingly difficult for MIA-U attacks to accurately infer the deleted individual information. Figure 5: Membership Inference Attack II. The AUC values of MIA-U at different deletion rates. The feature construction method employed in MIA-U is Direct Diff, with error bars representing 95% confidence intervals. The target model in the first plot is LR, while the second plot corresponds to a CNN. The findings show that lower deletion rates lead to privacy leakage, whereas higher deletion rates diminish MIA-U s attack performance, bringing it closer to random guessing. Figure 6: Membership Inference Attack III. ROC curves for MIA-U on Res Net18. The datasets used in the figure, from left to right, are CIFAR-10, LFW, and Celeb A. For large-scale setups, our method (red line) performs consistently with the retraining algorithm (light red line) under MIA-U. Published as a conference paper at ICLR 2025 (a) Sorted Dif (b) Direct Concat (c) Sorted Concat (d) Euc Dist Figure 7: Membership Inference Attack IV. ROC curve for different feature construction methods. The target model in the first row is overfitted CNN, whereas the second row is well-generalized LR. Different feature construction methods. As shown in Figure 7, we experiment with a wellgeneralized LR model and an overfitted CNN model, and the experimental results are consistent with the conclusion of Chen et al. (2021): sorting the posteriors improves attack performance; concatenation-based methods achieve higher attack performance on the overfitted CNN; and difference-based methods perform better on the well-generalized LR. In addition, we observe that perturbing the unlearning model can effectively defend the Euc Dist method. However, for other feature construction methods, adding noise to perturb the model after a certain level does not reduce the AUC of the attack model. This reveals a fundamental dilemma of most certified data removal processes developed so far: Although the goal is to achieve performance similar to a gold-standard retrained algorithm, the retrained model itself remains vulnerable to MIA-U, exposing the failure modes of certified unlearning. Nonetheless, as shown in Figure 5, removing at least 5% of the data when releasing the unlearned model can mitigate privacy leakage. However, we are not aware of any recent certified data removal mechanisms that can satisfy the deletion capacity exceeding 5% of the training set. Fortunately, one of the defense mechanisms proposed by Chen et al. (2021), which employs Differentially Private Stochastic Gradient Descent (DP-SGD), can effectively mitigate MIA-U but requires careful trade-off amongst privacy, unlearning efficiency, and model utility. E SENSITIVITY ANALYSIS AND ABLATION STUDIES In this section, we investigate the effects of different step sizes, epochs, stepsize decay rates, and gradient clipping on the proposed approach. We conducted simulations in both convex and nonconvex scenarios, using LR for the convex case and 2-layer CNN for the non-convex case, respectively. Takeaway. Let s first provide a brief summary of the main results of ablation studies. (1) Impact of Step Sizes. Firstly, a smaller step size enables our model to be closer to the retrained model. However, when the step size exceeds a certain threshold, the bound will become meaningless and result in significant approximation errors. (2) Impact of Epochs. Secondly, when the step size is sufficiently small, the approximation error is insensitive and exhibits slow growth with respect to the number of epochs, i.e. excessive iterations in this scenario do not lead to significant errors. (3) Impact of Stepsize Decay Rates. Thirdly, our results indicate that as the decay rate q decreases, the approximation error diminishes. Specifically, The stepsize decay strategy introduces a bound on approximation error, preventing a continuous increase in error as the size of epochs grows. (4) Impact of Gradient Clipping. To prevent gradient explosion, we introduced gradient clipping. We show that gradient clipping, which scales the gradient based on its norm, has a similar effect to using a smaller step size. Although, in theory, clipping the gradient may introduce errors, our experiments indicate that this impact is negligible when the threshold is set to a large value. Published as a conference paper at ICLR 2025 E.1 IMPACT OF STEP SIZES In this section, we investigated (i) the impact of step sizes and (ii) the impact of small stepsizes with more epochs. Additionally, we discussed the connection between learning generalization and our unlearning method. Configurations: (i) We investigate the impact of step sizes {0.005, 0.01, 0.05, 0.1, 0.3} on distance metrics in Table 7 and Table 8. For LR, training was performed for 15 epochs with a stepsize of 0.05 and batch size of 32. For CNN, training was carried out for 20 epochs with a step size of 0.05 and a batch size of 64. (ii) Furthermore, in Figure 9, to investigate the effect of epochs under small step sizes, LR was trained for 100 epochs and 200 epochs with a step size of 0.01 (initially 0.05 with 15 epochs). For CNN, training for 50 epochs and 500 epochs with a step size of 0.025 (originally 0.05 with 20 epochs) was performed. Impact of step sizes. As shown in Table 7 and Table 8, a smaller step size enables our model to be closer to the retrained model. Specifically, as Theorem 4 demonstrates that, in this process, the error arises from the scaled remainder term, i.e., the term o(w u e,b we,b) is scaled by the subsequent Me,b(u). Therefore, the approximation error mainly consists of two components: (1) the step size ηe,b and (2) the maximum eigenvalue λ1 of He,b. A smaller step size η contracts the eigenvalues of the matrix ηe,b He,b, thereby reducing the spectral radius ρ of the matrix I ηe,b He,b, and ultimately reducing the approximation error. Furthermore, when the step size is smaller than a certain threshold, increasing the step size does not lead to significant errors. However, when the step size exceeds this threshold (making ρ not less than or close to 1), the errors will grow exponentially, as demonstrated in Table 7 and Table 8 for step size η 0.3. In this scenario, our method leads to a complete breakdown of the model, rendering it unusable. However, it is noteworthy that theoretical predictions (Wu et al. (2018)) and empirical validation (Gilmer et al. (2022)) suggested that, across all datasets and models, successful training occurs only when optimization enters a stable region of parameter space where λ1 η 2. Therefore, achieving good optimization results typically does not require a strategy of using aggressive stepsizes. The connection between learning and unlearning. There is a connection between the sharpness of learning generalization studies and the unlearning method that we propose. Specifically, these studies aim to explore the sharpness of the Hessian of loss function He,b, and the term sharpness refers to the maximum eigenvalue of the Hessian, denoted as λ1 in some studies, such as Gilmer et al. (2022). A flat loss space is widely recognized as beneficial for gradient descent optimization, leading to improved convergence. Importantly, Gilmer et al. (2022) demonstrated the central role that λ1 plays in neural network optimization, emphasizing that maintaining a sufficiently small λ1 during optimization is a necessary condition for successful training (without causing divergence) at large step sizes, as shown in Figure 8. Models that train successfully enter a region where λ1 η 2 mid training (or fluctuate just above this bound). In our paper, we leverage the second-order information from the Hessian at each iteration during the training process to compute the approximators and achieve forgetting. As indicated in Theorem 4, a small λ1 of the Hessian is beneficial for reducing Table 7: Impact of Step Sizes on Convex Setting. Stepsize 0.05 represents the hyperparameter choice of experiments for LR in the main paper. It can be observed that as the step size decreases, the unlearned model is closer to the retrained model. (Seed 42) Stepsize Distance 1% 5% 20% 30% 0.005 0.018 0.040 0.092 0.128 0.01 0.023 0.0518 0.127 0.178 0.05 0.027 0.064 0.149 0.210 0.1 0.028 0.066 0.154 0.217 0.3 5.581 17.48 28.45 40.64 Table 8: Impact of step sizes on non-Convex setting. Stepsize 0.05 represents the hyperparameter choice of experiments for CNN in the main paper. It can be observed that as the step size decreases, the unlearned model is closer to the retrained model. (Seed 42) Stepsize Distance 1% 5% 20% 30% 0.005 0.004 0.015 0.058 0.084 0.01 0.017 0.069 0.257 0.372 0.05 0.167 0.380 0.584 0.825 0.1 0.437 0.795 1.722 2.72 0.3 122.2 393.4 1647 2455 Published as a conference paper at ICLR 2025 Figure 8: The connection between learning and unlearning. We conduct experiment on MNIST and keep the other hyperparameters fixed and adjust the learning rate from 0.001 to 0.31. Under a fixed iteration budget, we have: (1) When the step size is small (less than the threshold η1 = 0.01), the training is often insufficient, leading to lower test accuracy. (2) Increasing the step size ensures sufficient training, but when the learning rate becomes too large, overfitting and instability lead to a decrease in test accuracy. We also observe the following phenomena in the unlearning process: (1) When the step size is below a threshold η2 = 0.25, the growth of approximation error is extremely slow. (2) However, once the step size exceeds this threshold, the error grows exponentially. (a) Convex Setting (b) Non-Convex Setting Figure 9: Impact of Small Step Size with More Epochs. (a)(b) correspond to LR and CNN with 20% of the data to be forgotten, respectively. The red bar represents the distance metric when using a larger step size; the blue bar represents the distance metric with a small step size during the larger epochs. It can be observed that even when training more epochs with smaller stepsize, the approximation error remains lower compared to training with larger stepsize. (Seed 42) the approximation error. Therefore, effectively reducing λ1 during optimization, so that the training enters a flatter region, will not only improve the generalization ability of the learning stage model but also reduce the approximation error in the unlearning stage. As suggested in Gilmer et al. (2022), strategies like stepsize warm-up or initialization strategies for architectures can be employed to stabilize learning and potentially reduce unlearning errors by decreasing λ1, enabling training at larger step sizes. Impact of small step size with more epochs. Finally, as shown in Figure 9, with a small step size (0.005) during model training, we observed a lower approximation error. Even with large epoch sizes, the approximation error remains lower compared to cases with a large step size, such as 0.05. Therefore, selecting an appropriately small step size ensures smaller approximation errors, as these errors typically converge to a smaller value as the number of epochs increases. E.2 IMPACT OF EPOCHS In this subsection, we began investigating the distance metric as the number of epochs increases. Configurations: We investigated the distance by varying the epoch from 1 to 100. We conducted experiments with a high deletion rate of 20% and a low deletion rate of 5%. The selection of other hyperparameters remains consistent with the ones used previously. Published as a conference paper at ICLR 2025 (a) Convex Setting (b) Non-Convex Setting Figure 10: Impact of Epoch. (a) (b) illustrate that in the distance for LR and CNN by varying epoch from 1 and 100. As shown in the Figure (a) (b), the error accumulates with the increase in epochs. However, with an appropriate choice of step size, such growth is acceptable (Seed 42). Impact of epochs. As shown in Figure 10, the distance exhibits convergent behavior, ultimately stabilizing at a constant value. This phenomenon is consistent with our theoretical findings, as stated in Theorem 4, which demonstrates that the approximation error decreases with an increasing number of epochs for convex models. The contractive mapping ensures a gradual reduction in the final error. In the case of non-convex models, the error tends to increase with the number of epochs, ultimately stabilizing, a phenomenon that can be explained by the degeneracy of the Hessian in over-parameterized models (Sagun et al. (2018)), as we will discuss in further detail later. E.3 IMPACT OF STEPSIZE DECAY These experiments aim to study the impacts of stepsize decay. We consider two scenarios: one with stepsize decay (q = 0.995) and the other without stepsize decay (q = 1). Impact of decay rates of stepsize: We can observe from Figure 11 that the Euclidean distance for the stepsize decay strategy is generally smaller compared to that without the stepsize decay. This aligns with our earlier analysis of step size in Appendix E.1, i.e., a smaller step size leads to a smaller approximation error, and choosing an appropriate stepsize decay rate ensures a continuous reduction in step size, preventing the unbounded growth of the approximation error. Additionally, as in Figure 11 (b), It can be observed that the distance metric tends to stabilize even without stepsize decay. This result can be explained by the conclusions in Sagun et al. (2018; 2016). Specifically, Sagun et al. (2018; 2016) analyzed the spectral distribution of the Hessian eigenvalues during CNN training. It reveals that, initially, the Hessian eigenvalues are symmetrically distributed around 0. As training progresses, they converge towards 0, indicating Hessian degeneracy. Toward the end of training, only a small number of eigenvalues become negative, and a large portion of the Hessian eigenvalues approach 0 when the training approaches convergence. Recall that our previous analysis of the stepsize in Appendix E.1, where we demonstrate that the approximation error is mainly determined by ρ of I ηe,b He,b, and ρ is determined by ηe,b and λ1. As the number of iterations increases, the eigenvalues of the Hessian He,b tends to 0. This implies a decrease in λ1, which leads to a slower increase in error. Therefore, in Figure 11 (b), the distance gradually stabilizes even without a decay strategy due to the decrease in λ1 during training, and it still increases due to the existence of some larger outlier eigenvalues. E.4 IMPACT OF GRADIENT CLIPPING Stochastic gradient algorithms are often unstable when applied to functions that lack Lipschitz continuity and/or bounded gradients. It is well-established that exploding gradients pose a significant challenge in deep learning applications (Hanin (2018)). Since our proposed method relies on backtracking historical gradients, the approach may fail completely when gradients tend toward infinity. To circumvent this issue, we incorporated gradient clipping into all the aforementioned experiments. Gradient clipping is a widely adopted technique that ensures favorable performance in deep network training by restricting gradient norms from becoming excessively large (Mai & Johansson (2021); Chen et al. (2020); Zhang et al. (2020)). However, this operation indicates that Published as a conference paper at ICLR 2025 (a) Convex Setting (b) Non-Convex Setting Figure 11: Impacts of the stepsize decay rate on the Euclidean distances for (a) the LR and (b) the CNN with 20% data to be forgotten during 25 epochs. The red line represents learning with a decay rate of 0.995, while the blue line represents without decay. It can be observed that stepsize decay effectively reduces the approximation error of the unlearned model. (Seed 42) Figure 12: Impacts of the gradient clipping on the distance metric for the LR (orange) and the CNN (blue) with 20% data to be forgotten. (Seed 42) the gradients in the Taylor expansion of Equation (2) are not the true gradients but rather the clipped gradients, which introduces an approximation error. Our subsequent experiments will demonstrate that this error can be rendered negligible. It is worth noting that our method does not inherently rely on gradient clipping but instead imposes requirements on the stability of the learning process. Configurations: We investigate the impact of gradient clipping with thresholds set at {0.3, 0.5, 0.8, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 15, 20, 25, 30, 35}. We remove 20% of the data while keeping other hyperparameter settings consistent with those described in the main text. Impact of gradient clipping. As evidenced in Figure 12, the implementation of small clipping thresholds results in a scaling down of the gradient proportional to its norm, producing an effect similar to using a smaller step size. Consequently, the application of lower clipping thresholds maintains the error at a relatively minimal level. However, as the gradient clipping threshold surpasses a certain range, the error introduced by the clipping process becomes the predominant factor, culminating in an escalation of the overall error. Nonetheless, most gradients no longer require clipping during the optimization process. Therefore, the induced error diminishes until the threshold reaches a sufficiently elevated value, at which point no supplementary error is introduced. Once this threshold is attained, gradient clipping ceases to exert an influence on the performance of our proposed unlearning algorithm.