# visual_privacy_auditing_with_diffusion_models__20fb5557.pdf Published in Transactions on Machine Learning Research (03/2025) Visual Privacy Auditing with Diffusion Models Kristian Schwethelm k.schwethelm@tum.de Chair for AI in Healthcare and Medicine, Technical University of Munich (TUM) and TUM University Hospital Johannes Kaiser johannes.kaiser@tum.de Chair for AI in Healthcare and Medicine, Technical University of Munich (TUM) and TUM University Hospital Moritz Knolle moritz.knolle@tum.de Chair for AI in Healthcare and Medicine, Technical University of Munich (TUM) and TUM University Hospital Sarah Lockfisch sarah.lockfisch@tum.de Chair for AI in Healthcare and Medicine, Technical University of Munich (TUM) and TUM University Hospital Daniel Rückert daniel.rueckert@tum.de Chair for AI in Healthcare and Medicine, Technical University of Munich (TUM) and TUM University Hospital Department of Computing, Imperial College London, UK Munich Center for Machine Learning (MCML), Munich, Germany Alexander Ziller alex.ziller@tum.de Chair for AI in Healthcare and Medicine, Technical University of Munich (TUM) and TUM University Hospital Reviewed on Open Review: https: // openreview. net/ forum? id= D3DA7pgpvn Data reconstruction attacks on machine learning models pose a substantial threat to privacy, potentially leaking sensitive information. Although defending against such attacks using differential privacy (DP) provides theoretical guarantees, determining appropriate DP parameters remains challenging. Current formal guarantees on the success of data reconstruction suffer from overly stringent assumptions regarding adversary knowledge about the target data, particularly in the image domain, raising questions about their real-world applicability. In this work, we empirically investigate this discrepancy by introducing a reconstruction attack based on diffusion models (DMs) that only assumes adversary access to real-world image priors and specifically targets the DP defense. We find that (1) real-world data priors significantly influence reconstruction success, (2) current reconstruction bounds do not model the risk posed by data priors well, and (3) DMs can serve as heuristic auditing tools for visualizing privacy leakage. 1 Introduction The widespread collection of sensitive data including personal identities, private locations, and medical conditions has raised critical privacy concerns, particularly in machine learning (ML) where models can leak private information from their training data. While differential privacy (DP) (Dwork & Roth, 2014) has emerged as the gold standard for providing formal privacy guarantees, the practical effectiveness of these guarantees against real-world privacy attacks remains uncertain, especially in data reconstruction scenarios where adversaries attempt to recover complete data records. This uncertainty poses a substantial challenge for practitioners facing privacy-utility trade-offs, as stronger privacy protections typically reduce model performance. To deploy DP techniques effectively, practitioners need a clear understanding of how the mathematical privacy guarantees translate to practical protection of sensitive information. Published in Transactions on Machine Learning Research (03/2025) Less private DP-SGD with ε1 More private DP-SGD with ε2 (3) Denoised ε1 (3) Denoised ε2 (1) Gradient inversion xtε1 xtε1 1 x0 ... xtε2 xtε2 1 x0 ... (2) Reverse Diffusion Process Figure 1: (1) Our reconstruction attack first extracts a noisy image from a DP algorithm with privacy guarantee εn using, e.g., gradient inversion on DP-SGD. (2) Then, it employs a DM for reconstruction by initiating its reverse diffusion process from a specific intermediate state xtεn. (3) We demonstrate DMs strong utility for reconstruction and visual auditing, aiding communication with non-experts. In this example, it is possible to infer that ε1 offers little privacy protection, allowing accurate reconstruction, while ε2 safeguards certain details but still allows disclosure of high-level personal attributes. Attempts to address this issue have led to the development of formal upper bounds on data reconstruction success under DP, aiming to enable practitioners to assess the effect of DP guarantees on the maximum fidelity achievable by an adversary s reconstruction attack. Central to such bounds are the formalized threat models, which define the capabilities of potential adversaries and the scenarios under which the bounds hold. Due to the complexity of mathematically formalizing practical scenarios, overly pessimistic threat models that account for the most powerful (worst-case) attacks have been adopted (Balle et al., 2022; Guo et al., 2022). Although these bounds offer generality and hold against all possible attacks, they potentially overestimate the threat in real-world scenarios (Ziller et al., 2024a). For instance, while Hayes et al. (2023) demonstrate the near-tightness of reconstruction robustness (Re Ro) bounds (Balle et al., 2022), their analysis assumes a highly informed adversary with access to a prior that includes the complete target record a scenario that is unlikely to occur in practice. As a result, there is an ongoing effort to formulate practical reconstruction bounds tailored to realistic attack scenarios. Ziller et al. (2024b) derive formal bounds on error measures for a specific reconstruction attack on DP-SGD (Song et al., 2013; Abadi et al., 2016) training, assuming an uninformed adversary capable of selecting model architecture and hyperparameters but lacking prior knowledge about the data. However, this threat model may be overly optimistic in domains where data priors are common, particularly in the image domain, where the underlying structure and characteristics of the data are well understood. This raises critical questions: How do realistic data priors influence the effectiveness of data reconstruction attacks against differentially private machine learning, and can existing reconstruction bounds capture these threats? In this work, we empirically investigate the effectiveness of DP in defending against image reconstruction attacks that leverage real-world data priors (i.e., domain-specific knowledge about the underlying data distribution). We compare our findings with the theoretical guarantees given by the (worst-case) Re Ro bound (Hayes et al., 2023) and Ziller et al. (2024b). For our study, we extend a well-established gradient attack method that involves adversarial modification of the model architecture an approach proven effective in, e.g., federated learning (Fowl et al., 2022; Boenisch et al., 2023a) and backdooring pretrained models (Feng & Tramèr, 2024) and incorporate image priors approximating the underlying data distribution of the reconstruction target. Our attack extension builds upon a principal finding: under DP-SGD, this gradient attack is equivalent to reconstructing an image perturbed with additive noise (Boenisch et al., 2023b; Ziller et al., 2024b). We exploit this characteristic by leveraging strong image priors learned by diffusion models (DMs) (Ho et al., 2020; Dhariwal & Nichol, 2021) to denoise the reconstructions (see Fig. 1). This enables us to specifically target DP-SGD s reliance on noise perturbations and model adversarial access to real-world data priors. Our findings reveal the substantial influence of data priors on reconstruction success, with their impact varying depending on the strength of the prior (inherent distribution shift). We find that works on reconstruction Published in Transactions on Machine Learning Research (03/2025) bounds do not model these observations well. Instead, they rely on simplified assumptions not reflective of real-world scenarios. Beyond highlighting these theoretical limitations, we propose a practical application of our method: a visual auditing tool that complements formal DP guarantees, making privacy risks tangible and interpretable for non-technical stakeholders. By bridging theoretical bounds and practical privacy evaluation, our work contributes to a better understanding of DP s real-world implications and highlights crucial directions for developing more realistic privacy guarantees. Our main contributions can be summarized as follows: 1. We demonstrate that the efficacy of realistic reconstruction depends on the strength of the data prior, which is not adequately represented by current theoretical bounds. This highlights a significant gap between theory and practice in privacy guarantees. 2. We introduce an image reconstruction attack leveraging diffusion models to model adversaries with realistic data priors. Our results reveal the significant threat such priors pose in disclosing private information under DP-SGD. 3. We empirically identify privacy parameters necessary to defend against our attack, and demonstrate its efficacy as a heuristic tool for visually auditing privacy risks. 2 Background and Related Work 2.1 Differential Privacy Guarantees Differential privacy (DP) (Dwork & Roth, 2014) is a formal guarantee that provably bounds privacy leakage from computations on datasets. Definition 1. A randomized algorithm (mechanism) M satisfies (ε, δ)-DP if, for any pairs of adjacent datasets D D that differ in a single sample and all sets of outcomes S Range(M), it holds that: Pr[M(D) S] eεPr[M(D ) S] + δ. (1) Intuitively, DP limits the influence of an individual sample on the algorithm s outcome. In this work, we focus on the Gaussian mechanism (GM), a standard DP mechanism often used in machine learning (ML). GMs introduce controlled randomization to mask the contribution of a sample through the addition of i.i.d. Gaussian noise N(0, 2 2σ2I), where the variance of the noise is calibrated to the privacy guarantee using noise multiplier σ and the mechanism s global L2-sensitivity 2 (Balle & Wang, 2018). The standard DP threat model assumes an adversary with complete knowledge, except for the noise realization and whether D or D was used to compute the mechanism s outcome. The latter aspect naturally aligns DP with membership inference attacks, which aim to determine whether a specific individual s data was part of the dataset (Yeom et al., 2018). The prevailing approach to implementing DP in ML is DP-Stochastic Gradient Descent (DP-SGD) (Song et al., 2013; Abadi et al., 2016), which limits the privacy leakage from training. DP-SGD is a modified version of SGD that enforces an upper bound on sensitivity 2 = C by clipping per-sample gradients to an upper norm bound C and adding calibrated i.i.d. Gaussian noise N(0, C2σ2I). Interpreting DP Guarantees. In DP, the level of privacy preservation provided by an algorithm is typically quantified using parameters such as ε and δ. However, a more operationally interpretable way of quantifying privacy is by attributing practical risk against certain attacks under specific threat models. Recent work has advanced our understanding of practical DP effects through auditing techniques (Lokna et al., 2023; Nasr et al., 2023; Steinke et al., 2023; 2024), relaxed threat models (Nasr et al., 2021; Kaissis et al., 2023b; Ziller et al., 2024a), novel privacy attacks (Geiping et al., 2020; Boenisch et al., 2023a; Feng & Tramèr, 2024), and deployment strategies (Ponomareva et al., 2023; Cummings et al., 2024). However, the impact of adversarial access to real-world data priors on privacy risk remains largely unexplored. Furthermore, while membership inference attacks have been extensively studied, the threats posed by reconstruction Published in Transactions on Machine Learning Research (03/2025) attacks which aim to recover complete data records are less understood. Our work addresses these gaps by leveraging powerful data priors to evaluate reconstruction risks under DP, providing practitioners with empirical insights for navigating privacy-utility trade-offs. 2.2 Bounding Data Reconstruction Success Data reconstruction attacks on ML models pose a critical privacy risk by attempting to recover complete data records. While DP mechanisms primarily target membership inference protection, they inherently also defend against broader privacy breaches, including data reconstruction. However, in scenarios where membership information is considered insensitive or even public knowledge, practitioners might consider relaxing DP guarantees to improve model utility. This has motivated several theoretical bounds on reconstruction success (Guo et al., 2022; Stock et al., 2022; Balle et al., 2022; Ziller et al., 2024b). Yet, these bounds may not fully capture the threat of practical attack scenarios, particularly when adversaries possess prior knowledge about the data. Our work complements these theoretical results by providing an empirical framework to assess reconstruction risks in practical settings. Reconstruction Robustness (Re Ro). Re Ro (Balle et al., 2022; Hayes et al., 2023; Kaissis et al., 2023a) provides a formal upper bound on the probability of a successful data reconstruction attack. Definition 2. A randomized algorithm (mechanism) M satisfies (η, γ)-Re Ro if, for any reconstruction attack R on the algorithm s output ω, any dataset D {z}, where z denotes the reconstruction target sampled from prior π, fixed error function ρ, and baseline success probability κπ,ρ(η), it holds that: κπ,ρ(η) Pz π, ω M(D {z})(ρ(z, R(ω)) η) γ. (2) Re Ro adopts the DP threat model with the slight modification that only a fixed part of the dataset D is known to the adversary, while the added reconstruction target z is not. However, the adversary has some prior knowledge π about the target, which, informally, serves as a reference distribution for z. Given the difficulty in determining ρ, η, and π, as well as approximating κπ,ρ(η), Hayes et al. (2023) introduced a worst-case Re Ro definition ((0, γ)-Re Ro) based on sample matching. Let ρ = 1(z = R(ω)), η = 0, κπ,ρ(η) = 1/n, and π be a uniform distribution over a discrete set of n candidate samples {ztarget, z1, . . . , zn 1}. Then, the adversary s task reduces to re-identifying the target by matching the observation ω to the correct sample from the prior set, which results in a simplified reconstruction setting. Despite its limitations, we use (0, γ)-Re Ro as our theoretical baseline since it remains the only computationally viable implementation of Re Ro. For clarity, we use (0, γ)-Re Ro and Re Ro interchangeably. Uninformed Data Reconstruction. Ziller et al. (2024b) introduced formal bounds on error metrics for a specific data reconstruction attack on DP-SGD training. They assume an uninformed adversary with no prior knowledge about the data but with the ability to observe gradient updates and modify the model architecture and training hyperparameters. By exploiting these capabilities, they showed that a worst-case adversary can replace the architecture to maximize privacy vulnerabilities. Specifically, they reduce the model to a single fully connected layer without bias, where the output of the layer directly represents the loss: ℓ= Wx, with W denoting the weights and x the input data. This architecture allows direct reconstruction of the input by inverting the observed gradients xrec = ℓ W = x. However, the application of clipping and additive Gaussian noise on the gradients introduced by DP-SGD perturbs the reconstruction, leading to: xrec = x max(||x||2/C, 1) + ξ, with ξ = N(0, C2σ2I). (3) Leveraging the closed-form solution of this uninformed reconstruction attack, Ziller et al. formally analyze the reconstruction success and, e.g., bound the expected mean squared error (MSE): MSE(x, xrec) C2σ2. Published in Transactions on Machine Learning Research (03/2025) 2.3 Diffusion Models (DMs) Diffusion models, particularly the denoising diffusion probabilistic models (DDPMs) (Ho et al., 2020), have gained significant attention in recent years. DDPMs rely on a forward diffusion process that step-wise perturbs a signal (image) x0 q(x0) with additive i.i.d. Gaussian noise until the noise predominates. Mathematically, the forward process is described by: 1 βtxt 1 + p βtϵt 1, with ϵt 1 N(0, I), (4) where the noise schedule βt (0, 1) controls both the variance of the noise and the factor reducing the signal at step t = {1, 2, . . . , T}. By defining αt := 1 βt, αt := Qt s=1 αs, and given the underlying Markov chain q(x1, . . . , x T | x0) = QT t=1 q(xt | xt 1), the noisy latent variables xt can be conditioned on x0: xt = αtx0 + 1 αtϵ, with ϵ N(0, I). (5) The reverse process, used for generating new signals, employs a neural network to approximate the intractable distribution q(xt 1 | xt) and predict the sampled noise ϵ. Given a large number of steps T and well-behaved schedules of βt, x T converges to a standard Gaussian. Thus, a signal can be generated by initiating the reverse process with a standard Gaussian sample and iterative denoising. Image Denoising with Diffusion Models. Generative image denoising strategies leveraging diffusion models have demonstrated state-of-the-art perceptual quality in natural (Xie et al., 2023; Pearl et al., 2023; Yang et al., 2023) and medical imaging (Xiang et al., 2023b;a; Chung et al., 2023). Notably, in the broader field of image restoration, diffusion models have also demonstrated efficacy in tasks like super-resolution, colorization, and inpainting (Li et al., 2023). In contrast to these works, we do not aim to enhance image quality by removing some minor natural noise. Instead, we aim to recover private information from deliberately perturbed images with substantial noise scales introduced to provide DP guarantees. Image Reconstruction with Diffusion Models. Concurrent work (Huang et al., 2024; Liu et al., 2025) has also explored the use of DMs for maximizing attack success in image reconstruction. However, our study adopts a broader approach by investigating how data priors influence these attacks in the context of DP. We also analyze how such attacks align with or challenge existing theoretical reconstruction bounds, providing a deeper understanding of their implications for privacy guarantees. In this section, we present our methodology by formally introducing the problem and describing our approach to leveraging diffusion models (DMs) for image reconstruction. 3.1 Problem Definition Threat Model. We study a common attack scenario on DP-SGD training where an adversary can manipulate the model architecture, hyperparameters, and observe training gradients to reconstruct private images. Beyond these capabilities, we assume the adversary has access to realistic image priors, i.e., statistical knowledge about natural image features (such as textures, edges, and color gradients) or domain-specific patterns (like facial features or medical imaging characteristics). Base attack. Our work builds on the analytical attack introduced by Fowl et al. (2022), which enables near-perfect data reconstruction from training gradients by placing a fully connected (imprint) layer at the model s front. For a fully connected layer (y = Wx + b), where x is the input, W i denotes a weight row, and bi a bias parameter, the gradients of the loss L with respect to a single row of weights and the bias are: yi x, bi L = L Published in Transactions on Machine Learning Research (03/2025) Element-wise division of these gradients ( L yi = x) perfectly recovers the input, showing that gradients can directly encode the training data. When applied to DP-SGD, this attack yields a scaled, noisy version of the target image (see (Boenisch et al., 2023b; Ziller et al., 2024b)). While the original attack results in noise from a ratio distribution due to the division of Gaussian random variables, Ziller et al. (2024b) show how to modify the attack to achieve Gaussian noise instead. Adversarial Problem Statement. We consider a scenario where an adversary extracts a perturbed image from a differentially private computation. In a standard DP-SGD setting, this extraction can be achieved by either maliciously setting a batch size of 1 to directly obtain per-sample gradients or by applying common techniques to extract per-sample gradients from accumulated gradients (Fowl et al., 2022; Boenisch et al., 2023a;b). The privatized observation is represented as: λx + ξ, (7) where λ = max(||x||2/C, 1) denotes the clipping factor, x q(x) denotes the original image sampled from data distribution q(x), and ξ is sampled from i.i.d. Gaussian noise N(0, C2σ2I). The adversary s goal is to reconstruct the private information in x from xpriv. Following the attack scenario described above, the reconstruction task reduces to a denoising problem. This formulation allows us to assess the practical privacy leakage and determine sufficient noise levels for protection. 3.2 Private Image Reconstruction with Diffusion Models Diffusion models (DMs) learn powerful image priors that closely approximate complex data distributions by solving denoising tasks. Their ability to combine observed features with learned statistical patterns makes them highly effective at reconstructing corrupted information. Additionally, DMs can handle various noise levels without retraining, making them well-suited for our work, investigating the effectiveness of different noise perturbations. We leverage these strengths to develop a reconstruction attack that exploits DP-SGD s reliance on noise-based privacy mechanisms. Given the inverse problem in Eq. (7), we define the posterior over the observation as q(x | xpriv). We approximate this posterior using DMs and leverage their Markov chain to initiate the reverse process from a conditional intermediate state pθ(xt 1 |xpriv) instead of pure noise until the original image is recovered, i.e., x0 x. The easiest choice to integrate xpriv into the reverse process is adopting the Variance Exploding (VE) form of DMs (Song et al., 2021b): xt = x0 + σtϵ, with ϵ N(0, I), (8) with variance schedule1 {σ2 t }T t=1 and σ2 T . Notice that this formulation does not reduce the signal by αt, which is also the case in Eq. (7). However, the VE form complicates hyperparameter tuning since standard DMs employ the Variance Preserving (VP) form, where 1 αT 1 (see Eq. (5)). To utilize the VP form, we use the equivalence between the two forms (Kawar et al., 2022) and define the starting point of the reverse process as follows: xtstart = 1 q 1 + σ2 tstart xpriv = 1 q 1 + σ2 tstart λx + ξ , (9) where tstart denotes the starting step in the DM s noise schedule. Handling the Clipping Factor λ. An unknown parameter in Eq. (9) is the linear scalar λ introduced by the clipping operation of DP-SGD. This parameter scales down the image, reducing its brightness and value range. In a realistic scenario, the exact value of λ is unknown to the adversary. However, given that λ represents a single value and images are typically characterized by a constrained range of color values, the 1Note that σt denotes the noise variance relative to x0, which differs from βt and αt in the standard DM definition (see Sec. 2.3), as well as from σ in DP (see Sec. 2.1). Published in Transactions on Machine Learning Research (03/2025) adversary can easily approximate λ through normalization or trial-and-error (see Appendix A). Therefore, we assume the worst-case scenario, wherein the adversary successfully recovers the exact value of λ. We stress that knowing λ comes with little advantage to the adversary, as it only enables them to rescale the image after perturbation, which increases the noise sample ξ by factor λ. Thus, the signal-to-noise ratio remains unchanged. Combining our assumption with Eq. (9) yields: xtstart = λ q 1 + σ2 tstart xpriv = 1 q 1 + σ2 tstart (x + λξ). (10) Markov Chain Matching. The Markov chain of (discrete) DMs is based on a pre-defined noise schedule {βt}T t=1 and, therefore, does not contain a state for all possible noise variances. Thus, to initiate the reverse process from a perturbed image with variance ˆσ2 = C2σ2λ2, we must compute the variance schedule 1 Qt s=1(1 βt) 1 = r 1 αt 1 (11) and search the next largest state tstart under the condition σtstart > ˆσ. Enforcing Data Consistency. The stochastic generative process of DMs introduces randomness after each denoising step, increasing sample diversity. However, this is not desirable in reconstruction problems, where the results should closely resemble the original. Therefore, we enforce data consistency by adopting the deterministic generation process of denoising diffusion implicit models (DDIMs) (Song et al., 2021a), which has been shown to retain image features throughout the generation process: xt 1 αt ϵ(t) θ (xt) αt 1 αt 1 ϵ(t) θ (xt). Data consistency can also be enforced by conditioning every step of the reverse process on xpriv by, e.g., concatenating the low-quality sample to each latent state xt, yielding the posterior distribution pθ(xt 1 | xpriv, xt). However, our scenario considers extreme cases where the images are heavily perturbed. Conditioning with such low-quality images causes harmful effects on the generation of DMs (Li et al., 2023). Thus, we forgo such an approach. Algorithm 1 summarizes our method. Algorithm 1: Private Image Reconstruction with DMs Require: xpriv = 1/λx + ξ, with ξ N(0, C2σ2I), noise schedule αt, model θ 1 αt 1 Variance schedule 2: x priv = λxpriv Rescaling 3: tstart = arg min t (σt Cσλ) σt > Cσλ Markov chain matching 4: xtstart = 1 p 1+σ2 tstart xpriv Reparameterization 5: for t = tstart, . . . , 1 do Step-wise denoising 6: xt 1 = αt 1 xt 1 αt ϵ(t) θ (xt) αt + 1 αt 1 ϵ(t) θ (xt), Published in Transactions on Machine Learning Research (03/2025) 10 3 10 2 10 1 100 101 102 10 3 10 2 10 1 100 101 102 103 10 3 10 2 10 1 100 101 102 103 0 0.2 0.4 0.6 0.8 0 0.2 0.4 0.6 0.8 0 0.2 0.4 0.6 0.8 101 102 0 0.2 0.4 0.6 0.8 (a) CIFAR-10 (32 32) 101 102 0 0.2 0.4 0.6 0.8 (b) Celeb A-HQ (256 256) 101 102 0 0.2 0.4 0.6 0.8 (c) Image Net (256 256) Figure 2: Average similarity of image reconstructions. We compute the similarity between the original and the reconstructed images from our DM attack (blue ), the attack of (Ziller et al., 2024b) (red ), and the Re Ro attack (Hayes et al., 2023) (green ). For µ < 3, Celeb A-HQ and Image Net images exceed the maximal noise variance σT in the schedule; thus, no results can be given. The dashed line represents average similarity between test images, indicating at which point reconstructions become unrelated to the original. 4 Experiments This section compares the data reconstruction success of prevailing theoretical reconstruction bounds and our practical attack leveraging image priors learned by diffusion models (DMs). Additionally, it investigates the effectiveness of using DMs to specifically target the DP-SGD defense in scenarios with limited target access and weaker attack assumptions. For experimental details and ablation experiments, refer to Appendices B and D, respectively. 4.1 Experimental Setting Our experimentation includes three datasets: CIFAR-10 (Krizhevsky & Hinton, 2009), Celeb A-HQ (Karras et al., 2018), and Image Net-1K (Deng et al., 2009), with the latter two resized to 256 256. For evaluation, we randomly select a subset of 5,000 test images from each dataset and quantitatively measure the reconstruction success with mean squared error (MSE), VGG-based learned perceptual image patch similarity (LPIPS) (Simonyan & Zisserman, 2015; Zhang et al., 2018), and structural similarity index measure (SSIM) (Wang et al., 2004). We note that the employed DM s are not trained on test images. We report results with respect to µ = C/σ, where C denotes the clipping parameter and σ the noise multiplier of DP-SGD. It can be interpreted as a signal-to-noise ratio (SNR), where C bounds the signal amplitude and σ represents the noise. Analogously to the privacy parameter ε, a lower µ (SNR) makes reconstruction more difficult and, thus, corresponds to a higher privacy guarantee. We note that given a specific DP-SGD configuration (number of steps and sampling rate), µ can be converted to the (ε, δ) notion (see Appendix C). Published in Transactions on Machine Learning Research (03/2025) Original µ = 100 µ = 50 µ = 30 µ = 20 µ = 10 µ = 5 µ = 3 Figure 3: Reconstruction results under DP with respect to µ = C/σ. For each dataset, the reconstructed image from the base attack without prior knowledge (top) and our DM attack (bottom) are shown. The top images also represent the input of our attack. 4.2 Reconstruction Success under Different Data Priors We evaluate privacy leakage under reconstruction attacks with varying levels of prior knowledge about the target data. Our attack incorporates realistic priors learned by diffusion models that capture both general statistics of image features and domain-specific patterns. We compare our approach against the (0, γ)-Re Ro bound, which assumes access to the target image within a prior set of 256 images (Hayes et al., 2023) and the bound of (Ziller et al., 2024b), which assumes no prior knowledge. The results in Fig. 2 show that our reconstruction error falls between the theoretical bounds: higher than the Re Ro bound but lower than Ziller et al. s uninformed adversary bound2. As expected, the Re Ro bound is overly pessimistic, assuming a powerful attacker achieving mostly perfect reconstructions an unrealistic scenario, especially for the challenging Image Net dataset. Conversely, Ziller et al. are too optimistic and underestimate the threat of a realistic attacker. A crucial finding emerges when examining reconstruction performance across image scales: As image size increases, the gap between our and Ziller et al. s results widens, revealing a significant weakness in their method. Images with the same SNR show similar reconstruction difficulty, suggesting that image size should not substantially impact reconstruction success under constant µ. This aligns with both DP and Re Ro, which depend on the SNR ratio C/σ. The discrepancy with Ziller et al. s findings where Cσ is derived for specific error metrics indicates that directly bounding metrics with limited perceptual relevance may inadequately capture both reconstruction difficulty and actual privacy risk. This also highlights the challenge of formulating an appropriate error function for Re Ro that isn t based on matching, as in (0, γ)-Re Ro. 2In Fig. 2, LPIPS and SSIM for Ziller et al. s attack converge while the MSE does not. This discrepancy arises not from limitations in their method but from LPIPS and SSIM, which necessitate clipping color values between 0 and 1. Published in Transactions on Machine Learning Research (03/2025) Original µ = 200 µ = 150 µ = 100 µ = 50 µ = 30 µ = 20 µ = 10 µ = 5 Figure 4: Reconstruction results under distribution shift. The performance of the DM trained on CIFAR-10 and tested on CIFAR-100 (top), and the performance of the Image Net DM on Celeb A-HQ (middle) and Che Xpert (bottom) are shown. Regarding our reconstruction success, a phase transition becomes apparent. For µ 5, the similarity of our reconstructions to the original converges to the average similarity of test images (dashed line in Fig. 2), indicating the diffusion model generates plausible but unrelated images from the learned distribution. The qualitative results in Fig. 3 and Appendix E demonstrate the strong performance of our DM-based attack against the DP-SGD defense. While the base attack yields noisy images, the DM successfully recovers substantial original image content. Notably, we observe an additional phase transition at µ = 20, where reconstructed images start deviating from the original while still sharing similar high-level attributes such as dataset class, image color, or gender. Additionally, as observed in Fig. 2, for µ 5, the reconstructions become unrelated to the original, indicating good privacy protection. 4.3 Reconstruction Success under Distribution Shift Previously, we assumed the adversary has access to training data with the same underlying data distribution as the target (test) data, which enables them to learn a very strong data prior. To investigate how prior knowledge quality affects reconstruction success, we examine scenarios where the data prior does not stem from the same distribution, i.e., an out-of-distribution prior, which is weaker than in-distribution priors. We perform this experiment in three settings: (1) The DM is trained on CIFAR-10 and is used to reconstruct test images from CIFAR-100 (Krizhevsky & Hinton, 2009). These datasets are very similar, differing primarily in class number and diversity. (2) An Image Net DM reconstructs Celeb A-HQ face images, and (3) the Image Net DM reconstructs grayscale chest X-ray images from the Che Xpert dataset (Irvin et al., 2019). Intuitively, the greater the discrepancy between training and test data, the larger the distribution shift. The results in Figs. 4 and 5 show a clear trend: larger distribution shifts (weaker data priors) lead to decreased reconstruction success. This is particularly evident from the shift in the privacy guarantee (µvalue) at which the similarity of the reconstructions surpasses the average similarity of the test datasets (dashed line in Fig. 5). Irrespective of the error function, this serves as a good indicator for less-than-useful reconstructions, which are more similar to the training data than the target images. Our findings demonstrate the significant impact of data distribution shift on the reconstruction performance of DMs, especially in high privacy regimes. However, our method yields reasonable reconstructions for low privacy guarantees even in scenarios with significant distribution shifts. Published in Transactions on Machine Learning Research (03/2025) 101 102 0 0.2 0.4 0.6 0.8 Figure 5: Reconstruction success under distribution shift. The performance of the DM trained on CIFAR-10 and tested on CIFAR-100 (blue ), and the Image Net DM tested on Celeb A-HQ (green ) and Che Xpert (red ) are shown. The dashed line represents average similarity between test images of the datasets (same color). The results show the significant influence of distribution shift between the data prior and the reconstruction target. 101 102 0 0.2 0.4 0.6 0.8 Figure 6: Reconstruction success estimated from average similarity between multiple DDPM samples of CIFAR-10 (blue ), Celeb A (green ), and Image Net (red ) compared to the true success obtained by computing the average similarity between reconstructions and the original images (dashed line). The closeness between same colored lines shows that the original image is not required to estimate the reconstruction success well. 4.4 Estimating Reconstruction Success without Target Access DMs always generate a candidate reconstruction, even when the perturbed image lacks information for reconstruction. This implies that the resulting reconstructions may differ from the target images. While this is useful for data owners and practitioners who can directly compare the features of the reconstructions with the original images, adversaries lacking access to the original image (reconstruction target) may struggle to infer which features are made up by the DM. We propose that adversaries can overcome this challenge by generating multiple candidate reconstructions using the probabilistic generation process of, e.g., DDPMs and assess which features remain consistent across reconstructions. Such features are most likely to originate from the reconstruction target. This is analogous to a maximum a posteriori attack, where the mode of the empirically generated images is computed. Fig. 6 shows the average pairwise similarity between five DDPM generations from each of the 5,000 noisy images under different privacy levels, providing insights into estimating the reconstruction success using such an approach. It shows that, for all datasets, the true reconstruction success (dashed lines in Fig. 6) can be estimated well without access to the original image. Qualitative results in Supplementary Fig. 15 illustrate the shared features among different generations. In our example, gender can be inferred until µ = 5, and the hair color remains consistent until µ = 20. These findings highlight that visual insights from DMs reconstructions hold value not only for data owners comparing reconstructions with the original images but also for adversaries who only have access to the noisy image and multiple generations. 4.5 Reconstruction Success under Weaker Attack Assumptions While our previous experiments assumed ideal attack conditions to align with theoretical bounds, we now evaluate our method s effectiveness in a practical training scenario where adversaries can only access accumulated mini-batch gradients rather than per-sample gradients. This setting aligns with real-world DP-SGD implementations and follows the attack methodology of Fowl et al. (2022). The key challenge lies in extracting individual sample information from aggregated gradients. Our experimental setup employs mini-batches of size 64 and a Res Net-9 architecture (Klause et al., 2022) augmented with an imprint layer (Fowl et al., 2022). The imprint layer s parameters are carefully tuned to separate individual activations within the accumulated gradient using a binning technique (128 bins). For reconstruction, we process the binned weight and bias gradients according to Eq. (6), instead of relying on Published in Transactions on Machine Learning Research (03/2025) (a) Ground truth batch (b) Reconstructed (no DP) (c) Reconstructed (DP) (d) DM reconstructed (DP) Figure 7: Reconstruction results on Res Net-9 training using the attack of Fowl et al. (2022) under: (b) no privatization, (c) DP-SGD, and (d) DP-SGD including our DM attack strategy. DP-SGD is applied with µ = 1000. Cells are sorted by reconstructions success of (d). Empty cells (gray) indicate failed reconstructions, while green borders highlight successfully reconstructed images matching the original training data. the clipping factor λ approach used in prior experiments. This produces reconstructions with noise following a Gaussian ratio distribution, though the precise distribution may deviate due to the binning process. Our attack pipeline consists of four key steps: (1) apply binning to separate individual contributions from the observed accumulated gradient, (2) divide weight and bias gradients for each bin to yield the (perturbed) reconstruction, (3) determine the reverse diffusion starting point tstart by estimating the noise variance using scikit-image, and (4) apply our DM reconstruction method to each bin. This approach enables simultaneous recovery of all images in the batch from a single accumulated gradient. Results in Fig. 7 demonstrate the effectiveness of our attack under these non-ideal conditions. Without DP, most images are reconstructed with high fidelity (Fig. 7b). While DP-SGD protection initially reduces most bins to noise patterns (Fig. 7c), our DM-based method successfully recovers many recognizable features from these protected versions (Fig. 7d). This strongly supports the effectiveness of data priors in reducing DP s effects. Finally, it is important to note that µ can be significantly larger in this scenario due to other non-zero gradients in the Res Net architecture that are not used for reconstruction but add to the gradient norm. This effect decreases the signal for the relevant gradients while the noise scale remains constant, further challenging the reconstruction process. 5 Discussion Our investigation into real-world data priors reveals a significant gap between theoretical reconstruction bounds and empirical attack outcomes. We demonstrate that the strength of the data prior substantially influences the reconstruction success, positioning our attack between existing bounds. This finding highlights both the importance and challenge of incorporating realistic data priors into formal privacy guarantees. While our work primarily provides empirical evidence of enhanced reconstruction capabilities through diffusion models (DMs), establishing theoretical bounds for reconstruction attacks under such data priors remains an open challenge. The key difficulty lies in capturing the semantics of learned representations and their relationship to the private training data in a formal framework. While threat models assuming no prior (Ziller et al., 2024b) remain valuable where priors are unavailable, incorporating prior knowledge could substantially improve bound accuracy. We also recognize the flexibility of the Re Ro bound in formalizing different data priors. Nevertheless, addressing challenges related to defining an appropriate prior π and error functions ρ will be crucial for its effective implementation. The development of theoretical foundations for realistic data priors, such as those learned by DMs, represents an important direction for future research. Furthermore, we find that DMs excel at extracting information from heavily perturbed images beyond human visual capabilities. Our method substantially improves the reconstruction outcomes of previous methods (Fowl et al., 2022; Boenisch et al., 2023b; Ziller et al., 2024b) with a simple post-processing step. Given the Published in Transactions on Machine Learning Research (03/2025) widespread availability of pre-trained DMs across various data distributions, it is reasonable to assume that adversaries can profit from their capabilities. This accessibility broadens the scope of potential adversaries who could utilize such techniques, emphasizing the urgency for robust defenses to counter such threats. However, this same capability presents an opportunity: DMs can serve as powerful tools for visualizing reconstruction risk in privacy audits. Our reconstructions effectively capture the residual information after privatization, offering intuitive insights into privacy leakage that complement formal DP guarantees. Unlike abstract privacy parameters that are challenging to interpret for non-experts (Cummings et al., 2024), our approach offers tangible means of visualizing privacy leakage, thereby facilitating communication with stakeholders and enhancing their comprehension of privacy in machine learning. For instance, when reconstructions preserve class information while altering low-level features, it suggests that low-level features are privatized, whereas the class information can be disclosed (Sec. 4.4). This practical approach to privacy auditing aligns with recent developments in heuristic auditing methods (Steinke et al., 2024). We emphasize, however, that our method is intended as a communication tool and does not provide theoretical guarantees. Finally, since DP ensures consistent mathematical privacy guarantees regardless of prior knowledge even under ideal priors our findings suggest an interesting practical consideration: the standard noise levels in mechanisms like DP-SGD may be excessive. We demonstrate that prior-based post-processing of privatized gradients can increase the utility of data reconstruction attacks. The same approach can be used to get improved model utility in DP-SGD training by post-processing gradients before the model update, while obtaining the same mathematical guarantee. This insight aligns with recent advances in gradient denoising techniques (Nasr et al., 2020; Zhang et al., 2024), suggesting promising directions for future research. 6 Conclusion Our work empirically shows the critical role of data priors in reconstruction attacks, revealing limitations in current theoretical bounds. This gap between theory and practice highlights the need for reconstruction bounds that better capture real-world adversarial capabilities. We show both the threat and utility of DMs in privacy contexts: while they enhance reconstruction attacks, they also enable intuitive privacy auditing that bridges the gap between theoretical guarantees and practical understanding. Future work should focus on developing more adaptive privacy metrics and defenses that can address realistic capabilities of adversaries. Broader Impact Statement This work proposes a data reconstruction attack capable of disclosing sensitive information from real-world ML models. While our method can be used maliciously, we use it to highlight how to defend against data reconstruction attacks using privacy methods like DP and how our attack can be utilized as a tool for non-experts to select sufficient privacy guarantees. Furthermore, we only utilize publicly available images, thereby not exposing data that is not already available. Acknowledgments KS and DR received support from the Bavarian Collaborative Research Project PRIPREKI of the Free State of Bavaria Funding Programme Artificial Intelligence Data Science . AZ and DR received support from the German Ministry of Education and Research and the Medical Informatics Initiative as part of the Private AIM Project (grant no. 01ZZ2316C). JK received support from the European Union under Grant Agreement 101100633 (EUCAIM). Views and opinions expressed are however those of the author(s) only and do not necessarily reflect those of the European Union or the European Commission. Neither the European Union nor the granting authority can be held responsible for them. MK was supported by the DAAD programme Konrad Zuse Schools of Excellence in Artificial Intelligence, sponsored by the Federal Ministry of Education and Research. Published in Transactions on Machine Learning Research (03/2025) SL received support from the Research and Development Program Information and Communication Technology Bavaria, DIK0444/03. Martin Abadi, Andy Chu, Ian Goodfellow, H. Brendan Mc Mahan, Ilya Mironov, Kunal Talwar, and Li Zhang. Deep learning with differential privacy. In Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security, CCS 16, pp. 308 318, New York, NY, USA, 2016. Association for Computing Machinery. ISBN 9781450341394. doi: 10.1145/2976749.2978318. URL https://doi.org/10.1145/2976749.2978318. Borja Balle and Yu-Xiang Wang. Improving the gaussian mechanism for differential privacy: Analytical calibration and optimal denoising. In International Conference on Machine Learning, 2018. URL https: //api.semanticscholar.org/Corpus ID:21713075. Borja Balle, Giovanni Cherubin, and Jamie Hayes. Reconstructing training data with informed adversaries. In 2022 IEEE Symposium on Security and Privacy (SP), pp. 1138 1156. IEEE Computer Society, 2022. doi: 10.1109/SP46214.2022.9833677. URL https://doi.ieeecomputersociety.org/10.1109/SP46214. 2022.9833677. Leonard Berrada, Soham De, Judy Hanwen Shen, Jamie Hayes, Robert Stanforth, David Stutz, Pushmeet Kohli, Samuel L. Smith, and Borja Balle. Unlocking accuracy and fairness in differentially private image classification, 2023. F. Boenisch, A. Dziedzic, R. Schuster, A. Shahin Shamsabadi, I. Shumailov, and N. Papernot. When the curious abandon honesty: Federated learning is not private. In 2023 IEEE 8th European Symposium on Security and Privacy (Euro S&P), pp. 175 199, Los Alamitos, CA, USA, jul 2023a. IEEE Computer Society. doi: 10.1109/Euro SP57164.2023.00020. URL https://doi.ieeecomputersociety.org/ 10.1109/Euro SP57164.2023.00020. Franziska Boenisch, Adam Dziedzic, Roei Schuster, Ali Shahin Shamsabadi, Ilia Shumailov, and Nicolas Papernot. Reconstructing individual data points in federated learning hardened with differential privacy and secure aggregation. In 2023 IEEE 8th European Symposium on Security and Privacy (Euro S&P), pp. 241 257, 2023b. doi: 10.1109/Euro SP57164.2023.00023. S.G. Chang, Bin Yu, and M. Vetterli. Adaptive wavelet thresholding for image denoising and compression. IEEE Transactions on Image Processing, 9(9):1532 1546, 2000. doi: 10.1109/83.862633. Hyungjin Chung, Eun Sun Lee, and Jong Chul Ye. Mr image denoising and super-resolution using regularized reverse diffusion. IEEE Transactions on Medical Imaging, 42(4):922 934, 2023. doi: 10.1109/TMI.2022. 3220681. Rachel Cummings, Damien Desfontaines, David Evans, Roxana Geambasu, Yangsibo Huang, Matthew Jagielski, Peter Kairouz, Gautam Kamath, Sewoong Oh, Olga Ohrimenko, Nicolas Papernot, Ryan Rogers, Milan Shen, Shuang Song, Weijie Su, Andreas Terzis, Abhradeep Thakurta, Sergei Vassilvitskii, Yu-Xiang Wang, Li Xiong, Sergey Yekhanin, Da Yu, Huanyu Zhang, and Wanrong Zhang. Advancing Differential Privacy: Where We Are Now and Future Directions for Real-World Deployment. Harvard Data Science Review, jan 16 2024. https://hdsr.mitpress.mit.edu/pub/sl9we8gh. Kostadin Dabov, Alessandro Foi, Vladimir Katkovnik, and Karen Egiazarian. Image denoising by sparse 3-d transform-domain collaborative filtering. IEEE Transactions on Image Processing, 16(8):2080 2095, 2007. doi: 10.1109/TIP.2007.901238. Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248 255, 2009. doi: 10.1109/CVPR.2009.5206848. Published in Transactions on Machine Learning Research (03/2025) Prafulla Dhariwal and Alexander Quinn Nichol. Diffusion models beat GANs on image synthesis. In A. Beygelzimer, Y. Dauphin, P. Liang, and J. Wortman Vaughan (eds.), Advances in Neural Information Processing Systems, 2021. URL https://openreview.net/forum?id=AAWu Cvza Vt. David L Donoho and Iain M Johnstone. Ideal spatial adaptation by wavelet shrinkage. Biometrika, 81(3): 425 455, 09 1994. ISSN 0006-3444. doi: 10.1093/biomet/81.3.425. URL https://doi.org/10.1093/ biomet/81.3.425. Cynthia Dwork and Aaron Roth. The algorithmic foundations of differential privacy. Found. Trends Theor. Comput. Sci., 9(3 4):211 407, aug 2014. ISSN 1551-305X. doi: 10.1561/0400000042. URL https: //doi.org/10.1561/0400000042. Shanglun Feng and Florian Tramèr. Privacy backdoors: Stealing data with corrupted pretrained models. In Forty-first International Conference on Machine Learning, 2024. URL https://openreview.net/forum? id=7yix JXmzb8. Liam H Fowl, Jonas Geiping, Wojciech Czaja, Micah Goldblum, and Tom Goldstein. Robbing the fed: Directly obtaining private data in federated learning with modified models. In International Conference on Learning Representations, 2022. URL https://openreview.net/forum?id=fwz Ugo0FM9v. Jonas Geiping, Hartmut Bauermeister, Hannah Dröge, and Michael Moeller. Inverting gradients - how easy is it to break privacy in federated learning? In H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin (eds.), Advances in Neural Information Processing Systems, volume 33, pp. 16937 16947. Curran Associates, Inc., 2020. URL https://proceedings.neurips.cc/paper_files/paper/2020/file/ c4ede56bbd98819ae6112b20ac6bf145-Paper.pdf. Sivakanth Gopi, Yin Tat Lee, and Lukas Wutschitz. Numerical composition of differential privacy. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan (eds.), Advances in Neural Information Processing Systems, volume 34, pp. 11631 11642. Curran Associates, Inc., 2021. URL https://proceedings.neurips.cc/paper_files/paper/2021/file/ 6097d8f3714205740f30debe1166744e-Paper.pdf. Chuan Guo, Brian Karrer, Kamalika Chaudhuri, and Laurens van der Maaten. Bounding training data reconstruction in private (deep) learning. In Kamalika Chaudhuri, Stefanie Jegelka, Le Song, Csaba Szepesvari, Gang Niu, and Sivan Sabato (eds.), Proceedings of the 39th International Conference on Machine Learning, volume 162 of Proceedings of Machine Learning Research, pp. 8056 8071. PMLR, 17 23 Jul 2022. URL https://proceedings.mlr.press/v162/guo22c.html. Jamie Hayes, Borja Balle, and Saeed Mahloujifar. Bounding training data reconstruction in DP-SGD. In Thirty-seventh Conference on Neural Information Processing Systems, 2023. URL https://openreview. net/forum?id=7LZ4t Zr Ylx. Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler, and Sepp Hochreiter. Gans trained by a two time-scale update rule converge to a local nash equilibrium. In Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS 17, pp. 6629 6640, Red Hook, NY, USA, 2017. Curran Associates Inc. ISBN 9781510860964. Jonathan Ho, Ajay Jain, and Pieter Abbeel. Denoising diffusion probabilistic models. In Proceedings of the 34th International Conference on Neural Information Processing Systems, NIPS 20, Red Hook, NY, USA, 2020. Curran Associates Inc. ISBN 9781713829546. Tao Huang, Jiayang Meng, Hong Chen, Guolong Zheng, Xu Yang, Xun Yi, and Hua Wang. Gradient-guided conditional diffusion models for private image reconstruction: Analyzing adversarial impacts of differential privacy and denoising, 2024. URL https://arxiv.org/abs/2411.03053. Jeremy Irvin, Pranav Rajpurkar, Michael Ko, Yifan Yu, Silviana Ciurea-Ilcus, Chris Chute, Henrik Marklund, Behzad Haghgoo, Robyn Ball, Katie Shpanskaya, Jayne Seekins, David A. Mong, Safwan S. Halabi, Jesse K. Sandberg, Ricky Jones, David B. Larson, Curtis P. Langlotz, Bhavik N. Patel, Matthew P. Lungren, and Andrew Y. Ng. Chexpert: A large chest radiograph dataset with uncertainty labels and expert Published in Transactions on Machine Learning Research (03/2025) comparison. Proceedings of the AAAI Conference on Artificial Intelligence, 33(01):590 597, Jul. 2019. doi: 10.1609/aaai.v33i01.3301590. URL https://ojs.aaai.org/index.php/AAAI/article/view/3834. Georgios Kaissis, Jamie Hayes, Alexander Ziller, and Daniel Rueckert. Bounding data reconstruction attacks with the hypothesis testing interpretation of differential privacy, 2023a. Georgios Kaissis, Alexander Ziller, Stefan Kolek, Anneliese Riess, and Daniel Rueckert. Optimal privacy guarantees for a relaxed threat model: Addressing sub-optimal adversaries in differentially private machine learning. In Thirty-seventh Conference on Neural Information Processing Systems, 2023b. URL https: //openreview.net/forum?id=BRSg Vw85Mc. Tero Karras, Timo Aila, Samuli Laine, and Jaakko Lehtinen. Progressive growing of GANs for improved quality, stability, and variation. In International Conference on Learning Representations, 2018. URL https://openreview.net/forum?id=Hk99z Ce Ab. Bahjat Kawar, Michael Elad, Stefano Ermon, and Jiaming Song. Denoising diffusion restoration models. In Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho (eds.), Advances in Neural Information Processing Systems, 2022. URL https://openreview.net/forum?id=kx Xvopt9p WK. Helena Klause, Alexander Ziller, Daniel Rueckert, Kerstin Hammernik, and Georgios Kaissis. Differentially private training of residual networks with scale normalisation, 2022. Alex Krizhevsky and Geoffrey Hinton. Learning multiple layers of features from tiny images. Technical Report 0, University of Toronto, Toronto, Ontario, 2009. URL https://www.cs.toronto.edu/~kriz/ learning-features-2009-TR.pdf. Xin Li, Yulin Ren, Xin Jin, Cuiling Lan, Xingrui Wang, Wenjun Zeng, Xinchao Wang, and Zhibo Chen. Diffusion models for image restoration and enhancement a comprehensive survey, 2023. Xuan Liu, Siqi Cai, Qihua Zhou, Song Guo, Ruibin Li, and Kaiwei Lin. Mjolnir: Breaking the shield of perturbation-protected gradients via adaptive diffusion, 2025. URL https://arxiv.org/abs/2407. 05285. Johan Lokna, Anouk Paradis, Dimitar I Dimitrov, and Martin Vechev. Group and attack: Auditing differential privacy. In Proceedings of the 2023 ACM SIGSAC Conference on Computer and Communications Security, 2023. Ilya Mironov. Rényi differential privacy. In 2017 IEEE 30th Computer Security Foundations Symposium (CSF), pp. 263 275, 2017. doi: 10.1109/CSF.2017.11. Milad Nasr, Reza Shokri, and Amir houmansadr. Improving deep learning with differential privacy using gradient encoding and denoising, 2020. Milad Nasr, Shuang Song, Abhradeep Thakurta, Nicolas Papernot, and Nicholas Carlini. Adversary instantiation: Lower bounds for differentially private machine learning. In 2021 IEEE Symposium on Security and Privacy (SP), pp. 866 882. IEEE Computer Society, 2021. doi: 10.1109/SP40001.2021.00069. URL https://doi.ieeecomputersociety.org/10.1109/SP40001.2021.00069. Milad Nasr, Jamie Hayes, Thomas Steinke, Borja Balle, Florian Tramèr, Matthew Jagielski, Nicholas Carlini, and Andreas Terzis. Tight auditing of differentially private machine learning. In Proceedings of the 32nd USENIX Conference on Security Symposium, SEC 23, USA, 2023. USENIX Association. ISBN 978-1939133-37-3. Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas Köpf, Edward Yang, Zach De Vito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. Py Torch: an imperative style, high-performance deep learning library. Curran Associates Inc., Red Hook, NY, USA, 2019. Published in Transactions on Machine Learning Research (03/2025) Naama Pearl, Yaron Brodsky, Dana Berman, Assaf Zomet, Alex Rav Acha, Daniel Cohen-Or, and Dani Lischinski. Svnr: Spatially-variant noise removal with denoising diffusion, 2023. Natalia Ponomareva, Hussein Hazimeh, Alex Kurakin, Zheng Xu, Carson Denison, H. Brendan Mc Mahan, Sergei Vassilvitskii, Steve Chien, and Abhradeep Guha Thakurta. How to dp-fy ml: A practical guide to machine learning with differential privacy. Journal of Artificial Intelligence Research, 77:1113 1201, July 2023. ISSN 1076-9757. doi: 10.1613/jair.1.14649. URL http://dx.doi.org/10.1613/jair.1.14649. Olaf Ronneberger, Philipp Fischer, and Thomas Brox. U-net: Convolutional networks for biomedical image segmentation. In Nassir Navab, Joachim Hornegger, William M. Wells, and Alejandro F. Frangi (eds.), Medical Image Computing and Computer-Assisted Intervention MICCAI 2015, pp. 234 241, Cham, 2015. Springer International Publishing. ISBN 978-3-319-24574-4. Tim Salimans, Andrej Karpathy, Xi Chen, and Diederik P. Kingma. Pixel CNN++: Improving the pixel CNN with discretized logistic mixture likelihood and other modifications. In International Conference on Learning Representations, 2017. URL https://openreview.net/forum?id=BJr FC6ceg. Karen Simonyan and Andrew Zisserman. Very deep convolutional networks for large-scale image recognition. In International Conference on Learning Representations, 2015. Jiaming Song, Chenlin Meng, and Stefano Ermon. Denoising diffusion implicit models. In International Conference on Learning Representations, 2021a. URL https://openreview.net/forum?id=St1giar CHLP. Shuang Song, Kamalika Chaudhuri, and Anand D. Sarwate. Stochastic gradient descent with differentially private updates. In 2013 IEEE Global Conference on Signal and Information Processing, pp. 245 248, 2013. doi: 10.1109/Global SIP.2013.6736861. Yang Song, Jascha Sohl-Dickstein, Diederik P Kingma, Abhishek Kumar, Stefano Ermon, and Ben Poole. Score-based generative modeling through stochastic differential equations. In International Conference on Learning Representations, 2021b. URL https://openreview.net/forum?id=Px TIG12RRHS. Thomas Steinke, Milad Nasr, and Matthew Jagielski. Privacy auditing with one (1) training run. In Thirtyseventh Conference on Neural Information Processing Systems, 2023. URL https://openreview.net/ forum?id=f38EY21l Bw. Thomas Steinke, Milad Nasr, Arun Ganesh, Borja Balle, Christopher A. Choquette-Choo, Matthew Jagielski, Jamie Hayes, Abhradeep Guha Thakurta, Adam Smith, and Andreas Terzis. The last iterate advantage: Empirical auditing and principled heuristic analysis of differentially private sgd, 2024. URL https: //arxiv.org/abs/2410.06186. Pierre Stock, Igor Shilov, Ilya Mironov, and Alexandre Sablayrolles. Defending against reconstruction attacks with rényi differential privacy, 2022. Stéfan van der Walt, Johannes L. Schönberger, Juan Nunez-Iglesias, François Boulogne, Joshua D. Warner, Neil Yager, Emmanuelle Gouillart, Tony Yu, and the scikit-image contributors. scikit-image: image processing in Python. Peer J, 2:e453, 6 2014. ISSN 2167-8359. doi: 10.7717/peerj.453. URL https: //doi.org/10.7717/peerj.453. Patrick von Platen, Suraj Patil, Anton Lozhkov, Pedro Cuenca, Nathan Lambert, Kashif Rasul, Mishig Davaadorj, and Thomas Wolf. Diffusers: State-of-the-art diffusion models. https://github.com/ huggingface/diffusers, 2022. Zhou Wang, A.C. Bovik, H.R. Sheikh, and E.P. Simoncelli. Image quality assessment: from error visibility to structural similarity. IEEE Transactions on Image Processing, 13(4):600 612, 2004. doi: 10.1109/TIP. 2003.819861. Tiange Xiang, Mahmut Yurt, Ali B Syed, Kawin Setsompop, and Akshay Chaudhari. DDM$ 2$: Selfsupervised diffusion MRI denoising with generative diffusion models. In The Eleventh International Conference on Learning Representations, 2023a. URL https://openreview.net/forum?id=0vqjc50Hfc C. Published in Transactions on Machine Learning Research (03/2025) Tianqi Xiang, Wenjun Yue, Yiqun Lin, Jiewen Yang, Zhenkun Wang, and Xiaomeng Li. Diffcmr: Fast cardiac mri reconstruction with diffusion probabilistic models, 2023b. Yutong Xie, Minne Yuan, Bin Dong, and Quanzheng Li. Diffusion model for generative image denoising, 2023. Cheng Yang, Lijing Liang, and Zhixun Su. Real-world denoising via diffusion model, 2023. Samuel Yeom, Irene Giacomelli, Matt Fredrikson, and Somesh Jha. Privacy risk in machine learning: Analyzing the connection to overfitting. In 2018 IEEE 31st Computer Security Foundations Symposium (CSF), pp. 268 282, 2018. doi: 10.1109/CSF.2018.00027. R. Zhang, P. Isola, A. A. Efros, E. Shechtman, and O. Wang. The unreasonable effectiveness of deep features as a perceptual metric. In 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 586 595, Los Alamitos, CA, USA, jun 2018. IEEE Computer Society. doi: 10.1109/CVPR. 2018.00068. URL https://doi.ieeecomputersociety.org/10.1109/CVPR.2018.00068. Xinwei Zhang, Zhiqi Bu, Mingyi Hong, and Meisam Razaviyayn. Doppler: Differentially private optimizers with low-pass filter for privacy noise reduction, 2024. URL https://arxiv.org/abs/2408.13460. Alexander Ziller, Tamara T. Mueller, Simon Stieger, Leonhard F. Feiner, Johannes Brandt, Rickmer Braren, Daniel Rueckert, and Georgios Kaissis. Reconciling privacy and accuracy in ai for medical imaging. Nature Machine Intelligence, 6(7):764 774, 2024a. Alexander Ziller, Anneliese Riess, Kristian Schwethelm, Tamara T. Mueller, Daniel Rueckert, and Georgios Kaissis. Bounding reconstruction attack success of adversaries without data priors, 2024b. Published in Transactions on Machine Learning Research (03/2025) A Approximating the Clipping Factor λ Recall our assumption that the adversary has knowledge about the exact value of the clipping factor λ = max(||x||2/C, 1) in Eq. (7). In this section, we demonstrate the simplicity of yielding a good approximation of λ using trial-and-error. As an example, we take an image from the CIFAR-10 dataset and assume C = 1 (a standard value in DP-SGD practice (Ponomareva et al., 2023)) and µ = C/σ = 30. The example image has a L2-norm of ||x||2 = 27.24, thus, it will be clipped and λ = ||x||2/C = 27.24. The first step of an adversary could be to set a value range for λ. Given the standard range of color values xi [0, 1], the maximum L2-norm of a flattened image x RHW D is HWD, in this case, (||x||2)max = 32 32 3 = 55.43 and, therefore, λ [1, 55.43]. Now, without further assumptions, the adversary can repeat the reconstruction process with different values for λ and select the best result. Figure 8 shows some example results of the trial-and-error approach. Original Clipped λ = 10 λ = 25 λ = 40 ˆσ = 0.03 ˆσ = 0.3 ˆσ = 0.75 ˆσ = 1.2 Figure 8: Reconstruction results for different approximations of λ. The (scaled) perturbed image (top) and the DM s reconstruction (bottom) are shown. Additionally, the change in the standard deviation of the noise ˆσ = Cσλ resulting from rescaling is given. B Experimental Details Models. For the CIFAR-10 and Celeb A-HQ datasets, we utilize the diffusion models and exponential moving average (EMA) checkpoints from Ho et al. (2020). These checkpoints achieve a validation Fréchet Inception Distance (FID) (Heusel et al., 2017) of 3.17 for CIFAR-10, the FID for Celeb A-HQ was not reported. For the Image Net dataset, we employ the unconditional DM of Dhariwal & Nichol (2021), which achieves a validation FID of 12.00. All DMs are based on U-Net architectures (Ronneberger et al., 2015) and Pixel CNN++ (Salimans et al., 2017). Frameworks. We use the Diffusers library (von Platen et al., 2022) (based on Py Torch (Paszke et al., 2019)) to leverage state-of-the-art pre-trained DMs for implementing our reconstruction attack. Compute Reconstruction Performance. To evaluate the reconstruction performance of the considered attacks, we asses the average similarity between the reconstructed test images and their original counterparts. First, we execute the reconstruction attack proposed by Ziller et al. (2024b) on DP-SGD (as described in Sec. 2), obtaining their reconstruction performance. Then, we post-process the noisy images generated by Ziller et al. s attack using our proposed DM method (as described in Sec. 3), representing our attack s reconstruction performance. For the Re Ro lower bound, we implement the prior-aware attack proposed by Hayes et al. (2023) under identical DP-SGD settings and architectures as Ziller et al. (2024b). We compute reconstructions by matching the noisy and clipped gradients from DP-SGD with the clipped gradients derived from a prior set comprising Published in Transactions on Machine Learning Research (03/2025) 256 candidate images using the dot product. The resulting matched images are then considered as reconstructions and used to compute the similarity. Intuitively, successful matching by the Re Ro attack results in perfect reconstructions. C Interpreting µ for Standard DP-SGD Configurations 100 101 102 10 1 Figure 9: Conversion of µ to (ε, δ)-DP for typical DP-SGD configurations for CIFAR-10 (blue ) and Image Net (green ). Additionally, we adopt a worst-case configuration (red ). As discussed in Sec. 4.1, we report results using the signal-to-noise ratio (SNR), defined as µ = C/σ, where C is the clipping parameter and σ is the noise multiplier in DP-SGD. This metric provides a privacy measure that is independent of specific training hyperparameters, enabling generalization of our results across various training configurations. In DP, privacy guarantees are typically expressed in the (ε, δ)-DP framework. For a fully defined DP-SGD configuration (including the number of steps T, sampling probability p, and δ), µ can be converted into an (ε, δ) pair by using advanced privacy accountants (Mironov, 2017; Gopi et al., 2021). To facilitate comparisons with the (ε, δ) notion, we present ε estimates for representative DP-SGD configurations in Fig. 9. These values are illustrative and were not directly used in our experiments, as our setup does not rely on specific configurations. For CIFAR-10 with N = 50, 000 training samples, we adopt the parameters from (Klause et al., 2022), using T = 2, 400 steps, a sampling probability of p = 1, 024/50, 000, and δ = 10 5. For Image Net (N = 1, 281, 167), we employ the fine-tuning setting from (Berrada et al., 2023), with T = 1, 000, p = 262, 144/1, 281, 167, and δ = 8 10 7. Additionally, we consider a worst-case scenario where an adversary modifies the hyperparameters to T = 1 step and a batch size of 1, resulting in p = 1/50, 000. Our results reveal large differences in privacy guarantees across these configurations, underscoring the substantial impact of training hyperparameters on privacy protection. These findings support our choice of using µ as an independent, robust metric in our analysis. D Ablation Experiments In this section, we conduct a series of ablation experiments to assess the performance of our reconstruction attack under different settings and assumptions. Each ablation experiment evaluates the average similarity between the reconstructed images and the original images using CIFAR-10 and LPIPS. In all figures, the dashed line represents the average similarity of test images. 101 102 0 0.2 0.4 0.6 0.8 Figure 10: Image matching success γ, with prior set size of 256 using our reconstructed images (blue ) compared to the Re Ro lower (red ) and upper bound (green). Privacy Leakage from Re-Identification. In this experiment, we evaluate the capabilities of an adversary using our method for re-identification. For this, we employ the matching strategy introduced by Hayes et al. (2023) (see Sec. 2.1) and match the reconstructed image with the most similar image from a prior set using LPIPS and compute the ratio of correct matches. We compare our matching success with the (0,γ)-Re Ro bound. The results in Fig. 10 corroborate our expectation that our method achieves lower matching success than the Re Ro bound. This difference can be attributed to our attack solely assuming a general data prior, while Re Ro assumes access to the full underlying dataset. Consequently, our method relies on the DM-based reconstruction of the image, which may drop some information in the generation process. However, despite this limitation, our matching perfor- Published in Transactions on Machine Learning Research (03/2025) mance is notably close to the Re Ro bound, indicating our attack s ability to recover unique features even under strong perturbations. Once again, we observe that µ 5 serves as a threshold beyond which our attack cannot recover any unique features. 101 102 0 0.2 0.4 0.6 Figure 11: DDIM (blue ) and DDPM (red ) generation performance. Comparison between DDIM and DDPM Generation. Recall that we employ the deterministic generation process of DDIMs to enforce data consistency. In this experiment, we evaluate the effect of this design choice on reconstruction performance by comparing DDIM generation with the probabilistic generation process of DDPMs, which is usually used in implementations of DMs. The results in Fig. 11 show improved performance across various privacy levels with DDIM sampling, suggesting that DDIMs exhibit greater consistency and remain closer to the original image throughout the generation process. 101 102 0 0.2 0.4 0.6 Figure 12: Performance under known noise variance (blue ) and noise variance estimation (red ). Unknown Noise Variance. In our main experiments, we assume that the adversary knows the variance of the noise in the privatized image. However, this assumption may not always hold true in practical scenarios. In this experiment, we assess the impact of unknown noise variance on our reconstruction performance. For this, we approximate the noise variance using the waveletbased implementation in scikit-image (van der Walt et al., 2014) (restoration.estimate_sigma), which is described in Section 4.2 of (Donoho & Johnstone, 1994). The results in Fig. 12 show only a slight decrease in reconstruction performance, indicating that the noise variance can be accurately estimated and that our attack is robust against estimation errors. 101 102 0 0.2 0.4 0.6 0.81 Figure 13: Performance of traditional denoising methods based on wavelet transformation (green ) and BM3D (red ) compared to our DM method (blue ). Denoising without Learned Data Priors. In this experiment, we assess the effectiveness of structural image priors that only capture patterns inherent in images, and do not approximate a specific data distribution. For this, we employ traditional denoising methods based on wavelet transformation (Chang et al., 2000) and BM3D (Dabov et al., 2007) and compare their performance with our DM method approximating the underlying data distribution. The results in Fig. 13 show the limitations of traditional denoising methods when confronted with large noise perturbations. Specifically, the results reveal a large performance difference across all privacy levels. Published in Transactions on Machine Learning Research (03/2025) E Additional Reconstruction Results Figure 14: Reconstruction results of our DM attack with respect to µ = C/σ for CIFAR-10 (top), Celeb A-HQ (middle), and Image Net (bottom). Published in Transactions on Machine Learning Research (03/2025) Orig. Noisy Rec. 1 Rec. 2 Rec. 3 Rec. 4 Rec. 5 Figure 15: DDPM reconstruction results with respect to µ = C/σ for a Celeb A-HQ image. For each µvalue, the original image, the noisy image, and five reconstructions from the noisy image are shown. We observe that lower µ (SNR) lead to larger deviations between generations. The adversary is interested in the features that stay consistent across generations, as these likely originate from the original image. For example, hair color stays consistent until µ = 20, and gender can be inferred until µ = 5. This shows that the visual insights from reconstructions of DMs are not only valuable for data owners who can compare the reconstruction with the original image, but also for adversaries who only have access to the noisy image and the ability to compare different generations.