# drag_data_reconstruction_attack_using_guided_diffusion__871ed3cb.pdf DRAG: Data Reconstruction Attack using Guided Diffusion Wa-Kin Lei 1 Jun-Cheng Chen 2 Shang-Tse Chen 1 Abstract With the rise of large foundation models, split inference (SI) has emerged as a popular computational paradigm for deploying models across lightweight edge devices and cloud servers, addressing data privacy and computational cost concerns. However, most existing data reconstruction attacks have focused on smaller CNN classification models, leaving the privacy risks of foundation models in SI settings largely unexplored. To address this gap, we propose a novel data reconstruction attack based on guided diffusion, which leverages the rich prior knowledge embedded in a latent diffusion model (LDM) pre-trained on a large-scale dataset. Our method performs iterative reconstruction on the LDM s learned image prior, effectively generating high-fidelity images resembling the original data from their intermediate representations (IR). Extensive experiments demonstrate that our approach significantly outperforms state-of-the-art methods, both qualitatively and quantitatively, in reconstructing data from deeplayer IRs of the vision foundation model. The results highlight the urgent need for more robust privacy protection mechanisms for large models in SI scenarios. Code is available at: https: //github.com/ntuaislab/DRAG 1. Introduction The rapid development of deep learning has revolutionized various aspects of daily life from AI assistants to autonomous vehicles. However, the substantial computational resources required by these emerging models often hinder their deployment on edge devices. Therefore, offloading intensive computation to cloud servers has become a popular alternative. Following this paradigm, split inference (SI) (Kang et al., 2017) has emerged as one of the most 1National Taiwan University 2Research Center for Information Technology Innovation, Academia Sinica. Correspondence to: Shang-Tse Chen . Proceedings of the 42 nd International Conference on Machine Learning, Vancouver, Canada. PMLR 267, 2025. Copyright 2025 by the author(s). promising solutions, as it balances computational and privacy concerns. This approach enables efficient utilization of cloud resources, reduces the computational burden on local devices, and facilitates the integration of complex models into everyday technologies by partitioning neural network computations between edge devices and cloud servers, with data processed locally before being sent to the server. Despite its advantages, recent studies (He et al., 2019; Dong et al., 2022; Li et al., 2023; Xu et al., 2024; Sa et al., 2024) have uncovered significant privacy risks associated with SI, particularly in the form of data reconstruction attacks (DRA). In DRA, adversaries attempt to reconstruct clients input data by exploiting the exchanged intermediate representations (IR) between clients and servers, posing serious threats that break users privacy. While these risks are established, the growing adoption of more powerful models, such as Vision Transformers (Vi T) (Dosovitskiy et al., 2021), raises concerns about the effectiveness of existing defenses. Vi Ts have demonstrated superior performance across various vision tasks and are widely used in modern applications. Nevertheless, the privacy implications of deploying these models in SI settings remain underexplored. In this paper, we address this gap by investigating privacy leaks in vision transformers in the context of SI. We propose a novel attack based on guided diffusion that effectively utilizes the prior knowledge captured by large latent diffusion models (LDM) (Rombach et al., 2022) pre-trained on largescale datasets (e.g., Stable Diffusion) to reconstruct input data from deep-layer IR. Leveraging this prior knowledge, we successfully invert IR back to the original input data across various natural image datasets, revealing a critical privacy vulnerability in the SI framework. Additionally, we evaluate our attack on models equipped with existing defenses (Singh et al., 2021; Vepakomma et al., 2020) and show that input data can still be successfully reconstructed from deep-layer IR despite the defenses. Our key contributions are summarized as follows: We propose DRAG, a novel attack that exploits the prior knowledge captured by LDMs to reconstruct input data from deep-layer IR. Our attack can reconstruct high-quality images from DRAG: Data Reconstruction Attack using Guided Diffusion widely used vision foundation models, specifically CLIP (Radford et al., 2021) and DINOv2 (Oquab et al., 2024), demonstrating that the privacy threat exists even in widely used general-purpose vision encoders. We explore defense strategies tailored for vision transformers to mitigate the threat of privacy leakage. 2. Related Work 2.1. Split Inference Split inference (SI) (Kang et al., 2017) is a method aimed at speeding up inference and/or reducing power consumption in endpoint devices while ensuring data privacy. It has been widely studied in various applications, including computer vision tasks such as image classification, detection, and segmentation, as well as natural language processing tasks such as language understanding (Matsubara et al., 2022). In recent years, SI has also attracted attention for its role in generative AI, including LLMs and text-to-image generation (Ohta & Nishio, 2023). Unlike the traditional cloud-based inference approaches that require transmitting raw data to servers, SI preserves privacy by sending only transformed, non-trivially interpretable IR to the cloud. Specifically, the model f is partitioned into two parts: the client model fc : X H that maps input data x from input space X to IR space H, and the server model fs : H Y that maps IR to output space Y. The client model is deployed on the edge device, while the server model operates in the cloud. During inference, the private data x is first processed at the edge by fc, producing the smashed data h = fc(x ). This IR is then transmitted to the cloud, where the server completes the computation by inferring y = fs(h ). This approach, which provides a certain degree of privacy preservation for users, also leverages the abundant computational resources of cloud servers to accelerate inference, making it a feasible solution for applications requiring both privacy and low-latency predictions. 2.2. Data Reconstruction Attack (DRA) In the context of SI, an adversary may extract private information by reconstructing the original input x from h , as illustrated in Figure 1. Following He et al. (2019), we categorize existing DRAs based on the adversary s knowledge of the client model as follows: White-box attacks assume complete knowledge of the architecture and parameters of the client model. This assumption has become increasingly reasonable with the rise of large models leveraging frozen, publicly available vision encoders (Liu et al., 2023; Chen et al., 2023). The early work of He et al. (2019) framed reconstruction as regularized Maximum Likelihood Estimation (r MLE), which optimizes Figure 1: Privacy threats in split inference. a candidate input to match the target IR, with Total Variation (Rudin et al., 1992) serving as an image prior. Singh et al. (2021) improved reconstruction quality by adding a deep image prior (Ulyanov et al., 2018) in their Likelihood Maximization (LM) method. More recently, Li et al. (2023) introduced GAN-based Latent Space Search (GLASS), which constrains reconstructions using Style GAN2 (Karras et al., 2020b). This method yields high-fidelity images that successfully evade several defenses (He et al., 2019; Singh et al., 2021; Titcombe et al., 2021; Mireshghallah et al., 2020; Li et al., 2021; Osia et al., 2020). Black-box attacks require only query access to fc, typically utilizing inverse networks trained on input-output pairs (He et al., 2019). Recent work has enhanced this approach by incorporating diffusion models (Chen et al., 2024), which offer better reconstruction quality. Query-free attacks operate without access to fc. Instead, they use a collection of IRs to construct surrogate models f c that approximate fc, followed by applying reconstruction techniques to these surrogates. Beyond He et al. (2019), previous works (Pasquini et al., 2021; Erdo gan et al., 2022; Gao & Zhang, 2023; Xu et al., 2024) assume the adversary participates in the model training process to enhance the effectiveness of subsequent reconstruction attacks. Existing evaluations primarily focused on CNN architectures like Res Net18 (He et al., 2016). However, Vi Ts process images fundamentally differently through patch tokenization and attention mechanisms, and their vulnerability to reconstruction attacks remains largely unexplored. Our analysis in Section 5.5 reveals that Vi Ts exhibit token order invariance, a feature absent in CNNs, which significantly affects attack effectiveness. 2.3. Diffusion Models In recent years, diffusion models (Ho et al., 2020; Song et al., 2020) have demonstrated remarkable capabilities in generating realistic images by iteratively refining noise into coherent visuals. Conditional generation has been signif- DRAG: Data Reconstruction Attack using Guided Diffusion icantly advanced through methods such as classifier guidance (Dhariwal & Nichol, 2021) and classifier-free guidance (Ho & Salimans, 2021), which incorporate class information to guide the generation process. Specifically, classifier guidance (Dhariwal & Nichol, 2021) estimates the class probability pt(c|x) while maintaining a fixed unconditional distribution pt(x), enabling conditional generation through the Bayes rule: pt(x|c) pt(x) pt(c|x). Universal Guidance Diffusion (UGD) (Bansal et al., 2024) further expands the scope of conditional generation by integrating various neural networks to incorporate diverse conditions, such as object bounding boxes, segmentation masks, face identities, and stylistic attributes. While classifier guidance seeks to adapt a fixed unconditional model pt(x) by developing an appropriate conditional distribution pt(c|x), another line of research (Chung et al., 2023; Yang et al., 2024) assumes the availability of a predefined conditional distribution p0(c|x). This assumption positions diffusion models as promising tools for addressing a range of conditional generation tasks. Moreover, LDMs (Rombach et al., 2022) have further advanced diffusion models by enabling the generation of diverse, high-resolution, high-quality images within a latent space, thus improving computational efficiency. Subsequent work (Ramesh et al., 2022; Zhang et al., 2023) has extended control mechanisms within the latent diffusion framework, allowing more precise and hierarchical image manipulation. 3. Methodology 3.1. Threat Model Following previous work (He et al., 2019; Singh et al., 2021; Dong et al., 2022; Li et al., 2023), we consider an honestbut-curious server in SI that seeks to reconstruct x from h . Given the prevalence of frozen foundation models in downstream applications, we assume white-box access to the client model fc, providing the adversary with complete architectural and parameter knowledge. Under this threat model, we formulate the problem using two complementary approaches: (1) optimization-based and (2) learning-based. Optimization Based. The adversary aims to find input data x whose corresponding IR closely matches h by solving the following optimization problem: x = arg min x X d H(fc(x), h ) + λRI(x), (1) where d H measures the distance between the IRs, RI represents the regularization term ensuring perceptually realistic, and λ 0 controls the weight of regularization. Ultimately, the adversary s goal is to achieve x x . Learning Based. The adversary infers x by training an inverse network f 1 c : H X, where the network is trained using paired data (x, fc(x)) from a public dataset Dpublic: f 1 c = arg min f 1 c E ||f 1 c (fc(x)) x||2 . (2) The following subsections detail our reconstruction approach, as illustrated in Figure 2. 3.2. Data Reconstruction Attack using Guided Diffusion In this section, we propose leveraging guided diffusion for data reconstruction attacks. Pre-trained diffusion models serve as image priors RI, constraining reconstructions to the natural image manifold. Based on a noise predictor ϵθ and its corresponding noise scheduler {(αt, σt)}T t=1, unconditional DDIM sampling (Song et al., 2020) transitions sample x from timestep t to t 1 as follows: xt 1 = αt 1 xt 1 αt ϵθ(xt) αt 1 αt 1 σ2 t ϵθ(xt) + σt ϵt, (3) where ϵt N(0, I) denotes Gaussian noise sampled from a standard normal distribution. For conditional image generation, classifier guidance (Dhariwal & Nichol, 2021) modifies the predicted noise by adding the gradient of a classifier s log-probability to the predicted noise: ϵθ(xt, c) = ϵθ(xt) w 1 αt xt log pθ(c|xt), (4) where c denotes the condition, and w controls the strength of guidance. Extending classifier guidance, UGD (Bansal et al., 2024) replaces the classification loss with a general, differentiable loss L(xt, c). For data reconstruction, we define the objective function as: L(xt, c) = d H(fc(xt), h ). (5) Notably, directly feeding xt to fc for guidance computation is unreliable, as fc is typically trained only on clean images. To address this, Chung et al. (2023) and Bansal et al. (2024) estimate the guidance using the unique posterior mean ˆx0, a single-step denoised sample: L(ˆx0, c) = d H(fc(ˆx0), h ), (6) where ˆx0 is computed using Tweedie s formula with a single forward pass through the noise predictor ϵθ: ˆx0 = xt 1 αt ϵθ(xt) αt . (7) DRAG: Data Reconstruction Attack using Guided Diffusion Figure 2: Illustration of DRAG (Data Reconstruction Attack using Guided Diffusion). The diffusion model serves as an image prior, constraining the solution space of x in optimization-based DRAs. Moreover, DRAG can be extended by incorporating the Inverse Network (He et al., 2019): we first obtain an initial estimate xcoarse = f 1 c (h ), then refine it by diffusion-denoising process. We refer to this enhanced variant as DRAG++. Diffusion with Spherical Gaussian constraint (DSG) (Yang et al., 2024) reduces the required denoising steps while enhancing generation quality. DSG aligns ϵt with the guidance gt = xt L(ˆx0, c) through: ϵt r UNIT((1 w) σt ϵt + wr UNIT(gt)). (8) where r = nσt with n being the dimension of xt, and operator UNIT() normalizes vectors to unit norm. We adopt Equation (8) for our guided diffusion process. To enhance reconstruction quality, we refine the guidance gt using gradient clipping and historical guidance vector with optimizers such as Adam (Kingma & Ba, 2015). Since fc is typically non-convex, a single guidance step is insufficient for high-quality reconstruction. Therefore, we employ selfrecurrence (Bansal et al., 2024), projecting xt 1 back to xt via small-step DDPM diffusion (Ho et al., 2020) and iterating this denoising-diffusion cycle k times: αt/αt 1 xt 1 + p 1 αt/αt 1 ϵ. (9) We refer to this approach as DRAG Data Reconstruction Attack using Guided Diffusion. Algorithm 1 outlines our proposed attack in detail. Note that the problem differs from previous works (Chung et al., 2023; Yang et al., 2024) in three key aspects, introducing additional challenges: (1) client models fc are typically assumed to be non-convex, (2) defensive mechanisms may be deployed, forcing attackers to operate in adversarial settings, (3) clients can embed randomness into fc, further complicating the problem, as detailed in Section 5.5. 3.3. Extending DRAG with Inverse Networks To enhance the performance and efficiency of DRAG, we integrate an auxiliary Inverse Network (He et al., 2019). This network accelerates DRAG by providing a coarse reconstruction from the target IR, serving as a more effective initialization. We train this network on an auxiliary dataset of publicly available data Dpublic to project the IR back to image space and obtain xcoarse. This coarse reconstruction is further projected onto an editable manifold by adding random noise at a small timestep t = s T: xt = αt xcoarse+ 1 αt ϵ, where ϵ N(0, I), (10) where the strength parameter s satisfies 0 s 1. We refer to this enhanced method as DRAG++. 3.4. Adapting to Latent Diffusion Models As LDMs perform diffusion and denoising processing in the latent space Z instead of the pixel space X, we adapt our approach when leveraging LDMs as the image prior by replacing the noisy sample xt with the noisy latent zt. The mapping between X and Z is provided by the corresponding latent autoencoder E and D. All other components of the method remain unchanged. 4. Experimental Setups In this section, we first introduce the details of our experimental settings, including datasets, target models, compared methods, evaluation metrics, and attacker models. DRAG: Data Reconstruction Attack using Guided Diffusion Algorithm 1 DRAG // Noise ϵ is sampled from N(0, I) for every usage s { .m = 0, .v = 0, .i = 0 } z T N(0, I) for t = T to 1 do for n = 1 to k do ˆx0 D(TWEEDIESESTIMATION(zt)) gt zt(d H(fc(ˆx0), h ) + λℓ2Rℓ2(ˆx0)) gt CLIPNORM(gt, cmax) gt, s STATEUPDATE(gt, s) zt 1 GUIDEDSAMPLING(zt, gt) zt p αt/αt 1 zt 1 + p 1 αt/αt 1 ϵ end for end for return D(z0) // Refine gt via momentum such as Adam function STATEUPDATE(gt, s) s.m β1 s.m + (1 β1) gt s.v β2 s.v + (1 β2) g2 t s.i s.i + 1 ˆm, ˆv s.m/ 1 βi 1 , s.v/ 1 βi 2 return gt, s end function function GUIDEDSAMPLING(zt, gt) ϵt r UNIT((1 w) σt ϵ + wr UNIT(gt)) return DDIM(zt, ϵθ(zt), ϵt) end function 4.1. Datasets To evaluate our proposed methods, we sample 10 images from the official validation splits of each dataset: (1) MSCOCO (Lin et al., 2014), (2) FFHQ (Karras et al., 2019), and (3) Image Net-1K (Deng et al., 2009), constructing a collection of diverse natural images. All images are centercropped and resized to 224 224. We use Image Net-1K image classification as the primary task to quantitatively assess model utility. To simulate realistic conditions where the client and adversary have non-overlapping datasets, we randomly split the official training split of Image Net-1K into two distinct, equal-sized and non-overlapping subsets: a private portion Dprivate and a public portion Dpublic. The target model f is fine-tuned exclusively on Dprivate, while the inverse network f 1 c , as proposed in Section 3.3, is trained solely on Dpublic. 4.2. Target Models We aim to reconstruct data from the widely used CLIP-Vi TB/16, CLIP-RN50 (Radford et al., 2021), and DINOv2-Base (Oquab et al., 2024) vision encoder, known for its strong adaptability and zero-shot capabilities across vision tasks (Rao et al., 2022; Mokady et al., 2021). The reconstruction is performed after layers l = {0, 3, 6, 9, 12} for CLIP-Vi TB/16 and DINOv2-Base, while for CLIP-RN50, the attack is conducted after blocks l = {1, 2, 3, 4, 5}. We evaluated the attack in three configurations: (1) the model is frozen at the pre-trained checkpoint, (2) protected by DISCO (Singh et al., 2021), and (3) protected by No Peek (Vepakomma et al., 2020). The details of these two defenses can be found in Appendix D. These two defenses, highlighted in Li et al. (2023), have demonstrated superior privacy-preserving performance compared to other defenses. 4.3. Baseline and Metrics We compare our method with previous optimization-based DRAs: r MLE (He et al., 2019), LM (Singh et al., 2021) and GLASS (Li et al., 2023). Implementation details are provided in Appendix C. To quantify privacy leakage across these attacks, we evaluated reconstruction performance using three complementary metrics: MS-SSIM (Wang et al., 2003), LPIPS (Zhang et al., 2018), and image similarity measured by DINO Vi T-S/16 (Caron et al., 2021). These metrics capture both low-level fidelity and high-level semantic similarity, better reflecting privacy risks by aligning with human perceptual judgment compared to pixel-wise measures such as MSE or PSNR (Hor e & Ziou, 2010). 4.4. Attacker Models We use Stable Diffusion v1.5 (SDv1.5) as our image prior. GLASS (Li et al., 2023) uses domain-specific priors: Style GAN2-ADA (FFHQ) (Karras et al., 2020a) for facial images and Style GAN-XL (Image Net-1K) (Sauer et al., 2022) for others, assuming knowledge of the target distribution. This inherently advantages GLASS by matching priors to true data distribution, while our diffusion-based approach uses a single, domain-agnostic prior. For the architecture of the inverse network, we adopt the decoder architecture from He et al. (2022) to reconstruct images from the tokenized representations produced by Vi Ts. During training, we randomly replace rmask = 25% of patch tokens with [MASK] token to improve its generalization. 4.5. Distance Metrics and Regularization We use token-wise cosine distance as the distance metric d H for the data reconstruction process in Vi T-family models: d H(h1, h2) = 1 1 h1[i, :], h2[i, :] ||h1[i, :]|| ||h2[i, :]||, (11) where N denotes the number of tokens. For CLIP-RN50, the MSE loss is adopted. To prevent the latent zt from DRAG: Data Reconstruction Attack using Guided Diffusion Table 1: Reconstruction performance of optimization-based attacks across target models and split points without defenses. Bold indicates the best scores, while underlined indicate the second-best. CLIP-Vi T-B/16 DINOv2-Base CLIP-RN50 Method Split Point MS-SSIM LPIPS DINO MS-SSIM LPIPS DINO Split Point MS-SSIM LPIPS DINO 0.8888 0.0709 0.9712 0.9162 0.0504 0.9630 0.6832 0.1543 0.9111 LM 0.9638 0.0237 0.9903 0.9698 0.0227 0.9850 0.9769 0.0150 0.9919 GLASS 0.8700 0.1466 0.8289 0.9147 0.1076 0.8369 0.9052 0.0785 0.8485 DRAG 0.9588 0.0489 0.9259 0.9567 0.0440 0.9284 0.9316 0.0476 0.9454 0.8612 0.0914 0.9706 0.8371 0.1641 0.9302 0.5741 0.2618 0.9053 LM 0.9742 0.0206 0.9923 0.9682 0.0250 0.9890 0.9313 0.0382 0.9868 GLASS 0.9091 0.0556 0.9623 0.9340 0.0488 0.9580 0.7829 0.2185 0.7563 DRAG 0.9500 0.0372 0.9715 0.9570 0.0323 0.9728 0.9151 0.0604 0.9539 0.6888 0.2608 0.8875 0.6566 0.2562 0.9111 0.6745 0.2233 0.8986 LM 0.8604 0.0784 0.9734 0.7334 0.1676 0.9768 0.9028 0.0596 0.9769 GLASS 0.7113 0.1352 0.9326 0.7444 0.1495 0.9333 0.6785 0.2405 0.7877 DRAG 0.9028 0.0465 0.9784 0.9196 0.0455 0.9782 0.9118 0.0528 0.9662 0.4957 0.5131 0.7159 0.5855 0.4374 0.7663 0.4888 0.4198 0.7776 LM 0.6681 0.2138 0.9063 0.5281 0.3839 0.9555 0.5855 0.2576 0.9012 GLASS 0.3852 0.4310 0.6740 0.5404 0.3230 0.8467 0.4872 0.3568 0.7315 DRAG 0.7974 0.0967 0.9652 0.8483 0.0820 0.9719 0.7896 0.0898 0.9622 0.3884 0.5900 0.6524 0.4375 0.5680 0.6079 0.3980 0.5006 0.6739 LM 0.2560 0.6024 0.4248 0.3640 0.6190 0.8878 0.4432 0.3409 0.7614 GLASS 0.2396 0.5790 0.4553 0.4456 0.4076 0.7297 0.2917 0.4223 0.6811 DRAG 0.6735 0.1857 0.9331 0.7581 0.1443 0.9463 0.5206 0.2231 0.9001 deviating too far from the domain of the noise predictor ϵθ, we introduce ℓ2 regularization on ˆx0 to the objective (Equation (6)) to ensure it remains within the range [ 1, 1] during the reconstruction process: Rℓ2(ˆx0) = λℓ2 CHW ˆx2 0. (12) 5. Experimental Results In this section, we first compare the reconstruction performance of DRAG with prior methods on frozen, pre-trained foundation models. Next, we evaluate DRAG++, which integrates an auxiliary inverse network (IN) to improve reconstruction performance. Subsequently, we evaluate the generalization ability of DRAG on out-of-distribution data. We then examine its robustness against two defenses, DISCO and No Peek. Finally, we analyze the reconstruction performance when applying the token shuffling defense, a mechanism intrinsic to Vi Ts. Complete experimental results can be found in Appendix A, and further experiments on the effects of key hyperparameters appear in Appendix B. 5.1. Reconstruction from Frozen Foundation Models Table 1 presents quantitative reconstruction results, while Figure 3 visualizes results on CLIP-Vi T-B/16, highlighting qualitative differences. Our approach consistently outperforms prior methods in reconstructing data from deeper layers. Although r MLE and LM achieve competitive performance at shallow split points, their reconstruction quality degrades beyond layer 9 and layer 12, respectively. In contrast, DRAG outperforms the others at deeper split points across the evaluated metrics. Compared to GLASS, which also utilizes a data-driven image prior and achieves strong performance on the FFHQ dataset, our method attains comparable performance on FFHQ and generalizes robustly to MSCOCO and Image Net. Despite employing a GAN trained on Image Net, GLASS reconstructs images from Image Net with evident artifacts at deep split points. In contrast, DRAG maintains high-fidelity outputs free from such distortions, as shown in Figure 3. 5.2. Enhancing DRAG with Inverse Networks We evaluate DRAG++ on CLIP-Vi T-B/16 by comparing its performance to the inverse network (IN) and the original DRAG method. As shown in Table 2, DRAG++ achieves higher reconstruction performance in terms of LPIPS and DINO at split points beyond layer 3. While IN attains higher MS-SSIM, DRAG++ demonstrates notable improvements in LPIPS and DINO, with only a slight sacrifice in MS-SSIM compared to IN. This result suggests that the combined approach enhances perceptual quality in reconstruction. 5.3. Distribution Shift We evaluate the generalization capability of DRAG and DRAG++ through two experiments in out-of-distribution settings. In the first experiment, we reconstruct aerial images from the UCMerced Land Use dataset (Yang & DRAG: Data Reconstruction Attack using Guided Diffusion Figure 3: Reconstruction results for CLIP-Vi T-B/16 across split points without defenses. Table 2: Reconstruction performance comparison between inverse network alone and DRAG++ on CLIP-Vi T-B/16. Split Point Method MS-SSIM LPIPS DINO IN 0.9907 0.0112 0.9937 DRAG 0.9588 0.0489 0.9259 DRAG++ 0.9608 0.0485 0.9234 IN 0.9763 0.0351 0.9458 DRAG 0.9500 0.0372 0.9715 DRAG++ 0.9504 0.0349 0.9719 IN 0.9120 0.1799 0.7869 DRAG 0.9028 0.0465 0.9784 DRAG++ 0.9093 0.0457 0.9785 IN 0.8188 0.2993 0.7130 DRAG 0.7974 0.0967 0.9652 DRAG++ 0.8224 0.0875 0.9700 IN 0.7443 0.3618 0.6660 DRAG 0.6735 0.1857 0.9331 DRAG++ 0.7257 0.1685 0.9492 Newsam, 2010), using f 1 c trained on Image Net-1K. As shown in Figure 4, DRAG and DRAG++ successfully reconstruct these images, demonstrating their robustness across different domains. In the second experiment, we replace the diffusion model with a domain-specific model trained on the LSUN-bedroom dataset (Yu et al., 2015). We compare this configuration while keeping the other attack settings unchanged. The results shown in Table 3 indicate that DRAG achieves the best performance at layer 12 and ranks as the runner-up at layers 3, 6, and 9, further demonstrating its robustness even in an out-of-distribution setting. 5.4. Reconstruction from Privacy-Guarded Models We evaluate attacks on models protected by DISCO and No Peek defenses with varying hyperparameter settings that determine privacy protection strength, as detailed in Table 4. Table 3: Reconstruction performance using diffusion model trained on out-of-distribution data for CLIP-Vi T-B/16. Split Point Method MS-SSIM LPIPS DINO r MLE 0.8888 0.0709 0.9712 LM 0.9638 0.0237 0.9903 GLASS 0.8700 0.1466 0.8289 DRAG 0.9692 0.0275 0.9591 r MLE 0.8612 0.0914 0.9706 LM 0.9742 0.0206 0.9923 GLASS 0.9091 0.0556 0.9623 DRAG 0.9488 0.0427 0.9722 r MLE 0.6888 0.2608 0.8875 LM 0.8604 0.0784 0.9734 GLASS 0.7113 0.1352 0.9326 DRAG 0.8244 0.1574 0.9438 r MLE 0.4957 0.5131 0.7159 LM 0.6681 0.2138 0.9063 GLASS 0.3852 0.4310 0.6740 DRAG 0.5379 0.3937 0.8154 r MLE 0.3884 0.5900 0.6524 LM 0.2560 0.6024 0.4248 GLASS 0.2396 0.5790 0.4553 DRAG 0.3958 0.4938 0.7245 Based on diffusion model trained on LSUN bedroom dataset. Since Config-III and Config-VI represent the most challenging settings for adversaries, we highlight the results under these configurations in Table 5 and Figure 5. Our attack proves effective against weaker defense configurations, with comprehensive results across all configurations presented in Appendix A.2 and Figure 9. For DISCO, we assume the adversary lacks knowledge of the pruning model or the mask applied to h , as pruning is a dynamic, auxiliary component that can be decomposed from fc. The pruned channels mislead the reconstruction DRAG: Data Reconstruction Attack using Guided Diffusion (a) Reconstruction from DRAG (b) Reconstruction from DRAG++ Figure 4: Reconstruction results on out-of-distribution aerial images from the UCMerced Land Use dataset. Table 4: Defense parameters for DISCO and No Peek. Defense Config Defense Parameters I ρD = 0.95, rp = 0.1 II ρD = 0.75, rp = 0.2 III ρD = 0.95, rp = 0.5 IV ρN = 1.0 V ρN = 3.0 VI ρN = 5.0 process, leading to failures in the most challenging setting, Config-III. However, since pruned channels generally have smaller absolute values than unpruned counterparts, this discrepancy can be exploited by adaptive attacks. Specifically, the adversary can filter out channels with low mean absolute values when calculating d H, thereby mitigating the misleading influence of pruned channels. Our experimental results also show that DRAG performs the best after the adaptive attack filters out the pruned channels. For No Peek, we observe that d H is significantly lower than in unprotected models during the optimization process, consistent with findings in Li et al. (2023). Despite this, DRAG still reconstructs the target images with higher fidelity compared to the previous works. 5.5. Token Shuffling (and Dropping) Defense Vi Ts naturally exhibit an adaptive computation capability, enabling reduced inference time by discarding redundant tokens in the intermediate layers. Previous work (Yin et al., 2022) investigates strategies for dropping tokens in intermediate layers. From a privacy protection perspective, shuffling (and dropping) patch tokens hinders data reconstruction, as the distance metrics d H is sensitive to the token order. For tasks where token order is irrelevant (e.g., classification), shuffling patch tokens offers a straightforward defense against DRA. Additionally, this method is easy to implement, as it only requires memory copying. We evaluate the privacy risk against the token-shuffling (and dropping) defense. To simulate a token-dropping scenario, Table 5: Performance under DISCO (Config-III) and No Peek (Config-VI) for CLIP-Vi T-B/16 at layer 12. Method MS-SSIM LPIPS DINO r MLE 0.1686 0.8128 0.1129 LM 0.1638 0.7079 0.1654 GLASS 0.1479 0.6225 0.3878 DRAG 0.0788 0.7449 0.3201 Config-III - w/ adaptive attacks r MLE 0.2101 0.7822 0.2072 LM 0.1696 0.6953 0.1818 GLASS 0.2402 0.5401 0.4862 DRAG 0.4557 0.3778 0.7557 r MLE 0.2270 0.7048 0.3747 LM 0.1783 0.7917 0.2099 GLASS 0.1950 0.7248 0.3141 DRAG 0.4469 0.3836 0.8096 we propose the following protocol: the client shuffles the patch tokens and randomly drops rdrop N patch tokens before sending them to the server, where rdrop is the proportion of the dropped tokens. Specifically, shuffling is applied after the positional embedding layer, while the [CLS] token remains fixed in position. As noted in Darcet et al. (2024), tokens retain information about their original positions, which can be inferred by a classifier. Based on this observation, we train a 2-layer MLP classifier to predict the probability that a token h[i, :] was originally at position arg max pθ(h[i, :]). Once trained, the classifier enables us to reorder the patch tokens by solving an assignment problem, maximizing the joint probability using the Hungarian algorithm (Kuhn, 1955). We present the reconstruction results under three configurations: (1) patch tokens are randomly permuted, and the adversary is unaware of the permutation; (2) the adversary uses a token position classifier to reorder the tokens; and DRAG: Data Reconstruction Attack using Guided Diffusion (a) Config-III (b) Config-III (adaptive attack) (c) Config-VI Figure 5: Reconstruction results for CLIP-Vi T-B/16 at layer 12 with DISCO (Config-III) and No Peek (Config-VI). (a) rdrop = 0.0 (w/o reordering tokens) (b) rdrop = 0.0 (w/ reordering tokens) (c) rdrop = 0.5 (w/ reordering tokens) Figure 6: Reconstruction results for CLIP-Vi T-B/16 at layer 12 with token shuffling defense. (3) the client drops 50% of the patch tokens before sending them to the server, leaving the adversary to infer their correct placement. Experiments are conducted on CLIP-Vi TB/16 splitting at layer 12. The token position classification model achieves 21.40% top-1 accuracy in predicting token positions, with an average ℓ1 distance of 2.34 from the correct position on Image Net-1K. As shown in Figure 6, both r MLE and LM fail to reconstruct the target images, whereas the normal configuration succeeds. For GLASS and DRAG, reconstruction performance is degraded in the shuffled scenarios, but some reconstructed images still retain key features of the original inputs. 6. Conclusion This work reveals significant privacy risks in SI with large vision foundation models like CLIP-Vi T and DINOv2, extending beyond previous attacks on CNN models like Res Net18. We propose a novel data reconstruction attack leveraging LDMs pre-trained on large-scale datasets. Our method generates high-fidelity images from IR and outperforms stateof-the-art approaches in reconstructing data from deep-layer IR. These findings highlight the need for stronger defenses to protect privacy when deploying transformer-based models in SI settings. Acknowledgements This research is supported by National Science and Technology Council, Taiwan under grant numbers 113-2222-E002-004-MY3, 113-2634-F-002-007, and 113-2634-F-002001-MBK; by the Featured Area Research Center Program within the Higher Education Sprout Project of the Ministry of Education (grant 113L900903); and by Academia Sinica (grant AS-CDA-110-M09). We also thank the National Center for High-performance Computing of National Applied Research Laboratories in Taiwan for providing computational and storage resources. Impact Statement This work investigates the privacy risks of large vision foundation models in the split inference framework. As these models are widely adopted in downstream tasks, privacy concerns become crucial. Our proposed DRAG method shows that, unlike smaller CNN models studied previously, large vision models are also vulnerable to embedding reconstruction attacks, posing risks to applications like domestic robots and cyber-physical systems. These findings provide key insights for ML developers and users, urging a re-evaluation of privacy risks and the development of more robust, privacy-preserving architectures. DRAG: Data Reconstruction Attack using Guided Diffusion Bansal, A., Chu, H.-M., Schwarzschild, A., Sengupta, S., Goldblum, M., Geiping, J., and Goldstein, T. Universal guidance for diffusion models. In International Conference on Learning Representations (ICLR), 2024. Caron, M., Touvron, H., Misra, I., J egou, H., Mairal, J., Bojanowski, P., and Joulin, A. Emerging properties in self-supervised vision transformers. In International Conference on Computer Vision (ICCV), 2021. Chen, D., Li, S., Zhang, Y., Li, C., Kundu, S., and Beerel, P. A. DIA: Diffusion based inverse network attack on collaborative inference. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2024. Chen, J., Zhu, D., Shen, X., Li, X., Liu, Z., Zhang, P., Krishnamoorthi, R., Chandra, V., Xiong, Y., and Elhoseiny, M. Mini GPT-v2: large language model as a unified interface for vision-language multi-task learning. ar Xiv preprint ar Xiv:2310.09478, 2023. Chung, H., Kim, J., Mccann, M. T., Klasky, M. L., and Ye, J. C. Diffusion posterior sampling for general noisy inverse problems. In International Conference on Learning Representation (ICLR), 2023. Darcet, T., Oquab, M., Mairal, J., and Bojanowski, P. Vision transformers need registers. In International Conference on Learning Representations (ICLR), 2024. Deng, J., Dong, W., Socher, R., Li, L.-J., Li, K., and Fei-Fei, L. Image Net: A large-scale hierarchical image database. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2009. Dhariwal, P. and Nichol, A. Diffusion models beat GANs on image synthesis. In Advances in Neural Information Processing Systems (Neur IPS), 2021. Dong, X., Yin, H., Alvarez, J. M., Kautz, J., Molchanov, P., and Kung, H. Privacy vulnerability of split computing to data-free model inversion attacks. In 33rd British Machine Vision Conference (BMVC), 2022. Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., Uszkoreit, J., and Houlsby, N. An image is worth 16x16 words: Transformers for image recognition at scale. International Conference on Learning Representations (ICLR), 2021. Erdo gan, E., K upc u, A., and C ic ek, A. E. Unsplit: Dataoblivious model inversion, model stealing, and label inference attacks against split learning. In 21st Workshop on Privacy in the Electronic Society, 2022. Gao, X. and Zhang, L. PCAT: Functionality and data stealing from split learning by pseudo-client attack. In 32nd USENIX Security Symposium (USENIX Security 23), 2023. Hatamizadeh, A., Yin, H., Roth, H. R., Li, W., Kautz, J., Xu, D., and Molchanov, P. Grad Vi T: Gradient inversion of vision transformers. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022. He, K., Zhang, X., Ren, S., and Sun, J. Deep residual learning for image recognition. In IEEE conference on Computer Vision and Pattern Recognition (CVPR), 2016. He, K., Chen, X., Xie, S., Li, Y., Doll ar, P., and Girshick, R. Masked autoencoders are scalable vision learners. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022. He, Z., Zhang, T., and Lee, R. B. Model inversion attacks against collaborative inference. In Annual Computer Security Applications Conference (ACSAC), 2019. Ho, J. and Salimans, T. Classifier-free diffusion guidance. In Neur IPS Workshop on Deep Generative Models and Downstream Applications, 2021. Ho, J., Jain, A., and Abbeel, P. Denoising diffusion probabilistic models. Advances in Neural Information Processing Systems (Neur IPS), 2020. Hor e, A. and Ziou, D. Image quality metrics: PSNR vs. SSIM. In International Conference on Pattern Recognition (ICPR), 2010. Kang, Y., Hauswald, J., Gao, C., Rovinski, A., Mudge, T., Mars, J., and Tang, L. Neurosurgeon: Collaborative intelligence between the cloud and mobile edge. In International Conference on Architectural Support for Programming Languages and Operating Systems (ASPLOS), 2017. Karras, T., Laine, S., and Aila, T. A style-based generator architecture for generative adversarial networks. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019. Karras, T., Aittala, M., Hellsten, J., Laine, S., Lehtinen, J., and Aila, T. Training generative adversarial networks with limited data. In Advances in Neural Information Processing Systems (Neur IPS), 2020a. Karras, T., Laine, S., Aittala, M., Hellsten, J., Lehtinen, J., and Aila, T. Analyzing and improving the image quality of Style GAN. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020b. DRAG: Data Reconstruction Attack using Guided Diffusion Kingma, D. P. and Ba, J. Adam: A method for stochastic optimization. In International Conference on Learning Representations (ICLR), 2015. Kuhn, H. W. The hungarian method for the assignment problem. Naval research logistics quarterly, 2(1-2), 1955. Li, A., Guo, J., Yang, H., Salim, F. D., and Chen, Y. Deepobfuscator: Obfuscating intermediate representations with privacy-preserving adversarial learning on smartphones. In International Conference on Internet-of-Things Design and Implementation (Io TDI), 2021. Li, Z., Yang, M., Liu, Y., Wang, J., Hu, H., Yi, W., and Xu, X. GAN you see me? Enhanced data reconstruction attacks against split inference. In Advances in Neural Information Processing Systems (Neur IPS), 2023. Lin, T.-Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Doll ar, P., and Zitnick, C. L. Microsoft COCO: Common objects in context. In European Conference on Computer Vision (ECCV), 2014. Liu, H., Li, C., Wu, Q., and Lee, Y. J. Visual instruction tuning. In Advances in Neural Information Processing Systems (Neur IPS), 2023. Matsubara, Y., Levorato, M., and Restuccia, F. Split computing and early exiting for deep learning applications: Survey and research challenges. ACM Computing Surveys, 2022. Mireshghallah, F., Taram, M., Ramrakhyani, P., Jalali, A., Tullsen, D., and Esmaeilzadeh, H. Shredder: Learning noise distributions to protect inference privacy. In International Conference on Architectural Support for Programming Languages and Operating Systems (ASPLOS), 2020. Mokady, R., Hertz, A., and Bermano, A. H. Clipcap: Clip prefix for image captioning. ar Xiv preprint ar Xiv:2111.09734, 2021. Ohta, S. and Nishio, T. Λ-split: A privacy-preserving split computing framework for cloud-powered generative ai. ar Xiv preprint ar Xiv:2310.14651, 2023. Oquab, M., Darcet, T., Moutakanni, T., Vo, H. V., Szafraniec, M., Khalidov, V., Fernandez, P., HAZIZA, D., Massa, F., El-Nouby, A., et al. Dinov2: Learning robust visual features without supervision. Transactions on Machine Learning Research (TMLR), 2024. Osia, S. A., Shamsabadi, A. S., Sajadmanesh, S., Taheri, A., Katevas, K., Rabiee, H. R., Lane, N. D., and Haddadi, H. A hybrid deep learning architecture for privacypreserving mobile analytics. IEEE Internet of Things Journal (Io T-J), 2020. Pasquini, D., Ateniese, G., and Bernaschi, M. Unleashing the tiger: Inference attacks on split learning. In ACM SIGSAC Conference on Computer and Communications Security, 2021. Radford, A., Kim, J. W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., Sastry, G., Askell, A., Mishkin, P., Clark, J., et al. Learning transferable visual models from natural language supervision. In International Conference on Machine Learning (ICML), 2021. Ramesh, A., Dhariwal, P., Nichol, A., Chu, C., and Chen, M. Hierarchical text-conditional image generation with CLIP latents. ar Xiv preprint ar Xiv:2204.06125, 2022. Rao, Y., Zhao, W., Chen, G., Tang, Y., Zhu, Z., Huang, G., Zhou, J., and Lu, J. Dense CLIP: Language-guided dense prediction with context-aware prompting. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022. Rombach, R., Blattmann, A., Lorenz, D., Esser, P., and Ommer, B. High-resolution image synthesis with latent diffusion models. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022. Rudin, L. I., Osher, S., and Fatemi, E. Nonlinear total variation based noise removal algorithms. Physica D: Nonlinear Phenomena, 1992. Sa, C.-C., Cheng, L.-C., Chung, H.-H., Chiu, T.-C., Wang, C.-Y., Pang, A.-C., and Chen, S.-T. Ensuring bidirectional privacy on wireless split inference systems. IEEE Wireless Communications, 2024. Sauer, A., Schwarz, K., and Geiger, A. Style GAN-XL: Scaling Style GAN to large diverse datasets. In ACM SIGGRAPH conference proceedings, 2022. Singh, A., Chopra, A., Sharma, V., Garza, E., Zhang, E., Vepakomma, P., and Raskar, R. DISCO: Dynamic and invariant sensitive channel obfuscation for deep neural networks. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021. Song, J., Meng, C., and Ermon, S. Denoising diffusion implicit models. In International Conference on Learning Representations (ICLR), 2020. Titcombe, T., Hall, A. J., Papadopoulos, P., and Romanini, D. Practical defences against model inversion attacks for split neural networks. ar Xiv preprint ar Xiv:2104.05743, 2021. Ulyanov, D., Vedaldi, A., and Lempitsky, V. Deep image prior. In IEEE/CVF conference on Computer Vision and Pattern Recognition (CVPR), 2018. DRAG: Data Reconstruction Attack using Guided Diffusion Vepakomma, P., Singh, A., Gupta, O., and Raskar, R. No Peek: Information leakage reduction to share activations in distributed deep learning. In International Conference on Data Mining Workshops (ICDMW), 2020. Wang, Z., Simoncelli, E. P., and Bovik, A. C. Multiscale structural similarity for image quality assessment. In Asilomar Conference on Signals, Systems & Computers, 2003. Xu, X., Yang, M., Yi, W., Li, Z., Wang, J., Hu, H., Zhuang, Y., and Liu, Y. A stealthy wrongdoer: Feature-oriented reconstruction attack against split learning. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2024. Yang, L., Ding, S., Cai, Y., Yu, J., Wang, J., and Shi, Y. Guidance with spherical gaussian constraint for conditional diffusion. In International Conference on Machine Learning (ICML), 2024. Yang, Y. and Newsam, S. Bag-of-visual-words and spatial extensions for land-use classification. In SIGSPATIAL international conference on advances in geographic information systems, 2010. Yin, H., Vahdat, A., Alvarez, J. M., Mallya, A., Kautz, J., and Molchanov, P. Ada Vi T: Adaptive tokens for efficient vision transformer. In IEEE/CVF conference on Computer Vision and Pattern Recognition (CVPR), 2022. Yu, F., Zhang, Y., Song, S., Seff, A., and Xiao, J. LSUN: Construction of a large-scale image dataset using deep learning with humans in the loop. ar Xiv preprint ar Xiv:1506.03365, 2015. Zhang, L., Rao, A., and Agrawala, M. Adding conditional control to text-to-image diffusion models. In IEEE/CVF International Conference on Computer Vision (ICCV), 2023. Zhang, R., Isola, P., Efros, A. A., Shechtman, E., and Wang, O. The unreasonable effectiveness of deep features as a perceptual metric. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2018. DRAG: Data Reconstruction Attack using Guided Diffusion A.1. Reconstruction from Frozen Foundation Models Besides Table 1, we provide a figure in Figure 7 that visualizes the performance of each attack. Additionally, we present the attack success rate (ASR), which is defined as the proportion of images for which the reconstruction metrics exceed a specified threshold, as shown in Figure 7. Figure 8 illustrates the performance of attacks on the same target images at different split points for CLIP-Vi T-B/16. An image from each dataset was chosen for evaluation. 0 3 6 9 12 Split Point MS-SSIM ( ) 0 3 6 9 12 Split Point 1 - LPIPS ( ) 0 3 6 9 12 Split Point (a) CLIP-Vi T-B/16 (mean) 0 3 6 9 12 Split Point MS-SSIM ( ) 0 3 6 9 12 Split Point 1 - LPIPS ( ) 0 3 6 9 12 Split Point (b) DINOv2-Base (mean) 0 1 2 3 4 Split Point MS-SSIM ( ) 0 1 2 3 4 Split Point 1 - LPIPS ( ) 0 1 2 3 4 Split Point (c) CLIP-RN50 (mean) r MLE LM GLASS DRAG 0.0 0.2 0.4 0.6 0.8 1.0 MS-SSIM ( ) 0.0 0.2 0.4 0.6 0.8 1.0 1 - LPIPS ( ) 0.0 0.2 0.4 0.6 0.8 1.0 DINO ( ) (d) CLIP-Vi T-B/16 (ASR, Layer 12) 0.0 0.2 0.4 0.6 0.8 1.0 MS-SSIM ( ) 0.0 0.2 0.4 0.6 0.8 1.0 1 - LPIPS ( ) 0.0 0.2 0.4 0.6 0.8 1.0 DINO ( ) (e) DINOv2-Base (ASR, Layer 12) 0.0 0.2 0.4 0.6 0.8 1.0 MS-SSIM ( ) 0.0 0.2 0.4 0.6 0.8 1.0 1 - LPIPS ( ) 0.0 0.2 0.4 0.6 0.8 1.0 DINO ( ) (f) CLIP-RN50 (ASR, Block 5) r MLE LM GLASS DRAG Figure 7: Reconstruction quality metric and attack success rate (ASR) across target models without defenses. Figure 8: Reconstruction results for CLIP-Vi T-B/16 across split points without defenses. The same images are used as evaluation targets to compare the performance of previous attacks. DRAG: Data Reconstruction Attack using Guided Diffusion A.2. Reconstruction from Privacy Guarded Models Table 6 provides the model utility, which is measured by classification accuracy on Image Net-1K. The complete quantitative results for Section 5.4, conducted on CLIP-Vi T-B/16 and DINOv2-Base, as presented in Table 7 and Figure 9. Table 6: Model utility under privacy defenses, measured by Image Net-1K classification accuracy. Defense Config Parameters CLIP-Vi T-B/16 DINOv2-Base w/o defense 79.87% 83.81% I ρD = 0.95, rp = 0.1 79.20% 83.51% II ρD = 0.75, rp = 0.2 79.02% 83.46% III ρD = 0.95, rp = 0.5 78.04% 82.85% IV ρN = 1.0 79.28% 83.43% V ρN = 3.0 78.67% 83.39% VI ρN = 5.0 77.88% 83.22% (a) Config-I (b) Config-II (c) Config-III (d) Config-I (adaptive attacks) (e) Config-II (adaptive attacks) (f) Config-III (adaptive attacks) (g) Config-IV (h) Config-V (i) Config-VI Figure 9: Reconstruction results for CLIP-Vi T-B/16 under various defense configurations. DRAG: Data Reconstruction Attack using Guided Diffusion Table 7: Performance of the optimization-based attack against various defenses, split at layer 12. CLIP-Vi T-B/16 DINOv2-Base Config Method MS-SSIM LPIPS DINO MS-SSIM LPIPS DINO r MLE 0.1957 0.7881 0.1768 0.3840 0.5993 0.4710 LM 0.1749 0.6690 0.1965 0.3953 0.5112 0.7031 GLASS 0.2138 0.5568 0.4783 0.3881 0.4222 0.6667 DRAG 0.3309 0.4793 0.6497 0.7392 0.1602 0.9370 r MLE 0.1706 0.8145 0.1101 0.3840 0.5993 0.4710 LM 0.1800 0.6838 0.1872 0.3953 0.5112 0.7031 GLASS 0.2050 0.5627 0.4433 0.3881 0.4222 0.6667 DRAG 0.2704 0.5264 0.6015 0.7392 0.1602 0.9370 r MLE 0.1686 0.8128 0.1129 0.2469 0.6571 0.2463 LM 0.1638 0.7079 0.1654 0.3198 0.5770 0.4781 GLASS 0.1479 0.6225 0.3878 0.2699 0.4884 0.5310 DRAG 0.0788 0.7449 0.3201 0.5703 0.2880 0.8432 Config-I w/ adaptive r MLE 0.4062 0.5088 0.6880 0.4481 0.5796 0.5597 LM 0.1765 0.6996 0.1962 0.4336 0.4916 0.8097 GLASS 0.2321 0.5443 0.4770 0.4526 0.3849 0.6962 DRAG 0.5930 0.2489 0.8772 0.7918 0.1218 0.9573 Config-II w/ adaptive r MLE 0.3837 0.5498 0.6229 0.3851 0.6160 0.4820 LM 0.1833 0.6816 0.2056 0.4155 0.4939 0.8124 GLASS 0.2157 0.5630 0.4528 0.4242 0.3967 0.6954 DRAG 0.6076 0.2347 0.8830 0.7783 0.1329 0.9489 Config-III w/ adaptive r MLE 0.2101 0.7822 0.2072 0.3993 0.5794 0.5063 LM 0.1696 0.6953 0.1818 0.4068 0.5104 0.7899 GLASS 0.2402 0.5401 0.4862 0.4052 0.3971 0.6792 DRAG 0.4557 0.3778 0.7557 0.7551 0.1467 0.9415 r MLE 0.2799 0.6604 0.4729 0.2428 0.6822 0.2546 LM 0.1834 0.7159 0.2271 0.2852 0.6868 0.7270 GLASS 0.1837 0.6707 0.3381 0.3499 0.4664 0.6782 DRAG 0.5526 0.2925 0.8778 0.6252 0.2216 0.9243 r MLE 0.2348 0.6939 0.3863 0.2306 0.6653 0.3183 LM 0.1782 0.7377 0.2114 0.2546 0.7107 0.6272 GLASS 0.1715 0.6864 0.3043 0.3710 0.4447 0.6446 DRAG 0.4889 0.3369 0.8381 0.5235 0.2790 0.9106 r MLE 0.2270 0.7048 0.3747 0.2218 0.6821 0.3136 LM 0.1783 0.7917 0.2099 0.2410 0.7140 0.5797 GLASS 0.1950 0.7248 0.3141 0.3620 0.4339 0.6516 DRAG 0.4469 0.3836 0.8096 0.5010 0.2977 0.9093 DRAG: Data Reconstruction Attack using Guided Diffusion A.3. Execution Time The execution times for each attack algorithm are provided in Table 8. Table 8: Execution times for optimization-based attacks at deepest split points. Method # of Iterations CLIP-Vi T-B/16 DINOv2-Base CLIP-RN50 r MLE 20,000 6 min 42 s 7 min 13 s 6 min 04 s LM 20,000 24 min 20 s 24 min 32 s 22 min 27 s GLASS (Style GAN2-ADA-FFHQ) 20,000 21 min 28 s 22 min 02 s 19 min 45 s GLASS (Style GAN-XL-Image Net-1K) 20,000 1 hr 37 min 08 s 1 hr 37 min 02 s 1 hr 34 min 34 s DRAG (SDv1.5) 4,000 32 min 52 s 33 min 04 s 32 min 40 s B. Extended Experiments B.1. Scaling Reconstruction Schedule Increasing T or k improves reconstruction performance by allowing more refinement steps, especially in the deeper layer. However, it also raises computational overhead. Figure 10a and Figure 10b visualizes the attack performance for different values of T and k. Since T = 250 and k = 16 provides a satisfactory balance between performance and efficiency, we adopt it as the default hyperparameter setting. B.2. Guidance Strength w The guidance strength w balances feature matching and image prior during sampling. Increasing w enhances guidance by focusing more on minimizing the distance d H, but this may compromise the realistic property RI as defined in Equation (1). Conversely, an excessively low w results in an unsuccessful attack due to insufficient guidance. Figure 10c presents the relationship between w and reconstruction performance. We also tested Equation (8) with linear interpolation (lerp) and spherical linear interpolation (slerp), but found no significant performance differences. B.3. Importance of the Optimizer Figure 10d compares the performance of attacks on IR from layer 12 with and without the Adam optimizer. The figure demonstrates that smoothing gradients with non-convex optimization techniques significantly enhances attack performance, especially at the deeper layer. C. Baseline Attacks r MLE. (He et al., 2019) first proposed an optimization-based reconstruction attack that reconstructs x by optimizing a zero-initialized x to minimize d H(fc(x), h ). To improve reconstruction quality, the method incorporates Total Variation regularization (Rudin et al., 1992) as an image prior: x = arg min x X d H(fc(x), h ) + λTVRTV(x). (13) LM. (Singh et al., 2021) enhances reconstruction quality by applying a deep image prior (Ulyanov et al., 2018) to regularize x. Rather than updating x, they re-parameterize it as the output of a CNN-based image synthesis model fθ(ϵ), where the fixed input ϵ N(0, I) remain constant during optimization, while the model parameters θ are optimized: x = arg min θ d H(fc(fθ(ϵ)), h ) + λTVRTV(x). (14) Since Vi Ts divide images into non-overlapping patches, directly optimizing x often yields visible artifacts at patch boundaries. To alleviate this, we add the patch-smoothness prior Rpatch from Hatamizadeh et al. (2022) to both the r MLE and LM DRAG: Data Reconstruction Attack using Guided Diffusion objectives when reconstructing images from Vi Ts: Rpatch(x) = k=1 x[:, P k, :, :] x[:, P k 1, :, :] 2 + k=1 x[:, :, P k, :] x[:, :, P k 1, :] 2, (15) where P denotes the patch size of the Vi T model. GLASS. (Li et al., 2023) proposed a scenario in which the adversary has knowledge of the data distribution and access to auxiliary data for training a Style GAN. Instead of directly updating x, the adversary updates the latent code z Z or the style code w+ W+ to improve the quality of the generated image. In the first stage, the latent code z is randomly 0 3 6 9 12 Split Point MS-SSIM ( ) 0 3 6 9 12 Split Point 1 - LPIPS ( ) 0 3 6 9 12 Split Point T=50 T=100 T=125 T=250 T=500 (a) Effect of denoising steps T. 0 3 6 9 12 Split Point MS-SSIM ( ) 0 3 6 9 12 Split Point 1 - LPIPS ( ) 0 3 6 9 12 Split Point k=4 k=8 k=16 k=24 k=32 (b) Effect of internal iterations k. 0.10 0.15 0.20 0.25 0.30 Guidance strength w MS-SSIM ( ) 0.10 0.15 0.20 0.25 0.30 Guidance strength w 1 - LPIPS ( ) 0.10 0.15 0.20 0.25 0.30 Guidance strength w (c) Effect of guidance strength w at layer 12. 0 3 6 9 12 Split Point MS-SSIM ( ) 0 3 6 9 12 Split Point 1 - LPIPS ( ) 0 3 6 9 12 Split Point w/ Adam w/o Adam (d) Performance comparison of DRAG with and without the Adam optimizer. Figure 10: Hyperparameter sensitivity analysis for DRAG on CLIP-Vi T-B/16. DRAG: Data Reconstruction Attack using Guided Diffusion initialized from a standard normal distribution, z N(0, I), and is then update as follow: z = arg min z Z d H(fc(G(fmap(z))), h ) + λTVRTV(x) + λKLRKL(z), (16) where RKL is the Kullback-Leibler divergence that regularize the latent code z. The updated latent code z is then transformed to the style code w+ = fmap(z ), which is subsequently updated for fine-grained reconstruction: w+ = arg min w+ d H(fc(G(w+)), h ) + λTVRTV(x). (17) For class-conditioned GANs, we further optimize a zero-initialized class logits vector l R1000, which is normalized via the softmax function to produce class probabilities p = softmax(l) during the forward pass w+ = fmap(z , Ep). Then p are multiplied by the class embeddings E Rd 1000 to compute the class-specific latent code Ep. D. Defensive Algorithms D.1. Privacy Leakage Mitigation Methods DISCO. (Singh et al., 2021) introduces a method to mitigate privacy leakage by pruning a subset of the IRs channels before transmitting them to the server. Specifically, the pruning operation h = fp(h, rp) is performed using an auxiliary channel pruning module fp, where the pruning ratio rp controls the proportion of pruned channels. This ratio can be dynamically adjusted during model inference. The pruning module fp is trained in a min-max framework, where fp minimizes privacy leakage by maximizing the reconstruction loss, and the inverse network f 1 c minimizes the reconstruction loss: Lutil = E[ℓutil(fs(h ), y)], Lprivacy = E[||f 1 c (h ) x||2], min fp [max f 1 c Lprivacy + ρD min fc,fs Lutil]. (18) No Peek. (Vepakomma et al., 2020) aims to mitigate privacy leakage by training models to minimize the mutual information I(X; H) between the input data X and the intermediate representation H. Since directly calculating I(X; H) is challenging, the authors propose using distance correlation (d Cor) as a surrogate measure: min fc,fs E[ρN d Cor(fc(x), x) + ℓutil(fs(fc(x)), y)]. (19) While Vepakomma et al. (2020) assumes that users pre-train the target models f from scratch using distance correlation loss, our experiments differ by applying the loss during model fine-tuning. This adaptation allows us to leverage pre-existing model knowledge while still addressing privacy concerns. D.2. Implementation of Defense Mechanisms Since the target models are not pre-trained on Image Net-1K, they lack classification heads tailored for the Image Net-1K classification task. While it is possible to directly initialize random classification heads and train with DISCO or No Peek, this approach significantly degrades accuracy. To address this, we prepare target models f through a two-stage process. First, we perform linear probing by freezing the pre-trained backbone and training only the classification head on Dprivate. Then, we fine-tune the entire model using the selected defense mechanism. This strategy helps preserve classification performance during the defensive training phase. To ensure the effectiveness and robustness of DRAG, we adopt an informed defender threat model, assuming the client has full knowledge of DRAG. Under this assumption, defenders can design countermeasures by leveraging insights from the methodology of DRAG. Specifically, for DISCO, we employ an inverse network with the same architecture as described in Section 4.4. For No Peek, we adopt the same distance metric d H as defined in Equation (11), which reflects how an informed defender would calibrate a privacy-preserving model against DRAG. DRAG: Data Reconstruction Attack using Guided Diffusion E. Implementation Details We list the hyperparameters for various optimization-based and learning-based reconstruction attacks in Table 9 and Table 10, respectively. The experiments were conducted on a server equipped with 384 GB RAM, two Intel Xeon Gold 6226R CPUs, and eight NVIDIA RTX A6000 GPUs. The implementation of r MLE (He et al., 2019), LM (Singh et al., 2021), DISCO (Singh et al., 2021) and No Peek (Vepakomma et al., 2020) are adapted from prior works.1 Table 9: Default hyperparameters for the optimization-based reconstruction attacks. r MLE LM GLASS Optimizer Adam (lr = 0.05) Adam (lr = 0.01) Adam (lr = 0.01) Num of iters (n) 20,000 20,000 20,000 Pretrained model - - Style GAN2-ADA (FFHQ) Style GAN-XL (Image Net-1K) λTV 1.5 0.05 0 λpatch 0.001 0.001 0 λℓ2 0 0 0 λKL - - 1.0 DRAG DRAG++ Strength (s) 1.0 0.3 DDIM randomness (η) 1.0 1.0 Guidance strength (w) 0.2 0.2 Max grad norm (cmax) 0.02 0.02 Sampling steps (T) 250 250 Self-recurrence (k) 16 16 λTV 0 0 λpatch 0 0 λℓ2 0.01 0.01 Table 10: Hyperparameters for inverse network training across all experiments. Optimizer Adam (lr = 0.001) LR Scheduler Cosine annealing (linear warm-up in 5000 iterations) Number of iterations 100,000 Batch size 256 Mask ratio (rmask) 0.25 1https://github.com/aidecentralized/Inference Benchmark