# selfsupervised_dataset_distillation_for_transfer_learning__51190e51.pdf Published as a conference paper at ICLR 2024 SELF-SUPERVISED DATASET DISTILLATION FOR TRANSFER LEARNING Dong Bok Lee1 , Seanie Lee1 , Joonho Ko1, Kenji Kawaguchi2, Juho Lee1, Sung Ju Hwang1 KAIST1, National University of Singapore2 {markhi, lsnfamily02, joonho.ko, juholee, sjhwang82}@kaist.ac.kr kenji@comp.nus.edu.sg Dataset distillation aims to optimize a small set so that a model trained on the set achieves performance similar to that of a model trained on the full dataset. While many supervised methods have achieved remarkable success in distilling a large dataset into a small set of representative samples, however, they are not designed to produce a distilled dataset that can be effectively used to facilitate selfsupervised pre-training. To this end, we propose a novel problem of distilling an unlabeled dataset into a set of small synthetic samples for efficient self-supervised learning (SSL). We first prove that a gradient of synthetic samples with respect to a SSL objective in naive bilevel optimization is biased due to the randomness originating from data augmentations or masking for inner optimization. To address this issue, we propose to minimize the mean squared error (MSE) between a model s representations of the synthetic examples and their corresponding learnable target feature representations for the inner objective, which does not introduce any randomness. Our primary motivation is that the model obtained by the proposed inner optimization can mimic the self-supervised target model. To achieve this, we also introduce the MSE between representations of the inner model and the self-supervised target model on the original full dataset for outer optimization. We empirically validate the effectiveness of our method on transfer learning. Our code is available at https://github.com/db-Lee/selfsup dd. 1 INTRODUCTION As a consequence of collecting large-scale datasets and recent advances in parallel data processing, deep models have achieved remarkable success in various machine learning problems. However, some applications such as hyperparameter optimization (Franceschi et al., 2017), continual learning (Lopez-Paz & Ranzato, 2017), or neural architecture search (Liu et al., 2019) require repetitive training processes. In such scenarios, it is prohibitively costly to use all the examples from the huge dataset, which motivates the need to compress the full dataset into a small representative set of examples. Recently, many dataset distillation (or condensation) methods (Wang et al., 2018; Zhao et al., 2021; Zhao & Bilen, 2021; Nguyen et al., 2021a;b; Cazenavette et al., 2022; Zhou et al., 2022; Loo et al., 2022; Zhao & Bilen, 2023) have successfully learned a small number of examples on which we can train a model to achieve performance comparable to the one trained on the full dataset. Despite the recent success of dataset distillation methods, they are not designed to produce a distilled dataset that can be effectively transferred to downstream tasks (Figure 1-(a)). In other words, we may not achieve meaningful performance improvements when pre-training a model on the distilled dataset and fine-tuning it on the target dataset. However, condensing general-purpose datasets into a small set for transfer learning is crucial for some applications. For example, instead of using a large pre-trained model, we may need to search a hardware-specific neural architecture due to constraints on the device (Lee et al., 2021). To evaluate the performance of an architecture during the search process, we repeatedly pre-train a model with the architecture on large unlabeled dataset and fine-tune it on the target training datast, which is time consuming and expensive. If we distill the pre-training dataset into a small dataset at once, we can accelerate the architecture search by pre-training the model on the small set. Another example is target data-free knowledge distillation Equal contribution Published as a conference paper at ICLR 2024 (a) Previous Dataset Distillation SVH N condense (b) Proposed Self-supervised Dataset Distillation Large-scale Unlabeled Dataset Architecture Search Pool feature space target dataset Efficient Pre-training Transfer Learning Figure 1: (a): Previous supervised dataset distillation methods. (b): Our proposed method that distills unlabeled dataset into a small set that can be effectively used for pre-training and transfer to target datasets. (KD) (Lopes et al., 2017; Raikwar & Mishra, 2022), where we aim to distill a teacher into a smaller student without access to the target training data due to data privacy or intellectual property issues. Instead of the target dataset, we can employ a condensed surrogate dataset for KD (Kim et al., 2023). To obtain a small representative set for efficient pre-training, as illustrated in Figure 1-(b), we propose a self-supervised dataset distillation framework which distills an unlabeled dataset into a small set on which the pre-training will be done. Specifically, we formulate the unsupervised dataset distillation as a bilevel optimization problem: optimizing a small representative set such that a model trained on the small set can induce a latent representation space similar to the space of the model trained on the full dataset. Naively, we can replace the objective function of existing bilevel optimization methods for supervised dataset condensation with a SSL objective function that involves some randomness, such as data augmentations or masking inputs. However, we have empirically found that back-propagating through data augmentation or masking is unstable. Moreover, we prove that a gradient of the self-supervised learning (SSL) loss with randomly sampled data augmentations or masking is a biased estimator of the true gradient, explaining the instability. Based on this insight, we propose to use a mean squared error (MSE) for both inner and outer objective functions, which does not introduce any randomness due to the SSL objectives and thus contributes to stable optimization. First, we parameterize a pair of synthetic examples and target representations of the synthetic ones. For inner optimization, we train a model to minimize the MSE between the target representations and a model s representations of the synthetic examples. Then, we evaluate the MSE between the original data representation of the model trained with the inner optimization and that of the model pre-trained on the original full dataset with a SSL objective. Since we do not sample any data augmentations or input masks, we can avoid the biased gradient of the SSL loss. Lastly, similar to Zhou et al. (2022), we simplify the inner optimization to reduce computational cost. We decompose the model into a feature extractor and a linear head, and optimize only the linear head with kernel ridge regression during the inner optimization while freezing the feature extractor. With the linear head and the frozen feature extractor, we compute the meta-gradient of the synthetic examples and target representations with respect to the outer loss, and update them. To this end, we dub our proposed self-supervised dataset distillation method for transfer learning as Kernel Ridge Regression on Self-supervised Target (KRR-ST). We empirically show that our proposed KRR-ST significantly outperforms the supervised dataset distillation methods in transfer learning experiments, where we condense a source dataset, which is either CIFAR100 (Krizhevsky et al., 2009), Tiny Image Net (Le & Yang, 2015), or Image Net (Deng et al., 2009), into a small set, pre-train models with different architectures on the condensed dataset, and fine-tune all the models on target labeled datasets such as CIFAR10, Aircraft (Maji et al., 2013), Stanford Cars (Krause et al., 2013), CUB2011 (Wah et al., 2011), Stanford Dogs (Khosla et al., 2011), and Flowers (Nilsback & Zisserman, 2008). Our contributions are as follows: We propose a new problem of self-supervised dataset distillation for transfer learning, where we distill an unlabeled dataset into a small set, pre-train a model on it, and fine-tune it on target tasks. We have observed training instability when utilizing existing SSL objectives in bilevel optimization for self-supervised dataset distillation. Furthermore, we prove that a gradient of the SSL objectives with data augmentations or masking inputs is a biased estimator of the true gradient. To address the instability, we propose KRR-ST using MSE without any randomness at an inner loop. For the inner loop, we minimize MSE between a model representation of synthetic samples and target representations. For an outer loop, we minimize MSE between the original data representation of the model from inner loop and that of the model pre-trained on the original dataset. Published as a conference paper at ICLR 2024 We extensively validate our proposed method on numerous target datasets and architectures, and show that ours outperforms supervised dataset distillation methods. 2 RELATED WORK Dataset Distillation (or Condensation) To compress a large dataset into a small set, instead of selecting coresets (Borsos et al., 2020; Mirzasoleiman et al., 2020), dataset condensation optimizes a small number of synthetic samples while preserving information of the original dataset to effectively train high-performing deep learning models. Wang et al. (2018) propose the first dataset distillation (DD) method based on bilevel optimization, where the inner optimization is simply approximated by one gradient-step. Instead of one-step approximation, recent works propose a back-propagation through time for full inner optimization steps (Deng & Russakovsky, 2022), or implicit gradient based on implicit function theorem (Loo et al., 2023). This bilevel formulation, however, incurs expensive computational costs and hinders scaling to large datasets. To overcome this, several papers propose surrogate objective alternatives to bilevel optimization. Specifically, DSA (Zhao et al., 2021; Zhao & Bilen, 2021) minimizes the distance between gradients of original and synthetic samples for each training step. MTT (Cazenavette et al., 2022) proposes to match the parameter obtained on real data and the parameter optimized on the synthetic data. DM (Zhao & Bilen, 2023) matches the first moments of the feature distributions induced by the original dataset and synthetic dataset. As another line of work, Kernel Inducing Points (Nguyen et al., 2021a;b) propose DD methods based on kernel ridge regression, which simplifies inner optimization by back-propagating through Neural Tangent Kernel (NTK) (Lee et al., 2019). Due to the expensive cost of computing NTK for neural networks, FRe Po (Zhou et al., 2022) proposes kernel ridge regression on neural features sampled from a pool, and RFAD (Loo et al., 2022) proposes random feature approximation of the neural network Gaussian process. Despite the recent advances in DD methods, none of them have tackled unsupervised/self-supervised DD for transferable learning. Self-Supervised Learning A vast amount of works have proposed self-supervised learning (SSL) methods. A core idea of SSL is that we use large-scale unlabeled data to train a model to learn meaningful representation space that can be effectively transferred to downstream tasks. We introduce a few representative works. Sim CLR (Chen et al., 2020a) is one of the representative contrastive learning methods. It maximizes the similarity between two different augmentations of the same input while minimizing the similarity between two randomly chosen pairs. MOCO (He et al., 2020) constructs a dynamic dictionary using a moving average encoder and queue, and minimizes contrastive loss with the dictionary. On the other hand, several non-contrastive works achieve remarkable performance. BYOL (Grill et al., 2020) encodes two different views of an input with a student and teacher encoder, respectively, and minimizes the distance between those two representations. Barlow Twins (Zbontar et al., 2021) constructs a correlation matrix between two different views of a batch of samples with an encoder and trains the encoder to enforce the correlation matrix to the identity matrix. Lastly, MAE (He et al., 2022) learns a meaningful image representation space by masking an image and reconstructing the masked input. In this paper, we utilize Barlow Twins as an SSL framework to train our target model. 3.1 PRELIMINARIES Problem Definition Suppose that we are given an unlabeled dataset Xt = [x1 xn] Rn dx where each row xi Rdx is an i.i.d sample. We define the problem of self-supervised dataset distillation as the process of creating a compact synthetic dataset Xs = [ˆx1 ˆxm] Rm dx that preserves most of the information from the unlabeled dataset Xt for pre-training any neural networks, while keeping m n. Thus, after the dataset distillation process, we can transfer knowledge embedded in the large dataset to various tasks using the distilled dataset. Specifically, our final goal is to accelerate the pre-training of a neural network with any architectures by utilizing the distilled dataset Xs in place of the full unlabeled dataset Xt for pre-training. Subsequently, one can evaluate the performance of the neural network by fine-tuning it on various downstream tasks. Bilevel Optimization with SSL Recent success of transfer learning with self-supervised learning (SSL) (Chen et al., 2020a; He et al., 2020; Grill et al., 2020; Zbontar et al., 2021) is deeply rooted in the ability to learn meaningful and task-agnostic latent representation space. Inspired by the SSL, Published as a conference paper at ICLR 2024 we want to find a distilled dataset Xs such that a model ˆgθ : Rdx Rdy, trained on Xs with a SSL objective, achieves low SSL loss on the full dataset Xt. Here, θ denotes the parameter of the neural network ˆgθ. Similar to previous supervised dataset condensation methods (Wang et al., 2018; Zhao et al., 2021; Cazenavette et al., 2022; Deng & Russakovsky, 2022), estimation of Xs can be formulated as a bilevel optimization problem: minimize Xs LSSL(θ (Xs); Xt), where θ (Xs) = arg min θ LSSL(θ; Xs). (1) Here, LSSL(θ; Xs) denotes a SSL loss function with ˆgθ evaluated on the dataset Xs. The bilevel optimization can be solved by iterative gradient-based algorithms. However, it is computationally expensive since computing gradient with respect to Xs requires back-propagating through unrolled computational graphs of inner optimization steps. Furthermore, we empirically find out that backpropagating through data augmentations involved in SSL is unstable and challenging. 3.2 KERNEL RIDGE REGRESSION ON SELF-SUPERVISED TARGET Motivation We theoretically analyze the instability of the bilevel formulation for optimizing a condensed dataset with an SSL objective and motivate our objective function. Define dθ by θ Rdθ. Let us write LSSL(θ; Xs) = Eξ D[ℓξ(θ,Xs)] where ξ D is the random variable corresponding to the data augmentation (or input mask), and ℓξ is SSL loss with the sampled data augmentation (or input mask) ξ. Define ˆθ(Xs) = arg minθ ˆLSSL(θ; Xs) where ˆLSSL(θ; Xs) = 1 r Pr i=1 ℓζi(θ,Xs) and ζi D. In practice, we compute LSSL(ˆθ(Xs);Xt) Xs to update Xs. The use of the empirical estimate ˆLSSL(θ; Xs) in the place of the true SSL loss LSSL(θ; Xs) is justified in standard SSL without bilevel optimization because its gradient is always an unbiased estimator of the true gradient: i.e., Eζ[ ˆ LSSL(θ;Xs) θ ] = LSSL(θ;Xs) θ where ζ = (ζi)r i=1. However, the following theorem shows that this is not the case for bilevel optimization. This explains the empirically observed instability of the SSL loss in bilevel optimization. A proof is deferred to Appendix A. Theorem 1. The derivative LSSL(ˆθ(Xs);Xt) Xs is a biased estimator of LSSL(θ (Xs);Xt) i.e., Eζ[ LSSL(ˆθ(Xs);Xt) Xs ] = LSSL(θ (Xs);Xt) Xs , unless ( LSSL(θ;Xt) θ |θ=θ (Xs)) θ (Xs) Eζ[ LSSL(θ;Xt) θ |θ=ˆθ(Xs)]Eζ[ ˆθ(Xs) (Xs)ij ] + Pdθ k=1 Covζ[ LSSL(θ;Xt) θk |θ=ˆθ(Xs), ˆθ(Xs)k (Xs)ij ] for all (i,j) {1, . . . , m} {1, . . . , dx}. Regression on Self-supervised Target Based on the insight of Theorem 1, we propose to replace the inner objective function with a mean squared error (MSE) by parameterizing and optimizing both synthetic examples Xs and their target representations Ys = [ˆy1 ˆym] Rm dy as: Linner(θ; Xs, Ys) = 1 2 Ys ˆgθ(Xs) 2 F , (2) which avoids the biased gradient of SSL loss due to the absence of random variables ζ corresponding to data augmentation (or input mask) in the MSE. Here, F denotes a Frobenius norm. Similarly, we replace the outer objective with the MSE between the original data representation of the model trained with Linner(θ; Xs, Ys) and that of the target model gϕ : Rdx Rdy trained on the full dataset Xt with the SSL objective as follows: minimize Xs,Ys 1 2 gϕ(Xt) ˆgθ (Xs,Ys)(Xt) 2 F , where θ (Xs, Ys) = arg min θ Linner(θ; Xs, Ys). (3) Note that we first pre-train the target model gϕ on the full dataset Xt with the SSL objective, i.e., ϕ = arg minϕ LSSL(ϕ; Xt). After that, gϕ(Xt) is a fixed target which is considered to be constant during the optimization of Xs and Ys. Here, gϕ(Xt) = [gϕ(x1) gϕ(xn)] Rn dy and ˆgθ (Xs,Ys)(Xt) = [ˆgθ (Xs,Ys)(x1) ˆgθ (Xs,Ys)(xn)] Rn dy. The intuition behind the objective function is as follows. Assuming that a model trained with an SSL objective on a large-scale dataset generalizes to various downstream tasks (Chen et al., 2020b), we aim to ensure that the representation space of the model ˆgθ (Xs,Ys), trained on the condensed data, is similar to that of the self-supervised target model gϕ. Again, one notable advantage of using the MSE is that it removes the need for data augmentations or masking inputs for the evaluation of the inner objective. Furthermore, we can easily evaluate Published as a conference paper at ICLR 2024 Algorithm 1 Kernel Ridge Regression on Self-supervised Target (KRR-ST) 1: Input: Dataset Xt, batch size b, learning rate α, η, SSL objective LSSL, and total steps T. 2: Optimize gϕ with SSL loss on Xt using data augmentation: ϕ arg minϕ LSSL(ϕ; Xt). 3: Initialize Xs = [ˆx1 ˆxm] with a random subset of Xt. 4: Initialize Ys = [ˆy1 ˆym] with ˆyi = gϕ(ˆxi) for i = 1, . . . ,m. 5: Initialize a model pool M = {(ω1,W1, t1) . . . , (ωl, Wl, tl)} using Xs and Ys. 6: while not converged do 7: Uniformly sample a mini batch Xt = [ x1 xb] from the full dataset Xt. 8: Uniformly sample an index i from {1, . . . , l} and retrieve the model (ωi,Wi, ti) M. 9: Compute the outer objective Louter(Xs, Ys) with fωi in equation 4. 10: Update Xs and Ys: Xs Xs α Xs Louter(Xs,Ys), Ys Ys α Ys Louter(Xs, Ys). 11: if ti < T then 12: Set ti ti + 1 and evaluate MSE loss LMSE 1 2 Ys h Wi fωi(Xs) 2 F . 13: Update ωi and Wi with ωi ωi η ωi LMSE, Wi Wi η Wi LMSE. 14: else 15: Set ti 0 and randomly initialize ωi and Wi. 16: end if 17: end while 18: Output: Distilled dataset (Xs,Ys) the inner objective with full batch Xs since the size of Xs (i.e., m) is small enough and we do not need m m pairwise correlation matrix required for many SSL objectives (Chen et al., 2020a; He et al., 2020; Zbontar et al., 2021; Bardes et al., 2022). Consequently, the elimination of randomness enables us to get an unbiased estimate of the true gradient and contributes to stable optimization. Kernel Ridge Regression Lastly, following Zhou et al. (2022), we simplify the inner optimization to reduce the computational cost of bilevel optimization in equation 3. First, we decompose the function ˆgθ into a feature extractor fω : Rdx Rdh and a linear head h W : v Rdh 7 v W Rdy, where W Rdh dy, and θ = (ω,W) (i.e., ˆgθ = h W fω). Naively, we can train the feature extractor and linear head on Xs and Ys during inner optimization and compute the meta-gradient of Xs and Ys w.r.t the outer objective while considering the feature extractor constant. However, previous works (Cazenavette et al., 2022; Zhou et al., 2022; Zhao et al., 2023) have shown that using diverse models at inner optimization is robust to overfitting compared to using a single model. Based on this insight, we maintain a model pool M consisting of l different feature extractors and linear heads. To initialize each h W fω in the pool, we first sample t {1, . . . , T} and then optimize ω and W to minimize the MSE in equation 2 on randomly initialized Xs and Ys with full-batch gradient descent algorithms for t steps, where T is the maximum number of steps. Afterward, we sample a feature extractor fω from M for each meta-update. We then optimize another head h W on top of the sampled feature extractor fω which is fixed. Here, kernel ridge regression (Murphy, 2012) enables getting a closed form solution of the linear head as h W : v 7 v fω(Xs) (KXs,Xs + λIm) 1Ys, where λ > 0 is a hyperparameter for ℓ2 regularization, Im Rm m is an identity matrix, and KXs,Xs = fω(Xs)fω(Xs) Rm m with fω(Xs) = [fω(ˆx1) fω(ˆxm)] Rm d. Then, we sample a mini-batch Xt = [ x1 xb] Rb dx from the full set Xt and compute a meta-gradient of Xs and Ys with respect to the following outer objective function: Louter(Xs,Ys) = 1 gϕ( Xt) fω( Xt)fω(Xs) (KXs,Xs + λIm) 1Ys 2 where gϕ( Xt) = [gϕ( x1) gϕ( xb)] Rb dy and fω( Xt) = [fω( x1) fω( xb)] Rb dh. Finally, we update the distilled dataset Xs and Ys with gradient descent algorithms. After updating the distilled dataset, we update the selected feature extractor fω and its corresponding head h W with the distilled dataset for one step. Based on this, we dub our proposed method as Kernel Ridge Regression on Self-supervised Target (KRR-ST), and outline its algorithmic design in Algorithm 1. Transfer Learning We now elaborate on how we deploy the distilled dataset for transfer learning scenarios. Given the distilled dataset (Xs, Ys), we first pre-train a randomly initialized feature extractor fω and head h W : v Rdh 7 v W Rdy on the distilled dataset to minimize either MSE for our KRR-ST, KIP (Nguyen et al., 2021a), and FRe PO (Zhou et al., 2022), or cross-entropy loss for DSA (Zhao & Bilen, 2021), MTT (Cazenavette et al., 2022), and DM (Zhao & Bilen, 2023): minimize ω,W 1 2 fω(Xs)W Ys 2 F , or minimize ω,W i=1 ℓ(ˆyi, softmax(fω(ˆxi) W)), (5) Published as a conference paper at ICLR 2024 where ℓ(p, q) = Pc i=1 pi log qi for p = (p1, . . . , pc), q = (q1, . . . , qc) c 1, and c 1 is simplex over Rc. Then, we discard h W and fine-tune the feature extractor fω with a randomly initialized task-specific head h Q : v Rdh 7 softmax(v Q) c 1 on a target labeled dataset to minimize the cross-entropy loss ℓ. Here, Q Rdh c and c is the number of classes. Note that we can use any loss function for fine-tuning, however, we only focus on the classification in this work. 4 EXPERIMENTS In this section, we empirically validate the efficacy of our KRR-ST on various applications: transfer learning, architecture generalization, and target data-free knowledge distillation. 4.1 EXPERIMENTAL SETUPS Datasets We use either CIFAR100 (Krizhevsky et al., 2009), Tiny Image Net (Le & Yang, 2015), or Image Net (Deng et al., 2009) as a source dataset for dataset distillation, while evaluating the distilled dataset on CIFAR10 (Krizhevsky et al., 2009), Aircraft (Maji et al., 2013), Stanford Cars (Krause et al., 2013), CUB2011 (Wah et al., 2011), Stanford Dogs (Khosla et al., 2011), and Flowers (Nilsback & Zisserman, 2008). For Image Net, we resize the images into a resolution of 64 64, following the previous dataset distillation methods (Zhou et al., 2022; Cazenavette et al., 2022). We resize the images of the target dataset into the resolution of the source dataset, e.g., 32 32 for CIFAR100 and 64 64 for Tiny Image Net and Image Net, respectively. Baselines We compare the proposed method KRR-ST with 8 different baselines including a simple baseline without pre-training, 5 representative supervised baselines from the dataset condensation benchmark (Cui et al., 2022), and 2 kernel ridge regression baselines as follows: 1) w/o pre: We train a model on solely the target dataset without any pre-training. 2) Random: We pre-train a model on randomly chosen images of the source dataset. 3) Kmeans (Cui et al., 2022): Instead of 2) Random choice, we choose the nearest image for each centroid of kmeans-clustering (Lloyd, 1982) in feature space of the source dataset. 4 6) DSA (Zhao & Bilen, 2021), DM (Zhao & Bilen, 2023), and MTT (Cazenavette et al., 2022): These are representative dataset condensation methods based on surrogate objectives such as gradient matching, distribution matching, and trajectory matching, respectively. 7) KIP (Nguyen et al., 2021a;b): Kernel Inducing Points (KIP) is the first proposed kernel ridge regression method for dataset distillation. For transfer learning, we use the distilled datasets with standard normalization instead of ZCA-whitening. 8) FRe Po (Zhou et al., 2022): Feature Regression with Pooling (FRe Po) is a relaxed version of bilevel optimization, where inner optimization is replaced with the analytic solution of kernel ridge regression on neural features. Since FRe Po does not provide datasets distilled with standard normalization, we re-run the official code of Zhou et al. (2022) with standard normalization for transfer learning. Implementation Details Following Nguyen et al. (2021a;b); Zhou et al. (2022), we use convolutional layers consisting of batch normalization (Ioffe & Szegedy, 2015), Re LU activation, and average pooling for distilling a dataset. We choose the number of layers based on the resolution of images, i.e., 3 layers for 32 32 and 4 layers for 64 64, respectively. We initialize and maintain l = 10 models for the model pool M, and update the models in the pool using full-batch gradient descent with learning rate, momentum, and weight decay being set to 0.1, 0.9, and 0.001, respectively. The total number of steps T is set to 1,000. We meta-update our distilled dataset for 160,000 iterations using Adam W optimizer (Loshchilov & Hutter, 2019) with an initial learning rate of 0.001, 0.00001, and 0.00001 for CIFAR100, Tiny Image Net, and Image Net, respetively. The learning rate is linearly decayed. We use Res Net18 (He et al., 2016) as a self-supervised target model gϕ which is trained on a source dataset with Barlow Twins (Zbontar et al., 2021) objective. After distillation, we pre-train a model on the distilled dataset for 1,000 epochs with a mini-batch size of 256 using stochastic gradient descent (SGD) optimizer, where learning rate, momentum, and weight decay are set to 0.1, 0.9, and 0.001, respectively. For the baselines, we follow their original Published as a conference paper at ICLR 2024 Table 1: The results of transfer learning with CIFAR100. The data compression ratio for source dataset is 2%. Conv Net3 is pre-trained on a condensed dataset, and then fine-tuned on target datasets. We report the average and standard deviation over three runs. The best results are bolded. Source Target Method CIFAR100 CIFAR10 Aircraft Cars CUB2011 Dogs Flowers w/o pre 64.95 0.21 87.34 0.13 34.66 0.39 19.43 0.14 18.46 0.11 22.31 0.22 58.75 0.41 Random 65.23 0.12 87.55 0.19 33.99 0.45 19.77 0.21 18.18 0.21 21.69 0.18 59.31 0.27 Kmeans 65.67 0.25 87.67 0.09 35.08 0.69 20.02 0.44 18.12 0.15 21.86 0.18 59.58 0.18 DSA 65.48 0.21 87.21 0.12 34.38 0.17 19.59 0.25 18.08 0.33 21.90 0.24 58.50 0.04 DM 65.47 0.12 87.64 0.20 35.24 0.64 20.13 0.33 18.68 0.33 21.87 0.23 59.89 0.57 MTT 65.92 0.18 87.87 0.08 36.11 0.27 21.42 0.03 18.94 0.41 22.82 0.02 60.88 0.45 KIP 65.97 0.02 87.90 0.19 37.67 0.36 23.12 0.69 20.10 0.57 23.83 0.28 63.04 0.33 FRe Po 65.64 0.40 87.67 0.22 35.34 0.85 21.05 0.06 18.88 0.20 22.90 0.29 60.35 0.07 KRR-ST 66.81 0.11 88.72 0.11 41.54 0.37 28.68 0.32 25.30 0.37 26.39 0.08 67.88 0.18 Table 2: The results of transfer learning with Tiny Image Net. The data compression ratio for source dataset is 2%. Conv Net4 is pre-trained on a condensed dataset, and then fine-tuned on target datasets. We report the average and standard deviation over three runs. The best results are bolded. Source Target Method Tiny Image Net CIFAR10 Aircraft Cars CUB2011 Dogs Flowers w/o pre 49.57 0.18 88.74 0.10 43.81 0.56 23.42 0.16 22.19 0.27 24.74 0.49 59.48 0.37 Random 50.49 0.30 88.46 0.18 43.17 0.34 24.74 0.36 21.91 0.08 25.17 0.28 62.49 0.40 Kmeans 50.69 0.37 88.58 0.16 43.77 0.31 26.26 0.33 22.52 0.37 25.85 0.25 63.63 0.38 DSA 50.42 0.05 88.77 0.13 43.63 0.13 26.02 0.84 22.98 0.48 26.52 0.33 63.98 0.42 DM 49.81 0.10 88.48 0.08 42.14 0.38 25.68 0.44 22.48 0.62 25.05 0.28 63.45 0.14 MTT 50.92 0.18 89.20 0.05 48.21 0.30 30.35 0.22 25.95 0.14 28.53 0.26 66.07 0.18 FRe Po 49.71 0.06 88.28 0.09 47.59 0.59 29.25 0.41 24.81 0.66 27.68 0.37 62.91 0.47 KRR-ST 51.86 0.24 89.31 0.08 58.83 0.30 49.26 0.64 35.55 0.66 35.78 0.46 71.16 0.70 Table 3: The results of transfer learning with Image Net. The data compression ratio for the source dataset is 0.08%. Conv Net4 is pre-trained on a condensed dataset and then fine-tuned on target datasets. We report the average and standard deviation over three runs. The best results are bolded. Method CIFAR10 CIFAR100 Aircraft Cars CUB2011 Dogs Flowers w/o pre 88.66 0.09 66.62 0.32 42.45 0.46 23.62 0.70 22.00 0.09 24.59 0.46 59.39 0.29 Random 88.46 0.09 65.97 0.08 40.09 0.46 20.92 0.42 19.41 0.28 23.08 0.40 56.81 0.44 FRe Po 87.88 0.20 65.23 0.47 39.03 0.35 20.00 0.73 19.26 0.21 22.05 0.45 52.50 0.51 KRR-ST 89.33 0.19 68.04 0.22 57.17 0.16 46.95 0.37 35.66 0.56 35.51 0.45 70.45 0.34 experimental setup to pre-train a model on their condensed dataset. For fine-tuning, all the experimental setups are fixed as follows: we use the SGD optimizer with learning rate of 0.01, momentum of 0.9 and weight decay of 0.0005. We fine-tune the models for 10,000 iterations (CIFAR100, CIFAR10, and Tiny Imagenet), or 5,000 iterations (Aircraft, Cars, CUB2011, Dogs, and Flowers) with a mini-batch size of 256. The learning rate is decayed with cosine scheduling. 4.2 EXPERIMENTAL RESULTS AND ANALYSIS Transfer learning We investigate how our proposed KRR-ST can be effectively used for transfer learning. To this end, we pre-train a model on the distilled source dataset and fine-tune the model using a target training dataset. We report the average and standard deviation of the model s accuracy on the target test dataset over three runs. First, we use Conv Net3 (3-layer CNN) to distill the CIFAR100 dataset into 1,000 synthetic examples, which is equivalent to 2% of the original dataset. After distillation, we pre-train the model with the synthetic samples and fine-tune it on the target training datasets. As shown in Table 1, KRR-ST outperforms all the baselines, including those using labels for distillation. Next, we distill Tiny Image Net into 2,000 synthetic images, which constitute 2% of the original dataset. We pre-train Conv Net4 on the distilled dataset and fine-tune the model on the target datasets. As shown in Table 2, we observe that our unsupervised dataset distillation method outperforms all the baselines by a larger margin than in the previous experiments with Published as a conference paper at ICLR 2024 (a) CIFAR100 (b) Tiny Image Net Figure 2: Visualization of the distilled images, their feature representation and corresponding distilled labels in the output space of the target model. All distilled images are provided in Appendix B. Figure 3: The results of architecture generalization. Conv Net4 is utilized for condensing Tiny Image Net into 2,000 synthetic examples. Models with different architectures are pre-trained on the condensed dataset and fine-tuned on target datasets. We report the average and standard deviation over three runs. The above results are reported as a tabular format in Appendix C. the distilled CIFAR100 dataset. Lastly, we distill Image Net into 1,000 synthetic samples which are approximately 0.08% of the original full dataset using Conv Net4, and report experimental results in Table 3. For Image Net experiments, FRe Po is the only available supervised dataset distillation method since we can not run the other baselines due to their large memory consumption. These experimental results demonstrate the efficacy of our method in condensing an unlabeled dataset for pre-training a model that can be transferred to target datasets. Visualization In this paragraph, we analyze how our method distills images and their corresponding target representations in both pixel and feature space. We first present the distilled images Xs and visualize their representation gϕ(Xs) Rm 512 along with the learned target representation Ys Rm 512 and representations of the original full data gϕ(Xt) Rn 512, where gϕ is the target Res Net18 trained with SSL Barlow Twins objective on the original dataset Xt. For visualization, we employ t-SNE (Van der Maaten & Hinton, 2008) to project the high-dimensional representations to 2D vectors. Figure 2 demonstrates the distilled images and corresponding feature space of CIFAR100 and Tiny Image Net. As reported in Zhou et al. (2022), we have found that distilled images with our algorithm result in visually realistic samples, which is well-known to be a crucial factor for architecture generalization. Lastly, we observe that the distilled data points cover most of the feature space induced by the full dataset, even with either 1,000 or 2,000 synthetic samples which are only 2% of the full dataset. All distilled images are provided in Appendix B. Architecture Generalization To examine whether our method can produce a distilled dataset that can be generalized to different architectures, we perform the following experiments. First, we use Conv Net4 as ˆgθ in equation 3 to condense Tiny Image Net into 2,000 synthetic examples. Then, we pre-train models of VGG11 (Simonyan & Zisserman, 2015), Alex Net (Krizhevsky et al., 2012), Mobile Net (Howard et al., 2017), and Res Net10 (Gong et al., 2022) architectures on the condensed dataset. Finally, the models are fine-tuned on five target datasets Stanford Cars, Stanford Dogs, Aircraft, CUB2011, and Flowers dataset. We choose those architectures since they are lightweight and suitable for small devices, and pre-trained weights of those architectures for 64 64 resolution are rarely available on the internet. As shown in Figure 3 and Tables 5 to 8 from Appendix C, our method achieves significant improvements over baselines across different architectures except for one setting (Mobile Net on the Aircraft dataset). These results showcase that our method can effectively distill the source dataset into a small one that allows pre-training models with different architectures. Published as a conference paper at ICLR 2024 Table 4: The results of target data-free KD on CIFAR10. Conv Net4 is utilized for condensing Tiny Image Net into 2,000 synthetic examples. Models with different architectures are pre-trained on the condensed dataset and fine-tuned on CIFAR10 using KD loss. We report the average and standard deviation over three runs. Method Conv Net4 VGG11 Alex Net Mobile Net Res Net10 Gaussian 32.45 0.65 33.25 1.33 30.58 0.56 23.96 0.94 24.83 1.86 Random 49.98 0.73 51.47 1.10 51.16 0.22 44.27 0.92 40.63 0.54 Kmeans 52.00 1.10 53.86 0.92 53.31 0.65 48.10 0.46 40.90 0.42 DSA 45.64 1.28 47.97 0.13 47.42 0.70 39.67 0.92 37.33 1.70 DM 46.90 0.18 48.41 0.34 48.61 0.25 40.09 1.30 39.35 0.40 MTT 49.62 0.90 53.18 0.72 51.22 0.38 44.48 1.04 38.75 0.47 FRe Po 45.25 0.40 50.51 0.50 47.24 0.21 42.98 0.71 42.16 0.81 KRR-ST 58.31 0.94 62.15 1.08 59.80 0.69 54.13 0.49 52.68 1.47 Target Data-Free Knowledge Distillation One of the most challenging transfer learning scenarios is data-free knowledge distillation (KD) (Lopes et al., 2017; Yin et al., 2020; Raikwar & Mishra, 2022), where we aim to distill the knowledge of teacher into smaller student models without a target dataset due to data privacy or intellectual property issues. Inspired by the success of KD with a surrogate dataset (Orekondy et al., 2019; Kim et al., 2023), we utilize distilled Tiny Image Net dataset Xs as a surrogate dataset for KD instead of using the target dataset CIFAR10. Here, we investigate the efficacy of each dataset distillation method on this target data-free KD task. First, we are given a teacher model Tψ : Rdx c 1 which is trained on the target dataset CIFAR10, where c = 10 is the number of classes. We first pre-train a feature extractor fω, as demonstrated in equation 5. After that, we randomly initialize the task head h Q : v Rdh 7 softmax(v Q) c 1 with Q Rdh c, and fine-tune ω and Q with the cross-entropy loss ℓusing the teacher Tψ as a target: minimize ω,Q 1 m i=1 ℓ Tψ(ˆxi), softmax(fω(ˆxi) Q) . (6) In preliminary experiments, we have found that direct use of distilled dataset Xs for KD is not beneficial due to the discrepancy between the source and target dataset. To address this issue, we always use a mean and standard deviation of current mini-batch for batch normalization in both student and teacher models, even at test time, as suggested in Raikwar & Mishra (2022). We optimize the parameter ω and Q of the student model for 1,000 epochs with a mini-batch size of 512, using an Adam W optimizer with a learning rate of 0.0001. Besides the supervised dataset distillation baselines, we introduce another baseline (Raikwar & Mishra, 2022) referred to as Gaussian , which uses Gaussian noise as an input to the teacher and the student for computing the KD loss in equation 6, i.e., ˆxi N(0, Idx). Table 4 presents the results of target data-free KD experiments on CIFAR10. Firstly, we observe that utilizing a condensed surrogate dataset is more effective for knowledge distillation than using a Gaussian noise. Moreover, supervised dataset distillation methods (DSA, DM, MTT, and FRe PO) even perform worse than the baseline Random. On the other hand, our proposed KRR-ST consistently outperforms all the baselines across different architectures, which showcases the effectiveness of our method for target data-free KD. 5 CONCLUSION In this work, we proposed a novel problem of unsupervised dataset distillation where we distill an unlabeled dataset into a small set of synthetic samples on which we pre-train a model on, and fine-tune the model on the target datasets. Based on a theoretical analysis that the gradient of the synthetic samples with respect to existing SSL loss in naive bilevel optimization is biased, we proposed minimizing the mean squared error (MSE) between a model s representation of the synthetic samples and learnable target representations for the inner objective. Based on the motivation that the model obtained by the inner optimization is expected to imitate the self-supervised target model, we also introduced the MSE between representations of the inner model and those of the self-supervised target model on the original full dataset for outer optimization. Finally, we simplify the inner optimization by optimizing only a linear head with kernel ridge regression, enabling us to reduce the computational cost. The experimental results demonstrated the efficacy of our self-supervised data distillation method in various applications such as transfer learning, architecture generalization, and target data-free knowledge distillation. Published as a conference paper at ICLR 2024 Reproducibility Statement We use Pytorch (Paszke et al., 2019) to implement our self-supervised dataset distillation method, KRR-ST. First, we have provided the complete proof of Theorem 1 in Appendix A. Moreover, we have detailed our method in Algorithm 1 and specified all the implementation details including hyperaparameters in Section 4.1. Our code is available at https://github.com/db-Lee/selfsup dd. Ethics Statement Our work is less likely to bring about any negative societal impacts. However, we should be careful about bias in the original dataset, as this bias may be transferred to the distilled dataset. On the positive side, we can significantly reduce the search cost of NAS, which, in turn, can reduce the energy consumption when running GPUs. Acknowledgement This work was supported by Institute of Information & communications Technology Planning & Evaluation (IITP) grant funded by the Korea government(MSIT) (No.20190-00075, Artificial Intelligence Graduate School Program(KAIST)), Institute of Information & communications Technology Planning & Evaluation (IITP) grant funded by the Korea government(MSIT) (No.2020-0-00153, and No.2022-0-00713), KAIST-NAVER Hypercreative AI Center, Samsung Electronics (IO201214-08145-01), the National Research Foundation Singapore under the AI Singapore Programme (AISG Award No: AISG2-TC-2023-010-SGIL) and the Singapore Ministry of Education Academic Research Fund Tier 1 (Award No: T1 251RES2207). Adrien Bardes, Jean Ponce, and Yann Le Cun. VICReg: Variance-invariance-covariance regularization for self-supervised learning. In International Conference on Learning Representations, 2022. Zal an Borsos, Mojmir Mutny, and Andreas Krause. Coresets via bilevel optimization for continual learning and streaming. Advances in neural information processing systems, 33:14879 14890, 2020. George Cazenavette, Tongzhou Wang, Antonio Torralba, Alexei A Efros, and Jun-Yan Zhu. Dataset distillation by matching training trajectories. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4750 4759, 2022. Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. A simple framework for contrastive learning of visual representations. In International conference on machine learning, pp. 1597 1607. PMLR, 2020a. Ting Chen, Simon Kornblith, Kevin Swersky, Mohammad Norouzi, and Geoffrey E Hinton. Big self-supervised models are strong semi-supervised learners. Advances in neural information processing systems, 33:22243 22255, 2020b. Justin Cui, Ruochen Wang, Si Si, and Cho-Jui Hsieh. Dc-bench: Dataset condensation benchmark. Advances in Neural Information Processing Systems, 35:810 822, 2022. Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Fei-Fei Li. Imagenet: A large-scale hierarchical image database. In 2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2009), 20-25 June 2009, Miami, Florida, USA, pp. 248 255. IEEE Computer Society, 2009. Zhiwei Deng and Olga Russakovsky. Remember the past: Distilling datasets into addressable memories for neural networks. In Advances in Neural Information Processing Systems, 2022. Luca Franceschi, Michele Donini, Paolo Frasconi, and Massimiliano Pontil. Forward and reverse gradient-based hyperparameter optimization. In International Conference on Machine Learning, pp. 1165 1173. PMLR, 2017. Jiaming Gong, Wei Liu, Mengjie Pei, Chengchao Wu, and Liufei Guo. Resnet10: A lightweight residual network for remote sensing image classification. In International Conference on Measuring Technology and Mechatronics Automation (ICMTMA), pp. 975 978, 2022. Published as a conference paper at ICLR 2024 Jean-Bastien Grill, Florian Strub, Florent Altch e, Corentin Tallec, Pierre Richemond, Elena Buchatskaya, Carl Doersch, Bernardo Avila Pires, Zhaohan Guo, Mohammad Gheshlaghi Azar, et al. Bootstrap your own latent-a new approach to self-supervised learning. Advances in neural information processing systems, 33:21271 21284, 2020. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770 778, 2016. Kaiming He, Haoqi Fan, Yuxin Wu, Saining Xie, and Ross Girshick. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 9729 9738, 2020. Kaiming He, Xinlei Chen, Saining Xie, Yanghao Li, Piotr Doll ar, and Ross Girshick. Masked autoencoders are scalable vision learners. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 16000 16009, 2022. Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler, and Sepp Hochreiter. Gans trained by a two time-scale update rule converge to a local nash equilibrium. Advances in neural information processing systems, 30, 2017. Andrew G Howard, Menglong Zhu, Bo Chen, Dmitry Kalenichenko, Weijun Wang, Tobias Weyand, Marco Andreetto, and Hartwig Adam. Mobilenets: Efficient convolutional neural networks for mobile vision applications. ar Xiv preprint ar Xiv:1704.04861, 2017. Jeremy Howard. Imagenette: A smaller subset of 10 easily classified classes from imagenet. https://github.com/fastai/imagenette, March 2019. Sergey Ioffe and Christian Szegedy. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In International conference on machine learning, pp. 448 456. pmlr, 2015. Aditya Khosla, Nityananda Jayadevaprakash, Bangpeng Yao, and Li Fei-Fei. Novel dataset for finegrained image categorization. In First Workshop on Fine-Grained Visual Categorization, IEEE Conference on Computer Vision and Pattern Recognition, Colorado Springs, CO, June 2011. Byungjoo Kim, Suyoung Lee, Seanie Lee, Sooel Son, and Sung Ju Hwang. Margin-based neural network watermarking. In International Conference on Machine Learning, 2023. Jonathan Krause, Michael Stark, Jia Deng, and Li Fei-Fei. 3d object representations for fine-grained categorization. In 4th International IEEE Workshop on 3D Representation and Recognition (3d RR-13), Sydney, Australia, 2013. Alex Krizhevsky, Geoffrey Hinton, et al. Learning multiple layers of features from tiny images, 2009. Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. Imagenet classification with deep convolutional neural networks. Advances in neural information processing systems, 25, 2012. Ya Le and Xuan Yang. Tiny imagenet visual recognition challenge. CS 231N, 7(7):3, 2015. Hayeon Lee, Sewoong Lee, Song Chong, and Sung Ju Hwang. Hardware-adaptive efficient latency prediction for NAS via meta-learning. In Advances in Neural Information Processing Systems, 2021. Jaehoon Lee, Lechao Xiao, Samuel Schoenholz, Yasaman Bahri, Roman Novak, Jascha Sohl Dickstein, and Jeffrey Pennington. Wide neural networks of any depth evolve as linear models under gradient descent. Advances in neural information processing systems, 32, 2019. Hanxiao Liu, Karen Simonyan, and Yiming Yang. DARTS: Differentiable architecture search. In International Conference on Learning Representations, 2019. Stuart Lloyd. Least squares quantization in pcm. IEEE transactions on information theory, 28(2): 129 137, 1982. Published as a conference paper at ICLR 2024 Noel Loo, Ramin Hasani, Alexander Amini, and Daniela Rus. Efficient dataset distillation using random feature approximation. Advances in Neural Information Processing Systems, 35:13877 13891, 2022. Noel Loo, Ramin Hasani, Mathias Lechner, and Daniela Rus. Dataset distillation with convexified implicit gradients. ar Xiv preprint ar Xiv:2302.06755, 2023. Raphael Gontijo Lopes, Stefano Fenu, and Thad Starner. Data-free knowledge distillation for deep neural networks. ar Xiv preprint ar Xiv:1710.07535, 2017. David Lopez-Paz and Marc Aurelio Ranzato. Gradient episodic memory for continual learning. Advances in neural information processing systems, 30, 2017. Ilya Loshchilov and Frank Hutter. Decoupled weight decay regularization. In International Conference on Learning Representations, 2019. Subhransu Maji, Esa Rahtu, Juho Kannala, Matthew Blaschko, and Andrea Vedaldi. Fine-grained visual classification of aircraft. ar Xiv preprint ar Xiv:1306.5151, 2013. Baharan Mirzasoleiman, Jeff Bilmes, and Jure Leskovec. Coresets for data-efficient training of machine learning models. In International Conference on Machine Learning, pp. 6950 6960. PMLR, 2020. Kevin P Murphy. Machine learning: a probabilistic perspective. MIT press, 2012. Timothy Nguyen, Zhourong Chen, and Jaehoon Lee. Dataset meta-learning from kernel ridgeregression. In International Conference on Learning Representations, 2021a. Timothy Nguyen, Roman Novak, Lechao Xiao, and Jaehoon Lee. Dataset distillation with infinitely wide convolutional networks. Advances in Neural Information Processing Systems, 34:5186 5198, 2021b. Maria-Elena Nilsback and Andrew Zisserman. Automated flower classification over a large number of classes. In 2008 Sixth Indian conference on computer vision, graphics & image processing, pp. 722 729. IEEE, 2008. Tribhuvanesh Orekondy, Bernt Schiele, and Mario Fritz. Knockoff nets: Stealing functionality of black-box models. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 4954 4963, 2019. Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, et al. Pytorch: An imperative style, highperformance deep learning library. Advances in neural information processing systems, 32, 2019. Piyush Raikwar and Deepak Mishra. Discovering and overcoming limitations of noise-engineered data-free knowledge distillation. Advances in Neural Information Processing Systems, 35:4902 4912, 2022. Karen Simonyan and Andrew Zisserman. Very deep convolutional networks for large-scale image recognition. In International Conference on Learning Representations, ICLR 2015, 2015. Laurens Van der Maaten and Geoffrey Hinton. Visualizing data using t-SNE. Journal of machine learning research, 9(11), 2008. C. Wah, S. Branson, P. Welinder, P. Perona, and S. Belongie. The caltech-ucsd birds-200-2011 dataset. Technical Report CNS-TR-2011-001, California Institute of Technology, 2011. Tongzhou Wang, Jun-Yan Zhu, Antonio Torralba, and Alexei A Efros. Dataset distillation. ar Xiv preprint ar Xiv:1811.10959, 2018. Hongxu Yin, Pavlo Molchanov, Jose M Alvarez, Zhizhong Li, Arun Mallya, Derek Hoiem, Niraj K Jha, and Jan Kautz. Dreaming to distill: Data-free knowledge transfer via deepinversion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 8715 8724, 2020. Published as a conference paper at ICLR 2024 Jure Zbontar, Li Jing, Ishan Misra, Yann Le Cun, and St ephane Deny. Barlow twins: Self-supervised learning via redundancy reduction. In International Conference on Machine Learning, pp. 12310 12320. PMLR, 2021. Bo Zhao and Hakan Bilen. Dataset condensation with differentiable siamese augmentation. In International Conference on Machine Learning, 2021. Bo Zhao and Hakan Bilen. Dataset condensation with distribution matching. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 6514 6523, 2023. Bo Zhao, Konda Reddy Mopuri, and Hakan Bilen. Dataset condensation with gradient matching. In International Conference on Learning Representations, 2021. Ganlong Zhao, Guanbin Li, Yipeng Qin, and Yizhou Yu. Improved distribution matching for dataset condensation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 7856 7865, 2023. Yongchao Zhou, Ehsan Nezhadarya, and Jimmy Ba. Dataset distillation using neural feature regression. Advances in Neural Information Processing Systems, 35:9813 9827, 2022. Published as a conference paper at ICLR 2024 A PROOF OF THEOREM 1 Proof. Let (i,j) {1, . . . ,m} {1, . . . , dx}. By the chain rule, LSSL(θ (Xs); Xt) (Xs)ij = LSSL(θ; Xt) Similarly, LSSL(ˆθ(Xs); Xt) (Xs)ij = LSSL(θ; Xt) (Xs)ij By taking expectation and using the definition of the covariance, " LSSL(ˆθ(Xs); Xt) " LSSL(θ; Xt) LSSL(θ; Xt) " LSSL(θ; Xt) k=1 Eζ[vk]Eζ[αk] + k=1 Covζ[vk,αk], where vk = LSSL(θ;Xt) θk |θ=ˆθ(Xs) and αk = ˆθ(Xs)k (Xs)ij . By defining the vectors v = [v1,v2, . . . ,vdθ] Rdθ and α = [α1,α2, . . . ,αdθ] Rdθ, " LSSL(ˆθ(Xs); Xt) k=1 Eζ[vk]Eζ[αk] + k=1 Covζ[vk,αk] = [Eζ[v1] Eζ[vdθ]] Eζ[α1] ... Eζ[αdθ] k=1 Covζ[vk,αk] = Eζ [[v1 vdθ]] Eζ k=1 Covζ[vk,αk] = Eζ [v] Eζ [α] + k=1 Covζ[vk,αk]. Therefore, Eζ h LSSL(ˆθ(Xs);Xt) i = Eζ [v] Eζ [α]+Pdθ k=1 Covζ[vk,αk]. Comparing this with equation 7 proves the statement. Published as a conference paper at ICLR 2024 B VISUALIZATION OF DISTILLED IMAGES Figure 4: Visualization of the synthetic images distilled by our method in CIFAR100. Figure 5: Visualization of the synthetic images distilled by our method in Tiny Image Net. Figure 6: Visualization of the synthetic images distilled by our method in Image Net. Published as a conference paper at ICLR 2024 C EXPERIMENTAL RESULTS OF ARCHITECTURE GENERALIZATION Table 5: The results of transfer learning using VGG11. Conv4 is utilized for condensing Tiny Image Net into 2,000 synthetic examples. We report the average and standard deviation over three runs. Method Aircraft Cars CUB2011 Dogs Flowers w/o pre 59.17 0.39 32.41 0.44 32.62 0.11 32.24 0.34 60.88 0.48 Random 55.67 0.27 35.70 0.71 33.99 0.41 33.55 0.21 65.75 0.25 Kmeans 57.50 0.65 38.66 0.45 34.20 0.10 34.50 0.33 66.79 0.59 DSA 55.26 0.39 37.22 0.54 34.75 0.98 33.01 0.46 67.07 0.99 DM 56.40 0.18 39.13 0.99 36.01 0.12 34.65 0.37 68.82 0.60 MTT 61.54 0.51 46.44 0.42 37.95 0.33 38.35 0.24 70.97 0.50 FRe Po 49.13 1.23 30.72 0.51 31.57 0.92 30.93 0.66 64.66 0.18 KRR-ST 66.35 0.76 54.43 0.17 39.22 0.28 38.54 0.31 71.46 0.70 Table 6: The results of transfer learning using Alex Net. Conv4 is utilized for condensing Tiny Image Net into 2,000 synthetic examples. We report the average and standard deviation over three runs. Method Aircraft Cars CUB2011 Dogs Flowers w/o pre 52.29 0.55 21.83 0.29 24.69 0.31 23.95 0.19 55.46 0.33 Random 50.56 0.53 22.77 0.34 25.20 0.39 24.09 0.20 59.81 0.62 Kmeans 50.72 0.26 24.23 0.05 25.29 0.49 24.11 0.19 61.20 0.45 DSA 49.67 0.38 24.78 0.22 25.77 0.38 24.30 0.38 60.80 0.22 DM 48.01 0.43 22.84 0.46 23.88 0.14 23.04 0.12 59.33 0.33 MTT 54.24 0.45 29.93 0.48 28.59 0.17 27.74 0.34 64.40 0.35 FRe Po 51.39 0.31 24.86 1.08 25.19 0.72 24.68 0.70 61.08 1.58 KRR-ST 58.76 0.48 41.60 0.08 33.69 0.22 31.55 0.26 65.50 0.29 Table 7: The results of transfer learning using Mobile Net. Conv4 is utilized for condensing Tiny Image Net into 2,000 synthetic examples. We report the average and standard deviation over three runs. Method Aircraft Cars CUB2011 Dogs Flowers w/o pre 55.99 0.85 33.96 2.03 30.73 0.39 32.21 0.51 52.67 1.11 Random 58.71 0.75 40.69 0.62 30.03 1.10 32.52 0.27 56.17 0.92 Kmeans 59.59 1.02 42.90 1.43 30.82 0.24 33.09 0.45 55.52 0.38 DSA 60.32 0.47 41.44 1.23 32.10 0.48 33.58 0.88 57.27 0.30 DM 60.29 0.97 42.63 1.03 31.66 0.81 34.39 0.37 56.23 1.70 MTT 64.54 0.55 50.84 1.13 36.31 0.51 37.97 0.86 60.25 1.68 FRe Po 57.48 1.05 41.85 0.76 35.03 0.22 35.05 0.83 61.29 0.40 KRR-ST 63.08 0.49 54.89 0.06 37.65 0.63 40.81 0.14 67.03 1.03 Table 8: The results of transfer learning using Res Net10. Conv4 is utilized for condensing Tiny Image Net into 2,000 synthetic examples. We report the average and standard deviation over three runs. Method Aircraft Cars CUB2011 Dogs Flowers w/o pre 6.34 0.25 4.16 0.11 5.19 0.23 6.69 0.11 43.38 0.49 Random 39.96 1.48 28.80 0.33 18.55 0.33 29.34 0.61 57.29 0.49 Kmeans 41.67 0.78 30.12 0.28 18.63 0.56 29.34 0.11 57.46 0.18 DSA 42.43 1.26 31.66 0.70 22.17 0.62 31.58 1.12 59.47 0.54 DM 43.86 0.47 33.17 0.81 22.83 0.18 32.42 0.39 60.85 0.76 MTT 46.62 0.63 36.11 0.36 24.65 0.08 33.42 0.14 61.47 0.58 FRe Po 30.70 0.52 14.82 0.24 17.41 0.29 25.37 0.17 55.09 0.48 KRR-ST 59.91 1.24 51.55 0.90 32.55 1.55 40.05 0.74 66.61 1.38 Published as a conference paper at ICLR 2024 D IMAGENETTE EXPERIMENTS Table 9: The results of transfer learning using Conv5. Conv5 is utilized for condensing Image Nette (Howard, 2019) into 10 synthetic examples. Here, we use Res Net50 pre-trained on Image Net using Barlow Twins as the target model gϕ for KRR-ST. We report the average and standard deviation of test accuracy over three runs. Source Target Method Image Nette Aircraft Cars CUB2011 Dogs Flowers w/o pre 77.55 0.89 46.65 0.26 16.31 0.30 18.92 0.43 19.64 0.40 46.80 0.69 Random 62.22 3.12 21.15 1.54 5.77 0.91 7.03 1.10 10.10 0.52 22.31 1.34 DSA 63.11 0.99 23.07 1.19 6.37 0.50 7.89 0.91 11.12 0.37 24.70 1.71 DM 67.69 2.39 28.50 2.87 7.47 0.45 8.73 1.17 11.56 0.39 26.23 2.43 KRR-ST 82.64 0.62 54.34 0.22 16.65 0.03 22.15 0.01 23.29 0.12 47.86 0.05 Table 10: The results of transfer learning using Res Net18. Conv5 is utilized for condensing Image Nette (Howard, 2019) into 10 synthetic examples. Here, we use Res Net50 pre-trained on Image Net using Barlow Twins as the target model gϕ for KRR-st. We report the average and standard deviation of test accuracy over three runs. Source Target Method Image Nette Aircraft Cars CUB2011 Dogs Flowers w/o pre 77.16 0.12 6.69 0.18 4.18 0.12 5.02 0.30 5.24 0.06 41.51 0.63 Random 79.16 0.48 12.47 0.21 5.35 0.06 6.44 0.08 9.94 0.42 45.50 0.29 DSA 78.94 0.08 8.22 0.30 5.01 0.11 6.47 0.13 7.82 0.50 45.36 0.21 DM 79.13 0.16 8.79 0.75 5.21 0.17 6.71 0.14 8.96 0.52 44.71 0.14 KRR-ST 86.94 0.21 33.94 3.15 6.19 0.17 14.43 0.18 17.26 0.21 51.88 0.23 E ABLATION STUDY ON THE INITIALIZATION OF Ys Table 11: The results of ablation study on the initialization of the target representation Ys with Conv3. w/o init denotes the initialization of Ys = [ˆy1 ˆym] with ˆyi drawn from standard normal distribution (i.e., ˆyi iid N(0, Idy)) for i = 1, . . . , m, and w init denotes initializing it with the target model (i.e., ˆyi = gϕ(ˆxi) for i = 1, . . . ,m) as done in Algorithm 1, respectively. We report the average and standard deviation of test accuracy over three runs. Source Target Method CIFAR100 CIFAR10 Aircraft Cars CUB2011 Dogs Flowers w/o init 65.09 0.09 87.56 0.19 25.54 0.36 17.75 0.07 16.68 0.18 20.82 0.30 54.94 0.15 w init 66.46 0.12 88.87 0.05 32.94 0.21 23.92 0.09 24.38 0.18 23.39 0.24 64.28 0.16 F FID SCORE OF CONDENSED DATASET Table 12: The FID score (Heusel et al., 2017) of condensed dataset. Method CIFAR100 Tiny Image Net Random 34.49 20.31 DSA 113.62 90.88 DM 206.09 130.06 MTT 105.11 114.33 KRR-ST 76.28 47.54