# dataset_metalearning_from_kernel_ridgeregression__48f335d2.pdf DATASET META-LEARNING FROM KERNEL RIDGEREGRESSION Timothy Nguyen Zhourong Chen Jaehoon Lee Google Research {timothycnguyen, zrchen, jaehlee}@google.com One of the most fundamental aspects of any machine learning algorithm is the training data used by the algorithm. We introduce the novel concept of ϵapproximation of datasets, obtaining datasets which are much smaller than or are significant corruptions of the original training data while maintaining similar model performance. We introduce a meta-learning algorithm called Kernel Inducing Points (KIP ) for obtaining such remarkable datasets, inspired by the recent developments in the correspondence between infinitely-wide neural networks and kernel ridge-regression (KRR). For KRR tasks, we demonstrate that KIP can compress datasets by one or two orders of magnitude, significantly improving previous dataset distillation and subset selection methods while obtaining state of the art results for MNIST and CIFAR-10 classification. Furthermore, our KIP -learned datasets are transferable to the training of finite-width neural networks even beyond the lazy-training regime, which leads to state of the art results for neural network dataset distillation with potential applications to privacy-preservation. 1 INTRODUCTION Datasets are a pivotal component in any machine learning task. Typically, a machine learning problem regards a dataset as given and uses it to train a model according to some specific objective. In this work, we depart from the traditional paradigm by instead optimizing a dataset with respect to a learning objective, from which the resulting dataset can be used in a range of downstream learning tasks. Our work is directly motivated by several challenges in existing learning methods. Kernel methods or instance-based learning (Vinyals et al., 2016; Snell et al., 2017; Kaya & Bilge, 2019) in general require a support dataset to be deployed at inference time. Achieving good prediction accuracy typically requires having a large support set, which inevitably increases both memory footprint and latency at inference time the scalability issue. It can also raise privacy concerns when deploying a support set of original examples, e.g., distributing raw images to user devices. Additional challenges to scalability include, for instance, the desire for rapid hyper-parameter search (Shleifer & Prokop, 2019) and minimizing the resources consumed when replaying data for continual learning (Borsos et al., 2020). A valuable contribution to all these problems would be to find surrogate datasets that can mitigate the challenges which occur for naturally occurring datasets without a significant sacrifice in performance. This suggests the following Question: What is the space of datasets, possibly with constraints in regards to size or signal preserved, whose trained models are all (approximately) equivalent to some specific model? In attempting to answer this question, in the setting of supervised learning on image data, we discover a rich variety of datasets, diverse in size and human interpretability while also robust to model architectures, which yield high performance or state of the art (SOTA) results when used as training data. We obtain such datasets through the introduction of a novel meta-learning algorithm called Kernel Inducing Points (KIP ). Figure 1 shows some example images from our learned datasets. Myrlte LS 500 support Figure 1: (a) Learned samples of CIFAR-10 using KIP and its variant KIPρ, for which ρ fraction of the pixels are uniform noise. Using 1000 such images to train a 1 hidden layer fully connected network results in 49.2% and 45.0% CIFAR-10 test accuracy, respectively, whereas using 1000 original CIFAR-10 images results in 35.4% test accuracy. (b) Example of labels obtained by label solving (LS ) (left two) and the covariance matrix between original labels and learned labels (right). Here, 500 labels were distilled from the CIFAR-10 train dataset using the the Myrtle 10-layer convolutional network. A test accuracy of 69.7% is achieved using these labels for kernel ridge-regression. We explore KIP in the context of compressing and corrupting datasets, validating its effectiveness in the setting of kernel-ridge regression (KRR) and neural network training on benchmark datasets MNIST and CIFAR-10. Our contributions can be summarized as follows: 1.1 SUMMARY OF CONTRIBUTIONS We formulate a novel concept of ϵ-approximation of a dataset. This provides a theoretical framework for understanding dataset distillation and compression. We introduce Kernel Inducing Points (KIP ), a meta-learning algorithm for obtaining ϵapproximation of datasets. We establish convergence in the case of a linear kernel in Theorem 1. We also introduce a variant called Label Solve (LS ), which gives a closed-form solution for obtaining distilled datasets differing only via labels. We explore the following aspects of ϵ-approximation of datasets: 1. Compression (Distillation) for Kernel Ridge-Regression: For kernel ridge regression, we improve sample efficiency by over one or two orders of magnitude, e.g. using 10 images to outperform hundreds or thousands of images (Tables 1, 2 vs Tables A1, A2). We obtain state of the art results for MNIST and CIFAR-10 classification while using few enough images (10K) to allow for in-memory inference (Tables A3, A4). 2. Compression (Distillation) for Neural Networks: We obtain state of the art dataset distillation results for the training of neural networks, often times even with only a single hidden layer fully-connected network (Tables 1 and 2). 3. Privacy: We obtain datasets with a strong trade-off between corruption and test accuracy, which suggests applications to privacy-preserving dataset creation. In particular, we produce images with up to 90% of their pixels corrupted with limited degradation in performance as measured by test accuracy in the appropriate regimes (Figures 3, A3, and Tables A5-A10) and which simultaneously outperform natural images, in a wide variety of settings. We provide an open source implementation of KIP and LS , available in an interactive Colab notebook1. In this section we define some key concepts for our methods. 1https://colab.research.google.com/github/google-research/google-research/blob/master/kip/KIP.ipynb Definition 1. A dataset in Rd is a set of n distinct vectors in Rd for some n 1. We refer to each such vector as a datapoint. A dataset is labeled if each datapoint is paired with a label vector in RC, for some fixed C. A datapoint along with its corresponding label is a labeled datapoint. We use the notation D = (X, y), where X Rn d and y Rn C, to denote the tuple of unlabeled datapoints X with their corresponding labels y. We henceforth assume all datasets are labeled. Next, we introduce our notions of approximation, both of functions (representing learned algorithms) and of datasets, which are characterized in terms of performance with respect to a loss function rather than closeness with respect to a metric. A loss function ℓ: RC RC R is one that is nonnegative and satisfies ℓ(z, z) = 0 for all z. Definition 2. Fix a loss function ℓand let f, f : Rd RC be two functions. Let ϵ 0. 1. Given a distribution P on Rd RC, we say f and f are weakly ϵ-close with respect to (ℓ, P) if E(x,y) P ℓ(f(x), y) E(x,y) P ℓ( f(x), y) ϵ. (1) 2. Given a distribution P on Rd we say f and f are strongly ϵ-close with respect to (ℓ, P) if Ex P ℓ(f(x), f(x)) ϵ. (2) We drop explicit reference to (ℓ, P) if their values are understood or immaterial. Given a learning algorithm A (e.g. gradient descent with respect to the loss function of a neural network), let AD denote the resulting model obtained after training A on D. We regard AD as a mapping from datapoints to prediction labels. Definition 3. Fix learning algorithms A and A. Let D and D be two labeled datasets in Rd with label space RC. Let ϵ 0. We say D is a weak ϵ-approximation of D with respect to ( A, A, ℓ, P) if A D and AD are weakly ϵ-close with respect to (ℓ, P), where ℓis a loss function and P is a distribution on Rd RC. We define strong ϵ-approximation similarly. We drop explicit reference to (some of) the A, A, ℓ, P if their values are understood or immaterial. We provide some justification for this definition in the Appendix. In this paper, we will measure ϵapproximation with respect to 0-1 loss for multiway classification (i.e. accuracy). We focus on weak ϵ-approximation, since in most of our experiments, we consider models in the low-data regime with large classification error rates, in which case, sample-wise agreement of two models is not of central importance. On the other hand, observe that if two models have population classification error rates less than ϵ/2, then (2) is automatically satisfied, in which case, the notions of weak-approximation and strong-approximation converge. We list several examples of ϵ-approximation, with ϵ = 0, for the case when A = A are given by the following: Example 1: Support Vector Machines. Given a dataset D of size N, train an SVM on D and obtain M support vectors. These M support vectors yield a dataset D that is a strong 0-approximation to D in the linearly separable case, while for the nonseparable case, one has to also include the datapoints with positive slack. Asymptotic lower bounds asserting M = O(N) have been shown in Steinwart (2003).2 Example 2: Ridge Regression. Any two datasets D and D that determine the same ridge-regressor are 0-approximations of each other. In particular, in the scalar case, we can obtain arbitrarily small 0-approximating D as follows. Given training data D = (X, y) in Rd, the corresponding ridgeregressor is the predictor x 7 w x , (3) w = Φλ(X)y, (4) Φλ(X) = XT (XXT + λI) 1 (5) 2As a specific example, many thousands of support vectors are needed for MNIST classification (Bordes et al. (2005)). where for λ = 0, we interpret the inverse as a pseudoinverse. It follows that for any given w Rd 1, we can always find ( X, y) of arbitrary size (i.e. X Rn d, y Rn 1 with n arbitrarily small) that satisfies w = Φλ( X) y. Simply choose X such that w is in the range of Φλ( X). The resulting dataset ( X, y) is a 0-approximation to D. If we have a C-dimensional regression problem, the preceding analysis can be repeated component-wise in label-space to show 0-approximation with a dataset of size at least C (since then the rank of Φλ( X) can be made at least the rank of w Rd C). We are interested in learning algorithms given by KRR and neural networks. These can be investigated in unison via neural tangent kernels. Furthermore, we study two settings for the usage of ϵ-approximate datasets, though there are bound to be others: 1. (Sample efficiency / compression) Fix ϵ. What is the minimum size of D needed in order for D to be an ϵ-approximate dataset? 2. (Privacy guarantee) Can an ϵ-approximate dataset be found such that the distribution from which it is drawn and the distribution from which the original training dataset is drawn satisfy a given upper bound in mutual information? Motivated by these questions, we introduce the following definitions: Definition 4. (Heuristic) Let D and D be two datasets such that D is a weak ϵ-approximation of D, with | D| |D| and ϵ small. We call |D|/| D| the compression ratio. In other words, the compression ratio is a measure of how well D compresses the information available in D, as measured by approximate agreement of their population loss. Our definition is heuristic in that ϵ is not precisely quantified and so is meant as a soft measure of compression. Definition 5. Let Γ be an algorithm that takes a dataset D in Rd and returns a (random) collection of datasets in Rd. For 0 ρ 1, we say that Γ is ρ-corrupted if for any input dataset D, every datapoint3 drawn from the datasets of Γ(D) has at least ρ fraction of its coordinates independent of D. In other words, datasets produced by Γ have ρ fraction of its entries contain no information about the dataset D (e.g. because they have a fixed value or are filled in randomly). Corrupting information is naturally a way of enhancing privacy, as it makes it more difficult for an attacker to obtain useful information about the data used to train a model. Adding noise to the inputs to neural network or of its gradient updates can be shown to provide differentially private guarantees (Abadi et al. (2016)). 3 KERNEL INDUCING POINTS Given a dataset D sampled from a distribution P, we want to find a small dataset D that is an ϵapproximation to D (or some large subset thereof) with respect to ( A, A, ℓ, P). Focusing on A = A for the moment, and making the approximation E(x,y) P ℓ( A D(x), y) E(x,y) D ℓ( A D(x), y), (6) this suggests we should optimize the right-hand side of (6) with respect to D, using D as a validation set. For general algorithms A, the outer optimization for D is computationally expensive and involves second-order derivatives, since one has to optimize over the inner loop encoded by the learning algorithm A. We are thus led to consider the class of algorithms drawn from kernel ridgeregression. The reason for this are two-fold. First, KRR performs convex-optimization resulting in a closed-form solution, so that when optimizing for the training parameters of KRR (in particular, the support data), we only have to consider first-order optimization. Second, since KRR for a neural tangent kernel (NTK) approximates the training of the corresponding wide neural network (Jacot et al., 2018; Lee et al., 2019; Arora et al., 2019a; Lee et al., 2020), we expect the use of neural kernels to yield ϵ-approximations of D for learning algorithms given by a broad class of neural networks trainings as well. (This will be validated in our experiments.) 3We ignore labels in our notion of ρ-corrupted since typically the label space has much smaller dimension than that of the datapoints. Algorithm 1: Kernel Inducing Point (KIP ) Require: A target labeled dataset (Xt, yt) along with a kernel or family of kernels. 1: Initialize a labeled support set (Xs, ys). 2: while not converged do 3: Sample a random kernel. Sample a random batch ( Xs, ys) from the support set. Sample a random batch ( Xt, yt) from the target dataset. 4: Compute the kernel ridge-regression loss given by (7) using the sampled kernel and the sampled support and target data. 5: Backpropagate through Xs (and optionally ys and any hyper-parameters of the kernel) and update the support set (Xs, ys) by updating the subset ( Xs, ys). 6: end while 7: return Learned support set (Xs, ys) This leads to our first-order meta-learning algorithm KIP (Kernel Inducing Points), which uses kernel-ridge regression to learn ϵ-approximate datasets. It can be regarded as an adaption of the inducing point method for Gaussian processes (Snelson & Ghahramani, 2006) to the case of KRR. Given a kernel K, the KRR loss function trained on a support dataset (Xs, ys) and evaluated on a target dataset (Xt, yt) is given by L(Xs, ys) = 1 2 yt KXt Xs(KXs Xs + λI) 1ys 2, (7) where if U and V are sets, KUV is the matrix of kernel elements (K(u, v))u U,v V . Here λ > 0 is a fixed regularization parameter. The KIP algorithm consists of optimizing (7) with respect to the support set (either just the Xs or along with the labels ys), see Algorithm 1. Depending on the downstream task, it can be helpful to use families of kernels (Step 3) because then KIP produces datasets that are ϵ-approximations for a variety of kernels instead of a single one. This leads to a corresponding robustness for the learned datasets when used for neural network training. We remark on best experimental practices for sampling methods and initializations for KIP in the Appendix. Theoretical analysis for the convergence properties of KIP for the case of a linear kernel is provided by Theorem 1. Sample KIP -learned images can be found in Section F. KIP variations: i) We can also randomly augment the sampled target batches in KIP . This effectively enhances the target dataset (Xt, yt), and we obtain improved results in this way, with no extra computational cost with respect to the support size. ii) We also can choose a corruption fraction 0 ρ < 1 and do the following. Initialize a random ρ-percent of the coordinates of each support datapoint via some corruption scheme (zero out all such pixels or initialize with noise). Next, do not update such corrupted coordinates during the KIP training algorithm (i.e. we only perform gradient updates on the complementary set of coordinates). Call this resulting algorithm KIPρ. In this way, KIPρ is ρ-corrupted according to Definition 5 and we use it to obtain our highly corrupted datasets. Label solving: In addition to KIP , where we learn the support dataset via gradient descent, we propose another inducing point method, Label Solve (LS ), in which we directly find the minimum of (7) with respect to the support labels while holding Xs fixed. This is simple because the loss function is quadratic in ys. We refer to the resulting labels y s = Φ0 KXt Xs(KXs Xs + λI) 1 yt (8) as solved labels. As Φ0 is the pseudo-inverse operation, y s is the minimum-norm solution among minimizers of (7). If KXt Xs is injective, using the fact that Φ0(AB) = Φ0(B)Φ0(A) for A injective and B surjective (Greville (1966)), we can rewrite (8) as y s = (KXs Xs + λI)Φ0(KXt Xs) yt. 4 EXPERIMENTS We perform three sets of experiments to validate the efficacy of KIP and LS for dataset learning. The first set of experiments investigates optimizing KIP and LS for compressing datasets and achieving state of the art performance for individual kernels. The second set of experiments explores transferability of such learned datasets across different kernels. The third set of experiments investigate the transferability of KIP -learned datasets to training neural networks. The overall conclusion is that KIP -learned datasets, even highly corrupted versions, perform well in a wide variety of settings. Experimental details can be found in the Appendix. We focus on MNIST (Le Cun et al., 2010) and CIFAR-10 (Krizhevsky et al., 2009) datasets for comparison to previous methods. For LS , we also use Fashion-MNIST. These classification tasks are recast as regression problems by using mean-centered one-hot labels during training and by making class predictions via assigning the class index with maximal predicted value during testing. All our kernel-based experiments use the Neural Tangents library (Novak et al., 2020), built on top of JAX (Bradbury et al., 2018). In what follows, we use FCm and Convm to denote a depth m fully-connected or fully-convolutional network. Whether we mean a finite-width neural network or else the corresponding neural tangent kernel (NTK) will be understood from the context. We will sometimes also use the neural network Gaussian process (NNGP) kernel associated to a neural network in various places. By default, a neural kernel refers to NTK unless otherwise stated. RBF denotes the radial-basis function kernel. Myrtle-N architecture follows that of Shankar et al. (2020), where an N-layer neural network consisting of a simple combination of N 1 convolutional layers along with (2, 2) average pooling layers are inter-weaved to reduce internal patch-size. We would have used deeper and more diverse architectures for KIP , but computational limits, which will be overcome in future work, placed restrictions, see the Experiment Details in Section D. 4.1 SINGLE KERNEL RESULTS We apply KIP to learn support datasets of various sizes for MNIST and CIFAR-10. The objective is to distill the entire training dataset down to datasets of various fixed, smaller sizes to achieve high compression ratio. We present these results against various baselines in Tables 1 and 2. These comparisons occur cross-architecturally, but aside from Myrtle LS results, all our results involve the simplest of kernels (RBF or FC1), whereas prior art use deeper architectures (Le Net, Alex Net, Conv Net). We obtain state of the art results for KRR on MNIST and CIFAR-10, for the RBF and FC1 kernels, both in terms of accuracy and number of images required, see Tables 1 and 2. In particular, our method produces datasets such that RBF and FC1 kernels fit to them rival the performance of deep convolutional neural networks on MNIST (exceeding 99.2%). By comparing Tables 2 and A2, we see that, e.g. 10 or 100 KIP images for RBF and FC1 perform on par with tens or hundreds times more natural images, resulting in a compression ratio of one or two orders of magnitude. For neural network trainings, for CIFAR-10, the second group of rows in Table 2 shows that FC1 trained on KIP images outperform prior art, all of which have deeper, more expressive architectures. On MNIST, we still outperform some prior baselines with deeper architectures. This, along with the state of the art KRR results, suggests that KIP , when scaled up to deeper architectures, should continue to yield strong neural network performance. For LS , we use a mix of NNGP kernels4 and NTK kernels associated to FC1, Myrtle-5, Myrtle-10 to learn labels on various subsets of MNIST, Fashion-MNIST, and CIFAR-10. Our results comprise the bottom third of Tables 1 and 2 and Figure 2. As Figure 2 shows, the more targets are used, the better the performance. When all possible targets are used, we get an optimal compression ratio of roughly one order of magnitude at intermediate support sizes. 4.2 KERNEL TO KERNEL RESULTS Here we investigate robustness of KIP and LS learned datasets when there is variation in the kernels used for training and testing. We draw kernels coming from FC and Conv layers of depths 1-3, since such components form the basic building blocks of neural networks. Figure A1 shows that KIP - datasets trained with random sampling of all six kernels do better on average than KIP -datasets trained using individual kernels. 4For FC1, NNGP and NTK perform comparably whereas for Myrtle, NNGP outperforms NTK. 100 101 102 103 104 Support Set Size Test Accuracy Myrtle-10 Label Solve 100 101 102 103 104 Support Set Size Test Accuracy FC1 (NNGP) Label Solve 100 101 102 103 104 Support Set Size Test Accuracy Myrtle-5 (NNGP) Label Solve (Fashion-MNIST) 100 101 102 103 104 Support Set Size Test Accuracy Myrtle-5 Label Solve Target: 0.1k Target: 0.2k Target: 0.5k Target: 1.0k Target: 2.0k Target: 5.0k Target: 10.0k Target: 20.0k Target: 50.0k Raw Figure 2: LS performance for Myrtle-(5/10) and FC on CIFAR-10/Fashion-MNIST/MNIST. Results computed over 3 independent samples per support set size. Table 1: MNIST: KIP and LS vs baselines. Comparing KRR (kernel ridge-regression) and NN (neural network) algorithms using various architectures and dataset distillation methods on datasets of varying sizes (10 to 10K). Alg. Arch., Method 10 100 500 5000 10000 KRR RBF, KIP 89.60 0.09 97.31 0.09 98.29 0.06 98.70 0.04 98.74 0.04 KRR RBF, KIP (a + l)1 90.63 0.27 97.84 0.06 98.85 0.04 99.31 0.04 99.34 0.03 KRR FC1, KIP 89.30 0.01 96.64 0.08 97.64 0.06 98.52 0.04 98.59 0.05 KRR FC1, KIP (a + l) 85.46 0.04 97.15 0.11 98.36 0.08 99.18 0.04 99.26 0.03 NN FC1, KIP2 86.49 0.40 88.96 0.37 95.70 0.09 97.97 0.07 - NN Conv Net3, DC4 91.7 0.5 97.4 0.2 - - - NN Le Net, DC - 93.9 0.6 - - - NN Le Net, SLDD - 82.7 2.8 - - - NN Le Net, DD - 79.5 8.1 - - - KRR FC1, LS 61.0 0.28 87.2 0.71 94.4 0.16 97.5 0.06 97.9 0.09 KRR Myrtle-5 NNGP, LS 70.24 1.59 95.44 0.17 98.32 0.91 99.17 0.01 99.33 0.07 KRR Myrtle-5, LS 68.50 2.52 95.53 0.22 98.17 0.07 99.05 0.06 99.22 0.02 NN Le Net, LD 64.57 2.67 87.85 0.43 94.75 0.29 - - 1 (a + l) denotes KIP trained with augmentations and learning of labels 2 KIP images are trained using the same kernel (FC1) corresponding to the evaluation neural network. Likewise for KRR, the train and test kernels coincide. 3 Conv Net is neural network consisting of 3 convolutional blocks, where a block consists of convolution, instance normalization, and a (2,2) average pooling. See Zhao et al. (2020). 4 DC (Zhao et al., 2020), LD (Bohdal et al., 2020), SLDD (Sucholutsky & Schonlau, 2019), DD (Wang et al., 2018). For LS , transferability between FC1 and Myrtle-10 kernels on CIFAR-10 is highly robust, see Figure A2. Namely, one can label solve using FC1 and train Myrtle-10 using those labels and vice versa. There is only a negligible difference in performance in nearly all instances between data with transferred learned labels and with natural labels. 4.3 KERNEL TO NEURAL NETWORKS RESULTS Significantly, KIP -learned datasets, even with heavy corruption, transfer remarkably well to the training of neural networks. Here, corruption refers to setting a random ρ fraction of the pixels of each image to uniform noise between 1 and 1 (for KIP , this is implemented via KIPρ )5. The deterioriation in test accuracy for KIP -images is limited as a function of the corruption fraction, especially when compared to natural images, and moreover, corrupted KIP -images typically outperform uncorrupted natural images. We verify these conclusions along the following dimensions: Robustness to dataset size: We perform two sets of experiments. (i) First, we consider small KIP datasets (10, 100, 200 images) optimized using multiple kernels (FC1-3, Conv1-2), see Tables A5, A6. We find that our in-distribution transfer (the downstream neural network has its neural kernel included among the kernels sampled by KIP ) performs re- 5Our images are preprocessed so as to be mean-centered and unit-variance per pixel. This choice of corruption, which occurs post-processing, is therefore meant to (approximately) match the natural pixel distribution. Table 2: CIFAR-10: KIP and LS vs baselines. Comparing KRR (kernel ridge-regression) and NN (neural network) algorithms using various architectures and dataset distillation methods on datasets of various sizes (10 to 10K). Notation same as in Table 1. Alg. Arch., Method 10 100 500 5000 10000 KRR RBF, KIP 39.9 0.9 49.3 0.3 51.2 0.8 - - KRR RBF, KIP (a + l) 40.3 0.5 53.8 0.3 60.1 0.2 65.6 0.2 66.3 0.2 KRR FC1, KIP 39.3 1.6 49.1 1.1 52.1 0.8 54.5 0.5 54.9 0.5 KRR FC1, KIP (a + l) 40.5 0.4 53.1 0.5 58.6 0.4 63.8 0.3 64.6 0.2 NN FC1, KIP 36.2 0.1 45.7 0.3 46.9 0.2 50.1 0.4 51.7 0.4 NN Conv Net, DC 28.3 0.5 44.9 0.5 - - - NN Alex Net, DC - 39.1 1.2 - - - NN Alex Net, SLDD - 39.8 0.8 - - - NN Alex Net, DD - 36.8 1.2 - - - KRR FC1 NNGP, LS 27.5 0.3 40.1 0.3 46.4 0.4 53.5 0.2 55.1 0.3 KRR Myrtle-10 NNGP, LS + ZCA5 31.7 0.2 56.0 0.5 69.8 0.1 80.2 0.1 82.3 0.1 KRR Myrtle-10, LS 28.8 0.4 45.8 0.7 58.0 0.3 69.6 0.2 72.0 0.2 NN Alex Net, LD 25.69 0.72 38.33 0.44 43.16 0.47 - - 5 We apply regularized ZCA whitening instead of standard preprocessing to the images, see Appendix D for further details. markably well, with both uncorrupted and corrupted KIP images beating the uncorrupted natural images of corresponding size. Out of distribution networks (Le Net (Le Cun et al., 1998) and Wide Resnet (Zagoruyko & Komodakis, 2016)) have less transferability: the uncorrupted images still outperform natural images, and corrupted KIP images still outperform corrupted natural images, but corrupted KIP images no longer outperform uncorrupted natural images. (ii) We consider larger KIP datasets (1K, 5K, 10K images) optimized using a single FC1 kernel for training of a corresponding FC1 neural network, where the KIP training uses augmentations (with and without label learning), see Tables A7-A10 and Figure A3. We find, as before, KIP images outperform natural images by an impressive margin: for instance, on CIFAR-10, 10K KIP -learned images with 90% corruption achieves 49.9% test accuracy, exceeding 10K natural images with no corruption (acc: 45.5%) and 90% corruption (acc: 33.8%). Interestingly enough, sometimes higher corruption leads to better test performance (this occurs for CIFAR-10 with cross entropy loss for both natural and KIP -learned images), a phenomenon to be explored in future work. We also find that KIP with label-learning often tends to harm performance, perhaps because the labels are overfitting to KRR. Robustness to hyperparameters: For CIFAR-10, we took 100 images, both clean and 90% corrupted, and trained networks on a wide variety of hyperparameters for various neural architectures. We considered both neural networks whose corresponding neural kernels were sampled during KIP - training those that were not. We found that in both cases, the KIP -learned images almost always outperform 100 random natural images, with the optimal set of hyperparameters yielding a margin close to that predicted from the KRR setting, see Figure 3. This suggests that KIP -learned images can be useful in accelerating hyperparameter search. 5 RELATED WORK Coresets: A classical approach for compressing datasets is via subset selection, or some approximation thereof. One notable work is Borsos et al. (2020), utilizing KRR for dataset subselection. For an overview of notions of coresets based on pointwise approximatation of datasets, see Phillips (2016). Neural network approaches to dataset distillation: Maclaurin et al. (2015); Lorraine et al. (2020) approach dataset distillation through learning the input images from large-scale gradient-based metalearning of hyper-parameters. Properties of distilled input data was first analyzed in Wang et al. (2018). The works Sucholutsky & Schonlau (2019); Bohdal et al. (2020) build upon Wang et al. 0.1 0.2 0.3 0.4 Original Test Accuracy KIP Test Accuracy FC3 (In Distribution) 0.1 0.2 0.3 0.4 Original Test Accuracy Conv2 (In Distribution) 0.1 0.2 0.3 0.4 Original Test Accuracy Conv8 (Out-of-Distribution) 0.1 0.2 0.3 0.4 Original Test Accuracy WRes Net (Out-of-Distribution) 0.2 0.4 Corrupted Input Test Accuracy KIP Test Accuracy FC3 90% Corruption (In Distribution) 0.2 0.4 Corrupted Input Test Accuracy Conv2 90% Corruption (In Distribution) 0.2 0.4 Corrupted Input Test Accuracy Conv8 90% Corruption (Out-of-Distribution) 0.2 0.4 Corrupted Input Test Accuracy WRes Net 90% Corruption (Out-of-Distribution) Figure 3: KIP learned images transfers well to finite neural networks. Test accuracy on CIFAR10 comparing natural images (x-axis) and KIP -learned images (y-axis). Each scatter point corresponds to varying hyperparameters for training (e.g. learning rate). Top row are clean images, bottom row are 90% corrupted images. KIP images were trained using FC1-3, Conv1-2 kernels. (2018) by distilling labels. More recently, Zhao et al. (2020) proposes condensing training set by gradient matching condition and shows improvement over Wang et al. (2018). Inducing points: Our approach has as antecedant the inducing point method for Gaussian Processes (Snelson & Ghahramani, 2006; Titsias, 2009). However, whereas the latter requires a probabilistic framework that optimizes for marginal likelihood, in our method we only need to consider minimizing mean-square loss on validation data. Low-rank kernel approximations: Unlike common low-rank approximation methods (Williams & Seeger, 2001; Drineas & Mahoney, 2005), we obtain not only a low-rank support-support kernel matrix with KIP , but also a low-rank target-support kernel matrix. Note that the resulting matrices obtained from KIP need not approximate the original support-support or target-support matrices since KIP only optimizes for the loss function. Neural network kernels: Our work is motivated by the exact correspondence between infinitelywide neural networks and kernel methods (Neal, 1994; Lee et al., 2018; Matthews et al., 2018; Jacot et al., 2018; Novak et al., 2019; Garriga-Alonso et al., 2019; Arora et al., 2019a). These correspondences allow us to view both Bayesian inference and gradient descent training of wide neural networks with squared loss as yielding a Gaussian process or kernel ridge regression with neural kernels. Instance-Based Encryption: A related approach to corrupting datasets involves encrypting individual images via sign corruption (Huang et al. (2020)). 6 CONCLUSION We introduced novel algorithms KIP and LS for the meta-learning of datasets. We obtained a variety of compressed and corrupted datasets, achieving state of the art results for KRR and neural network dataset distillation methods. This was achieved even using the simplest of kernels and neural networks (shallow fully-connected networks and purely-convolutional networks without pooling), which notwithstanding their limited expressiveness, outperform most baselines that use deeper architectures. Follow-up work will involve scaling up KIP to deeper architectures with pooling (achievable with multi-device training) for which we expect to obtain even more highly performant datasets, both in terms of overall accuracy and architectural flexibility. Finally, we obtained highly corrupt datasets whose performance match or exceed natural images, which when developed at scale, could lead to practical applications for privacy-preserving machine learning. ACKNOWLEDGMENTS We would like to thank Dumitru Erhan, Yang Li, Hossein Mobahi, Jeffrey Pennington, Si Si, Jascha Sohl-Dickstein, and Lechao Xiao for helpful discussions and references. Martin Abadi, Andy Chu, Ian Goodfellow, H. Brendan Mc Mahan, Ilya Mironov, Kunal Talwar, and Li Zhang. Deep learning with differential privacy. Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security, Oct 2016. doi: 10.1145/2976749.2978318. URL http://dx.doi.org/10.1145/2976749.2978318. Sanjeev Arora, Simon S Du, Wei Hu, Zhiyuan Li, Russ R Salakhutdinov, and Ruosong Wang. On exact computation with an infinitely wide neural net. In Advances in Neural Information Processing Systems, pp. 8141 8150. Curran Associates, Inc., 2019a. Sanjeev Arora, Simon S Du, Zhiyuan Li, Ruslan Salakhutdinov, Ruosong Wang, and Dingli Yu. Harnessing the power of infinitely wide deep nets on small-data tasks. ar Xiv preprint ar Xiv:1910.01663, 2019b. Ondrej Bohdal, Yongxin Yang, and Timothy Hospedales. Flexible dataset distillation: Learn labels instead of images. ar Xiv preprint ar Xiv:2006.08572, 2020. Antoine Bordes, Seyda Ertekin, Jason Weston, and L eon Bottou. Fast kernel classifiers with online and active learning. Journal of Machine Learning Research, 6(Sep):1579 1619, 2005. Zal an Borsos, Mojm ır Mutn y, and Andreas Krause. Coresets via bilevel optimization for continual learning and streaming. ar Xiv preprint ar Xiv:2006.03875, 2020. James Bradbury, Roy Frostig, Peter Hawkins, Matthew James Johnson, Chris Leary, Dougal Maclaurin, and Skye Wanderman-Milne. JAX: composable transformations of Python+Num Py programs, 2018. URL http://github.com/google/jax. Petros Drineas and Michael W Mahoney. On the nystr om method for approximating a gram matrix for improved kernel-based learning. journal of machine learning research, 6(Dec):2153 2175, 2005. Adri a Garriga-Alonso, Laurence Aitchison, and Carl Edward Rasmussen. Deep convolutional networks as shallow gaussian processes. In International Conference on Learning Representations, 2019. T. N. E. Greville. Note on the generalized inverse of a matrix product. SIAM Review, 8(4):518 521, 1966. ISSN 00361445. URL http://www.jstor.org/stable/2027337. Trevor Hastie, Andrea Montanari, Saharon Rosset, and Ryan J. Tibshirani. Surprises in highdimensional ridgeless least squares interpolation, 2019. Yangsibo Huang, Zhao Song, Kai Li, and Sanjeev Arora. Instahide: Instance-hiding schemes for private distributed learning, 2020. Arthur Jacot, Franck Gabriel, and Clement Hongler. Neural tangent kernel: Convergence and generalization in neural networks. In Advances in Neural Information Processing Systems, 2018. Ibrahim Jubran, Alaa Maalouf, and Dan Feldman. Introduction to coresets: Accurate coresets. ar Xiv preprint ar Xiv:1910.08707, 2019. T. Kato. Perturbation Theory of Linear Operators. Springer-Verlag, Berlin, 2 edition, 1976. Mahmut Kaya and H.s Bilge. Deep metric learning: A survey. Symmetry, 11:1066, 08 2019. doi: 10.3390/sym11091066. Alex Krizhevsky, Geoffrey Hinton, et al. Learning multiple layers of features from tiny images. 2009. Yann Le Cun, L eon Bottou, Yoshua Bengio, and Patrick Haffner. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278 2324, 1998. Yann Le Cun, Corinna Cortes, and CJ Burges. Mnist handwritten digit database. ATT Labs [Online]. Available: http://yann.lecun.com/exdb/mnist, 2, 2010. Jaehoon Lee, Yasaman Bahri, Roman Novak, Sam Schoenholz, Jeffrey Pennington, and Jascha Sohldickstein. Deep neural networks as gaussian processes. In International Conference on Learning Representations, 2018. Jaehoon Lee, Lechao Xiao, Samuel S. Schoenholz, Yasaman Bahri, Roman Novak, Jascha Sohl Dickstein, and Jeffrey Pennington. Wide neural networks of any depth evolve as linear models under gradient descent. In Advances in Neural Information Processing Systems, 2019. Jaehoon Lee, Samuel S Schoenholz, Jeffrey Pennington, Ben Adlam, Lechao Xiao, Roman Novak, and Jascha Sohl-Dickstein. Finite versus infinite neural networks: an empirical study. ar Xiv preprint ar Xiv:2007.15801, 2020. Jonathan Lorraine, Paul Vicol, and David Duvenaud. Optimizing millions of hyperparameters by implicit differentiation. In International Conference on Artificial Intelligence and Statistics, pp. 1540 1552. PMLR, 2020. Dougal Maclaurin, David Duvenaud, and Ryan Adams. Gradient-based hyperparameter optimization through reversible learning. In International Conference on Machine Learning, pp. 2113 2122, 2015. Julien Mairal, Piotr Koniusz, Zaid Harchaoui, and Cordelia Schmid. Convolutional kernel networks. In Advances in Neural Information Processing Systems, 2014. Alexander G. de G. Matthews, Jiri Hron, Mark Rowland, Richard E. Turner, and Zoubin Ghahramani. Gaussian process behaviour in wide deep neural networks. In International Conference on Learning Representations, 2018. Radford M. Neal. Priors for infinite networks (tech. rep. no. crg-tr-94-1). University of Toronto, 1994. Roman Novak, Lechao Xiao, Jaehoon Lee, Yasaman Bahri, Greg Yang, Jiri Hron, Daniel A. Abolafia, Jeffrey Pennington, and Jascha Sohl-Dickstein. Bayesian deep convolutional networks with many channels are gaussian processes. In International Conference on Learning Representations, 2019. Roman Novak, Lechao Xiao, Jiri Hron, Jaehoon Lee, Alexander A. Alemi, Jascha Sohl-Dickstein, and Samuel S. Schoenholz. Neural tangents: Fast and easy infinite neural networks in python. In International Conference on Learning Representations, 2020. URL https://github.com/ google/neural-tangents. Jeff M Phillips. Coresets and sketches. ar Xiv preprint ar Xiv:1601.00617, 2016. Vaishaal Shankar, Alex Chengyu Fang, Wenshuo Guo, Sara Fridovich-Keil, Ludwig Schmidt, Jonathan Ragan-Kelley, and Benjamin Recht. Neural kernels without tangents. In International Conference on Machine Learning, 2020. Sam Shleifer and Eric Prokop. Using small proxy datasets to accelerate hyperparameter search. ar Xiv preprint ar Xiv:1906.04887, 2019. Jake Snell, Kevin Swersky, and Richard Zemel. Prototypical networks for few-shot learning. In I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett (eds.), Advances in Neural Information Processing Systems 30, pp. 4077 4087. Curran Associates, Inc., 2017. URL http://papers.nips.cc/paper/ 6996-prototypical-networks-for-few-shot-learning.pdf. Edward Snelson and Zoubin Ghahramani. Sparse gaussian processes using pseudo-inputs. In Advances in neural information processing systems, pp. 1257 1264, 2006. Jascha Sohl-Dickstein, Roman Novak, Samuel S Schoenholz, and Jaehoon Lee. On the infinite width limit of neural networks with a standard parameterization. ar Xiv preprint ar Xiv:2001.07301, 2020. Ingo Steinwart. Sparseness of support vector machines. Journal of Machine Learning Research, 4 (Nov):1071 1105, 2003. Ilia Sucholutsky and Matthias Schonlau. Soft-label dataset distillation and text dataset distillation. ar Xiv preprint ar Xiv:1910.02551, 2019. Michalis Titsias. Variational learning of inducing variables in sparse gaussian processes. In Artificial Intelligence and Statistics, pp. 567 574, 2009. Oriol Vinyals, Charles Blundell, Timothy P. Lillicrap, Koray Kavukcuoglu, and Daan Wierstra. Matching networks for one shot learning. Co RR, abs/1606.04080, 2016. URL http://arxiv. org/abs/1606.04080. Tongzhou Wang, Jun-Yan Zhu, Antonio Torralba, and Alexei A Efros. Dataset distillation. ar Xiv preprint ar Xiv:1811.10959, 2018. Christopher KI Williams and Matthias Seeger. Using the nystr om method to speed up kernel machines. In Advances in neural information processing systems, pp. 682 688, 2001. Sergey Zagoruyko and Nikos Komodakis. Wide residual networks. Co RR, abs/1605.07146, 2016. URL http://arxiv.org/abs/1605.07146. Bo Zhao, Konda Reddy Mopuri, and Hakan Bilen. Dataset condensation with gradient matching. ar Xiv preprint ar Xiv:2006.05929, 2020. A REMARKS ON DEFINITION OF ϵ-APPROXIMATION Here, we provide insights into the formulation of Definition 3. One noticeable feature of our definition is that it allows for different algorithms A and A when comparing datasets D and D. On the one hand, such flexibility is required, since for instance, a mere preprocessing of the dataset (e.g. rescaling it), should be regarded as producing an equivalent (0-approximate) dataset. Yet such a rescaling may affect the hyperparameters needed to train an equivalent model (e.g. the learning rate). Thus, one must allow the relevant hyperparameters of an algorithm to vary when the datasets are also varying. On the other hand, it would be impossible to compare two datasets meaningfully if the learned algorithms used to train them differ too significantly. For instance, if D is a much larger dataset than D, but A is a much less expressive algorithm than A, then the two datasets may be ϵ-approximations of each other, but it would be strange to compare D and D in this way. Thus, we treat the notion of what class of algorithms to consider informally, and leave its specification as a practical matter for each use case. In practice, the pair of algorithms we use to compare datasets should be drawn from a family in which some reasonable range of hyperparameters are varied, the ones typically tuned when learning on an unknown dataset. The main case for us with differing A and A is when we compare neural network training alongside kernel ridge-regression. Another key feature of our definition is that datapoints of an ϵ-approximating dataset must have the same shape as those of the original dataset. This makes our notion of an ϵ-approximate dataset more restrictive than returning a specialized set of extracted features from some initial dataset. Analogues of our ϵ-approximation definition have been formulated in the unsupervised setting, e.g. in the setting of clustering data (Phillips, 2016; Jubran et al., 2019). Finally, note that the loss function ℓused for comparing datasets does not have to coincide with any loss functions optimized in the learning algorithms A and A. Indeed, for kernel ridge-regression, training mimimizes mean square loss while ℓcan be 0-1 loss. B TUNING KIP Sampling: When optimizing for KRR performance with support dataset size N, it is best to learn a support set D of size N and sample this entire set during KIP training. It is our observation that subsets of size M < N of D will not perform as well as optimizing directly for a size M dataset through KIP . Conversely, sampling subsets of size M from a support dataset of size N during KIP will not lead to a dataset that does as well as optimizing for all N points. This is sensible: optimizing for small support size requires a resourceful learning of coarse features at the cost of learning fine-grained features from many support datapoints. Conversely, optimizing a large support set means the learned support set has leveraged higher-order information, which will degrade when restricted to smaller subsets. For sampling from the target set, which we always do in a class-balanced way, we found larger batch sizes typically perform better on the test set if the train and test kernels agree. If the train and test kernels differ, then smaller batch sizes lead to less overfitting to the train kernel. Initialization: We tried two sets of initializations. The first ( image init ) initializes (Xs, ys) to be a subset of (Xt, yt). The second ( noise init ) initializes Xs with uniform noise and ys with mean-centered, one-hot labels (in a class-balanced way). We found image initialization to perform better. Regularization: The regularization parameter λ in (7) can be replaced with 1 nλ tr(KXs Xs), where n is the number of datapoints in Xs. This makes the loss function invariant with respect to rescaling of the kernel function K and also normalizes the regularization with respect to support size. In practice, we use this scale-invariant regularization with λ = 10 6. Number of Training Iterations: Remarkably, KIP converges very quickly in all experimental settings we tried. After only on the order of a hundred iterations, independently of the support size, kernel, and corruption factor, the learned support set has already undergone the majority of its learning (test accuracy is within more than 90% of the final test accuracy). For the platforms available to us, using a single V100 GPU, one hundred training steps for the experiments we ran involving target batch sizes that were a few thousand takes on the order of about 10 minutes. When we add augmentations to our targets, performance continues to improve slowly over time before flattening out after several thousands of iterations. C THEORETICAL RESULTS Here, we analyze convergence properties of KIP in returning an ϵ-approximate dataset. In what follows, we refer to gradient-descent KIP as the case when we sample from the entire support and train datasets for each update step to KIP . We also assume that the distribution P used to evaluate ϵ-approximation is supported on inputs x Rd with x 1 (merely to provide a convenient normalization when evaluating loss on regression algorithms). For the case of a linear kernel, we prove the below convergence theorem: Theorem 1. Let D = (Xt, yt) Rnt d Rnt C be an arbitrary dataset. Let wλ Rd C be the coefficients obtained from training λ ridge-regression (λ-RR) on (Xt, yt), as given by (4). 1. For generic6 initial conditions for the support set (Xs, ys) Rns d Rns C and sufficiently small λ > 0, gradient descent KIP with target dataset D converges to a dataset D. 2. The dataset D is a strong ϵ-approximation to D with respect to algorithms (λ-RR, 0-RR) and loss function equal to mean-square loss, where 2 w w0 2 2 (A1) and w Rd C are the coefficients of the linear classifier obtained from training λ-RR on D. If the size of D is at least C, then w is also a least squares classifier for D. In particular, if D has a unique least squares classifier, then ϵ = 0. Proof. We discuss the case where Xs is optimized, with the case where both (Xs, ys) are optimized proceeding similarly. In this case, by genericity, we can assume ys = 0, else the learning dynamics is trivial. Furthermore, to simplify notation for the time being, assume the dimensionality of the label space is C = 1 without loss of generality. First, we establish convergence. For a linear kernel, we can write our loss function as 2 yt Xt XT s (Xs XT s + λI) 1ys 2, (A2) defined on the space Mns d of ns d matrices. It is the pullback of the loss function LRd ns (Φ) = 1 2 yt XtΦys 2, Φ Rd ns (A3) under the map Xs 7 Φλ(Xs) = XT s (Xs XT s + λI) 1. The function (A3) is quadratic in Φ and all its local minima are global minima given by an affine subspace M Md ns. Moreover, each point of M has a stable manifold of maximal dimension equal to the codimension of M. Thus, the functional L has global minima given by the inverse image Φ 1 λ (M) (which will be nonempty for sufficiently small λ). Next, we claim that given a fixed initial (Xs, ys), then for sufficiently small λ, gradient-flow of (A2) starting from (Xs, ys) cannot converge to a non-global local minima. We proceed as follows. If X = UΣV T is a singular value decomposition of X, with Σ a ns ns diagional matrix of singular values (and any additional zeros for padding), then Φ(X) = V φ(Σ)U T where φ(Σ) denotes the diagonal matrix with the map φ : R 0 R 0 (A4) φ(µ) = µ µ2 + λ (A5) 6A set can be generic by either being open and dense or else having probability one with respect to some measure absolutely continuous with respect to Lebesgue measure. In our particular case, generic refers to the complement of a submanifold of codimension at least one. applied to each singular value of Σ. The singular value decomposition depends analytically on X (Kato (1976)). Given that φ : R 0 R 0 is a local diffeomorphism away from its maximum value at µ = µ := λ1/2, it follows that Φλ : Mns d Md ns is locally surjective, i.e. for every X, there exists a neighborhood U of X such that Φλ(U) contains a neighborhood of Φλ(X). Thus, away from the locus of matrices in Mns d that have a singular value equaling µ , the function (A2) cannot have any non-global local minima, since the same would have to be true for (A3). We are left to consider those matrices with some singular values equaling µ . Note that as λ 0, we have φ(µ ) . On the other hand, for any initial choice of Xs, the matrices Φλ(Xs) have uniformly bounded singular values as a function of λ. Moreover, as Xs = Xs(t) evolves, Φλ(Xs(t)) never needs to be larger than some large constant times Φλ(Xs(0)) + yt µ+ ys , where µ+ is the smallest positive singular value of Xt. Consequently, Xs(t) never visits a matrix with singular value µ for sufficiently small λ > 0; in particular, we never have to worry about convergence to a non-global local minimum. Thus, a generic gradient trajectory γ of L will be such that Φλ γ is a gradient-like7 trajectory for LRd ns that converges to M. We have to show that γ itself converges. It is convenient to extend φ to a map defined on the one-point compactification [0, ] R 0, so as to make φ a two-to-one map away from µ . Applying this compactification to every singular value, we obtain a compactification M ns d of Mns d, and we can naturally extend Φλ to such a compactification. We have that γ converges to an element of M := Φ 1 λ (M) M ns d, where we need the compactification to account for the fact that when Φλ γ converges to a matrix that has a zero singular value, γ may have one of its singular values growing to infinity. Let M0 denote the subset of M with a zero singular value. Then γ converges to an element of Mns d precisely when γ does not converge to an element of M := Φ 1 λ (M0) (M ns d \Mns d). However, M0 M has codimension one and hence so does M Φ 1 λ (M0). Thus, the stable set to M has codimension one in M ns d, and hence its complement is nongeneric. Hence, we have generic convergence of a gradient trajectory of L to a (finite) solution. This establishes the convergence result of Part 1. For Part 2, the first statement is a general one: the difference of any two linear models, when evaluated on P, can be pointwise bounded by the spectral norm of the difference of the model coefficient matrices. Thus D is a strong ϵ-approximation to D with respect to (λ-RR, 0-RR) where ϵ is given by (A1). For the second statement, observe that L is also the pullback of the loss function LRd C(w) = 1 2 yt Xtw 2, w Rd C. (A6) under the map Xs 7 w(Xs) = Φλ(Xs)ys. The function LRd C(w) is quadratic in w and has a unique minimum value, with the space of global minima being an affine subspace W of Rd given by the least squares classifiers for the dataset (Xt, yt). Thus, the global minima of L are the preimage of W under the map w(Xs). For generic initial (Xs, ys), we have ys Rns C is full rank. This implies, for ns C, that the range of all possible w(Xs) for varying Xs is all of Rd C, so that the minima of (A6) and (A3) coincide. This implies the final parts of Part 2. We also have the following result about ϵ-approximation using the label solve algorithm: Theorem 2. Let D = (Xt, yt) Rnt d Rnt C be an arbitrary dataset. Let wλ Rd C be the coefficients obtained from training λ ridge-regression (λ-RR) on (Xt, yt), as given by (4). Let Xs Rns d be an arbitrary initial support set and let λ 0. Define y s = y s(λ) via (8). Then (Xs, y s) yields a strong ϵ(λ)-approximation of (Xt, yt) with respect to algorithms (λ-RR, 0-RR) and mean-square loss, where 2 w (λ) w0 2 2 (A7) and w (λ) is the solution to w (λ) = argminw W yt Xtw 2, W = im Φλ(Xs) : ker XtΦλ(Xs) Rd C . 7A vector field v is gradient-like for a function f if v grad(f) 0 everywhere. Moreover, for λ = 0, if rank(Xs) = rank(Xt) = d, then w (λ) = w0. This implies y s = Xsw0, i.e. y s coincides with the predictions of the 0-RR classifier trained on (Xt, yt) evaluated on Xs. Proof. By definition, y s is the minimizer of 2 yt XtΦλ(Xs)ys 2, with minimum norm. This implies y s ker XtΦλ(Xs) and that w (λ) = Φλ(Xs)y s satisfies (A8). At the same time, w (λ) = Φλ(Xs)y s are the coefficients of the λ-RR classifier trained on (Xs, y s). If rank(Xs) = rank(Xt) = d, then Φ0(Xs) is surjective and Xt is injective, in which case ω (0) = Φ0(Xs)y s = Φ0(Xs)Φ0(XtΦ0(Xs))yt = Φ0(Xs)XsΦ0(Xt)yt = ω0. The results follow. For general kernels, we make the following simple observation concerning the optimal output of KIP . Theorem 3. Fix a target dataset (Xt, yt). Consider the family of all subspaces S of Rnt given by {im KXt Xs : Xs Rns d}, i.e. all possible column spaces of KXt Xs. Then the infimum of the loss (7) over all possible (Xs, ys) is equal to inf S S 1 2 Π S yt 2 where Π S is orthogonal projection onto the orthogonal complement of S (acting identically on each label component). Proof. Since ys is trainable, (KXs Xs + λ) 1ys is an arbitrary vector in Rns C. Thus, minimizing the training objective corresponds to maximizing the range of the linear map KXt Xs over all possible Xs. The result follows. D EXPERIMENT DETAILS In all KIP trainings, we used the Adam optimizer. All our labels are mean-centered 1-hot labels. We used learning rates 0.01 and 0.04 for the MNIST and CIFAR-10 datasets, respectively. When sampling target batches, we always do so in a class-balanced way. When augmenting data, we used the Image Generator class from Keras, which enables us to add horizontal flips, height/width shift, rotatations (up to 10 degrees), and channel shift (for CIFAR-10). All datasets are preprocessed using channel-wise standardization (i.e. mean subtraction and division by standard-deviation). For neural (tangent) kernels, we always use weight and bias variance σ2 w = 2 and σ2 b = 10 4, respectively. For both neural kernels and neural networks, we always use Re LU activation. Convolutional layers all use a (3, 3) filter with stride 1 and same padding. Compute Limitations: Our neural kernel computations, implemented using Neural Tangents libraray (Novak et al., 2020) are such that computation scales (i) linearly with depth; (ii) quadratically in the number of pixels for convolutional kernels; (iii) quartically in the number of pixels for pooling layers. Such costs mean that, using a single V100 GPU with 16GB of RAM, we were (i) only able to sample shallow kernels; (ii) for convolutional kernels, limited to small support sets and small target batch sizes; (iii) unable to use pooling if learning more than just a few images. Scaling up KIP to deeper, more expensive architectures, achievable using multi-device training, will be the subject of future exploration. Kernel Parameterization: Neural tangent kernels, or more precisely each neural network layer of such kernels, as implemented in Novak et al. (2020) can be parameterized in either the NTK parameterization or standard parameterization Sohl-Dickstein et al. (2020). The latter depends on the width of a corresponding finite-width neural network while the former does not. Our experiments mix both these parameterizations for variety. However, because we use a scale-invariant regularization for KRR (Section B), the choice of parameterization has a limited effect compared to other more significant hyperparameters (e.g. the support dataset size, learning rate, etc.)8. Single kernel results: (Tables 1 and 2) For FC, we used kernels with NTK parametrization. For RBF, our rbf kernel is given by rbf(x1, x2) = exp( γ x1 x2 2/d) (A9) where d is the dimension of the inputs and γ = 1. We found that treating γ as a learnable parameter during KIP had mixed results9 and so keep it fixed for simplicity. For MNIST, we found target batch size equal to 6K sufficient. For CIFAR-10, it helped to sample the entire training dataset of 50K images per step (hence, along with sampling the full support set, we are doing full gradient descent training). When support dataset size is small or if augmentations are employed, there is no overfitting (i.e. the train and test loss/accuracy stay positively correlated). If the support dataset size is large (5K or larger), sometimes there is overfitting when the target batch size is too large (e.g. for the RBF kernel on CIFAR10, which is why we exclude in Table 2 the entries for 5K and 10K). We could have used a validation dataset for a stopping criterion, but that would have required reducing the target dataset from the entire training dataset. We train KIP for 10-20k iterations and took 5 random subsets of images for initializations. For each such training, we took 5 checkpoints with lowest train and loss and computed the test accuracy. This gives 25 evaluations, for which we can compute the mean and standard deviation for our test accuracy numbers in Tables 1 and 2. Kernel transfer results: For transfering of KIP images, both to other kernels and to neural networks, we found it useful to use smaller target batch sizes (either several hundred or several thousand), else the images overfit to their source kernel. For random sampling of kernels used in Figure A1 and producing datasets for training of neural networks, we used FC kernels with width 1024 and Conv kernels with width 128, all with standard parametrization. Neural network results: Neural network trainings on natural data with mean-square loss use meancentered one-hot labels for consistency with KIP trainings. For cross entropy loss, we use one-hot labels. For neural network trainings on KIP -learned images with label learning, we transfer over the labels directly (as with the images), whatever they may be. For neural network transfer experiments occurring in Table 1, Table 2, Figure 3, Table A5, and Table A6, we did the following. First, the images were learned using kernels FC1-3, Conv1-2. Second, we trained for a few hundred iterations, after which optimal test performance was achieved. On MNIST images, we trained the networks with constant learning rate and Adam optimizer with cross entropy loss. Learning rate was tuned over small grid search space. For the FC kernels and networks, we use width of 1024. On CIFAR-10 images, we trained the networks with constant learning rate, momentum optimizer with momentum 0.9. Learning rate, L2 regularization, parameterization (standard vs NTK) and loss type (mean square, softmax-cross-entropy) was tuned over small grid search space. Vanilla networks use constant width at each layer: for FC we use width of 1024, for Conv2 we use 512 channels, and for Conv8 we use 128 channels. No pooling layers are used except for the Wide Res Net architecture, where we follow the original architecture of Zagoruyko & Komodakis (2016) except that our batch normalization layer is stateless (i.e. no exponential moving average of batch statistics are recorded). For neural network transfer experiments in Figure A3, Tables A7-A10, we did the following. Our KIP -learned images were trained using only an FC1 kernel. The neural network FC1 has an increased width 4096, which helps with the larger number of images. We used learning rate 4 10 4 and the Adam optimizer. The KIP learned images with only augmentations used target batch size equal to half the training dataset size and were trained for 10k iterations, since the use of augmentations allows for continued gains after longer training. The KIP learned images with augmentations 8All our final readout layers use the fixed NTK parameterization and all our statements about which parameterization we are using should be interpreted accordingly. This has no effect on the training of our neural networks while for kernel results, this affects the recursive formula for the NTK at the final layer if using standard parameterization (by the changing the relative scales of the terms involved). Since the train/test kernels are consistently parameterized and KIP can adapt to the scale of the kernel, the difference between our hybrid parameterization and fully standard parameterization has a limited affect. 9On MNIST it led to very slight improvement. For CIFAR10, for small support sets, the effect was a small improvement on the test set, whereas for large support sets, we got worse performance. Table A1: Accuracy on random subsets of MNIST. Standard deviations over 20 resamplings. # Images \ Kernel Linear RBF FC1 10 44.6 3.7 45.3 3.9 45.8 3.9 20 51.9 3.1 54.6 2.9 54.7 2.8 40 59.4 2.4 66.9 2.0 66.0 1.9 80 62.6 2.7 75.6 1.6 74.3 1.7 160 62.2 2.1 82.7 1.4 81.1 1.6 320 52.3 1.9 88.1 0.8 86.9 0.9 640 41.9 1.4 91.8 0.5 91.1 0.5 1280 71.0 0.9 94.2 0.3 93.6 0.3 2560 79.7 0.5 95.7 0.2 95.3 0.2 5000 83.2 0.4 96.8 0.2 96.4 0.2 10000 84.9 0.4 97.5 0.2 97.2 0.2 and label learning used target batch size equal to a tenth of the training dataset size and were trained for 2k iterations (the learned data were observed to overfit to the kernel and have less transferability if larger batch size were used or if trainings were carried out longer). All neural network trainings were run with 5 random initializations to compute mean and standard deviation of test accuracies. In Table 2, regularized ZCA preprocessing was used for a Myrtle-10 kernel (denoted with ZCA) on CIFAR-10 dataset. Shankar et al. (2020) and Lee et al. (2020) noticed that for neural (convolutional) kernels on image classification tasks, regularized ZCA preprocessing can improve performance significantly compared to standard preprocessing. We follow the prepossessing scheme used in Shankar et al. (2020), with regularization strength of 10 5 without augmentation. E TABLES AND FIGURES E.1 KERNEL BASELINES We report various baselines of KRR trained on natural images. Tables A1 and A2 shows how various kernels vary in performance with respect to random subsets of MNIST and CIFAR-10. Linear denotes a linear kernel, RBF denotes the rbf kernel (A9) with γ = 1, and FC1 uses standard parametrization and width 1024. Interestingly enough, we observe non-monotonicity for the linear kernel, owing to double descent phenomenon Hastie et al. (2019). We include additional columns for deeper kernel architectures in Table A2, taken from Shankar et al. (2020) for reference. Comparing Tables 1, 2 with Tables A1, A2, we see that 10 KIP -learned images, for both RBF and FC1, has comparable performance to several thousand natural images, thereby achieving a compression ratio of over 100. This compression ratio narrows as the support size increases towards the size of the training data. Next, Table A3 compares FC1, RBF, and other kernels trained on all of MNIST to FC1 and RBF trained on KIP -learned images. We see that our KIP approach, even with 10K images (which fits into memory), leads to RBF and FC1 matching the performance of convolutional kernels on the original 60K images. Table A4 shows state of the art of FC kernels on CIFAR-10. The prior state of the art used kernel ensembling on batches of augmented data in Lee et al. (2020) to obtain test accuracy of 61.5% (32 ensembles each of size 45K images). By distilling augmented images using KIP , we are able to obtain 64.7% test accuracy using only 10K images. E.2 KIP AND LS TRANSFER ACROSS KERNELS Figure A1 plots how KIP (with only images learned) performs across kernels. There are seven training scenarios: training individually on FC1, FC2, FC3, Conv1, Conv2, Conv3 NTK kernels and random sampling from among all six kernels uniformly (Avg All). Datasets of size 10, 100, 200 Table A2: Accuracy on random subsets of CIFAR-10. Standard deviations over 20 resamplings. # Images \ Kernel Linear RBF FC1 CNTK Myrtle10-G 10 16.2 1.3 15.7 2.1 16.4 1.8 15.33 2.43 19.15 1.94 20 17.1 1.6 17.1 1.7 18.0 1.9 18.79 2.13 21.65 2.97 40 17.8 1.6 19.7 1.8 20.6 1.8 21.34 1.91 27.20 1.90 80 18.6 1.5 23.0 1.5 23.9 1.6 25.48 1.91 34.22 1.08 160 18.5 1.4 25.8 1.4 26.5 1.4 30.48 1.17 41.89 1.34 320 18.1 1.1 29.2 1.2 29.9 1.1 36.57 0.88 50.06 1.06 640 16.8 0.8 32.8 0.9 33.4 0.8 42.63 0.68 57.60 0.48 1280 15.1 0.5 35.9 0.7 36.7 0.6 48.86 0.68 64.40 0.48 2560 13.0 0.5 39.1 0.7 40.2 0.7 - - 5000 17.8 0.4 42.1 0.5 43.7 0.6 - - 10000 24.9 0.6 45.3 0.6 47.7 0.6 - - Conv14 kernel with global average pooling (Arora et al., 2019b) Myrtle10-Gaussian kernel (Shankar et al., 2020) Table A3: Classification performance on MNIST. Our KIP -datasets, fit to FC1 or RBF kernels, outperform non-convolutional kernels trained on all training images. Kernel Method Accuracy FC1 Base1 98.6 Arc Cosine Kernel2 Base 98.8 Gaussian Kernel Base 98.8 FC1 KIP (a+l)3, 10K images 99.2 Le Net-5 (Le Cun et al., 1998) Base 99.2 RBF KIP (a+l), 10K images 99.3 Myrtle5 Kernel (Shankar et al., 2020) Base 99.5 CKN (Mairal et al., 2014) Base 99.6 1 Base refers to training on entire training dataset of natural images. 2 Non RBF/FC numbers taken from (Shankar et al., 2020) 3 (a + l) denotes KIP with augmentations and label learning during training. are thereby trained then evaluated by averaging over all of FC1-3, Conv1-3, both with the NTK and NNGP kernels for good measure. Moreover, the FC and Conv train kernel widths (1024 and 128) were swapped at test time (FC width 128 and Conv width 1024), as an additional test of robustness. The average performance is recorded along the y-axis. Avg All leads to overall boost in performance across kernels. Another observation is that Conv kernels alone tend to do a bit better, averaged over the kernels considered, than FC kernels alone. Avg All FC1 FC2 FC3 Conv1 Conv2 Conv3 0.0 Average Test Accuracy 0.39 0.36 0.35 0.37 0.40 0.39 0.38 0.37 0.40 0.41 0.40 KIP Transfer to the Other Kernels Support Set Size 10 100 200 Figure A1: Studying transfer between kernels. Table A4: CIFAR-10 test accuracy for FC/RBF kernels. Our KIP -datasets, fit to RBF/FC1, outperform baselines with many more images. Notation same as in Table A3. Kernel Method Accuracy FC1 Base 57.6 FC3 Ensembling (Lee et al., 2020) 61.5 FC1 KIP (a+l), 10k images 64.7 RBF Base 52.7 RBF KIP (a+l), 10k images 66.3 100 102 104 Support Set Size Test Accuracy Myrtle-10 -> FC: 1k target 100 102 104 Support Set Size Myrtle-10 -> FC: 2k target 100 102 104 Support Set Size Myrtle-10 -> FC: 5k target 100 102 104 Support Set Size Myrtle-10 -> FC: 10k target 100 102 104 Support Set Size Myrtle-10 -> FC: 50k target 100 102 104 Support Set Size Test Accuracy FC -> Myrtle-10: 1k target 100 102 104 Support Set Size FC -> Myrtle-10: 5k target 100 102 104 Support Set Size FC -> Myrtle-10: 10k target 100 102 104 Support Set Size FC -> Myrtle-10: 20k target 100 102 104 Support Set Size FC -> Myrtle-10: 50k target Figure A2: Label Solve transfer between Myrtle-10 and FC for CIFAR10. Top row: LS labels using Myrtle-10 applied to FC1. Bottom row: LS labels using FC1 applied to Myrtle-10. Results averaged over 3 samples per support set size. In all these plots, NNGP kernels were used and Myrtle-10 used regularized ZCA preprocessing. In Figure A2, we plot how LS learned labels using Myrtle-10 kernel transfer to the FC1 kernel and vice versa. We vary the number of targets and support size. We find remarkable stability across all these dimensions in the sense that while the gains from LS may be kernel-specific, LS -labels do not perform meaningfully different from natural labels when switching the train and evaluation kernels. E.3 KIP TRANSFER TO NEURAL NETWORKS AND CORRUPTION EXPERIMENTS Table A5: KIP transfer to NN vs NN baselines on MNIST. For each group of four experiments, the best number is marked boldface, while the second best number is in italics. Corruption refers to 90% noise corruption. KIP images used FC1-3, Conv1-2 kernel during training. Method 10 uncrpt 10 crpt 100 uncrpt 100 crpt 200 uncrpt 200 crpt FC1, KIP 73.57 1.51 44.95 1.23 86.84 1.65 79.73 1.10 89.55 0.94 83.38 1.37 FC1, Natural 42.28 1.59 35.00 2.33 72.65 1.17 45.39 2.25 81.70 1.03 54.20 2.61 Le Net, KIP 59.69 8.98 38.25 6.42 87.85 1.46 69.45 3.99 91.08 1.65 70.52 4.39 Le Net, Natural 48.69 4.10 30.56 4.35 80.32 1.26 59.99 0.95 89.03 1.13 62.00 0.94 Table A6: KIP transfer to NN vs NN baselines on CIFAR-10. Notation same as in Table A5. Method 100 uncrpt 100 crpt FC3, KIP 43.09 0.20 37.71 0.38 FC3, Natural 24.48 0.15 18.92 0.61 Conv2, KIP 43.68 0.46 37.08 0.48 Conv2, Natural 26.23 0.69 17.10 1.33 Wide Res Net, KIP 33.29 1.14 23.89 1.30 Wide Res Net, Natural 27.93 0.75 19.00 1.01 Table A7: MNIST. KIP and natural images on FC1. MSE Loss. Test accuracy of image datasets of size 1K, 5K, 10K, trained using FC1 neural network using mean-square loss. Dataset size, noise corruption percent, and dataset type are varied: natural refers to natural images, KIP refers to KIP - learned images with either augmentations only (a) or both augmentations with label learning (a + l). Only FC1 kernel was used for KIP . For each KIP row, we place a * next to the most corrupt entry whose performance exceeds the corresponding 0% corrupt natural images. For each dataset size, we boldface the best performing entry. Dataset 0% crpt 50% crpt 75% crpt 90% crpt Natural 1000 92.8 0.4 87.3 0.5 82.3 0.9 74.3 1.4 KIP (a) 1000 94.5 0.4 95.9 0.1 94.4 0.2* 92.0 0.3 KIP (a+l) 1000 96.3 0.2 95.9 0.3 95.1 0.3 94.6 1.9* Natural 5000 96.4 0.1 92.8 0.2 88.5 0.5 80.0 0.9 KIP (a) 5000 97.0 0.6 97.1 0.6 96.3 0.2 96.6 0.4* KIP (a+l) 5000 97.6 0.0* 95.8 0.0 94.5 0.4 91.4 2.3 Natural 10000 97.3 0.1 93.9 0.1 90.2 0.1 81.3 1.0 KIP (a) 10000 97.8 0.1* 96.1 0.2 95.8 0.2 96.0 0.2 KIP (a+l) 10000 97.9 0.1* 95.8 0.1 94.7 0.2 88.1 3.5 Table A8: MNIST. KIP and natural images on FC1. Cross Entropy Loss. Test accuracy of image datasets trained using FC1 neural network using cross entropy loss. Notation same as in Table A7. Dataset 0% crpt 50% crpt 75% crpt 90% crpt Natural 1000 91.3 0.4 86.3 0.3 81.9 0.5 75.0 1.3 KIP (a) 1000 95.9 0.1 95.0 0.1 93.5 0.3* 90.9 0.3 Natural 5000 95.8 0.1 91.9 0.2 87.3 0.3 80.4 0.5 KIP (a) 5000 98.3 0.0 96.8 0.8* 95.5 0.3 95.1 0.2 Natural 10000 96.9 0.1 93.8 0.1 89.6 0.2 81.3 0.5 KIP (a) 10000 98.8 0.0 97.0 0.0* 95.2 0.2 94.7 0.3 Table A9: CIFAR-10. KIP and natural images on FC1. MSE Loss. Test accuracy of image datasets trained using FC1 neural network using mean-square loss. Notation same as in Table A7. Dataset 0% crpt 50% crpt 75% crpt 90% crpt Natural 1000 34.1 0.5 34.3 0.4 31.7 0.5 27.7 0.8 KIP (a) 1000 48.0 0.5 46.7 0.2 45.7 0.5 44.3 0.5* KIP (a+l) 1000 47.5 0.3 46.7 0.8 44.3 0.4 41.6 0.5* Natural 5000 41.4 0.6 41.3 0.4 37.2 0.2 32.5 0.7 KIP (a) 5000 51.4 0.4 50.0 0.4 48.8 0.6 47.5 0.3* KIP (a+l) 5000 50.6 0.5 48.5 0.9 44.7 0.6 43.4 0.5* Natural 10000 44.5 0.3 43.2 0.2 39.5 0.2 34.3 0.2 KIP (a) 10000 53.3 0.8 50.5 1.3 49.4 0.2 48.2 0.6* KIP (a+l) 10000 51.9 0.4 50.0 0.5 46.5 1.0 43.8 1.3* Table A10: CIFAR-10. KIP and natural images on FC1. Cross Entropy Loss. Test accuracy of image datasets trained using FC1 neural network using cross entropy loss. Notation same as in Table A7. Dataset 0% crpt 50% crpt 75% crpt 90% crpt Natural 1000 35.4 0.3 35.4 0.3 31.7 0.9 27.2 0.8 KIP (a) 1000 49.2 0.8 47.6 0.4 47.4 0.4 45.0 0.3* Natural 5000 43.1 0.8 42.0 0.2 38.0 0.4 31.7 0.6 KIP (a) 5000 44.5 1.0 51.5 0.3 51.0 0.4 48.9 0.4* Natural 10000 45.3 0.2 44.8 0.1 40.6 0.3 33.8 0.2 KIP (a) 10000 46.9 0.4 54.0 0.3 52.1 0.3 49.9 0.2* 0 0.5 0.75 0.9 corruption fraction test accuracy MNIST. KIP vs Natural Images (MSE Loss) 1000,natural 5000,natural 10000,natural 1000,kip+aug 5000,kip+aug 10000,kip+aug 1000,kip+aug+label 5000,kip+aug+label 10000,kip+aug+label 0 0.5 0.75 0.9 corruption fraction test accuracy MNIST. KIP vs Natural Images (XENT Loss) 0 0.5 0.75 0.9 corruption fraction test accuracy CIFAR-10. KIP vs Natural Images (MSE Loss) 0 0.5 0.75 0.9 corruption fraction test accuracy CIFAR-10. KIP vs Natural Images (XENT Loss) Figure A3: KIP vs natural images, FC1. Data plotted from Tables A7-A10, showing natural images vs. KIP images for FC1 neural networks across dataset size, corruption type, dataset type, and loss type. For instance, the upper right figure shows that on MNIST using cross entropy loss, 1k KIP + aug learned images with 90% corruption achieves 90.9% test accuracy, comparable to 1k natural images (acc: 91.3%) and far exceeding 1k natural images with 90% corruption (acc: 75.0%). Similarly, the lower right figure shows on CIFAR10 using cross entropy loss, 10k KIP + aug learned images with 90% corruption achieves 49.9%, exceeding 10k natural images (acc: 45.3%) and 10k natural images with 90% corruption (acc: 33.8%). F EXAMPLES OF KIP LEARNED SAMPLES Figure A4: KIP learned images (left) vs natural MNIST images (right). Samples from 100 learned images. Top row: 0% corruption. Bottom row: 90% noise corruption. Figure A5: KIP learned images (left) vs natural CIFAR-10 images (right). Samples from 100 learned images. Top row: 0% corruption. Bottom row: 90% noise corruption.