# tasklevel_differentially_private_meta_learning__6391a6be.pdf Task-level Differentially Private Meta Learning Xinyu Zhou Department of Computer Science & Engineering The Ohio State University zhou.3542@buckeyemail.osu.edu Raef Bassily Department of Computer Science & Engineering and TDAI Institute The Ohio State University bassily.1@osu.edu We study the problem of meta-learning with task-level differential privacy. Metalearning has received increasing attention recently because of its ability to enable fast generalization to new task with small number of data points. However, the training process of meta learning likely involves exchange of task specific information, which may pose privacy risk especially in some privacy-sensitive applications. Therefore, it is important to provide strong privacy guarantees such that the learning process will not reveal any task sensitive information. To this end, existing works have proposed meta learning algorithms with record-level differential privacy, which is not sufficient in many scenarios since it does not protect the aggregated statistics based on the task dataset as a whole. Moreover, the utility guarantees in the prior work are based on assuming the loss function satisfies both smoothness and quadratic growth conditions, which do not necessarily hold in practice. To address these issues, we propose meta learning algorithms with task-level differential privacy; that is, our algorithms protect the entire dataset for each task. In the case when a single meta model is trained, we give both privacy and utility guarantees assuming only that the loss is convex and Lipschitz. Moreover, we propose a new private clustering-based meta-learning algorithm that enables private meta learning of multiple meta models. This can provide significant accuracy gains over the single meta model paradigm, especially when the tasks distribution cannot be well represented by a single meta model. Finally, we conduct several experiments demonstrating the effectiveness of our proposed algorithms. 1 Introduction Meta learning has received increasing attention recently due to its successful applications in a variety of machine learning domains. Meta learning is the process in which existing learning tasks are used to infer a learning rule that enables faster adaptation to new tasks from the same environment. Specifically, meta learning takes in a collection of tasks (datasets) sampled from an unknown distribution. Each task defines a learning problem with respect to an input dataset. The goal of meta learning is to output a meta-learner which can extract the shared knowledge among the tasks (datasets) and generalize well on a newly sampled task (dataset) with a small number of data points. The performance of a meta learner is measured by transfer risk, defined as the expected risk of the output of the algorithm w.r.t. a newly sampled task. Machine learning algorithms typically rely on gradient-based optimization methods. The optimizationbased meta learning framework allows such methods to converge to a solution using a small number of samples. This is done by learning either a good parameter initialization or a regularizer. In meta 36th Conference on Neural Information Processing Systems (Neur IPS 2022). initialization methods like MAML [1] and Reptile [2], the meta leaner obtains a meta initialization w0 such that one or a few gradient updates of w0 produces a good task-specific model for new tasks. Another class of meta learning algorithms are based on an approach known as meta-regularization [3 6] in which a regularizer is first learned based on all tasks datasets, then a task-specific model is obtained for a new task by minimizing the regularized loss. Meta-regularization has received more attention recently due to its formal utility guarantees and computational efficiency. In meta-regularization, the goal is to find a good bias vector h such that a local learner (base learner) can converge to a good task-specific model by minimizing the regularized empirical loss: minw W RZ(w) + λ 2 w h 2. In [7], the authors formalize the connection between meta-initialization and meta-regularization under the framework of online convex optimization. Meanwhile, the nature of the meta learning requires the exchange of task-specific information, which may pose serious privacy risks, especially in distributed settings where all tasks (users) are participating into a meta training process by communicating meta-model updates among themselves, typically with the assistance of a central server. Therefore it is important to ensure that the sensitive information in every task (dataset) remains private through the meta learning process. Despite its importance, the privacy aspect of meta learning is still underexplored. In [8], using the meta-initialization framework, the authors propose a meta-learning algorithm with record-level differential privacy (DP) and give an accuracy guarantee in terms of the transfer risk. Record-level DP is a weak privacy guarantee for a wide range of scenarios. Imagine training an email spam filter for each user (task). Even if we ensure differential privacy w.r.t. each record of a user s email, aggregate statistics across the entirety of a user s records (e.g., the frequency of certain words or sentences) may still reveal some sensitive information which the user doesn t want to share. Moreover, the accuracy guarantees in [8] rely on relatively strong assumptions concerning the loss function used for training. In particular, the loss function therein is assumed to satisfy both smoothness and quadratic growth conditions, which do not hold simultaneously for many loss functions commonly used in practice. In this work, we propose new meta-learning algorithms with task-level DP and derive formal transfer risk bounds. Unlike [8], our algorithms are based on the meta-regularization framework. We now highlight our main technical contributions: We propose a task-level differentially private (task-level DP) meta learning algorithm based on the meta-regularization approach. We observe that when the meta-loss is properly defined, one can view each task dataset as a data point and the meta-loss as its corresponding loss function. Therefore, applying existing DP optimization techniques like Noisy SGD on the newly defined "data point" and loss function achieves task-level differential privacy naturally. We give a bound for the statistical performance for our algorithm in terms of its excess transfer risk. Our statistical accuracy guarantee holds under standard assumptions; namely, assuming only that loss function is convex and Lipshitz. With sufficiently large number of tasks, the bound asserts improved performance over local training, where a model is trained locally based on a single task dataset. When the task distribution becomes more complex such that a single meta model is not enough to capture the relationship among tasks, we propose a task-level DP clustering-based meta learning algorithm that can learn multiple meta models. This approach can provide significant improvements in the statistical accuracy (transfer risk) over the standard single-model case, especially when the tasks distribution is better captured by multiple meta models. We empirically evaluate our algorithms to demonstrate the effectiveness of our proposed algorithms1. Related work: The most related work to ours is [8], where the authors propose a meta-learning algorithm with record-level DP via the framework of meta-initialization. Their analysis relies on assuming that the loss function satisfies a quadratic growth condition, which we do not require in our analysis. Furthermore, we consider a stronger notion of the excess transfer risk than the one in [8]. To be more specific, in [8], the achieved transfer risk is compared to the transfer risk of the optimal global model, while the transfer risk in our work is compared to that of the optimal personalized model. This is because we allow different tasks to have different parameters. Since 1Our code is available online at https://github.com/xyzhou055/Meta NSGD the optimal personalized model always outperforms global model, we have a stronger notion of the excess transfer risk. Consequently, our utility guarantee also holds under the notion studied in [8]. Our results are also related to the line of work in private model personalization. The goal of model personalization is to train a personalized model for each user that generalizes well on the user s distribution. Reference [9] proposes a private personalization algorithm that solves a stochastic optimization problem involving both local and global models. The authors of [10] develop an algorithm for mean-regularized multi-task learning under the joint differential privacy guarantee and study its application in model personalization. The work of [11] presents private personalization algorithms that learn a common embedding model among users using the alternating minimization framework, where convergence to the optimal solution is guaranteed for linear regression models with Gaussian data. There are also a variety of works, mainly in the non-private regime, exploring the connection between meta learning and model personalization. For example, [12] proposes a personalization mechanism based on MAML, which aims at finding a good initialization parameter for the users. The authors of [13] study the connection between meta learning and personalization based on the observation that Fed Avg [14] can be seen as a meta learning algorithm. Therefore, it is possible to extend the results in this work to the domain of model personalization. There are also prior works studying learning with task-level or user-level differential privacy but under different learning paradigms. In [15, 16], the authors provide user-level differential privacy under the framework of federated learning, in which case all users share a single global model. This is different from our setting because we allow each user to have its own task-specific model. The authors in [17] develop algorithms for a variety of learning tasks under user-level differential privacy. They consider a limited heterogeneity setting where all users distributions are close to each other in total variation distance. On the other hand, in this work we focus on the meta learning paradigm, where we only assume small variation across users models. 2 Preliminaries Distributional setup: We consider a meta distribution ρ and each task is sampled from ρ which will induce a task specific distribution µ over the data domain Z, that is µ1, . . . µK i.i.d ρ for K tasks. After the sampling of µi, a dataset Zi with n datapoints is subsequently sampled from µi written as Zi = {z1, . . . zn} i.i.d µn i . The collection of K task datasets is denoted as Z = {Z1, . . . ZK} In this paper, we focus on distributed settings such that the datasets {Zi}K i=1 are decentralized across multiple locations. All tasks jointly participate into a model training process coordinated by a trusted central server. The data of each task is stored and processed locally, only the model updates are shared to the server or other tasks. Task population and empirical losses:Given a closed convex parameter space W, we consider learning task parametrized by vector w Rd. We let ℓ: Rd Z R be a loss function, where ℓ(w, z) is the loss incurred by a parameter vector w on a data point z. For any task drawn from a distribution µ and observed via a dataset Z = (z1, . . . zn), we denote the population risk of w W as Rµ(w) = Ez µℓ(w, z) and the empirical risk as RZ(w) = 1 n Pn i=1 ℓ(w, zi). Meta learning: Given K learning tasks with datasets Z, a meta-learning process aims to learn an algorithm A that enables fast adaption on newly sampled tasks from the meta-distribution ρ. Such adaptation is measured by the notion of transfer risk defined as En(A; ρ) = Eµ ρEZ µn Rµ(A(Z)). (1) The transfer risk En(A; ρ) can be considered as the expected loss of the parameter returned by applying the algorithm A on a dataset Z Zn sampled from the meta distribution ρ. In summary, the whole process starts with a meta-learning algorithm taking Z as input and outputting a learning algorithm A. After that, new learning task with distribution µ ρ trains its corresponding dataset with the learned algorithm A and apply the resulting parameter vector w to new data drawn from µ. For any task specific distribution µ, we denote the minimizer wµ as wµ minw W Rµ(w) and the expected minimum error over the meta-distribution ρ as Eρ = Eµ ρRµ(wµ). Note that, there might be more than one minimizers of Rµ(w), but Eρ remains same regardless of the selection of wµ. Furthermore, we define the excess transfer risk as En(A; ρ) = En(A; ρ) Eρ (2) Task similarity metric: To achieve low transfer risk, the meta learning algorithm leverages the shared similarity among tasks and output the optimized inner algorithm. In this work, the task similarity is measured by the quantity: V 2 = min ϕ W Eµ ρ ϕ Proj Wµ(ϕ) 2 (3) where Wµ is the set of optimal parameters for task µ written as Wµ = arg minw W Rµ(w), and Proj Wµ denotes the Euclidean projection on Wµ. V can be roughly seen as the expected variation of task parameters. We denote w as the minimizer of (3) such that w W and V 2 = w Proj Wµ( w) 2 . Task-level Differential Privacy [18]: A randomized algorithm M is (ϵ, δ)-differential private if for any pair of neighboring sets Z and Z and for any measurable subset O of Range(M), we have P(M( Z) O) eϵP(M( Z ) O) + δ In this work, we consider task-level neighboring relationship. Let Z = {Z1, . . . ZK} and Z = {Z 1, . . . Z K}, and we say Z and Z are neighboring if, for some i [K], Z and Z differ only on the dataset from the ith task, i.e., Zj = Z j for all j except i. Therefore, a task-level differential private algorithm ensures that any adversary cannot tell if the dataset of a particular task was used in the algorithm given its output. Our analysis involves the following fairly standard assumption: Assumption 1. (Convex-Lipschitz-Bounded models) For any z Z, the loss function ℓ(w, z) is convex and L-Lipshitz with respect to w. Bounded parameter space: W 2 = maxw1,w2 W w1 w2 M 3 Learning with Meta Regularization Base learner:A base learner solves a regularized loss minimization problem w.r.t. a biased regularizer parameterized by a bias vector h.In particular, given a bias vector h and a task dataset Z, the base learner solves the following problem: min w W RZ,λ(w; h) = min w W RZ(w) + λ where λ is the regularization parameter. In the sequel, we define wh(Z) as the minimizer of RZ,λ(w; h), that is wh(Z) = arg minw W RZ,λ(w; h). Meta algorithm: Given a collection of base learners, the goal of a meta-algorithm is to find a good regularization bias h that yields a small transfer risk. First, let s rewrite the transfer risk defined in (1) as a function h: En(h; ρ) = Eµ ρEZ µn Rµ(wh(Z)) (4) and hence the excess transfer risk can be expressed as En(h; ρ) = En(h; ρ) Eρ. Therefore, the goal of the meta algorithm is to find an (approximate) minimizer h of En(h; ρ) in (4), i.e., the goal is to solve: min h W En(h; ρ) = Eµ ρEZ µn Rµ(wh(Z)) (5) Intuitively speaking, the base learner aims to figure out the best parameter vector around the bias vector h. Therefore, if the meta algorithm can find a good bias vector h that is close to the solution space of the tasks, the base learner can converge to its solution space quickly by only searching around h. 3.1 Meta Loss and proxy transfer risk Directly optimizing over (5) above is usually infeasible. A common strategy to get around this difficulty is to propose a proxy loss function which is related to the transfer risk but easier to handle. In this section, we follow the definitions in [3, 4] and define our meta loss as LZ(h) = min w W RZ,λ(w; h) = min w W RZ(w) + λ and the proxy to the transfer risk is defined as ˆEn(h; ρ) = Eµ ρEZ µn LZ(h) The goal of the meta-algorithm is to optimize the proxy risk over h, that is, min h W ˆEn(h; ρ) = Eµ ρEZ µn LZ(h) (6) Note that LZ(h) is in fact the Moreau envelope of the empirical loss RZ(w). In Proposition 2, we review some useful properties of LZ(h) which will help our analysis. Proposition 2. (follows from [19, 20]) Under Assumption 1, we have 1. LZ is convex, 2L-Lipshitz and λ-smooth. 2. For any h W, LZ(h) = λ(wh(Z) h) where wh(Z) = arg minw W RZ,λ(w; h). A key observation is that all datasets go through the same stochastic process under our distributional setup and thus can be seen as i.i.d. sequence drawn from a distribution PZ. Therefore, (6) can be rewritten as min h W ˆEn(h; ρ) = min h W EZ PZLZ(h) (7) Combined with Proposition (2), the objective of the meta algorithm is formulated as a stochastic convex optimization (SCO) problem which has been well studied under privacy constraints [19, 21, 22]. The notion of task-level differential privacy just comes up naturally when we see each task dataset Z as a data point and LZ(h) as its loss function, therefore the conventional record-level neighboring relationship can be readily translated to a task-level neighboring relationship. However, one challenge in adapting existing differentially private SCO methods pertains to the gradient computation. In standard SCO settings, we are expected to be able to compute the gradient of the loss function efficiently and precisely. However, in our case, as seen in Proposition 2, computing LZ(h) for each h requires computing wh(Z) which further involves solving a λ-strongly convex optimization problem. Therefore, finding the exact value of wh(Z) for general convex loss ℓis not feasible, alternatively what we can do is to approximate wh(Z) using some iterative methods up to some error b. Therefore, it is important to incorporate this approximation error into our convergence analysis. Moreover, due to the distributed nature of the algorithm, the model iterates are assumed to be visible to multiple tasks who are selected to participate in each round. Therefore, the privacy should be defined over all model iterates as well as the final model, and we cannot use the lineartime algorithm for DP smooth stochastic optimization [22] despite the fact that the objective of the meta-algorithm can be formulated as a stochastic convex optimization problem with a smooth loss. 4 Private Meta Learning with Noisy SGD In this section, we present our meta learning algorithm adapted from the mini batch noisy SGD method in [19] described in Algorithm 1. In iteration t, given the current iterate ht, each sampled task computes an approximation of the gradient of the meta loss LZ(ht), denoted as ˆ LZ(ht). After that, the server aggregates theses approximate gradients to update the model. The privacy is achieved by adding Gaussian noise scaled with the l2 sensitivity of each task to the model iterate. We will first describe the privacy and utility guarantee of Algorithm 1, and then show the resulting transfer risk. Theorem 3. (Privacy guarantee) Let Assumption 1 holds and let {h1, . . . h T 1} and h T be the output, Algorithm 1 is (ϵ, δ)-differentially private on task level. Proof. Since the computed gradient ˆ LZ(ht) is only an approximation to LZ(ht), then we have ˆ LZ(ht) LZ(ht) λ ˆwht(Z) wht(Z) b and because LZ(h) is 2L-Lipshitz by Proposition 2, we have ˆ LZ(ht) LZ(ht) + b 2L + b Algorithm 1: Ameta-NSGD: meta learning with mini-batch noisy SGD input :Initial iterate w0 W, number of iterations T, estimation error b, regularization parameter λ, step size η. 1 Set noise variance σ2 = 8T (2L+b)2 log(1/δ) K2ϵ2 and batch size m = max K p 2 for t = 1, . . . , T 1 do 3 Server samples a batch of tasks Bt = {Zi(t,1) . . . Zi(t,m)} Z and sends ht to tasks in Bt 4 for each task Z in Bt do 5 Solve RZ,λ(w; ht) and compute an estimate ˆwht(Z) with ˆwht(Z) wht(Z) b λ 6 Send ˆ LZ(ht) = λ( ˆwht(Z) ht) back to server 8 Server updates the model with ht+1 = Proj W ˆ LZ(ht) + Gt where Proj W denotes the Euclidean projection onto W, and Gt N(0, σ2Id) drawn independently each iteration. 9 end output : h T = 1 T PT t=1 ht Note that Algorithm 1 has the similar structure of the mini-batch noisy SGD method in [19] if we consider ˆ LZ(ht) as the model updates and ht as model iterates. By adding Gaussian noise scaled with (2L + b), a direct application of Theorem 3.1 in [19] completes the proof. Since the objective of the base learner RZ,λ(w; ht) is a λ-strongly convex problem, one can use gradient decent to bound the approximation error at step 5 in Algorithm 1. Alternatively, the base learner can use other stochastic optimization method i.e. SGD to achieve a high probability bound on the approximation error, such that ˆwht(Z) wht(Z) b λ with probability over 1 δ . By letting the overall failure probability m Tδ on the same order of the privacy parameter δ, the privacy guarantee still holds. For the simplicity of the analysis, we assume the approximation error ˆwht(Z) wht(Z) b λ always satisfies. In the next theorem, we will give the excess risk of Ameta-NSGD Theorem 4. (Utility guarantee) Under Assumptions 1, suppose λ 2L 2d log(1/δ) Let T = min K 8 , ϵ2K2 32d log(1/δ) , η = M 2L T and b = 2L E[ ˆEn( h T ; ρ)] min h W ˆEn(h; ρ) O where h T is the final model output by Algorithm 1, and the expectation is taken over the collection of datasets Z and the inner randomness of the algorithm. Note that the excess rate in Theorem 4 is with respect to ˆEn(h; ρ), the proxy to the transfer risk. It remains to show how it can be related to the excess rate of the true transfer risk En(h; ρ). In the next theorem, we give the excess transfer risk of Algorithm 1. The proof follows from [Theorem 6, [3]] and decomposes the excess transfer rate into three components. We bound each component separately and one of them is based on the excess rate achieved in Theorem 4. Theorem 5. (Transfer risk) Recall the settings in Theorem 4, we have E[En( h T ; ρ)] Eρ 2L2 By letting λ = 2L 1 n, we have E[En( h T ; ρ)] Eρ 2LV n + O where h T is the output of Ameta-NSGD and the expectation is taken w.r.t. Z and the inner randomness of the algorithm. Theorem 5 shows that when the number of tasks K is sufficiently large, the excess transfer risk becomes O LV n . Recall that V is a measure of task similarity, hence when the optimal task parameters are close to each other, V can be much smaller than the diameter of the parameter space which indicates better performance than local training where each task trains its own model independently. Note one can also give a direct bound of the transfer risk of the output of Algorithm 1 based on the fact that h T wµ W = M for any task µ with optimal parameter wµ, standard derivation gives that E[En( h T ; ρ)] Eρ O LM n which is the same as the excess risk for local training. Therefore, Algorithm 1 is at least as good as local training, in the worst case, it recover the performance of local training. The rate achieved in Theorem 5 is on the same order of the non-private rate O V L n + ML from [3] in the regime most common in practice where d = O(K) and ϵ = θ(1). Note that [8] obtains O 1 K convergence on the number of tasks K under record-level privacy guarantee, but their analysis relies on the assumption for the loss function to be α-quadratic growth, which we do not require for our analysis. Furthermore, we have different definitions of excess transfer risk from [8], given the meta learner Ah, the excess transfer risk in [8] is defined as En(Ah; ρ) minw W Eµ ρRµ(w) while ours is defined as En(Ah; ρ) Eµ ρ [minw W Rµ(w)]. Since minw W Eµ ρRµ(w) Eµ ρ [minw W Rµ(w)], our bound holds under a stronger definition for the excess transfer risk. Hence, our bound is a valid upper bound for the notion defined in [8]. The detailed proofs of Theorems 4 and 5 can be found in the appendix. Communication-efficient, private meta Learning with DP-FTMRL Although Algorithm 1 gives a transfer risk comparable with the non-private rate, it requires O min n K, K3/2/d o rounds of communication to each task on average. This can be expensive when the task number K is large but each task only has limited computation and communication resource, which is a common setting in distributed scenarios like federated learning. Therefore, we introduce a communication efficient DP meta algorithm Differentially Private Follow-the-Meta-Regularized-Leader (DP-FTMRL) which is based on the DP-FTRL algorithm in [23]. DP-FTMRL is a one-pass algorithm which means that every task will be communicated by the server for only once. Furthermore, it processes each task sequentially thus allows tasks to arrive in a stream. With high probability over the algorithm inner randomness, DP-FTMRL obtains transfer risk of O LV n + ML 1 . This rate is suboptimal compared with the one achieved in Theorem 5, but DP-FTMRL provides significant savings on the communication cost. More details of DP-FTMRL can be found in the appendix. Note: To get the excess transfer risk guarantee in Theorem 5, we need to set λ = 2L V 1 n which requires the knowledge of the value of V , the task similarity. In fact, we can privately approximate V using its empirical counterpart; in particular, we can find a private counterpart for ˆV 2 minϕ W 1 K PK i=1 ϕ ˆw(Zi) 2 where ˆw(Zi) is a minimizer of the empirical loss of task Zi. This can be done using standard DP convex optimization methods such Noisy Gradient Descent. An alternative approach to choose a setting for λ, which is quite common in practice, is to treat λ as a hyperparameter [4] and tune it privately. This can also be done using existing methods in the literature of DP [24, 25]. 5 Private Clustering-Based Meta Learning In this section, we consider the case where the tasks are sampled from distribution with multiple concentrations (e.g., a multi-modal distribution) where a single meta model is not enough to capture the relationship among tasks. A natural extension is to cluster the tasks into several groups and train a meta model for each group. To provide a more formal treatment, we let q denote the number of possible meta models (clusters), each is parametrized with a bias vector hi. These meta models are denoted as (h1, h2, . . . hq). The task similarity in the presence of q meta modals is hence modified to V 2 c = min h1,...hq W Eµ ρ min i [q] hi Proj Wµ(hi) 2 (8) Recall that Wµ is the set of optimal parameters for task µ. Note that Vc is no larger than V from single meta learner case defined in (3). In the case when {Wµ} contains multiple disjoint and far apart concentrations, Vc can be much smaller compared with V . In particular, inspired by the idea of hypothesis based clustering from [26], we define the objective of the meta-algorithm with q meta learners as min h1,...hq W ˆEn(h, Z)) = 1 j=1 min i [q] LZ(hi) (9) where LZ( ) is the meta loss. Given h = (h1 . . . hq), the base learner with dataset set Z first chooses the hi with lowest meta loss, that is, C(Z) = arg mini [q] LZ(hi) and output the corresponding minimizer wh C(Z)(Z) = minw W RZ,λ(w; h C(Z)) = minw W RZ(w) + λ 2 w h C(Z) 2. Therefore, the transfer risk is defined as En(h; ρ) = Eµ ρEZ µn Rµ wh C(Z)(Z) We present our meta-clustering algorithm described in Algorithm 2, which is inspired by the HYPCLUSTER algorithm in [26]. Algorithm 2 is a heuristic algorithm aiming at minimizing the objective (9) over h1, . . . hq. Each iteration consists of two steps, the model selection step and model update step. At the beginning of one iteration, each sampled task selects the model from {hi}q i=1 that incurs the lowest loss. After the assignment is done, the server will update each model with its corresponding tasks. We formally show the privacy guarantee of Algorithm 2. The performance of this algorithm in terms of its transfer risk will be evaluated empirically since convergence to a true optima is not guaranteed due to the non-convexity of ˆEn(h, Z)) in h. Algorithm 2: Ameta-cluster: private clustering-based meta learning algorithm input :q randomly initialized iterates h1 0, h2 0, . . . hq 0, number of iterations T, batch size m 1 Set noise variance σ2 = 8T (2L+b)2 log(1/δ) 2 for t = 1, . . . , T 1 do 3 Server samples a batch of tasks Bt of size m and sends {hi t}q i=1 to Bt 4 for each task Z in Bt do 5 For i [q], compute approximate minimizers of RZ,λ(w, hi t) as ˆwi h(Z) with approximation error b λ 6 Select the model with smallest meta loss C(Z) = arg mini [q] RZ,λ( ˆwi h(Z), hi t) 7 Compute the gradient ˆ LZ(h C(Z) t ) = λ ˆw C(Z) Z h C(Z) t 8 Send (C(Z), ˆ LZ(h C(Z) t )) back to server 10 For every i [q], denote the set Ci t = {Z Bt|C(Z) = i}, server updates hi t as hi t+1 = Proj W ˆ LZ(hi t) + Gi t where Gi t N(0, σ2Id) drawn independently each iteration t and meta model i. 11 end output :{h1 T , h2 T , . . . hq T } Theorem 6. (Privacy guarantee) Let {h1 t, . . . hq t}t T be the output, Algorithm 2 is (ϵ, δ)-differentially private on task level. Proof. Similar to Theorem 3, we have the norm of ˆ LZ(h C(Z) t ) is bounded by (2L + b) for any choice of C(Z). Since each task only selects one meta model to update in each iteration, we consider {hi t}q i=1 as the model iterate and ˆ LZ(hi t) as the model update from the task, we can use the same argument in Theorem 3 to complete the proof. 6 Experiments In this section, we empirically evaluate the performance of our algorithms under various privacy regimes. On the one hand, we demonstrate the performance improvement of our algorithms over local training. Meanwhile, when the tasks are sampled from multi-modal distribution, we show that our clustering based algorithm achieves lower transfer risk than the single meta model paradigm. (a) Single-cluster case (b) Multi-clusters case Figure 1: The transfer risk of Ameta-cluster in linear regression task with respect to the number of tasks. ϵ is the privacy parameter and q is the number of meta models (clusters). We consider linear regression task with mean square loss. The weight vector wµ is d-dimentional vectors that concentrates around one or more common vectors. We let d = 30 for all experiments. The feature vector x is uniformly sampled from unit ball, and the label y = wu, x + n where n is the zero mean Gaussian noise with standard deviation equal to 0.5. Single cluster case: For each task µ, its d-dimensional weight vector wµ is sampled from Gaussian distribution with mean h and with variance σ2, denoted as N( h, σ2Id). In the experiment, σ = 1 and h is a vector with all entries equal to 4. Each task contains 10 datapoints (xi, yi)10 i=1. The privacy parameters ϵ are chosen from {1, 3, 10} and δ is set as 10 5. We set the clipping norm to 2 and use the functions from Tensor Flow privacy[27] to choose the noise multiplier to fit the privacy budget. In Figure 1, the transfer risks of our meta learning algorithm are substantially better than the transfer risk of local training even in high privacy regimes when ϵ = 1. Furthermore, by comparing the transfer risks for different ϵ s, we show that the performance gaps are much less significant compared with the gap between local training. In modest privacy regimes when ϵ = 3 or 10, the corresponding transfer risks are quite close to the non-private rate (ϵ = ). Multi clusters case: The weight vector is sampled from the distribution N( h, σ2Id) where h is uniformly sampled from the set h = { h1, . . . ht}. In the experiment, t, the number of cluster in the underlying distribution, is set to be 3, σ = 0.5 and we choose { h1, h2, h3} to be three orthogonal vectors with different norms. More specifically, h1 = [2 . . . 2, 0, . . . 0] , h2 = [0, . . . , 0, 4 . . . , 4, 0, . . . 0] and h3 = [0, . . . , 0, 6, . . . 6], and each wi has 10 non-zero components. The privacy parameters ϵ are chosen from {3, 10} and δ is set as 10 5, the clipping norm is set to 1. Furthermore, we set q, the number of meta models in Algorithm 2, from 1 to 3 and report the transfer risk with each q. The plots in Figure 2 confirms our claim that we can achieve better performance by adding the number of meta models. By choosing different number of meta models, the performance gap can be significant. Meanwhile, similar to the single cluster case, with the same number of clusters, we see relatively small performance difference under various privacy parameters. In Appendix C, we present additional experiments to evaluate our algorithms on Omniglot [28] few-shot classification tasks, where we demonstrate an improved performance over existing DP meta learning methods. 7 Conclusion and Future Work In this work, we studied the meta learning problem with task-level privacy. When a single meta model is trained, we have shown, both theoretically and empirically, that our private meta algorithm achieves better performance than local training when the variation among the tasks is relatively small. To deal with more complex task distribution, we also proposed a private clustering-based meta learning algorithm that outputs multiple meta models. This approach can provide noticeable performance gains over the single meta model paradigm as demonstrated by our empirical results. An interesting direction for future work is to formally study the accuracy guarantees of our clusteringbased algorithm. Due to the non-convexity of the objective function, theoretical analysis on the transfer risk may be challenging. However, it is still possible to obtain a guarantee to characterize the generalization property of the clustering-based algorithm. Acknowledgments and Disclosure of Funding This work is supported by NSF AI-Institute Award 2112471. [1] Chelsea Finn, Pieter Abbeel, and Sergey Levine. Model-agnostic meta-learning for fast adaptation of deep networks. In International conference on machine learning, pages 1126 1135. PMLR, 2017. [2] Alex Nichol, Joshua Achiam, and John Schulman. On first-order meta-learning algorithms. ar Xiv preprint ar Xiv:1803.02999, 2018. [3] Giulia Denevi, Carlo Ciliberto, and Riccardo Grazzi. Learning-to-learn stochastic gradient descent with biased regularization. In International Conference on Machine Learning, pages 1566 1575. PMLR, 2019. [4] Pan Zhou, Xiaotong Yuan, Huan Xu, Shuicheng Yan, and Jiashi Feng. Efficient meta learning via minibatch proximal update. Advances in Neural Information Processing Systems, 32, 2019. [5] Aravind Rajeswaran, Chelsea Finn, Sham M Kakade, and Sergey Levine. Meta-learning with implicit gradients. Advances in neural information processing systems, 32, 2019. [6] Weisen Jiang, James Kwok, and Yu Zhang. Effective meta-regularization by kernelized proximal regularization. Advances in Neural Information Processing Systems, 34, 2021. [7] Maria-Florina Balcan, Mikhail Khodak, and Ameet Talwalkar. Provable guarantees for gradient-based meta-learning. In International Conference on Machine Learning, pages 424 433. PMLR, 2019. [8] Jeffrey Li, Mikhail Khodak, Sebastian Caldas, and Ameet Talwalkar. Differentially private meta-learning. In International Conference on Learning Representations, 2019. [9] Alberto Bietti, Chen-Yu Wei, Miroslav Dudik, John Langford, and Steven Wu. Personalization improves privacy-accuracy tradeoffs in federated learning. In International Conference on Machine Learning, pages 1945 1962. PMLR, 2022. [10] Shengyuan Hu, Zhiwei Steven Wu, and Virginia Smith. Private multi-task learning: Formulation and applications to federated learning. ar Xiv preprint ar Xiv:2108.12978, 2021. [11] Prateek Jain, John Rush, Adam Smith, Shuang Song, and Abhradeep Guha Thakurta. Differentially private model personalization. Advances in Neural Information Processing Systems, 34:29723 29735, 2021. [12] Alireza Fallah, Aryan Mokhtari, and Asuman Ozdaglar. Personalized federated learning: A meta-learning approach. ar Xiv preprint ar Xiv:2002.07948, 2020. [13] Yihan Jiang, Jakub Koneˇcn y, Keith Rush, and Sreeram Kannan. Improving federated learning personalization via model agnostic meta learning. ar Xiv preprint ar Xiv:1909.12488, 2019. [14] Brendan Mc Mahan, Eider Moore, Daniel Ramage, Seth Hampson, and Blaise Aguera y Arcas. Communication-efficient learning of deep networks from decentralized data. In Artificial intelligence and statistics, pages 1273 1282. PMLR, 2017. [15] H Brendan Mc Mahan, Daniel Ramage, Kunal Talwar, and Li Zhang. Learning differentially private recurrent language models. In International Conference on Learning Representations, 2018. [16] Robin C Geyer, Tassilo Klein, and Moin Nabi. Differentially private federated learning: A client level perspective. ar Xiv preprint ar Xiv:1712.07557, 2017. [17] Daniel Levy, Ziteng Sun, Kareem Amin, Satyen Kale, Alex Kulesza, Mehryar Mohri, and Ananda Theertha Suresh. Learning with user-level privacy. Advances in Neural Information Processing Systems, 34, 2021. [18] Cynthia Dwork, Frank Mc Sherry, Kobbi Nissim, and Adam Smith. Calibrating noise to sensitivity in private data analysis. In Theory of cryptography conference, pages 265 284. Springer, 2006. [19] Raef Bassily, Vitaly Feldman, Kunal Talwar, and Abhradeep Thakurta. Private stochastic convex optimization with optimal rates. In Proceedings of the 33rd International Conference on Neural Information Processing Systems, pages 11282 11291, 2019. [20] Yu Nesterov. Smooth minimization of non-smooth functions. Mathematical programming, 103(1):127 152, 2005. [21] Raef Bassily, Adam Smith, and Abhradeep Thakurta. Private empirical risk minimization: Efficient algorithms and tight error bounds. In 2014 IEEE 55th Annual Symposium on Foundations of Computer Science, pages 464 473. IEEE, 2014. [22] Vitaly Feldman, Tomer Koren, and Kunal Talwar. Private stochastic convex optimization: optimal rates in linear time. In Proceedings of the 52nd Annual ACM SIGACT Symposium on Theory of Computing, pages 439 449, 2020. [23] Peter Kairouz, Brendan Mc Mahan, Shuang Song, Om Thakkar, Abhradeep Thakurta, and Zheng Xu. Practical and private (deep) learning without sampling or shuffling. In International Conference on Machine Learning, pages 5213 5225. PMLR, 2021. [24] Jingcheng Liu and Kunal Talwar. Private selection from private candidates. In Proceedings of the 51st Annual ACM SIGACT Symposium on Theory of Computing, pages 298 309, 2019. [25] Nicolas Papernot and Thomas Steinke. Hyperparameter tuning with renyi differential privacy. ar Xiv preprint ar Xiv:2110.03620, 2021. [26] Yishay Mansour, Mehryar Mohri, Jae Ro, and Ananda Theertha Suresh. Three approaches for personalization with applications to federated learning. ar Xiv preprint ar Xiv:2002.10619, 2020. [27] Tensorflow-privacy. https://github.com/tensorflow/privacy, 2022. [28] Brenden Lake, Ruslan Salakhutdinov, Jason Gross, and Joshua Tenenbaum. One shot learning of simple visual concepts. In Proceedings of the annual meeting of the cognitive science society, volume 33, 2011. [29] Shai Shalev-Shwartz and Shai Ben-David. Understanding machine learning: From theory to algorithms. Cambridge university press, 2014. [30] H Brendan Mc Mahan. A survey of algorithms and analysis for adaptive online learning. The Journal of Machine Learning Research, 18(1):3117 3166, 2017. [31] Elad Hazan et al. Introduction to online convex optimization. Foundations and Trends in Optimization, 2 (3-4):157 325, 2016. [32] Shai Shalev-Shwartz, Ohad Shamir, Nathan Srebro, and Karthik Sridharan. Stochastic convex optimization. In COLT, volume 2, page 5, 2009. [33] Huang Gabriel. reptile-pytorch. https://github.com/gabrielhuang/reptile-pytorch, 2018. 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] (b) Did you describe the limitations of your work? [Yes] (c) Did you discuss any potential negative societal impacts of your work? [N/A] There are no foreseeable negative societal impacts for this work. (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [Yes] (b) Did you include complete proofs of all theoretical results? [Yes] 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] Included pseudocode and synthetic datasets details. (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [Yes] (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [No] 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [N/A] (b) Did you mention the license of the assets? [N/A] We use publicly available datasets. (c) Did you include any new assets either in the supplemental material or as a URL? [N/A] (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [N/A] (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [N/A] 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]