# private_identity_testing_for_highdimensional_distributions__44f93d3c.pdf Private Identity Testing for High-Dimensional Distributions Clément Canonne IBM Research, Almaden ccanonne@cs.columbia.edu Gautam Kamath Cheriton School of Computer Science University of Waterloo g@csail.mit.edu Audra Mc Millan Khoury College of Computer Sciences, Northeastern University Department of Computer Science, Boston University audramarymcmillan@gmail.com Jonathan Ullman Khoury College of Computer Sciences Northeastern University jullman@ccs.neu.edu Lydia Zakynthinou Khoury College of Computer Sciences Northeastern University zakynthinou.l@northeastern.edu We present novel differentially private identity (goodness-of-fit) testers for natural and widely studied classes of multivariate product distributions: product distributions over { 1}d and Gaussians in Rd with known covariance. Our testers have improved sample complexity compared to those derived from previous techniques, and are the first testers whose sample matches the order-optimal minimax sample complexity of O(d 1/2/α2) in many parameter regimes. We construct two types of testers, exhibiting tradeoffs between sample complexity and computational complexity. Finally, we provide a two-way reduction between testing a subclass of multivariate product distributions and testing univariate distributions, thereby obtaining upper and lower bounds for testing this subclass of product distributions. 1 Introduction A foundation of statistical inference is hypothesis testing: given two disjoint sets of probability distributions H0 and H1, we want to design an algorithm T that takes a random sample X from some distribution P H0 H1 and, with high probability, determines whether P is in H0 or H1. Hypothesis tests formalize yes-or-no questions about an underlying population given a random sample from that population, and are ubiquitous in the physical, life, and social sciences, where hypothesis tests with high confidence are the gold standard for publication in top journals. In many of these applications clinical trials, social network analysis, or demographic studies, to name a few this sample contains sensitive data belonging to individuals, in which case it is crucial for the hypothesis test to respect these individuals privacy. It is particularly desirable to guarantee differential privacy [33], which has become the de facto standard for the analysis of private data. Differential privacy is used as a measure of privacy for data analysis systems at Google [35], Apple [31], and the U.S. Census Bureau [26]. Differential privacy and related notions of algorithmic stability are also crucial for statistical validity even when confidentiality is not a direct concern, as they provide generalization guarantees in an adaptive setting [32, 8, 58]. 34th Conference on Neural Information Processing Systems (Neur IPS 2020), Vancouver, Canada. While differentially private hypothesis testing has been extensively studied (see Section 1.3), almost all this work has focused on low-dimensional distributions. Our main contribution is to give novel algorithms for hypothesis testing of high-dimensional distributions with improved sample complexity. In particular, we give differentially private algorithms for the following fundamental problems: 1. Given samples from a product distribution P over { 1}d, decide if P is the uniform distribution or is α-far from the uniform distribution in total variation distance. Or, equivalently, decide if E[P] = 0 or if E[P] 2 α. 2. Given samples from a product distribution P over {0, 1}d, decide if P is equal to some given extremely biased distribution Q with mean E[Q] O( 1 d) or is α-far from Q in total variation distance. In this case our tester achieves the provably optimal sample complexity. 3. Given samples from a multivariate Gaussian P in Rd whose covariance is known to be the identity Id d, decide if P is N(0, Id d) or is α-far from N(0, Id d) in total variation distance. Or, equivalently, decide if E[P] = 0 or if E[P] 2 α. Although we will focus on the first contribution since it highlights the main technical contributions, we note that the third contribution, (private) hypothesis testing on the mean of a Gaussian, is one of the most fundamental statistical primitives (see, e.g., the Z-test and Student s t-test). The main challenge in solving these high-dimensional testing problems privately is that the only known non-private test statistics for these problems have high worst-case sensitivity. That is, these test statistics can potentially be highly brittle to changing even a single one of the samples. We overcome this challenge by identifying two methods for reducing the sensitivity of the test statistic without substantially changing its average-case behavior on typical datasets sampled from the distributions we consider. The first is based on a novel private filtering method, which gives a computationally efficient tester. The second combines the method of Lipschitz extensions [13, 52] with recursive preconditioning, which yields an exponential-time tester, but with improved sample complexity. 1.1 Background: Private Hypothesis Testing We start by giving some background on private hypothesis testing. First, we say that an algorithm A: X {0, 1} is a differentially private hypothesis tester for a pair H0, H1 over domain X if 1. A is ε-differentially private in the worst case. That is, for every pair of samples X, X X n differing on one sample, A(X) and A(X ) are ε-close. In general, A(X) and A(X ) are (ε, δ)- close if for b {0, 1}, Pr[A(X) = b] eε Pr[A(X ) = b] + δ. 2. A distinguishes H0 from H1 on average. If X = (X(1), . . . , X(n)) is drawn i.i.d. from some P H0 then A(X) outputs 0 with high probability, and similarly if P H1. The minimum number of samples n such that A distinguishes H0 and H1 is the sample complexity of A. Note that we want testers that are private in the worst case, yet accurate on average. It is important for privacy to be a worst-case notion, rather than contingent on the assumption that the data is sampled i.i.d. from some P H0 H1 because we have no way of verifying this assumption (which may be a modeling simplification), and once privacy is lost it cannot be recovered. Worst-case privacy notions also enjoy strong composition and generalization properties not shared by average-case privacy. There exists a black-box method for obtaining a differentially private tester from any non-private tester A using the sample-and-aggregate framework [56]. Specifically, given any tester A with sample complexity n, we can obtain an ε-differentially private tester with sample complexity O(n/ε). When ε is a constant, the reduction is within a constant factor of the optimal sample complexity; however, this overhead factor in the sample complexity blows up as ε 0. One can often obtain stronger results using a white-box approach. For example, suppose P is a Bernoulli random variable and we aim to test if P = Ber(1/2) or P = Ber(1/2 + α). Nonprivately, Θ(1/α2) samples are necessary and sufficient. Thus the black-box approach gives a sample complexity of Θ(1/α2ε). However, if we work directly with the test statistic T(X) = 1 n Pn j=1 X(j), we can obtain privacy by computing T(X) + Z where Z is drawn from an appropriate distribution with standard deviation O(1/εn). One can now show that this private test succeeds using | {z } non-private sc | {z } overhead for privacy samples, which actually matches the non-private sample complexity up to a factor of 1 + o(1) unless ε is very small, namely ε = o(α). Our main contribution is to achieve qualitatively similar results for the high-dimensional testing problem we have described above. 1.2 Our Results Theorem 1.1 (Informal). There is a linear-time, ε-differentially private tester A that distinguishes the uniform distribution over { 1}d from any product distribution over { 1}d that is α-far in total variation distance using n = n(d, α, ε) samples for | {z } non-private sc | {z } overhead for privacy The sample complexity in Theorem 1.1 has an appealing form. One might even conjecture that this sample complexity is optimal by analogy with the case of privately estimating a product distribution over { 1}d or a Gaussian in Rd with known covariance, for which the sample complexity is Θ(d/α2+d/αε) in both cases [51]. However, the next result shows that there is in fact an exponentialtime private tester that has even lower sample complexity in some range of parameters. Theorem 1.2 (Informal). There is an exponential-time, ε-differentially private tester A that distinguishes the uniform distribution over { 1}d from any product distribution over { 1}d that is α-far in total variation distance using n = n(d, α, ε) samples for | {z } non-private sc αε1/2 + d1/3 α4/3ε2/3 + 1 | {z } overhead for privacy We remark that the sample complexity in (3) is precisely, up to logarithmic factors, the optimal sample complexity for testing uniformity of discrete distributions on the domain {1, . . . , d} [3], which hints that it may be optimal (especially in view of Theorem 1.3 below). The current best lower bound we can prove is n = Ω( d1/2 α2 + 1 αϵ). Bridging the gap between this lower bound and the upper bound seems like it would require developing fundamentally different lower bound techniques. Lower bounds for multivariate distribution testing appear to be much more challenging to prove than in the univariate case, due to the necessity of maintaining independence of marginals in the construction of a coupling, which is the standard technique for proving lower bounds for private distribution testing. The expression in (3) is rather complex and somewhat difficult to interpret and compare to (2). One way to simplify the comparison is to consider the range of the privacy parameter ε where privacy comes for free, meaning the sample complexity is dominated by the non-private term Θ(d1/2/α2). For the efficient algorithm, privacy comes for free roughly when ε = Ω(α). For the computationally inefficient algorithm, however, one can show that privacy comes for free roughly when ε = Ω(α2 + α/d1/4), which is better if both 1/α and d are superconstant. Using a simple reduction, we show that Theorems 1.1 and 1.2 extend, with a constant-factor loss in sample complexity, to identity testing for balanced product distributions. That is, suppose Q is a product distribution such that every coordinate of E[Q] is bounded away from 1 and +1. Then with a constant-factor loss in sample complexity, we can distinguish whether (i) a product distribution P over { 1}d is either equal to Q, or (ii) whether P is far from Q in total variation distance. We then focus on a specific class of Boolean product distributions, which we refer to as extreme. Informally, a product distribution is extreme if each of its marginals is O(1/d)-close to constant. For this restricted class of product distributions, we provide a two-way reduction showing that identity testing is equivalent to the identity testing in the univariate setting. This allows us to transfer known lower bounds on private univariate identity testing to our extreme product distribution class, which gives us the first non-trivial lower bounds for privately testing identity of product distributions. Theorem 1.3 (Informal). The sample complexity of privately testing identity of univariate distributions over [d] and the sample complexity of privately testing identity of extreme product distributions over { 1}d are equal, up to constant factors. Finally, we can obtain analogous results for identity testing of multivariate Gaussians with known covariance by reduction to uniformity testing of Boolean product distributions. Theorem 1.4 (Informal). There is a linear-time, ε-differentially private tester A that distinguishes the standard normal N(0, Id d) from any normal N(µ, Id d) that is α-far in total variation distance (or, equivalently, µ 2 α) using n = n(d, α, ε) samples for n = O( d1/2 α2 ) + O( d1/2 αε ). Theorem 1.5 (Informal). There is an exponential-time ε-differentially private tester A that distinguishes the standard normal N(0, Id d) from any normal N(µ, Id d) that is α-far in total variation distance using n = n(d, α, ε) samples for n = O( d1/2 α2 ) + O( d1/2 αε1/2 + d1/3 α4/3ε2/3 + 1 αε). We note that we can also obtain these results directly by extending the techniques of the Boolean case and constructing the tester. We demonstrate this for the computationally efficient test. We highlight some tools in this paper which are useful for hypothesis testing, even without privacy constraints: (1) A reduction from testing the mean of a Gaussian, to testing uniformity of a product distribution. (2) A reduction from testing identity to a balanced product distribution, to testing uniformity of a product distribution.(3) An equivalence between testing identity over a domain of size d, and testing identity to an extreme product distribution in d dimensions. 1.3 Related Work We focus on the most relevant related work here, broader coverage appears in the supplement. Hypothesis testing with a focus on minimax rates was initiated by Ingster and coauthors in the statistics community [46, 47, 48], and the work of Batu et al. [10, 9] in computer science (arising as a subfield of property testing [42, 43]). Works on testing in the multivariate setting include testing of independence [9, 5, 60, 54, 2, 30, 21], and testing on graphical models [20, 28, 27, 38, 1, 11]. We note that graphical models (both Ising models and Bayesian Networks) include the product distribution case we study in this paper. Surveys and more thorough coverage include [59, 19, 41, 7, 50]. Minimax study in the private was setting initiated by Cai et al. [17], and we now have close to optimal algorithms and lower bounds for the univariate setting [3, 4]. Other recent works focus on testing of simple hypotheses [25, 22]. Awan and Slavkovic [6] give a universally most powerful (UMP) test when the domain size is two, however Brenner and Nissim [16] shows that UMP tests don t exist for larger domains. A line of work complementary to the minimax setting [64, 63, 65, 37, 53, 49, 18, 62, 23] designs differentially private versions of popular test statistics. The Lipschitz-extension technique that we build upon was introduced in [13, 52], who also gave efficient Lipschitz extensions for graph statistics such as edge density and number of triangles in sparse graphs. Later work constructed efficient Lipschitz extensions for richer classes of graph statistics [57, 61]. [24] introduced a variant of the Lipschitz-extension machinery to give efficient algorithms for statistics like the median and trimmed mean. Recent results [15, 14] prove the existence of Lipschitz extensions for all differentially private algorithms, though efficiency is not a focus. 1.4 Techniques To avoid certain technicalities involving continuous and unbounded data, we will describe our tester for product distributions over { 1}d, rather than for Gaussians over Rd. The Gaussian case can be handled by using a reduction to the Boolean case, or directly, by using a nearly identical approach. First Attempts. A natural starting point is to study the asymptotically-optimal non-private tester of Canonne et al. [20]. Let P be a distribution over { 1}d and X = (X(1), . . . , X(n)) { 1}n d be n i.i.d. samples from P. Define the test statistic T(X) = X 2 2 nd where X = j=1 X(j) (4) The analysis of Canonne et al. shows that if P is the uniform distribution then E[T(X)] = 0, while if P is a product distribution that is α-far from uniform then E[T(X)] = Ω(α2n2). Moreover, the variance of T(X) is bounded so that, for n = O(d1/2/α2), we can distinguish between the two. To obtain a private tester, we need to add noise to T(X) with smaller magnitude than the α2n2 gap. The standard approach is to add noise to T(X) calibrated to the global sensitivity: GST = max X X (T(X) T(X )), where X X denotes that X and X are neighboring samples that differ on at most one sample. To ensure privacy it is then sufficient to compute a noisy statistic T(X) + Z where Z is chosen from an appropriate distribution (commonly, a Laplace distribution) with mean 0 and standard deviation O(GST /ε). One can easily see that the global sensitivity of T(X) is O(nd), so it suffices to add noise O(nd/ε). The tester will require n = Ω(d/(α2ε)) to succeed, giving an undesirable linear dependence in d. Note that this tester is dominated by the sample-and-aggregate approach. In view of the above, one way to find a better private tester would be to identify an alternative statistic for testing uniformity of product distributions with lower global sensitivity. However, we do not know of any other such test statistic that has asymptotically optimal non-private sample complexity; a prerequisite to achieving optimal private sample complexity. Intuition: High-Sensitivity is Atypical. Since we need privacy to hold in the worst case for every dataset, any private algorithm for computing T must have error proportional to GST on some dataset X. However, the utility of the tester only applies on average to typical datasets drawn from product distributions. Thus, we can try to find some alternative test statistic ˆT that is close to T on these typical datasets, but has lower global sensitivity. To this end, it will be instructive to look at the sensitivity of T at a particular pair of datasets X X , differing in samples X(j) and X (j). T(X) T(X ) = 2( X(j), X X (j), X ) (5) Notice that T(X) and T(X ) can only differ by a large amount when one of the two datasets contains a point X(j) such that X(j), X is large. Thus, if we could somehow restrict attention to datasets in C( ) = n X { 1}n d : j = 1, . . . , n | X(j), X | o the sensitivity would be at most 4 . As we will show, for typical datasets drawn from product distributions (except in some corner cases), the data will lie in C( ) for nd. For example, if we draw X(1), . . . , X(n) uniformly at random in { 1}d, then we have E max j [n] | X(j), X | d + (nd)1/2. More generally, if X(1), . . . , X(n) are drawn from any product distribution, then we show that max j [n] | X(j), X | 1 n X 2 2 + X 2. (6) Note that, except in pathological cases, the right-hand side will be much smaller than nd. Thus, for typical datasets drawn from product distributions, the sensitivity of T(X) should be much lower than its global sensitivity. We can exploit this in two different ways to obtain improved testers. A Sample-Efficient Tester via Lipschitz Extensions Suppose we fix some value of and we are only concerned with estimating T(X) accurately on nice datasets that lie in C( ). Even in this case, it would not suffice to add noise to T(X) proportional to /ε, since we need privacy for all datasets. However, we can compute T(X) accurately on C( ) while guaranteeing privacy for all datasets using the beautiful machinery of privacy via Lipschitz extensions introduced by Blocki et al. [13] and Kasiviswanathan et al. [52] in the context of node-differential-privacy for graph data. Specifically, for a function T defined on domain X, and a set C X, a Lipschitz extension of T from C is a function ˆT defined on all datasets such that: (1) ˆT(X) = T(X) for every X C. (2) The global sensitivity of ˆT on all of X is at most the sensitivity of T restricted to C, namely GS ˆT max X,X C T(X) T(X ). Perhaps surprisingly, such a Lipschitz extension exists for every real-valued function T and every subset of the domain C [55]! Once we have the Lipschitz extension, we can achieve privacy for all datasets by adding noise to ˆT proportional to the sensitivity of T on C. In general, the Lipschitz extension can be computed in time polynomial in |X|, which, in our case, is 2nd. Thus, if we knew a small value of such that X C( ) then we could compute a Lipschitz extension of T from C( ) and we would be done. From (6), we see that a good choice of is approximately 1 n X 2 2 + X 2. But X 2 2 is precisely the test statistic we wanted to compute! We untie this knot using a recursive approach. We start with some weak upper bound (m) such that we know X C( (m)). For example, we use the worst-case (1) = 4nd as a base case. Using the Lipschitz extension onto C( (m)), we can get a weak private estimate of the test statistic T(X) = X 2 2 nd with error (m)/ε. At this point, we may already be able to conclude that X was not sampled from the uniform distribution and reject. Otherwise, we can certify that X 2 2 nd + (m) ε .If X was indeed sampled from a product distribution, then (6) tells us that X C( (m+1)) for (m+1) X 2 2 n + X 2 d + (m) Then, as long as (m+1) is significantly smaller than (m) we can recurse, and get a better private estimate of X 2 2. Once the recurrence converges, we can stop and make a final decision whether to accept or reject. One can analyze this recurrence and show that it will converge rapidly to some = O(d+ nd+1/ε). Thus the final test will distinguish uniform from far-from-uniform provided /ε α2n2, which occurs at the sample complexity we claim. This recursive approach is loosely similar in spirit to methods in [51, 36], whereby we obtain more and more accurate private estimates, necessitating less noise addition at each step. However, our technique is very different from theirs, and the first to interact with the Lipschitz extension framework. We believe our work adds more evidence for the broad applicability of this perspective. A Computationally Efficient Tester via Private Filtering. The natural way to make the test above computationally efficient is to explicitly construct the Lipschitz extension ˆT onto C( ). Although we do not know how to do so, we can start with the following filtering approach: To compute ˆT(X), throw out every X(j) such that X(j), X > and replace it with a fresh uniform sample ˆX(j). Let the resulting dataset be Y and output T(Y ). Obviously this will agree with T(X) for X C( ), but it is not clear that it has sensitivity O( ). The reason is that the decision to throw out X(j) or not depends on the global quantity X = P j X(j). Thus, we could have one dataset X where no points are thrown out so Y = X, and a neighboring dataset X where, because X is slightly different from X, all points are thrown out so Y = . Then the difference between the test statistic on Y, Y could be much larger than . We solve this problem by modifying the algorithm to throw out points based on X(j), X for some private quantity X X. Although the proof is somewhat subtle, we can use the fact that X is private to argue that filtering based on X and then adding noise calibrated to preserves privacy. In order to compensate for the fact that X is only an approximation of X, we cannot take too small, which is why this tester has larger sample complexity than the non-computationally efficient one. Organisation. Due to space constraints, we focus in this extended abstract on Theorem 1.2, which illustrates the main ideas and challenges in our approach, and provide a detailed outline of its proof in Section 2. All omitted details and sections can be found in the supplementary material. 2 Uniformity Testing for Product-of-Bernoulli Distributions In this section, we provide the outline of our computationally inefficient algorithm for uniformity testing, Algorithm 2. We will prove the following main theorem (corresponding to Theorem 1.2). Theorem 2.1. Algorithm 2 is ε-differentially private. Furthermore, Algorithm 2 can distinguish between P = Ud and P Ud 1 α with probability at least 2/3 and has sample complexity αε1/2 + d1/3 α4/3ε2/3 + 1 αε . We first focus on the privacy guarantee of Theorem 2.1, which is based on the existence of Lipschitz extensions, as established by [55]. For our dataset X and any > 0, let us define C( ) = X { 1}n d j [n], | X(j), X | . The main element that we need for the privacy proof is the bound on the sensitivity of T on C( (m)) for all m [M]. This would ensure that T is 4 (m)-Lipschitz on C( (m)), so the 4 (m)-Lipschitz extensions ˆT exist in all the rounds and line 3 in LIPEXTTEST and line 11 in Algorithm 2 add enough noise to maintain privacy. Note that the algorithm would be private regardless of the choice of values (m). Recall the definition of the test statistic of (4): T(X) = X 2 2 nd for X = Pn j=1 X(j). The next lemma holds by a calculation of the sensitivity which led to equation (5) and the definition of C( ). Lemma 2.2 (Sensitivity of T). For any bound > 0, for two neighboring datasets X, X C( ), |T(X) T(X )| 4 . First, we note that LIPEXTTEST(X, ε , , β) is ε -DP, since it only accesses X via the Laplace mechanism in line 3. Since Algorithm 2 is a composition of M 1 invocations of LIPEXTTEST with DP parameter ε = ε/M, and of the Laplace mechanism in line 11, its privacy guarantee is a straightforward application of DP composition and post-processing. Lemma 2.3 (Privacy). Algorithm 2 is ε-differentially private. Algorithm 1 LIPEXTTEST Require: Sample X = (X(1), . . . , X(n)). Parameters ε, > 0, β (0, 1]. 1: Define the set C( ) = X { 1}n d j [n], | X(j), X | . 2: Let ˆT( ) be a 4 -Lipschitz extension of T from C( ) to all of { 1}n d. 3: Sample noise r Lap(4 /ε) and let z ˆT(X) + r. 4: if z > 10n d + 4 ln(1/β)/ε then return reject. 5: return accept. Algorithm 2 Private Uniformity Testing via Lipschitz Extension Require: Sample X = (X(1), . . . , X(n)) { 1}n d drawn from P n. Parameters ε, α, β > 0. 1: Let M log n , ε ε/M, β 1/(10n). 2: Let (1) nd and 1000 max(d, nd, ln(1/β)/ε ) ln(1/β). 3: for m 1 to M 1 do 4: if (m) then 5: Let (M) (m) and exit the loop. 6: else 7: if LIPEXTTEST(X, ε , (m), β) returns reject then return reject 8: Set (m+1) 11 d + 9: Define the set C( (M)) = X { 1}n d j [n], | X(j), X | (M) . 10: Let ˆT( ) be a 4 (M)-Lipschitz extension of T from C( (M)) to all of { 1}n d. 11: Sample noise r Lap(4 (M)/ε ) and let z ˆT(X) + r. 12: if z > 1 4n(n 1)α2 then return reject 13: return accept. We now turn our attention to the utility guarantee. The next lemma will help us determine how the bound on the sensitivity of the statistic decreases in each round, as stated in (6). It is established by a concentration bound on the sum of the independent r.v. s Yi = xi Xi | X, around its mean X 2 2/n. Lemma 2.4. If X is drawn i.i.d. from a product distribution, then, with probability at least 1 2nβ, x X, | x, X | X 2 2 n + ln(1/β). (7) As in Algorithm 2, we let = 1000 max(d, nd, M ln(1/β)/ε) ln(1/β), where β = 1/(10n) and M = log(n) . Based on the previous lemma, we now analyze the key subroutine, LIPEXTTEST: 1. If X is drawn from a product distribution and, for some , LIPEXTTEST returns accept, then X C( ), where , determined as in line 8 of Algorithm 2 (Lemma 2.5); 2. If X is drawn from the uniform distribution, LIPEXTTEST will return accept (Lemma 2.6). Lemma 2.5 (Sensitivity reduction). Fix some 0, and let = 11(d + nd + /(nε) + p /ε) ln(1/β). If (i) X is drawn from a product distribution, (ii) X satisfies (7), (iii) X C( ), and (iv) LIPEXTTEST(X, ε, , β) returns accept, then X C( ) with probability at least 1 β. To prove the lemma, we first observe that, by condition (iii), ˆT(X) = T(X) = X 2 2 nd. So condition (iv) and the Laplace mechanism guarantee that X 2 2 11(nd + ln(1/β)/ε), with probability 1 β. By conditions (i) and (ii), (7) implies that x X | x, X | for the given . Lemma 2.6. If (i) X is drawn from the uniform distribution Ud, satisfies T(X) 10n d, and (7), and (ii) , then LIPEXTTEST(X, ε, , β) returns accept with probability at least 1 β. Proof. By condition (i), X 2 2 = T(X) + nd 11nd. Plugging this in (7), we get that x X, | x, X | 11nd/n + 2 p nd ln(1/β) 22 max(d, p nd ln(1/β)) , where the last inequality holds by condition (ii). Thus, X C( ) so ˆT(X) = T(X) in line 2 and we can analyze line 4 by bounding ˆT(X) + r = T(X) + r 10n d + r. By the Laplace mechanism, |r| 4 ln(1/β)/ε, with probability 1 β. Hence, LIPEXTTEST accepts with probability 1 β. With the analysis of this subroutine, we turn to the utility guarantee, which, unlike the privacy, only needs to hold for datasets drawn from a product distribution. The crux of the proof is the following: 1. If X is drawn i.i.d. from a product distribution and passes line 3 in round m, then it belongs in C( (m+1)) with high probability. Thus in every round that the dataset has not been rejected, we have that with high probability ˆT(X) = T(X). 2. If P = Ud, then with high probability X passes all the steps in the loop at line 3. 3. The number of rounds M is sufficient to guarantee that the sensitivity and thus the noise added to ˆT(X) in line 11 is small enough that one distinguishes between the two hypotheses. For the latter, we will need to bound (M), i.e., the sensitivity of T(X) restricted to C( (M)). The proof follows by observing that in each round where > , is reduced by 1/2, based on line 8. Lemma 2.7. For M = log(n) , ε = ε/M, and n = Ω(log(1/β)/ε ), (M) . We can now complete our proof outline, showing that the last test will give the expected guarantees. Since our test relies on a private adaptation of the test from [20], we state a version of its guarantees. Lemma 2.8 (Non-private Test Guarantees). For the test statistic T(X) = X 2 2 nd: If P = Ud then E[T(X)] = 0, whereas if P Ud 1 α then E[T(X)] > 1 Var(T(X)) 2n2d + 4n E[T(X)]. Proof of Theorem 2.1. The privacy guarantee was established in Lemma 2.3. First, note that given (1) = nd, we have C( (1)) = { 1}n d, so X C( (1)). Completeness: Suppose X is drawn from Ud. By the guarantees of the non-private test and Chebyshev s inequality, with probability 49/50, T(X) 10n d. Moreover, since Ud is a fortiori a product distribution, X satisfies (7) with probability at least 1 2nβ. Assume both of these events hold. By Lemma 2.6 and a union bound over the M 1 calls in line 7, we get that with probability at least 1 Mβ every call returns accept. Assume this holds. Invoking Lemma 2.5, this implies that for all rounds m, X C( (m)) except with probability at most Mβ, and in particular X C( (M)). In this case, we have ˆT(X) = T(X) in line 11. It remains to show that with high probability, z = T(X) + r n(n 1)α2/4 in line 12. Since r Lap 4 (M)/ε , with probability at least 49/50, |r| 4 (M)(ln 50)/ε . Again, condition on this, and suppose for now that n satisfies |r| 4 (M) ln 50 Then, by Chevyshev s and the non-private test guarantees, we can bound the probability that Algorithm 2 rejects in line 11 as Pr z > n(n 1)α2/4 1/50 for n = Ω(d1/2/α2). Therefore, by an overall union bound and β = 1/(10n), the algorithm rejects with probability at most 1/3. It remains to determine the constraint on n for inequality (8) to hold. By Lemma 2.7, it suffice for n2α2 = Ω( /ε ) = Ω max(d, nd, log(1/β)/ε ) log(1/β)/ε . Recalling our choice of ε = ε/M, β = 1/(10n), and M = log(n) , it suffices to choose αε1/2 + d1/3 α4/3ε2/3 + 1 Soundness: Suppose X is drawn from a product distribution which is α-far from uniform. By Lemma 2.4, X satisfies (7) with probability at least 1 2nβ. Suppose that the algorithm does not return reject in any of the M 1 calls in line 7 (otherwise, we are done). Conditioning on this, by Lemma 2.5 and an immediate induction, with probability 1 Mβ, X C( (m)) for all rounds m, and in particular X C( (M)). The rest of the proof follows exactly as the completeness, relying on (8), the guarantees of the Laplace mechanism and the non-private test, and Chebyshev s inequality. Combining equation (9) with the non-private bound Ω( d1/2 α2 ) yields the stated sample complexity. Broader Impact This work focuses on privacy-preserving statistics: the task of performing statistical learning while maintaining the privacy of data participants. The right to privacy is a fundamental right, and works in this area allow for the facilitation of valuable social science and economic research, while preserving this fundamental right. There are many examples where releases of statistical data, even aggregate statistics, without formal privacy guarantees has resulted in blatant privacy violations. For example, Homer et al. [45] demonstrated that it is possible to identify individuals who participated in a GWAS study based on certain released test statistics.This can be especially harmful when resolving individuals who participated in trials pertaining to a socially stigmatized disease (e.g., HIV/AIDS). This finding was considered so significant that the NIH immediately revoked open access to a number of statistical results from medical trials, including χ2-statistics and p values. Hypothesis tests are applied to sensitive data in a myriad of studies, e.g. on voting behavior [44] or attitude toward abortion [34], highlighting urgent need for hypothesis tests which respect the privacy of participant data. Some specific applications of goodness-of-fit testing are [12, 39, 40]. The study of uniformity testing of product distributions is also equivalent to the well-motivated question of mean testing of categorical data, which has applications to high-dimensional statistical analysis. Further, uniformity testing is also deeply related to mean testing of high-dimensional normal distributions, as exemplified by our reduction from the latter to the former. Non-DP mean testing in high-dimensional settings itself has a long history in statistical hypothesis testing; see for example, [29]. Our objective is to initiate the study of high-dimensional hypothesis testing under privacy constraints. Although we do not close the chapter on identity testing, we believe our results to be an important, non-trivial, and natural first step, and a prerequisite to any further investigation (e.g., general identity testing for non-spherical Gaussians, graphical models, or sub-Gaussian distributions). Our work exposes and addresses a number of core challenges that arise in private high-dimensional data analysis. Indeed, the problems we study capture a challenge that is widespread in differential privacy: that solving high dimensional testing problems privately is difficult because the known non-private testers have high global sensitivity. We propose a novel way of dealing with this issue by reducing to the typical sensitivity. We demonstrate the effectiveness of this technique on identity testing, but we conjecture that it will find applications to other testing problems and beyond. Acknowledgments and Disclosure of Funding Part of this work was done while GK, AM, JU, and LZ were visiting the Simons Institute for the Theory of Computing, and while CC was a Motwani Postdoctoral Fellow at Stanford University. GK was supported as a Microsoft Research Fellow, as part of the Simons-Berkeley Research Fellowship program. Some of this work was completed while visiting Microsoft Research, Redmond. AM was supported by NSF grant CCF-1763786, a Sloan Foundation Research Award, and a postdoctoral fellowship from BU s Hariri Institute for Computing. JU and LZ were supported by NSF grants CCF1718088, CCF-1750640, and CNS-1816028. Subsequently, LZ has been supported by a Facebook fellowship. CC was supported by a Goldstine Fellowship. The authors would like to thank Jayadev Acharya and Himanshu Tyagi for discussions that contributed to the reduction of identity testing of multivariate Gaussians to uniformity testing of Boolean product distributions. [1] Jayadev Acharya, Arnab Bhattacharyya, Constantinos Daskalakis, and Saravanan Kandasamy. Learning and testing causal models with interventions. In Advances in Neural Information Processing Systems 31, Neur IPS 18. Curran Associates, Inc., 2018. [2] Jayadev Acharya, Constantinos Daskalakis, and Gautam Kamath. Optimal testing for properties of distributions. In Advances in Neural Information Processing Systems 28, NIPS 15, pages 3577 3598. Curran Associates, Inc., 2015. [3] Jayadev Acharya, Ziteng Sun, and Huanyu Zhang. Differentially private testing of identity and closeness of discrete distributions. In Advances in Neural Information Processing Systems 31, Neur IPS 18, pages 6878 6891. Curran Associates, Inc., 2018. [4] Maryam Aliakbarpour, Ilias Diakonikolas, and Ronitt Rubinfeld. Differentially private identity and closeness testing of discrete distributions. In Proceedings of the 35th International Conference on Machine Learning, ICML 18, pages 169 178. JMLR, Inc., 2018. [5] Noga Alon, Alexandr Andoni, Tali Kaufman, Kevin Matulef, Ronitt Rubinfeld, and Ning Xie. Testing k-wise and almost k-wise independence. In Proceedings of the 39th Annual ACM Symposium on the Theory of Computing, STOC 07, pages 496 505, New York, NY, USA, 2007. ACM. [6] Jordan Awan and Aleksandra Slavkovi c. Differentially private uniformly most powerful tests for binomial data. In Advances in Neural Information Processing Systems 31, Neur IPS 18, pages 4208 4218. Curran Associates, Inc., 2018. [7] Sivaraman Balakrishnan and Larry Wasserman. Hypothesis testing for high-dimensional multinomials: A selective review. The Annals of Applied Statistics, 12(2):727 749, 2018. [8] Raef Bassily, Kobbi Nissim, Adam Smith, Thomas Steinke, Uri Stemmer, and Jonathan Ullman. Algorithmic stability for adaptive data analysis. In Proceedings of the 48th Annual ACM Symposium on the Theory of Computing, STOC 16, pages 1046 1059, New York, NY, USA, 2016. ACM. [9] Tu gkan Batu, Eldar Fischer, Lance Fortnow, Ravi Kumar, Ronitt Rubinfeld, and Patrick White. Testing random variables for independence and identity. In Proceedings of the 42nd Annual IEEE Symposium on Foundations of Computer Science, FOCS 01, pages 442 451, Washington, DC, USA, 2001. IEEE Computer Society. [10] Tu gkan Batu, Lance Fortnow, Ronitt Rubinfeld, Warren D. Smith, and Patrick White. Testing that distributions are close. In Proceedings of the 41st Annual IEEE Symposium on Foundations of Computer Science, FOCS 00, pages 259 269, Washington, DC, USA, 2000. IEEE Computer Society. [11] Ivona Bezakova, Antonio Blanca, Zongchen Chen, Daniel Štefankoviˇc, and Eric Vigoda. Lower bounds for testing graphical models: Colorings and antiferromagnetic Ising models. In Proceedings of the 32nd Annual Conference on Learning Theory, COLT 19, pages 283 298, 2019. [12] Aaron Blair, P Decoufle, and Dan Grauman. Causes of death among laundry and dry cleaning workers. American journal of public health, 69:508 11, 06 1979. [13] Jeremiah Blocki, Avrim Blum, Anupam Datta, and Or Sheffet. Differentially private data analysis of social networks via restricted sensitivity. In Proceedings of the 4th Conference on Innovations in Theoretical Computer Science, ITCS 13, pages 87 96, New York, NY, USA, 2013. ACM. [14] Christian Borgs, Jennifer Chayes, Adam Smith, and Ilias Zadik. Private algorithms can always be extended. ar Xiv preprint ar Xiv:1810.12518, 2018. [15] Christian Borgs, Jennifer Chayes, Adam Smith, and Ilias Zadik. Revealing network structure, confidentially: Improved rates for node-private graphon estimation. In Proceedings of the 59th Annual IEEE Symposium on Foundations of Computer Science, FOCS 18, pages 533 543, Washington, DC, USA, 2018. IEEE Computer Society. [16] Hai Brenner and Kobbi Nissim. Impossibility of differentially private universally optimal mechanisms. SIAM Journal on Computing, 43(5):1513 1540, 2014. [17] Bryan Cai, Constantinos Daskalakis, and Gautam Kamath. Priv it: Private and sample efficient identity testing. In Proceedings of the 34th International Conference on Machine Learning, ICML 17, pages 635 644. JMLR, Inc., 2017. [18] Zachary Campbell, Andrew Bray, Anna Ritz, and Adam Groce. Differentially private ANOVA testing. In Proceedings of the 2018 International Conference on Data Intelligence and Security, ICDIS 18, pages 281 285, Washington, DC, USA, 2018. IEEE Computer Society. [19] Clément L. Canonne. A survey on distribution testing: Your data is big. but is it blue? Electronic Colloquium on Computational Complexity (ECCC), 22(63), 2015. [20] Clément L. Canonne, Ilias Diakonikolas, Daniel M. Kane, and Alistair Stewart. Testing Bayesian networks. In Proceedings of the 30th Annual Conference on Learning Theory, COLT 17, pages 370 448, 2017. [21] Clément L. Canonne, Ilias Diakonikolas, Daniel M. Kane, and Alistair Stewart. Testing conditional independence of discrete distributions. In Proceedings of the 50th Annual ACM Symposium on the Theory of Computing, STOC 18, pages 735 748, New York, NY, USA, 2018. ACM. [22] Clément L. Canonne, Gautam Kamath, Audra Mc Millan, Adam Smith, and Jonathan Ullman. The structure of optimal private tests for simple hypotheses. In Proceedings of the 51st Annual ACM Symposium on the Theory of Computing, STOC 19, New York, NY, USA, 2019. ACM. [23] Simon Couch, Zeki Kazan, Kaiyan Shi, Andrew Bray, and Adam Groce. Differentially private nonparametric hypothesis testing. In Proceedings of the 2019 ACM Conference on Computer and Communications Security, CCS 19, New York, NY, USA, 2019. ACM. [24] Rachel Cummings and David Durfee. Individual sensitivity preprocessing for data privacy. In Proceedings of the 31st Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 20, Philadelphia, PA, USA, 2020. SIAM. [25] Rachel Cummings, Sara Krehbiel, Yajun Mei, Rui Tuo, and Wanrong Zhang. Differentially private change-point detection. In Advances in Neural Information Processing Systems 31, Neur IPS 18. Curran Associates, Inc., 2018. [26] Aref N. Dajani, Amy D. Lauger, Phyllis E. Singer, Daniel Kifer, Jerome P. Reiter, Ashwin Machanavajjhala, Simson L. Garfinkel, Scot A. Dahl, Matthew Graham, Vishesh Karwa, Hang Kim, Philip Lelerc, Ian M. Schmutte, William N. Sexton, Lars Vilhuber, and John M. Abowd. The modernization of statistical disclosure limitation at the U.S. census bureau, 2017. Presented at the September 2017 meeting of the Census Scientific Advisory Committee. [27] Constantinos Daskalakis, Nishanth Dikkala, and Gautam Kamath. Testing Ising models. In Proceedings of the 29th Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 18, pages 1989 2007, Philadelphia, PA, USA, 2018. SIAM. [28] Constantinos Daskalakis and Qinxuan Pan. Square Hellinger subadditivity for Bayesian networks and its applications to identity testing. In Proceedings of the 30th Annual Conference on Learning Theory, COLT 17, pages 697 703, 2017. [29] A. P. Dempster. A high dimensional two sample significance test. Ann. Math. Statist., 29(4):995 1010, 12 1958. [30] Ilias Diakonikolas and Daniel M. Kane. A new approach for testing properties of discrete distributions. In Proceedings of the 57th Annual IEEE Symposium on Foundations of Computer Science, FOCS 16, pages 685 694, Washington, DC, USA, 2016. IEEE Computer Society. [31] Differential Privacy Team, Apple. Learning with privacy at scale. https: //machinelearning.apple.com/docs/learning-with-privacy-at-scale/ appledifferentialprivacysystem.pdf, December 2017. [32] Cynthia Dwork, Vitaly Feldman, Moritz Hardt, Toniann Pitassi, Omer Reingold, and Aaron Roth. The reusable holdout: Preserving validity in adaptive data analysis. Science, 349(6248):636 638, 2015. [33] Cynthia Dwork, Frank Mc Sherry, Kobbi Nissim, and Adam Smith. Calibrating noise to sensitivity in private data analysis. In Proceedings of the 3rd Conference on Theory of Cryptography, TCC 06, pages 265 284, Berlin, Heidelberg, 2006. Springer. [34] Helen Ebaugh and C. Haney. Church attendance and attitudes toward abortion: Differentials in liberal and conservative churches. Journal for the Scientific Study of Religion, 17:407, 12 1978. [35] Úlfar Erlingsson, Vasyl Pihur, and Aleksandra Korolova. RAPPOR: Randomized aggregatable privacy-preserving ordinal response. In Proceedings of the 2014 ACM Conference on Computer and Communications Security, CCS 14, pages 1054 1067, New York, NY, USA, 2014. ACM. [36] Vitaly Feldman, Tomer Koren, and Kunal Talwar. Private stochastic convex optimization: Optimal rates in linear time. In Proceedings of the 52nd Annual ACM Symposium on the Theory of Computing, STOC 20, New York, NY, USA, 2020. ACM. [37] Marco Gaboardi, Hyun-Woo Lim, Ryan M. Rogers, and Salil P. Vadhan. Differentially private chi-squared hypothesis testing: Goodness of fit and independence testing. In Proceedings of the 33rd International Conference on Machine Learning, ICML 16, pages 1395 1403. JMLR, Inc., 2016. [38] Reza Gheissari, Eyal Lubetzky, and Yuval Peres. Concentration inequalities for polynomials of contracting Ising models. Electronic Communications in Probability, 23(76):1 12, 2018. [39] J Gill, J Endres-Brooks, P Bauer, W.J. Marks, and Robert Montgomery. The effect of abo blood group on the diagnosis of von willebrand disease. Blood, 69:1691 5, 07 1987. [40] William A. Glaser. The family and voting turnout. The Public Opinion Quarterly, 23(4):563 570, 1959. [41] Oded Goldreich. Introduction to Property Testing. Cambridge University Press, 2017. [42] Oded Goldreich, ShafiGoldwasser, and Dana Ron. Property testing and its connection to learning and approximation. In Proceedings of the 37th Annual IEEE Symposium on Foundations of Computer Science, FOCS 96, pages 339 348, Washington, DC, USA, 1996. IEEE Computer Society. [43] Oded Goldreich and Dana Ron. On testing expansion in bounded-degree graphs. Electronic Colloquium on Computational Complexity (ECCC), 7(20), 2000. [44] Anthony G Greenwald, Catherine G Carnot, Rebecca Beach, and Barbara Young. Increasing voting behavior by asking people if they expect to vote. Journal of Applied Psychology, 72(2):315, 1987. [45] Nils Homer, Szabolcs Szelinger, Margot Redman, David Duggan, Waibhav Tembe, Jill Muehling, John V. Pearson, Dietrich A. Stephan, Stanley F. Nelson, and David W. Craig. Resolving individuals contributing trace amounts of DNA to highly complex mixtures using high-density SNP genotyping microarrays. PLo S Genetics, 4(8):1 9, 2008. [46] Yuri Izmailovich Ingster. Minimax detection of a signal in ℓp metrics. Journal of Mathematical Sciences, 68(4):503 515, 1994. [47] Yuri Izmailovich Ingster. Adaptive chi-square tests. Zapiski Nauchnykh Seminarov POMI, 244:150 166, 1997. [48] Yuri Izmailovich Ingster and Irina A. Suslina. Nonparametric Goodness-of-fit Testing Under Gaussian Models, volume 169 of Lecture Notes in Statistics. Springer, 2003. [49] Kazuya Kakizaki, Jun Sakuma, and Kazuto Fukuchi. Differentially private chi-squared test by unit circle mechanism. In Proceedings of the 34th International Conference on Machine Learning, ICML 17, pages 1761 1770. JMLR, Inc., 2017. [50] Gautam Kamath. Modern Challenges in Distribution Testing. Ph D thesis, Massachusetts Institute of Technology, September 2018. [51] Gautam Kamath, Jerry Li, Vikrant Singhal, and Jonathan Ullman. Privately learning highdimensional distributions. In Proceedings of the 32nd Annual Conference on Learning Theory, COLT 19, pages 1853 1902, 2019. [52] Shiva Prasad Kasiviswanathan, Kobbi Nissim, Sofya Raskhodnikova, and Adam Smith. Analyzing graphs with node differential privacy. In Proceedings of the 10th Conference on Theory of Cryptography, TCC 13, pages 457 476, Berlin, Heidelberg, 2013. Springer. [53] Daniel Kifer and Ryan M. Rogers. A new class of private chi-square tests. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, AISTATS 17, pages 991 1000. JMLR, Inc., 2017. [54] Reut Levi, Dana Ron, and Ronitt Rubinfeld. Testing properties of collections of distributions. Theory of Computing, 9(8):295 347, 2013. [55] Edward J. Mc Shane. Extension of range of functions. Bulletin of the American Mathematical Society, 40(12):837 842, 1934. [56] Kobbi Nissim, Sofya Raskhodnikova, and Adam Smith. Smooth sensitivity and sampling in private data analysis. In Proceedings of the 39th Annual ACM Symposium on the Theory of Computing, STOC 07, pages 75 84, New York, NY, USA, 2007. ACM. [57] Sofya Raskhodnikova and Adam D. Smith. Lipschitz extensions for node-private graph statistics and the generalized exponential mechanism. In Proceedings of the 57th Annual IEEE Symposium on Foundations of Computer Science, FOCS 16, pages 495 504, Washington, DC, USA, 2016. IEEE Computer Society. [58] Ryan Rogers, Aaron Roth, Adam Smith, and Om Thakkar. Max-information, differential privacy, and post-selection hypothesis testing. In Proceedings of the 57th Annual IEEE Symposium on Foundations of Computer Science, FOCS 16, pages 487 494, Washington, DC, USA, 2016. IEEE Computer Society. [59] Ronitt Rubinfeld. Taming big probability distributions. XRDS, 19(1):24 28, 2012. [60] Ronitt Rubinfeld and Ning Xie. Testing non-uniform k-wise independent distributions over product spaces. In Proceedings of the 37th International Colloquium on Automata, Languages, and Programming, ICALP 10, pages 565 581, 2014. [61] Adam Sealfon and Jonathan Ullman. Efficiently estimating Erdos-Renyi graphs with node differential privacy. In Advances in Neural Information Processing Systems 32, Neur IPS 19. Curran Associates, Inc., 2019. [62] Marika Swanberg, Ira Globus-Harris, Iris Griffith, Anna Ritz, Adam Groce, and Andrew Bray. Improved differentially private analysis of variance. Proceedings on Privacy Enhancing Technologies, 2019(3), 2019. [63] Caroline Uhler, Aleksandra Slavkovi c, and Stephen E. Fienberg. Privacy-preserving data sharing for genome-wide association studies. The Journal of Privacy and Confidentiality, 5(1):137 166, 2013. [64] Duy Vu and Aleksandra Slavkovi c. Differential privacy for clinical trial data: Preliminary evaluations. In 2009 IEEE International Conference on Data Mining Workshops, ICDMW 09, pages 138 143. IEEE, 2009. [65] Yue Wang, Jaewoo Lee, and Daniel Kifer. Revisiting differentially private hypothesis tests for categorical data. ar Xiv preprint ar Xiv:1511.03376, 2015.