# the_broad_optimality_of_profile_maximum_likelihood__01d73a05.pdf The Broad Optimality of Profile Maximum Likelihood Yi Hao Dept. of Electrical and Computer Engineering University of California, San Diego yih179@ucsd.edu Alon Orlitsky Dept. of Electrical and Computer Engineering University of California, San Diego alon@ucsd.edu We study three fundamental statistical learning problems: distribution estimation, property estimation, and property testing. We establish the profile maximum likelihood (PML) estimator as the first unified sample-optimal approach to a wide range of learning tasks. In particular, for every alphabet size k and desired accuracy ε: Distribution estimation Under ℓ1 distance, PML yields optimal Θ(k/(ε2 log k)) sample complexity for sorted distribution estimation, and a PML-based estimator empirically outperforms the Good-Turing estimator on the actual distribution; Additive property estimation For a broad class of additive properties, the PML plug-in estimator uses just four times the sample size required by the best estimator to achieve roughly twice its error, with exponentially higher confidence; α-Rényi entropy estimation For an integer α > 1, the PML plug-in estimator has optimal k1 1/α sample complexity; for non-integer α > 3/4, the PML plug-in estimator has sample complexity lower than the state of the art; Identity testing In testing whether an unknown distribution is equal to or at least ε far from a given distribution in ℓ1 distance, a PML-based tester achieves the optimal sample complexity up to logarithmic factors of k. With minor modifications, most of these results also hold for a near-linear-time computable variant of PML. 1 Introduction A distribution p over a discrete alphabet X of size k corresponds to an element of the simplex x X p(x) = 1 A distribution property is a mapping f : X R associating a real value with each distribution. A distribution property f is symmetric if it is invariant under domain-symbol permutations. A symmetric property is additive, i.e., additively separable, if it can be written as f(p) := P x f(p(x)), where for simplicity we use f to denote both the property and the corresponding real function. Many important symmetric properties are additive. For example, Support size S(p) := P x 1p(x)>0, a fundamental quantity arising in the study of vocabulary size [29, 53, 67], population estimation [34, 52], and database studies [37]. Support coverage Cm(p) := P x(1 (1 p(x))m), where m is a given parameter, the expected number of distinct elements observed in a sample of size m, arising in biological [17, 49] and ecological [17 19, 23] research; Shannon entropy H(p) := P x p(x) log p(x), the primary measure of information [24, 66] with numerous applications to machine learning [14, 22, 63] and neuroscience [30, 51]; 33rd Conference on Neural Information Processing Systems (Neur IPS 2019), Vancouver, Canada. Distance to uniformity D(p) := p pu 1, where pu is the uniform distribution over X , a property being central to the field of distribution property testing [10, 12, 15, 65]. Besides being additive and symmetric, these four properties have yet another attribute in common. Under the appropriate interpretation, they are also all 1-Lipschitz. Specifically, for two distributions p, q X , let Γp,q be the collection of distributions over X X with marginals p and q on the first and second factors respectively. The relative earth-mover distance [70], between p and q is R(p, q) := inf γ Γp,q E (X,Y ) γ One can verify [70, 71] that H, D, and Cm := Cm/m are all 1-Lipschitz on the metric space ( X , R), and S := S/k is 1-Lipschitz over ( 1/k, R), the set of distributions in X whose nonzero probabilities are at least 1/k. We will study all such Lipschitz properties in later sections. An important symmetric non-additive property is Rényi entropy, a well-known measure of randomness with numerous applications to unsupervised learning [44, 77] and image registration [50, 54]. For a distribution p X and a non-negative real parameter α = 1, the α-Rényi entropy [64] of p is Hα(p) := (1 α) 1 log (P x pα x). In particular, denoted by H1(p) := limα 1 Hα(p), the 1-Rényi entropy is exactly Shannon entropy [64]. 1.1 Problems of interest In this work, we consider three fundamental statistical learning problems concerning the estimation and testing of distributions and their properties. (Sorted) distribution estimation A natural learning problem is to estimate an unknown distribution p X from an i.i.d. sample Xn p. For any two distributions p, q X , let ℓ(p, q) be the loss when we approximate p by q. A distribution estimator ˆp : X X associates every sequence xn X with a distribution ˆp(xn). We measure the performance of an estimator by its sample complexity n(ˆp, ε, δ) := min{n: p X , Pr Xn p (ℓ(p, ˆp(Xn)) ε) δ}, the smallest sample size that ˆp requires to estimate all distributions in X to a desired accuracy ε > 0, with error probability δ (0, 1). The sample complexity of distribution estimation over X is n(ε, δ) := min{n(ˆp, ε, δ): ˆp : X X }, the lowest sample complexity of any estimator. For simplicity, we will omit δ when δ = 1/3. For a distribution p X , we denote by {p} the multiset of its probabilities. The sorted ℓ1 distance between two distributions p, q X is ℓ< 1(p, q) := min p X :{p }={p} p q 1 , the smallest ℓ1 distance between q and any sorted version of p. As illustrated in Section 7.1 of the supplementary material, this is essentially the 1-Wasserstein distance between uniform measures on the probability multisets {p} and {q}. We will consider both the sorted and unsorted ℓ1 distances. Property estimation Often we would like to estimate a given property f of an unknown distribution p X based on a sample Xn p. A property estimator is a mapping ˆf : X R. Analogously, the sample complexity of ˆf in estimating f over a set P X is nf( ˆf, P, ε, δ) := min{n: p P, Pr Xn p(| ˆf(Xn) f(p)| ε) δ}, the smallest sample size that ˆf requires to estimate f with accuracy ε and confidence 1 δ, for all distributions in P. The sample complexity of estimating f over P is nf(P, ε, δ) := min{nf( ˆf, P, ε, δ): ˆf : X R}, the lowest sample complexity of any estimator. For simplicity, we will omit P when P = X , and omit δ when δ = 1/3. The standard median trick" shows that log(1/δ) nf(P, ε) Ω(nf(P, ε, δ)). By convention, we say an estimator ˆf is sample-optimal if nf( ˆf, P, ε) = Θ(nf(P, ε)). Property testing: Identity testing A closely related problem is distribution property testing, of which identity testing is the most fundamental and well-studied [15, 32]. Given an error parameter ε, a distribution q, and a sample Xn from an unknown distribution p, identity testing aims to distinguish between the null hypothesis and the alternative hypothesis H1 : p q 1 ε. A property tester is a mapping ˆt : X {0, 1}, indicating whether H0 or H1 is accepted. Analogous to the two formulations above, the sample complexity of ˆt is nq(ˆt, ε, δ) := min{n: i {0, 1} and p Hi, Pr Xn p ˆt(Xn) = i δ}, and the sample complexity of identity testing with respect to q is nq(ε, δ) := min{n(ˆt, ε, δ): ˆt : X {0, 1}}. Again, when δ = 1/3, we will omit δ. For q = pu, the problem is also known as uniformity testing. 1.2 Profile maximum likelihood The multiplicity of a symbol x X in a sequence xn := x1, . . . , xn X is µx(xn) := |{j : xj = x, 1 j n}|, the number of times x appears in xn. These multiplicities induce an empirical distribution pµ(xn) that associates a probability µx(xn)/n with each symbol x X. The prevalence of an integer i 0 in xn is the number ϕi(xn) of symbols appearing i times in xn. For known X, the value of ϕ0 can be deduced from the remaining multiplicities, hence we define the profile of xn to be ϕ(xn) = (ϕ1(xn), . . . , ϕn(xn)), the vector of all positive prevalences. For example, ϕ(alfalfa) = (0, 2, 1, 0, 0, 0, 0). Note that the profile of xn also corresponds to the multiset of multiplicities of distinct symbols in xn. For a distribution p X , let p(xn) := Pr Xn p(Xn = xn) be the probability of observing a sequence xn under i.i.d. sampling from p, and let yn:ϕ(yn)=ϕ p(yn) be the probability of observing a profile ϕ. While the sequence maximum likelihood estimator maps a sequence to its empirical distribution, which maximizes the sequence probability p(xn), the profile maximum likelihood (PML) estimator [58] over a set P X maps each profile ϕ to a distribution pϕ := arg max p p(ϕ) that maximizes the profile probability. Relaxing the optimization objective, for any β (0, 1), a βapproximate PML estimator [4] maps each profile ϕ to a distribution pβ ϕ such that pβ ϕ(ϕ) β pϕ(ϕ). Originating from the principle of maximum likelihood, PML was proved [2, 4, 6, 7, 25, 58] to possess a number of useful attributes, such as existence over finite discrete domains, majorization by empirical distributions, consistency for distribution estimation under both sorted and unsorted ℓ1 distances, and competitiveness to other profile-based estimators. Let ε be an error parameter and f be one of the four properties in the introduction. Set n := nf(ε). Recently, Acharya et al. [4] showed that for some absolute constant c > 0, if c < c and ε n c, then a plug-in estimator for f, using an exp( n1 Θ(c))-approximate PML, is sample-optimal. Motivated by this result, Charikar et al. [20] constructed an explicit exp( O(n2/3 log3 n))-approximate PML (APML) whose computation time is near-linear in n. Combined, these results provide a unified, sample-optimal, and near-linear-time computable plug-in estimator for the four properties. 2 New results and implications 2.1 New results Additive property estimation Recall that for any property f, the expression nf(ε) denotes the smallest sample size required by any estimator to achieve accuracy ε with confidence 2/3, for all distributions in X . Let f be an additive symmetric property that is 1-Lipschitz on ( X , R). Let ε > 0 and n nf(ε) be error and sampling parameters. For an absolute constant c (10 2, 10 1), if ε n c, Theorem 1. The PML plug-in estimator, when given a sample of size 4n from any distribution p X , will estimate f(p) up to an error of (2 + o(1))ε, with probability at least 1 exp ( 4 n). For a different c > 0, Theorem 1 also holds for APML, which is near-linear-time computable [20]. Rényi entropy estimation For X of finite size k and any p X , it is well-known that Hα(p) [0, log k]. The following theorems characterize the performance of the PML plug-in estimator in estimating Rényi entropy. For any distribution p X , error parameter ε (0, 1), absolute constant λ (0, 0.1), and sampling parameter n, draw a sample Xn p and denote its profile by ϕ. Then for sufficiently large k, Theorem 2. For α (3/4, 1), if n = Ωα(k1/α/(ε1/α log k)), Pr (|Hα(pϕ) Hα(p)| ε) exp( n). Theorem 3. For non-integer α > 1, if n = Ωα(k/(ε1/α log k)), Pr (|Hα(pϕ) Hα(p)| ε) exp( n1 λ). Theorem 4. For integer α > 1, if n = Ωα(k1 1/α(ε2 log(1/ε)) (1+α)) and Hα(p) (log n)/4, Pr(|Hα(pϕ) Hα(p)| ε) 1/3. Replacing 3/4 by 5/6, Theorem 2 also holds for APML with a better probability bound exp( n2/3). In addition, Theorem 3 holds for APML without any modifications. Sorted distribution estimation Let c be the absolute constant defined just prior to Theorem 1. For any distribution p X , error parameter ε (0, 1), and sampling parameter n, draw a sample Xn p and denote its profile by ϕ. Theorem 5. If n = Ω(n(ε)) = Ω k/(ε2 log k) and ε n c, Pr(ℓ< 1(pϕ, p) ε) exp( Ω( n)). For a different c > 0, Theorem 5 also holds for APML with a better probability bound exp( n2/3). Identity testing The recent works of Diakonikolas and Kane [26] and Goldreich [31] provided a procedure reducing identity testing to uniformity testing, while modifying the desired accuracy and alphabet size by only absolute constant factors. Hence below we consider uniformity testing. The uniformity tester TPML shown in Figure 1 is purely based on PML and satisfies Theorem 6. If ε = Ω(k 1/4) and n = Ω( k/ε2), then the tester TPML(Xn) will be correct with probability at least 1 k 2. The tester also distinguishes between p = pu and p pu 2 ε/ The Ω( ) notation only hides logarithmic factors of k. The tester TPML is near-optimal as for uniform distribution pu, the results in [28] yield an Ω( k log k/ε2) lower bound on npu(ε, k 2). For space considerations, we postpone proofs and additional results to the supplementary material. The rest of the paper is organized as follows. Section 2.2 presents several immediate implications of the above theorems. Section 3 and Section 4 illustrate PML s theoretical and practical advantages by comparing it to existing methods for a variety of learning tasks. Section 5 concludes the paper and outlines multiple promising future directions. Input: parameters k, ε, and a sample Xn p with profile ϕ. if maxxµx(Xn) 3 max{1, n/k} log k then return 1; elif pϕ pu 2 3ε/(4 k) then return 1; else return 0. Figure 1: Uniformity tester TPML 2.2 Implications Several immediate implications are in order. We say that a plug-in estimator is universally sample-optimal for estimating symmetric properties if there exist absolute positive constants c1, c2 and c3, such that for any 1-Lipschitz property on ( X , R), with probability 9/10, the plug-in estimator uses just c1 times the sample size n required by the minimax estimator to achieve c2 times its error, whenever this error is at least n c3. Note that the 1-Lipschitz property class can be replaced by other general property classes, but not by those containing only a few specific properties, since universal means applicable to all cases . Theorem 1 makes PML the first plug-in estimator that is universally sample-optimal for a broad class of distribution properties. In particular, Theorem 1 also covers the four properties considered in [4]. To see this, as mentioned in the introduction, Cm, H, and D are 1-Lipschitz on ( X , R); as for S, the following result [4] relates it to Cm for distributions in 1/k, and proves PML s optimality. Lemma 1. For any ε > 0, m = k log(1/ε), and p 1/k, | S(p) Cm(p) log (1/ε) | ε. The theorem also applies to many other properties. As an example [70], given an integer s > 0, let fs(x) := min{x, |x 1/s|}. Then to within a factor of two, fs(p) := P x fs(px) approximates the ℓ1 distance between any distribution p and the closest uniform distribution in X of support size s. In Section 3.2 we compare Theorem 1 with existing results and present more of its implications. Theorem 2 and 3 imply that for all non-integer α > 3/4 (resp. α > 5/6), the PML (resp. APML) plug-in estimator achieves a sample complexity better than the best currently known [5]. This makes both the PML and APML plug-in estimators the state-of-the-art algorithms for estimating non-integer order Rényi entropy. See Section 3.3 for an introduction of known results, and see Section 3.4 for a detailed comparison between existing methods and ours. Theorem 4 shows that for all integer α > 1, the sample complexity of the PML plug-in estimator has optimal k1 1/α dependence [5, 55] on the alphabet size. Theorem 5 makes APML the first distribution estimator under sorted ℓ1 distance that is both nearlinear-time computable and sample-optimal for a range of desired accuracy ε beyond inverse polylogarithmic of n. In comparison, existing algorithms [2, 38, 72] either run in polynomial time in the sample sizes, or are only known to achieve optimal sample complexity for ε = Ω(1/ log n), which is essentially different from the applicable range of ε n Θ(1) in Theorem 5. We provide a more detailed comparison in Section 3.6. Theorem 6 provides the first PML-based uniformity tester with near-optimal sample complexity. As stated, the tester also distinguishes between p = pu and p pu 2 ε/ k. This is a stronger guarantee since by the Cauchy-Schwarz inequality, p pu 1 ε implies p pu 2 ε/ Note that several other uniformity testers in the literature (see Section 3.7) also provide the same ℓ2 testing guarantee, since all of them are essentially counting sample collisions, i.e., the number of location pairs such that the sample points at those locations are equal. 3 Related work and comparisons 3.1 Additive property estimation The study of additive property estimation dates back at least half a century [16, 34, 35] and has steadily grown over the years. For any additive symmetric property f and sequence xn, the simplest and most widely-used approach uses the empirical (plug-in) estimator ˆf E(xn) := f(pµ(xn)) that evaluates f at the empirical distribution. While the empirical estimator performs well in the large-sample regime, modern data science applications often concern high-dimensional data, for which more involved methods have yielded property estimators that are more sample-efficient. For example, for relatively large k and for f being S, Cm, H, or D, recent research [45, 59, 69, 70, 75, 76] showed that the empirical estimator is optimal up to logarithmic factors, namely nf(P, ε) = Θε(nf( ˆf E, P, ε)/log nf( ˆf E, P, ε)), where P is 1/k for S, and is X for the other properties. Below we classify the methods for deriving the corresponding sample-optimal estimators into two categories: plug-in and approximation, and provide a high-level description. For simplicity of illustration, we assume that ε (0, 1]. The plug-in approach essentially estimates the unknown distribution multiset, which suffices for computing any symmetric properties. Besides the empirical and PML estimators, Efron and Thisted [29] proposed a linear-programming approach that finds a multiset estimate consistent with the sample s profile. This approach was then adapted and analyzed by Valiant and Valiant [69, 72], yielding plug-in estimators that achieve near-optimal sample complexities for H and S, and optimal sample complexity for D, when ε is relatively large. The approximation approach modifies non-smooth segments of the probability function to correct the bias of empirical estimators. A popular modification is to replace those non-smooth segments by their low-degree polynomial approximations and then estimate the modified function. For several properties including the above four and power sum Pα(p) := P x pα x, where α is a given parameter, this approach yields property-dependent estimators [45, 59, 75, 76] that are sample-optimal for all ε. More recently, Acharya et al. [4] proved the aforementioned results on PML estimator and made it the first unified, sample-optimal plug-in estimator for S, Cm, H and D and relatively large ε. Following these advances, Han et al. [38] refined the linear-programming approach and designed a plug-in estimator that implicitly performs polynomial approximation and is sample-optimal for H, S, and Pα with α < 1, when ε is relatively large. 3.2 Comparison I: Theorem 1 and related property-estimation work In terms of the estimator s theoretical guarantee, Theorem 1 is essentially the same as Valiant and Valiant [70]. However, for each property, k, and n, [70] solves a different linear program and constructs a new estimator, which takes polynomial time. On the other hand, both the PML estimator and its near-linear-time computable variant, once computed, can be used to accurately estimate exponentially many properties that are 1-Lipschitz on ( X , R). A similar comparison holds between the PML method and the approximation approach, while the latter is provably sample-optimal for only a few properties. In addition, Theorem 1 shows that the PML estimator often achieves the optimal sample complexity up to a small constant factor, which is a desired estimator attribute shared by some, but not all approximation-based estimators [45, 59, 75, 76]. In term of the method and proof technique, Theorem 1 is closest to Acharya et al. [4]. On the other hand, [4] establishes the optimality of PML for only four properties, while our result covers a much broader property class. In addition, both the above mentioned small constant factor attribute, and the confidence boost from 2/3 to 1 exp( 4 n) are unique contributions of this work. The PML plug-in approach is also close in flavor to the plug-in estimators in Valiant and Valiant [69, 72] and their refinement in Han et al. [38]. On the other hand, as pointed out previously, these plug-in estimators are provably sample-optimal for only a few properties. More specifically, for estimating H, S, and Cm, the plug-in estimators in [69, 72] achieve sub-optimal sample complexities with regard to the desired accuracy ε; and the estimation guarantee in [38] is provided in terms of the approximation errors of O( n) polynomials that are not directly related to the optimal sample complexities. 3.3 Rényi entropy estimation Motivated by the wide applications of Rényi entropy, heuristic estimators were proposed and studied in the physics literature following [36], and asymptotically consistent estimators were presented and analyzed in the statistical learning literature [46, 78]. For the special case of 1-Rényi (or Shannon) entropy, the works of [69, 70] determined the sample complexity to be nf(ε) = Θ(k/(ε log k)). For general α-Rényi entropy, the best-known results in Acharya et al. [5] state that for integer and noninteger α values, the corresponding sample complexities nf(ε, δ) are Oα(k1 1/α log(1/δ)/ε2) and Oα(kmin{1/α,1} log(1/δ)/(ε1/α log k)), respectively. The upper bounds for integer α are achieved by an estimator that corrects the bias of the empirical plug-in estimator. To achieve the upper bounds for non-integer α values, one needs to compute some best polynomial approximation of zα, whose degree and domain both depend on n, and construct a more involved estimator using the approximation approach [45, 75] mentioned in Section 3.1. 3.4 Comparison II: Theorem 2 to 4 and related Rényi-entropy-estimation work Our result shows that a single PML estimate suffices to estimate the Rényi entropy of different orders α. Such adaptiveness to the order parameter is a significant advantage of PML over existing methods. For example, by Theorem 3 and the union bound, one can use a single APML or PML to accurately approximate exponentially many non-integer order Rényi entropy values, yet still maintains an overall confidence of 1 exp( k0.9). By comparison, the estimation heuristic in [5] requires different polynomial-based estimators for different α values. In particular, to construct each estimator, one needs to compute some best polynomial approximation of zα, which is not known to admit a closed-form formula for α Z. Furthermore, even for a single α and with a sample size k times larger, such estimator is not known to achieve the same level of confidence as PML or APML. As for the theoretical guarantees, the sample-complexity upper bounds in both Theorem 2 and 3 are better than those mentioned in the previous section. More specifically, for any α (3/4, 1) and δ exp( k 0.5), Theorem 2 shows that nf(ε, δ) = Oα(k1/α/(ε1/α log k)). Analogously, for any non-integer α > 1 and δ exp( k 0.9), Theorem 3 shows that nf(ε, δ) = Oα(k/(ε1/α log k)). Both bounds are better than the best currently known by a log(1/δ) factor. 3.5 (Sorted) distribution estimation Estimating large-alphabet distributions from their samples is a fundamental statistical learning tenet. Over the past few decades, distribution estimation has found numerous applications, ranging from natural language modeling [21] to biological research [8], and has been studied extensively. Under the classical ℓ1 and KL losses, existing research [13, 47] showed that the corresponding sample complexities n(ε) are Θ(k/ε2) and Θ(k/ε), respectively. Several recent works have investigated the analogous formulation under sorted ℓ1 distance, and revealed a lower sample complexity of n(ε) = Θ(k/(ε2 log k)). Specifically, under certain conditions, Valiant and Valiant [72], Han et al. [38] derived sample-optimal estimators using linear programming, and Acharya et al. [2], Das. [25] showed that PML achieves a sub-optimal O(k/(ε2.1 log k)) sample complexity for relatively large ε. 3.6 Comparison III: Theorem 5 and related distribution-estimation work We compare our results with existing ones from three different perspectives. Applicable parameter ranges: As shown by [38], for ε n 1/3, the simple empirical estimator is already sample-optimal. Hence we consider the parammeter range ε = Ω(n 1/3). For the results in [2, 25] and [72] to hold, we would need ε to be at least Ω(1/ log n). On the other hand, Theorem 5 shows that PML and APML are sample-optimal for ε larger than n Θ(1). Here, the gap is exponentially large. The result in [38] applies to the whole range ε = Ω(n 1/3), which is larger than the applicable range of our results. Time complexity: Both the APML and the estimator in [72] are near-linear-time computable in the sample sizes, while the estimator in [38] would require polynomial time to be computed. Statistical confidence: The PML and APML achieve the desired accuracy with an error probability at most exp( Ω( n)). On the contrary, the estimator in [38] is known to achieve an error probability that decreases only as O(n 3). The gap is again exponentially large. The estimator in [72] admits a better error probability bound of exp( n0.02), which is still far from ours. 3.7 Identity testing Initiated by the work of [33], identity testing is arguably one of the most important and widely-studied problem in distribution property testing. Over the past two decades, a sequence of works [3, 11, 26 28, 33, 61, 68] have addressed the sample complexity of this problem and proposed testers with a variety of guarantees. In particular, applying a coincidence-based tester, Paninski [61] determined the sample complexity of uniformity testing up to constant factors; utilizing a variant of the Pearson chi-squared statistic, Valiant and Valiant [68] resolved the general identity testing problem. For an overview of related results, we refer interested readers to [15] and [32]. The contribution of this work is mainly showing that PML, is a unified sample-optimal approach for several related problems, and as shown in Theorem 6, also provides a near-optimal tester for this important testing problem. 4 Experiments and distribution estimation under ℓ1 distance A number of different approaches have been taken to computing the PML and its approximations. Among the existing works, Acharya et al. [1] considered exact algebraic computation, Orlitsky et al. [57, 58] designed an EM algorithm with MCMC acceleration, Vontobel [73, 74] proposed a Bethe approximation heuristic, Anevski et al. [7] introduced a sieved PML estimator and a stochastic approximation of the associated EM algorithm, and Pavlichin et al. [62] derived a dynamic programming approach. Notably and recently, for a sample size n, Charikar et al. [20] constructed an explicit exp( O(n2/3 log3 n))-approximate PML whose computation time is near-linear in n. In Section 4 of the supplementary material, we introduce a variant of the MCMC-EM algorithm in [60] and demonstrate the exceptional efficacy of PML on a variety of learning tasks through experiments. In particular, we derive a new distribution estimator for (unsorted) ℓ1 distance by combining the proposed PML computation algorithm with the denoising procedure in [71] and a novel missing mass estimator. As shown below, the proposed distribution estimator has the state-of-the-art performance. Figure 2: Distribution estimation under ℓ1 distance In Figures 2, samples are generated according to six distributions of the same support size k = 5,000. Details about these distributions can be found in Section 4.2 of the supplementary material. The sample size n (horizontal axis) ranges from 10,000 to 100,000, and the vertical axis reflects the (unsorted) ℓ1 distance between the true distribution and the estimates, averaged over 30 independent trials. We compare our estimator with three different ones: the improved Good-Turing estimator in [56, 41], which is provably instance-by-instance near-optimal [56], the empirical estimator, serving as a baseline, and the empirical estimator with a larger n log n sample size. Note that log n is roughly 11. As shown in [56], the improved Good-Turing estimator substantially outperformed other estimators such as the Laplace (add-1) estimator, the Braess-Sauer estimator [13], and the Krichevsky-Trofimov estimator [48]. Hence we do not include those estimators here. The following plots showed that our proposed estimator further outperformed the improved Good-Turing estimator in all the experiments. 5 Conclusion and future directions We studied three fundamental problems in statistical learning: distribution estimation, property estimation, and property testing. We established the profile maximum likelihood (PML) as the first universally sample-optimal approach for several important learning tasks: distribution estimation under the sorted ℓ1 distance, additive property estimation, Rényi entropy estimation, and identity testing. Several future directions are promising. We believe that neither the factor of 4 in the sample size in Theorem 1, nor the lower bounds on ε in Theorem 1, 5, and 6 are necessary. In other words, the PML approach is universally sample-optimal for these tasks in all ranges of parameters. It is also of interest to extend the PML s optimality to estimating symmetric properties not covered by Theorem 1 to 4, such as generalized distance to uniformity [9, 39], the ℓ1 distance between the unknown distribution and the closest uniform distribution over an arbitrary subset of X. Another important direction is competitive (or instance-optimal) property estimation. It should be noted that all the referenced works including this paper are of the worst-case nature, namely, designing estimators with near-optimal worst-case performances. On the contrary, practical and natural distributions often possess simple structures, and are rarely the worst possible. To address this discrepancy, the recent work [40, 43] took a competitive approach and constructed estimators whose performances are adaptive to the simplicity of the underlying distributions. Specifically, for any property in a broad class and every distribution in X , the expected error of the proposed estimator with a sample of size n/ log n is at most that of the empirical estimator with a sample of size n, pluses a distribution-free vanishing function of n. These results not only cover S, Cm, H, and D, for which the log n-factor is optimal up to constants, but also apply to any non-symmetric additive property P x fx(px) where fx is 1-Lipschitz for all x X, such as the ℓ1-distance to a given distribution. It would be of interest to study the optimality of the PML approach under this formulation as well. Readers interested in estimating non-symmetric properties may also find the paper [42] helpful. Acknowledgments We are grateful to the National Science Foundation (NSF) for supporting this work through grants CIF-1564355 and CIF-1619448. [1] J. Acharya, H. Das, H. Mohimani, A. Orlitsky, and S. Pan. Exact calculation of pattern probabilities. In Proceedings of 2010 IEEE International Symposium on Information Theory (ISIT), pages 1498 1502. IEEE, 2010. [2] J. Acharya, H. Das, A. Jafarpour, A. Orlitsky, and S. Pan. Estimating multiple concurrent processes. In Proceedings 2012 IEEE International Symposium on Information Theory, pages 1628 1632. IEEE, 2012. [3] J. Acharya, C. Daskalakis, and G. Kamath. Optimal testing for properties of distributions. In Advances in Neural Information Processing Systems, pages 3591 3599. Society for Industrial and Applied Mathematics, 2015. [4] J. Acharya, H. Das, A. Orlitsky, and A. T. Suresh. A unified maximum likelihood approach for estimating symmetric properties of discrete distributions. In International Conference on Machine Learning, pages 11 21, 2017. [5] J. Acharya, A. Orlitsky, A. T. Suresh, and H. Tyagi. Estimating Rényi entropy of discrete distributions. IEEE Transactions on Information Theory, 63(1):38 56, 2017. [6] J. Acharya, Y. Bao, Y. Kang, and Z. Sun. Improved bounds for minimax risk of estimating missing mass. In 2018 IEEE International Symposium on Information Theory (ISIT), pages 326 330. IEEE, 2018. [7] D. Anevski, R. D. Gill, and S. Zohren. Estimating a probability mass function with unknown labels. The Annals of Statistics, 45(6):2708 2735, 2017. [8] R. Armaanzas, I. Inza, R. Santana, Y. Saeys, J. L. Flores, J. A. Lozano, . . . , and P. Larrañaga. A review of estimation of distribution algorithms in bioinformatics. Bio Data Mining, 1(6), 2008. [9] T. Batu and C. L. Canonne. Generalized uniformity testing. In 2017 IEEE 58th Annual Symposium on Foundations of Computer Science, pages 880 889. IEEE, 2017. [10] T. Batu, L. Fortnow, R. Rubinfeld, W. D. Smith, and P. White. Testing that distributions are close. In Proceedings 41st Annual Symposium on Foundations of Computer Science, pages 259 269. IEEE, 2000. [11] T. Batu, E. Fischer, Kumar Fortnow, L., R. R., Rubinfeld, and P. White. Testing random variables for independence and identity. In Proceedings 2001 IEEE International Conference on Cluster Computing, pages 442 451. IEEE, 2001. [12] T. Batu, E. Fischer, L. Fortnow, R. Kumar, R. Rubinfeld, and P. White. Testing random variables for independence and identity. In Proceedings of 42nd IEEE Symposium on Foundations of Computer Science, pages 442 451. IEEE, 2001. [13] D. Braess and T. Sauer. Bernstein polynomials and learning theory. Journal of Approximation Theory, 128(2):187 206, 2004. [14] G. Bresler. Efficiently learning ising models on arbitrary graphs. In Proceedings of the 47th Annual ACM Symposium on Theory of Computing, pages 771 782. ACM, 2015. [15] C. L. Canonne. A survey on distribution testing. Your Data is Big. But is it Blue., 2017. [16] A. Carlton. On the bias of information estimates. Psychological Bulletin, 71(2):108, 1969. [17] A. Chao. Nonparametric estimation of the number of classes in a population. Scandinavian Journal of Statistics, pages 265 270, 1984. [18] A. Chao and C. H. Chiu. Species richness: estimation and comparison. Wiley Stats Ref: Statistics Reference Online, pages 1 26, 2014. [19] A. Chao and S. M. Lee. Estimating the number of classes via sample coverage. Journal of the American Statistical Association, 87(417):210 217, 1992. [20] M. S. Charikar, K. Shiragur, and A. Sidford. Efficient profile maximum likelihood for universal symmetric property estimation. In Proceedings of the 51st Annual ACM SIGACT Symposium on Theory of Computing, pages 780 791. ACM, 2019. [21] S. F. Chen and J. Goodman. An empirical study of smoothing techniques for language modeling. Computer Speech & Language, 13(4):359 394, 1999. [22] C. Chow and C. Liu. Approximating discrete probability distributions with dependence trees. IEEE Transactions on Information Theory, 14(3):462 467, 1968. [23] R. K. Colwell, A. Chao, N. J. Gotelli, S. Y. Lin, C. X. Mao, R. L. Chazdon, and J. T. Longino. Models and estimators linking individual-based and sample-based rarefaction, extrapolation and comparison of assemblages. Journal of Plant Ecology, 5(1):3 21, 2012. [24] T. M. Cover and J. A. Thomas. Elements of information theory. John Wiley & Sons, 2012. [25] H. Das. Competitive tests and estimators for properties of distributions. (Doctoral dissertation, UC San Diego), 2012. [26] I. Diakonikolas and D. M. Kane. A new approach for testing properties of discrete distributions. In 2016 IEEE 57th Annual Symposium on Foundations of Computer Science (FOCS), pages 685 694. IEEE, 2016. [27] I. Diakonikolas, D. M. Kane, and V. Nikishkin. Testing identity of structured distributions. In Proceedings of the Twenty-Sixth Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), pages 1841 1854. Society for Industrial and Applied Mathematics, 2015. [28] I. Diakonikolas, T. Gouleakis, J. Peebles, and E. Price. Sample-optimal identity testing with high probability. In 45th International Colloquium on Automata, Languages, and Programming (ICALP 2018). Schloss Dagstuhl-Leibniz-Zentrum fuer Informatik, 2018. [29] B. Efron and R. Thisted. Estimating the number of unseen species: How many words did Shakespeare know? Biometrika, 63(3):435 447, 1976. [30] W. Gerstner and W. M. Kistler. Spiking neuron models: Single neurons, populations, plasticity. Cambridge University Press, 2002. [31] O. Goldreich. The uniform distribution is complete with respect to testing identity to a fixed distribution. In Electronic Colloquium on Computational Complexity (ECCC), volume 23, page 1. IEEE, 2016. [32] O. Goldreich. Introduction to property testing (chapter 11). Cambridge University Press, 2017. [33] O. Goldreich and D. Ron. On testing expansion in bounded-degree graphs. In Technical Report TR00-020, Electronic Colloquium on Computational Complexity (ECCC), 2000. [34] I. J. Good. The population frequencies of species and the estimation of population parameters. Biometrika, 40(3-4):237 264, 1953. [35] I. J. Good and G. H. Toulmin. The number of new species, and the increase in population coverage, when a sample is increased. Biometrika, 43(1-2):45 63, 1956. [36] P. Grassberger. Finite sample corrections to entropy and dimension estimates. Physics Letters A, 128(6-7):369 373, 1988. [37] P. J. Haas, J. F. Naughton, S. Seshadri, and L. Stokes. Sampling-based estimation of the number of distinct values of an attribute. VLDB, 95:311 322, 1995. [38] Y. Han, J. Jiao, and T. Weissman. Local moment matching: A unified methodology for symmetric functional estimation and distribution estimation under Wasserstein distance. In Conference on Learning Theory, pages 3189 3221, 2018. [39] Y. Hao and A. Orlitsky. Adaptive estimation of generalized distance to uniformity. In 2018 IEEE International Symposium on Information Theory, pages 1076 1080. IEEE, 2018. [40] Y. Hao and A. Orlitsky. Data amplification: Instance-optimal property estimation. In ar Xiv preprint ar Xiv:1903.01432., 2019. [41] Y. Hao and A. Orlitsky. Doubly-competitive distribution estimation. In International Conference on Machine Learning, pages 2614 2623, 2019. [42] Y. Hao and A. Orlitsky. Unified sample-optimal property estimation in near-linear time. In Advances in Neural Information Processing Systems, 2019. [43] Y. Hao, A. Orlitsky, A. T. Suresh, and Y. Wu. Data amplification: A unified and competitive approach to property estimation. In Advances in Neural Information Processing Systems, pages 8848 8857, 2018. [44] R. Jenssen, K. E. Hild, D. Erdogmus, J. C. Principe, and T. Eltoft. Clustering using Rényi s entropy. In Proceedings of the International Joint Conference on Neural Networks, volume 1, pages 523 528. IEEE, 2003. [45] J. Jiao, K. Venkat, Y. Han, and T. Weissman. Minimax estimation of functionals of discrete distributions. IEEE Transactions on Information Theory, 61(5):2835 2885, 2015. [46] D. Källberg, N. Leonenko, and O. Seleznjev. Statistical inference for Rényi entropy functionals. In Conceptual Modelling and Its Theoretical Foundations. Springer, Berlin, Heidelberg., pages 36 51, 2012. [47] S. Kamath, A. Orlitsky, D. Pichapati, and A. T. Suresh. On learning distributions from their samples. In Conference on Learning Theory, pages 1066 1100, 2015. [48] R. Krichevsky and V. Trofimov. The performance of universal encoding. IEEE Transactions on Information Theory, 27(2):199 207, 1981. [49] I. Kroes, P. W. Lepp, and D. A. Relman. Bacterial diversity within the human subgingival crevice. Proceedings of the National Academy of Sciences, 96(25):14547 14552, 1999. [50] B. Ma, A. Hero, J. Gorman, and O. Michel. Image registration with minimum spanning tree algorithm. In Proceedings 2000 International Conference on Image Processing (Cat. No. 00CH37101), volume 1, pages 481 484. IEEE, 2000. [51] Z. F. Mainen and T. J. Sejnowski. Reliability of spike timing in neocortical neurons. Science, 268(5216):1503 1506, 1995. [52] C. X. Mao and B. G. Lindsay. Estimating the number of classes. The Annals of Statistics, pages 917 930, 2007. [53] D. R. Mc Neil. Estimating an author s vocabulary. Journal of the American Statistical Association, 68(341):92 96, 1973. [54] H. Neemuchwala, A. Hero, S. Zabuawala, and P. Carson. Image registration methods in highdimensional space. International Journal of Imaging Systems and Technology, 16(5):130 145, 2006. [55] M. Obremski and M. Skorski. Rényi entropy estimation revisited. In Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques (APPROX/RANDOM 2017). Schloss Dagstuhl-Leibniz-Zentrum fuer Informatik, 2017. [56] A. Orlitsky and A. T. Suresh. Competitive distribution estimation: Why is Good-Turing good. In Advances in Neural Information Processing Systems, pages 2143 2151, 2015. [57] A. Orlitsky, S. Sajama, N. P. Santhanam, K. Viswanathan, and J. Zhang. Algorithms for modeling distributions over large alphabets. In Proceedings of 2004 IEEE International Symposium on Information Theory (ISIT), pages 304 304. IEEE, 2004. [58] A. Orlitsky, N. P. Santhanam, K. Viswanathan, and J. Zhang. On modeling profiles instead of values. In Proceedings of the 20th Conference on Uncertainty in Artificial Intelligence, pages 426 435. AUAI Press, 2004. [59] A. Orlitsky, A. T. Suresh, and Y. Wu. Optimal prediction of the number of unseen species. Proceedings of the National Academy of Sciences, 113(47):13283 13288, 2016. [60] S. Pan. On the theory and application of pattern maximum likelihood. (Doctoral dissertation, UC San Diego), 2012. [61] L. Paninski. A coincidence-based test for uniformity given very sparsely sampled discrete data. IEEE Transactions on Information Theory, 54(10):4750 4755, 2008. [62] D. S. Pavlichin, J. Jiao, and T. Weissman. Approximate profile maximum likelihood. ar Xiv preprint, ar Xiv:1712.07177, 2017. [63] C. J. Quinn, N. Kiyavash, and T. P. Coleman. Efficient methods to compute optimal tree approximations of directed information graphs. IEEE Transactions on Signal Processing, 61 (12):3173 3182, 2013. [64] A. Rényi. On measures of entropy and information. In Proceedings of the Fourth Berkeley Symposium on Mathematical Statistics and Probability, Volume 1: Contributions to the Theory of Statistics. The Regents of the University of California, 1961. [65] D. Ron. Algorithmic and analysis techniques in property testing. Number 5 in 2. Foundations and Trends in Theoretical Computer Science, 2010. [66] C. E. Shannon. A mathematical theory of communication. Bell System Technical Journal, 27 (3):379 423, 1948. [67] R. Thisted and B. Efron. Did Shakespeare write a newly-discovered poem? Biometrika, 74(3): 445 455, 1987. [68] G. Valiant and P. . Valiant. An automatic inequality prover and instance optimal identity testing. SIAM Journal on Computing, 46(1):429 455, 2017. [69] G. Valiant and P. Valiant. Estimating the unseen: An n/log(n)-sample estimator for entropy and support size, shown optimal via new CLTs. In Proceedings of the 43rd Annual ACM Symposium on Theory of Computing, pages 685 694. ACM, 2011. [70] G. Valiant and P. Valiant. The power of linear estimators. In IEEE 52nd Annual Symposium on Foundations of Computer Science, pages 403 412. IEEE, 2011. [71] G. Valiant and P. Valiant. Instance optimal learning of discrete distributions. In Proceedings of the 48th Annual ACM Symposium on Theory of Computing, pages 142 155. ACM, 2016. [72] G. Valiant and P. Valiant. Estimating the unseen: Improved estimators for entropy and other properties. Journal of the ACM (JACM), 64(6):37, 2017. [73] P. O. Vontobel. The Bethe approximation of the pattern maximum likelihood distribution. In Proceedings of 2012 IEEE International Symposium on Information Theory (ISIT). IEEE, 2012. [74] P. O. Vontobel. The Bethe and Sinkhorn approximations of the pattern maximum likelihood estimate and their connections to the Valiant-Valiant estimate. In 2014 Information Theory and Applications Workshop (ITA), pages 1 10. IEEE, 2014. [75] Y. Wu and P. Yang. Minimax rates of entropy estimation on large alphabets via best polynomial approximation. IEEE Transactions on Information Theory, 62(6):3702 3720, 2016. [76] Y. Wu and P. Yang. Chebyshev polynomials, moment matching, and optimal estimation of the unseen. The Annals of Statistics, 47(2):857 883, 2019. [77] D. Xu. Energy, entropy and information potential for neural computation. Doctoral dissertation, University of Florida, 1999. [78] D. Xu and D. Erdogmuns. Rényi s entropy, divergence and their nonparametric estimators. In Information Theoretic Learning. Springer, New York, NY., pages 47 102, 2010.