# samecluster_querying_for_overlapping_clusters__54d66dac.pdf Same-Cluster Querying for Overlapping Clusters Wasim Huleihel Department of Electrical Engineering Tel-Aviv University Tel-Aviv, Israel 6997801 wasimh@mail.tau.ac.il Arya Mazumdar College of Information & Computer Sciences University of Massachusetts Amherst Amherst, MA 01003 arya@cs.umass.edu Muriel Médard Electrical Engineering & Computer Science Massachusetts Institute of Technology Cambridge, MA 02139 medard@mit.edu Soumyabrata Pal College of Information & Computer Sciences University of Massachusetts Amherst Amherst, MA 01003 soumyabratap@umass.edu Overlapping clusters are common in models of many practical data-segmentation applications. Suppose we are given n elements to be clustered into k possibly overlapping clusters, and an oracle that can interactively answer queries of the form do elements u and v belong to the same cluster? The goal is to recover the clusters with minimum number of such queries. This problem has been of recent interest for the case of disjoint clusters. In this paper, we look at the more practical scenario of overlapping clusters, and provide upper bounds (with algorithms) on the sufficient number of queries. We provide algorithmic results under both arbitrary (worst-case) and statistical modeling assumptions. Our algorithms are parameter free, efficient, and work in the presence of random noise. We also derive information-theoretic lower bounds on the number of queries needed, proving that our algorithms are order optimal. Finally, we test our algorithms over both synthetic and real-world data, showing their practicality and effectiveness. 1 Introduction Recently, semi-supervised models of clustering that allow active querying during data segmentation have become quite popular. This includes active learning, as well as data labeling by amateurs via crowdsourcing. Clever implementation of interactive querying framework can improve the accuracy of clustering and help in inferring labels of large amount of data by issuing only a small number of queries. Interactions can easily be implemented (e.g., via captcha), especially if queries involve few data points, like pairwise queries of whether two points belong to the same cluster or not [1, 2, 4, 9, 13, 15, 18, 19, 21, 22, 23, 25]. Until now, the querying model and algorithms/lower bounds are highly tailored towards flat clustering that produces a partition of the data. Consider the problem of clustering from pairwise queries such as above when an element can be part of multiple clusters. Such overlapping clustering instances are ubiquitous across areas and many time are more practical model of data segmentation, see [3, 5, 20, 26]. Indeed, overlapping models are quite natural for communities in social networks or topic models [16]. In the supervised version of the problem every element (or data-point) can have multiple labels, and we would like to know all the labels. To see how the querying might work here consider the following input: {Tiger Shark, Grizzly Bear, Blue Whale, Bush Dog, Giant Octopus, Ostrich, Komodo Dragon}. This set 33rd Conference on Neural Information Processing Systems (Neur IPS 2019), Vancouver, Canada. can be clustered into the mammals {Grizzly Bear, Blue Whale, Bush Dog}, marinelife {Tiger Shark, Blue Whale, Giant Octopus}, non-mammals {Tiger Shark, Giant Octopus, Ostrich, Komodo Dragon}, land-dwellers {Grizzly Bear, Bush Dog, Ostrich, Komodo Dragon}. Quite clearly, this ideal clustering (without labels) is overlapping. If a query of whether two elements belong to the same cluster or not is made then the answer should be yes if there exists a cluster where they can appear together. If we form a response matrix of size 7 7 with rows and columns indexed by the order they appeared above in the list and entries being the binary answers to queries, then the matrix would be following: TS GB BW BD GO Os KD TS 0 1 0 1 1 1 GB 0 1 1 0 1 1 BW 1 1 1 1 0 0 BD 0 1 1 0 1 1 GO 1 0 1 0 1 1 Os 1 1 0 1 1 1 KD 1 1 0 1 1 1 What is the minimum number of adaptive queries that we should make to the above matrix so that it is possible to recover the clusters? In the case when the clusters are not overlapping, the answer is nk, where n is the number of elements and k is the number of possible clusters [19]. For the case of overlapping clusters it is not clear whether there is a unique clustering that explains the responses. For this, certain extra constraints must be placed: for example, a reasonable assumption is that an element can only be part of clusters, among the total k possible clusters, k. Just like the response matrix above, it is possible to form a similarity matrix. The (i, j)th entry of this matrix simply is the number of clusters where the ith and the jth elements coexists. It is clear that the response matrix is just a quantized version of the similarity matrix. Even when the entire similarity matrix is given, there is no guarantee on uniqueness of overlapping clustering, unless further assumptions are made. In this paper, we primarily aim to recover the clustering from a limited number of adaptive queries to the response matrix. However, in terms of uniqueness guarantees, we often have to stop at the uniqueness of the similarity matrix. Main results and Techniques. Recovery of overlapping clusters from budgeted same-cluster queries is widely open, and significantly more challenging than the flat clustering counterpart. In fact, none of the techniques proposed in prior literature as mentioned above extends easily to the case when the clusters may overlap. In this paper we tackle this problem for various types of responses. Specifically, in our setting there is an oracle having access to the similarity matrix AAT , where A is the n k clustering matrix whose ith row is the indicator vector of the cluster membership of ith element. In its most powerful mode, when queried the oracle provides the number of clusters where ith and the jth elements coexists, namely the values of the entries of the matrix AAT . It turns out, however, that even if one knows the matrix AAT perfectly, it is not always possible to recover A up to a permutation of columns.1 In fact, if no assumption is imposed on the clustering matrix A, then recovering A from AAT is intractable in general. Indeed, even just finding conditions on A such that the factorization is unique (up to permutations) is related to the famous unsolved question of for which orders a finite projective plane exists?" [11]. It is then clear that we need to impose some assumptions. We tackle this inherent problem in two different approaches. First, in Sections 3.1 and 3.2, we propose two generative models for A: 1) a uniform ensemble where a given element can only be part of 1 clusters,2 among the total of k clusters, and its membership is drawn uniformly at random among all possible k possible placements, 2) the matrix A is generated i.i.d. with Bernoulli entries. Then, for these two ensembles, we investigate the above fundamental question, and derive sufficient conditions under which this factorization is unique, along with quasi-polynomial worst-case complexity algorithms for recovering A from AAT . The main immediate implication of this result is that, under certain conditions, the clustering recovery problem reduces to recovering the similarity matrix, placing this objective to be our main task. While the above generative models allow us to obtain elegant and neat theoretical results, one might argue that they may not capture many challenges existing in real-world overlapping clustering problems. To this end, in Section 3.3, we go beyond the above generative models and analyze a 1Given AAT , it is only possible to recover the clustering matrix A up to a permutation of columns, see Section 2. 2The case where different items may belong to different numbers of clusters can be handled using the same techniques developed in this paper. general worst-case model with no statistical assumptions. Then, under certain realistic assumptions on the clustering matrix, we provide and analyze algorithms solving the recovery problem. In practice, however, the aforementioned value oracle responses might be quite expensive. Accordingly, we study also quantized and noisy variants of these responses. For example, instead of getting direct values from AAT , the oracle only supplies the learner with (possibly noisy) binary answers on whether arbitrarily picked pair of elements (i, j) appear together in some cluster or not ( same-cluster query ). We consider also the case of dithered oracle, where noise is injected before quantization. For these scenarios (and others), we provide both lower and upper bounds on the number of queries needed for exact recovery. Our lower bounds are obtained using standard information-theoretic results, such as Fano s inequality. For the upper bounds, we design novel randomized algorithms for recovering the similarity matrix, and further show that these algorithms can work when the noise parameter is not given in advance. For example, when k = O(log n) and k, we show that the sufficient and necessary number of quantized queries is Θ(n log n), for the uniform ensemble. Finally, we test our algorithms over both synthetic and real-world data, showing the practicality and effectiveness of our algorithms. Related Work. As mentioned above, there is a series of applied and theoretical works studying the query complexity of same-cluster queries for objective-based clustering (such as, k-means) and clustering with statistical generative models. In all the cases though, the clusters are assumed to be non-overlapping. From a practical standpoint, entity resolution via crowdsourced pairwise sameentity queries were studied in [13, 17, 24, 25, 22, 27]. The effect of (possibly noisy) same-cluster queries in similarity matrix based clustering has been studied in [18, 19, 1]. On the other hand, the effect of same-cluster queries in the efficiency of k-means clustering was initiated in [4] - which was then subsequently further studied in [27, 2, 9]. In our approach, we crucially use the low-rank structure of the similarity matrix to recover the clustering from a bounded number of responses. Low-rank matrix completion is a well-studied topic in statistics and data science [7, 14]. It is possible to obtain weaker version of some of our results by relying on the results of low-rank matrix completion as black-boxes. However, the specific structure of the similarity matrix under consideration allows us to obtain stronger results. The response matrix is a quantized version of a low-rank matrix. Querying entries of the response matrix can be seen as a so called 1-bit matrix completion problem [12]. However, most of the recovery guarantees of 1-bit matrix completion depends crucially on certain dither noise (see, [12]), which may not be what is allowed in our setting. Finally, we mention [6] where the problem of overlapping clustering was considered from a different point of view. Organization. The remaining part of this paper is organized as follows. The model and the learning problem are provided in Section 2. Our main results on the query complexity are presented in Section 3. In particular, we provide upper bounds (with algorithms) on the sufficient number of queries for each of the scenarios investigated in this paper. These results are also accompanied with information-theoretic lower bounds on the necessary number of queries, which are presented in the appendix due to page limitation. Finally, Section 4 is devoted for a numerical study, where our main results are illustrated empirically. Detailed proofs of all the theoretical results can be found in the supplementary material. 2 Model and Learning Problem Consider a set of elements N [n] with k latent clusters Ni, for i = 1, 2, . . . , k, such that each element in N belongs to at least one cluster. This data is represented by an n k matrix A, where Ai,j = 1 if the ith elements is in the jth cluster. We will denote the k-dimensional binary vector representing the cluster membership of the ith element by Ai (i.e., the ith row of A), and will henceforth refer to it as the ith membership vector. In our setting there is an oracle O : N N D that when queried with a pair of elements (i, j) N N , returns a natural number L D N according to some pre-defined rule. We shall refer to O as the oracle map. The queries Ω N N can be done adaptively. Our goal is to find the set Ω N N such that |Ω| is minimum, and it is possible to recover {Ni}k i=1 from the oracle answers. More specifically, the oracle have access to the similarity matrix AAT , and when queried with Ω, answers according to O. Given (i, j) N N, we consider the following oracle maps O, capturing several aspects of the problem: Direct responses: The oracle response is Odirect(i, j) = AT i Aj, namely, the number of clusters that elements i and j belong to simultaneously. Note that when the clusters are disjoint, the output is simply an answer to the question do elements i and j belong to the same cluster?". Quantized (noisy) responses: The oracle response is Oquantized(i, j) = Q AT i Aj Wi,j, where Q(x) 1 for x > 0, and 0, otherwise, and Wij Bernoulli(q), with 0 q 1, independent over pairs (i, j). In the noiseless case, i.e., q = 0, the oracle response is whether elements i and j appears together in at least one cluster or not. In the noisy case, the oracle response is the quantized response with probability 1 q, and flipped with probability q. This can be interpreted as if the quantized responses are further sent through a binary symmetric channel BSC(q). Dithered responses: The oracle response is Odithered(i, j) = Q AT i Aj + Zi,j , where Zij Normal(0, σ2), independent over pairs (i, j). In other words, the oracle outputs a dithered and quantized version of the direct responses. For simplicity of notation, throughout the rest of this paper we will denote the oracle response to query (i, j) N by Yij, irrespective of the oracle model, which will be clear from the context. If we permute the columns of A, the gram matrix AAT will remain the same. Therefore, finding A is only possible up to a permutation of columns. Unfortunately, it turns out that even if we know AAT perfectly it is not always possible to find A up to a permutation of columns, namely, the factorization may not be unique. As an example, consider the following matrices 1 1 0 0 0 1 1 0 0 0 1 1 1 0 1 0 0 0 1 1 0 1 0 1 1 1 0 0 1 0 0 1 which have the same gram matrix but evidently are not column permutations of each other. Hence, even if we observe all the entries of the gram matrix, it is not possible to distinguish between these two matrices. We tackle this inherent problem in two different approaches. Generative Models. We consider two generative models (random ensembles) for A; uniform and i.i.d. ensembles, defined as follows. Given k, N, define the set Tk( ) {c {0, 1}k : w H(c) = }, (1) as the set of all k-length binary sequence with Hamming weight (w H) . Then, we say that A belongs to the uniform ensemble if A is formed by drawing independently its n rows from Tk( ). In the latter ensemble, the matrix A is an i.i.d. matrix, with each entry being a Bernoulli(p) random variable, where 0 p 1. As mentioned above we are interested in exact recovery of the clusters {N}k i=1, or equivalently, the clustering matrix A up to a permutation of the columns of A. More precisely, we define the average probability of error associated with an algorithm which outputs an estimate ˆA of A by Perror P T π Sk{ ˆA = APπ} , where Pπ is the permutation matrix corresponding to the permutation π : [k] [k], and Sk is the symmetric group acting on [k]. Accordingly, we say that an algorithm properly recovered A if Perror is small. This recovery criterion follows from the fact that clustering is invariant to a permutation of the labels. In contrast to the above negative example where two different matrices have the same gram matrix, under certain weak conditions presented in Appendix A (see, Lemmas 3 and 4), we show that if the matrix A is generated according to the either one of the above random ensembles, the factorization is unique up to column permutations. This results have a straightforward implication: under the conditions of the Lemmas 3 and 4, the clustering recovery problem (i.e., recovering A) reduces to the recovery of the similarity matrix AAT , given partial observations Ωof its entries through the oracle O. To actually recover A from AAT we propose Algorithms 4 and 5, appear in Appendix A, for the uniform and i.i.d. ensembles, respectively. Worst-case Model. While the above generative models allow us to obtain elegant theoretical results, they may be too idealistic for real-world clustering applications. To this end, we also consider a general clustering model where each element in N can be belong to at most k clusters. Note that here each element may belong to different number of clusters. We show that under certain geometric conditions on the clusters, recovery is possible using a simple, efficient, and parameter free algorithm. Algorithm 1 Findmembership The algorithm for extracting the clustering matrix via queries to oracle. Require: Number of elements: N, number of clusters k, oracle responses Odirect(i, j) for query (i, j) Ω, where i, j [N]. 1: Choose a set S of elements drawn uniformly at random from N, and perform all pairwise queries corresponding to these |S| elements. 2: Extract the membership of all the |S| elements and find representatives T for the k clusters. 3: Query each of the remaining n |S| elements with all elements present in T . 4: Return the clusters. Algorithm 2 Find Similarity The algorithm for extracting the similarity matrix AAT via queries to oracle. Require: Number of elements: N, number of clusters k, oracle responses Odirect(i, j) for query (i, j) Ω, where i, j [N]. 1: Choose a set S of elements drawn uniformly at random from N, and perform all pairwise queries corresponding to these |S| elements. 2: Extract a valid membership of all the |S| elements by rank factorization of ASAT S. Then, find a set T S that forms a basis of Rk. 3: Query each of the remaining n |S| elements with all elements present in T . Subsequently solve for the membership vector of the unknown element. 4: Return the similarity matrix AAT . 3 Algorithms and Their Performance Guarantees In this section, we present our main theoretical results. Specifically, in Subsections 3.1 and 3.2 our main results about generative models are given. Subsection 3.3 is devoted to the worst-case model. Due to space limitations, a summary table showing our lower and upper bounds on the sample complexities, theoretical results concerning dithered oracle responses, information-theoretic lower bounds, as well as most of our pseudo-codes algorithms, appear in the appendices. 3.1 Direct Responses As a warm up we start with the case of disjoint clusters. In this case, no statistical generative assumption on A is needed. The simple algorithm (Algorithm 1) for this serves as a building block for the other more complicated scenarios considered in this paper. Proposition 1. There exists a poly-time algorithm, i.e. Algorithm 1, which with probability at least 1 n ε recovers exactly the set of clusters N1, . . . , Nk, Ni Nj = , for i = j, using |Ω| k (n m) + m 2 queries, m = (n/nmin) log(knε), where ε > 0 and nmin is the size of the smallest cluster. Proof Outline: Pick m elements uniformly at random from N, and perform all m 2 pairwise queries among these m elements. It can be shown that if m (n/nmin) log(knε), then with probability 1 n ε, among these m elements there will exist at least one element (representative) from each cluster. Finally, for the remaining (n m) items, we perform at most k queries to decide which cluster they belong to. From Prop. 1, when nmin = Ω(n/k), the number of queries needed are Ω(kn). This result should be contrasted with standard matrix completion results with uniform sampling, which state that O(kn log n) queries are needed [8]. Next, we consider the overlapping case, where Ni Nj = . In this case the similarity matrix AAT is not binary anymore. For a set S [n], with m = |S|, we let AS be the m k projection matrix formed by the rows of A that correspond to the indices in S. We have the following result. Theorem 1. There exists a polynomial-time algorithm, given in Algorithm 2, which with probability at least 1 poly(n 1) recovers exactly the set of (overlapping) clusters N1, . . . , Nk, using |Ω| |S| 2 + k (n |S|) queries, where |S| > Suniform ( k ) ( k 1)[1 + c1 log k + c2 log n], for the uniform ensemble, and |S| > Si.i.d. k 1 log k+c3 log n log max(p,1 p), for the i.i.d. ensemble, with c1, c2, c3 > 0 arbitrary positive numbers. Let us explain the main idea behind Theorem 1. It is evident from Algorithm 2 that as long as we get a valid subset of elements T S whose membership vectors form a basis of Rk, then querying a particular element i N with all elements in T gives k linearly independent equations in k variables that denote the membership of ith element to the different clusters. Subsequently, we can solve this system of equations uniquely to obtain the membership vector of ith element. Hence, if we choose |S| such that there exists a valid subset of S forming a basis of Rk with high probability, then we will be done and the sample complexity will be |S| 2 + k(n |S|). Lemmas 5 and 6 (see, Appendix B), respectively, show that if |S| > Suniform for the uniform ensemble, and |S| > Si.i.d. for the i.i.d. ensemble, then the above property holds. Remark 2. Note that in the second step of Algorithm 2 we perform a rank factorization of the matrix ASAT S. However, this factorization is not guaranteed to be unique, and accordingly, the resultant rank factorized matrix might be wrong. However, we show in the supplementary material, that even if this is the case, Algorithm 2 will nevertheless recover the true similarity matrix. 3.2 Quantized Noisy Responses We next move to the case where the oracle responses are quantized and noisy, namely, when queried with (i, j), the oracle output is Oquantized(i, j) = Q AT i Aj Wi,j, where Wi,j Bernoulli(q). We start with the uniform ensemble, for which we have the following result. Theorem 2. Assume that A was generated according to the uniform ensemble, with k 3 . Then, there exists a polynomial-time algorithm, given in Algorithm 6, which with probability 1 n ϵ, recovers the similarity matrix AAT , using |Ω| |S| 2 + |S| (n |S|) queries, where for any ε > 0, |S| > 2(1 2q) 4 k 2 log(2n2+ε). (2) The main idea behind Algorithm 6 is the following: we first choose a random subset S N of elements, such that (2) holds, and perform all pairwise queries among these elements. Using the resultant queries we infer the unquantized inner products of AT i Aj, for any (i, j) S. To this end, we count the number of elements which are similar to both the profile of elements i and j (see, the definition of Tij in Algorithm 6). Intuitively, it makes sense that the more similar the two elements i and j themselves are, the more the number of elements should be which are similar to both of them. We show that the condition in (2) suffices to make the count highly concentrated around its mean, and accordingly, outputs the true value of AT i Aj. Finally, the remaining (n |S|) elements are queried with the elements in S, and then we apply the above inferring procedure once again. We emphasize here that the exponential dependency of the upper bounds on is inherent, as the information-theoretic lower bounds in Appendix I suggest. It turns out that the above idea is capable to handle the other scenarios considered in this paper, albeit with certain technical modifications. Indeed, for the i.i.d. ensemble, we need an additional step before we can use the idea mentioned above. This is mainly because of the fact that analyzing the aforementioned count statistic requires the knowledge of support size of Ai and Aj (which is fixed in the uniform ensemble). An easy way around this problem is to infer first the ℓ0-norm of every element by counting the number of other elements that are similar. As before, under certain conditions, this count behaves differently for different values of the actual ℓ0-norm value and therefore we can infer the correct value. Once this step is done, everything else falls into place. Due to space limitation we relegate the pseudo-algorithm for the i.i.d. setting to the appendices. We have the following result. Theorem 3. Assume that A was generated according to the i.i.d. ensemble. Then, there exists a polynomial-time algorithm, given in Algorithm 9, which with probability 1 n ϵ, recovers the similarity matrix AAT , using |Ω| |S| 2 + |S| (n |S|) queries, where for any ε > 0, |S| > 2p 2(1 2q) 4(1 p)2 2k log(2n2+ε). (3) In practice, the value of the noise parameter q might be unknown to the learner. In this case, we will not know the expected values of the triangle counts under the different hypotheses a-priori, and thus our previous algorithms cannot be used directly. Fortunately, however, it turns out that with a simple modification, our algorithms can be used also when q is unknown. We have the following result stated for the uniform ensemble. A similar result can be obtained also for the i.i.d. ensemble. Theorem 4. Assume that A was generated according to the uniform ensemble with k 3 , and n > 10 k log n. Then, there exists a polynomial-time algorithm, given in Algorithm 14, independent of the noise parameter q, which with probability 1 n ϵ, recovers the similarity matrix AAT , using |Ω| |S| 2 + |S| (n |S|) queries, where for any ε > 0, |S| > 18(1 2q) 4 k 2 log(2n2+ε). (4) Comparing Theorems 2 and 4, we notice that the query complexity grows by a multiplicative constant factor only. Note that the additional technical condition n > 10 k log n is rather weak, and naturally satisfied, for example, in the regime k = O(log n). Finally, note that since we deal with quantized responses without any continuous dithering, matrix completion results cannot be used. In fact, without dithering matrix completion algorithms will fail on quantized data [12], as they do not exploit the discrete structure of the data, which is the main source for the success of our algorithms. 3.3 Beyond Generative Models: Arbitrary Worst-Case Instances In this subsection, we consider the worst-case model, where we do not impose any statistical assumptions, and assume that each element belong to at most clusters. We focus on noiseless quantized oracle responses, but also discuss the direct responses scenario in Section 4. For this case, we propose Algorithm 17. We have the following result. Theorem 5. Let Ni be the set of elements which belong to the i th cluster. If, for every cluster i [k], we have |Ni \{S j:j =i Nj}| > α n, for some α > 0, then by using Algorithm 3, |S| 2 +|S| (n |S|) queries are sufficient to recover the clusters, where α |S| = log k + log n. As mentioned above, Algorithm 3 is parameter free, do not require the knowledge of , and efficient. For the special case of = 2, we show in Appendix J (see, Theorem 9) that the same result holds under less restrictive conditions than those in Theorem 5. In fact, in Appendix K we conjecture that Theorem 5 holds true under a similar assumption as in Theorem 9. Depending on the dataset, the scaling of α in Theorem 5 w.r.t. ( , k, n) may vary widely. For example, in the non-overlapping case, α = kmin/n 1/k, where kmin is the size of the smallest cluster, which implies that the query complexity in the best scenario is O(nk log n), which is consistent with our results in the previous section. In the worst-case, a positive α could be as small as 1/n (unreasonable in real-world datasets), which implies a query complexity of O(n2). This is much higher than our average case results, as expected. More generally, note that α decreases as a function of , which implies that the query complexity increases with . For example, consider the example of 3 equally-sized clusters A, B and C. Suppose = 1 and in that case |A \ B C| = |A| = n/3, implying that α = 1/3. Now suppose that = 2. In this case A B and A C are non-empty and therefore |A \ B C| = |A| |A B| |A C| < n/3, namely, α is less than 1/3. Algorithm 3 Worst-case quantized responses Require: N, k, and oracle responses Oquantized(i, j) for every query (i, j) Ω. 1: Choose a set S of elements drawn uniformly at random from [N], and perform all pairwise queries corresponding to these |S| elements. 2: Construct a graph G = (V, E) where the vertices are the |S| sampled elements. There exist an edge between elements (i, j) only if they are determined to be similar by the oracle. 3: Construct the maximal cliques of the graph G such that all edges in E are covered. Each maximal clique forms a cluster. 4: Query each of the remaining n |S| elements with all elements present in S. For each cluster, if an element is similar with all the elements in that particular cluster, then assign the element to that cluster. Return the obtained clusters. Table 1: Sample complexities for k = O(log n) and k Oracle Type Lower-Bound Upper-Bound Direct responses (disjoint) O(nk) Ω(nk) Direct responses (overlapping) O(nk) Ω(nk) Quantized responses O(n polylog n) Ω(n polylog n) Quantized responses (worst-case, α = n c) NA Ω(n1+c polylog n) (a) Mean, median and maximum errors for = 2. (b) Number of failures for = 2. (c) Mean, median and maximum errors for = 3. (d) Number of failures for = 3. Figure 1: Results of our techniques on Movie Lens dataset. 3.4 Summary Table Table 1 summarizes the scaling of our lower and upper bounds on the sample complexities for each of the different oracle types considered in this paper. In the table, we opted to focus on the regime where k = O(log n) and k, as we found it to be the most interesting one. We also assume that the noise parameter q is fixed. Note, however, that our theoretical results are general and apply for any scaling of k and . Also, since the scalings of the sample complexities associated with the uniform and the i.i.d. ensembles, as well as when the noise parameter q is unknown, are similar, we choose to combine them together. We also present the worst-case scenario in Theorem 5, assuming that α = n c, for some c (0, 1). For simplicity of presentation, we do not explicitly present the scaling of the lower and upper bound on polylog factors. For the regime above, we can see that the scaling of the upper and lower bounds w.r.t. n is the same up to constants for direct responses. For quantized responses there is a polylog factor difference between the obtain upper and lower bounds. 4 Experimental Results We provide only a summary of our experimental results here, deferring synthetic data results to Appendix L. We focus on real-world data from the popular Movielens dataset (https://grouplens. org/datasets/movielens/) for our experiments. The dataset we used describes 5-star rating and free-text tagging activity from Movielens (http://movielens.org), a movie recommendation service. It contains 100836 ratings and 3683 tag applications across 9742 movies. Quantized Query Responses. In order to establish our results we chose the following categories Mystery, Drama, Sci-Fi, Horror, and Crime, and first selected only those movies that belong to at most two categories and at least one category (i.e., = 2). The total number of such movies were 3470 and in accordance to the statement in Theorem 9, we have α = 0.0152 (there are 53 movies that belong to Mystery but does not belong to Sci-fi and Horror). The total number of possible queries is about 1.2 107 and the number of queries that are sufficient theoretically is 2948935 (theoretical value of |S| is 245). We ran Algorithm 3 (running Algorithm 17 requires parsing all possible clique covers which is computationally hard) with different values of |S| (number of movies randomly chosen in the first step) and since the movies are sampled randomly, we ran 50 trials for each value of |S|. Finally, after the final clustering is provided by the algorithm, we calculated the gram matrix from the resulting clustering and compared it with the gram matrix of the ground truth clustering. Figure 1a shows the mean, median, and maximum error as a function of the total number of queries accrued by Algorithm 3. Here, the error refers to the total number of different entries in the estimated and true gram matrices. We can observe that the mean error almost reaches zero around 3 105 queries (about 2.5% of total). Figure 1b presents the total number of failures in perfect clustering (trials when error is larger than 1) among the 50 trials for each value of |S| we have chosen. We obtain perfect clustering in all the 50 trials first using 1.2 106 queries ( 10% of total). Note that since Theorem 5 gives a sufficient condition on T only, in practice we can take smaller values for T and still guarantee recovery. Of course, the smaller the size of S is, the sample and computational complexities are smaller as well. We repeated the experiment with the same set of categories as in the previous one but this time, we included movies that belonged to at most three clusters at the same time i.e ( = 3). The total number of such movies is 5082 and therefore the total number of possible queries is about 2.59 107. Again, we conducted 50 trials for each chosen value of |S| and as before, we plotted the mean, median and maximum error in Figure 1c and the number of failures in perfect clustering in Figure 1d for this setting. Notice that the mean error drops almost to zero at about 6 105 queries (2.32% of total) and perfect clustering over all 50 trials is achieved at 1.5 106 queries (5.8% of total). We would like to point out over here that Algorithm 3 is parameter free, and provides a non-trivial solution even when the number of queries are far below the theoretical threshold limit. It turns out that the experimental threshold is better than the theoretical threshold on queries for perfect clustering. Moreover, Algorithm 3 is efficient in partial clustering as well since the error drops very fast as the number of queries is increased. Unquantized Query Responses. In this experiment, we used the Movielens dataset again and chose 5 classes Mystery, Drama, IMAX, Sci-Fi, and Horror, and selected those elements who belonged to at most two categories and at least one category (i.e., = 2). The total number of such movies are 3270, and for each query involving two movies, we obtain back the unquantized similarity (total number of categories they both belong to). We follow Algorithm Find Similarity very closely but with a small modification. Indeed, note that Algorithm 2 is designed so that the guarantees hold under a specific stochastic assumption. More concisely, the necessary size of S is not defined for arbitrary real-world datasets. Note, however, that the main objective in the first part of the algorithm is to select a number of elements so that the gram matrix is of full rank. Therefore, for a real-world dataset, instead of sampling a fixed number of movies a-priori, we randomly select k = 5 movies and make all pairwise queries restricted to those 5 movies. Then, we check if the 5 5 gram matrix (with the (i, j) entry being the unquantized similarity between the ith and jth movies) is of rank 5 and if yes, then we will use that matrix for further calculations. If not, we sample again until we succeed. We then proceed to factorize the obtained 5 5 gram matrix into the form of BBT , where B is binary. Finally, we query every movie with the 5 movies already sampled (and clustered). This provides us with five linearly independent equations in five variables (each corresponding to whether the movie belongs to a particular cluster). Solving the equations for each movie, we finally obtain the categories each movie belongs to. Hence, the number of queries is at most Number of trials to obtain a rank 5 matrix 25 + 3270 5. Since the algorithm is randomized, we simulated this 50 times and we found that the Mean query complexity is 232126 (with a standard deviation of 269315.36) which is only 4.34% of the total number of possible queries. Acknowledgements: This research is supported in part by NSF Grants CCF 1642658, 1642550, 1618512, and 1909046. [1] K. Ahn, K. Lee, and C. Suh. Community recovery in hypergraphs. In Communication, Control, and Computing (Allerton), 2016 54th Annual Allerton Conference on, pages 657 663. IEEE, 2016. [2] N. Ailon, A. Bhattacharya, R. Jaiswal, and A. Kumar. Approximate clustering with same-cluster queries. ar Xiv preprint ar Xiv:1704.01862, 2017. [3] P. Arabie, J. D. Carroll, W. De Sarbo, and J. Wind. Overlapping clustering: A new method for product positioning. Journal of Marketing Research, pages 310 317, 1981. [4] H. Ashtiani, S. Kushagra, and S. Ben-David. Clustering with same-cluster queries. NIPS, 2016. [5] A. Banerjee, C. Krumpelman, J. Ghosh, S. Basu, and R. J. Mooney. Model-based overlapping clustering. In Proceedings of the eleventh ACM SIGKDD international conference on Knowledge discovery in data mining, pages 532 537. ACM, 2005. [6] F. Bonchi, A. Gionis, and A. Ukkonen. Overlapping correlation clustering. Knowledge and Information Systems, 35:1 32, 04 2013. [7] E. J. Candès and B. Recht. Exact matrix completion via convex optimization. Foundations of Computational mathematics, 9(6):717, 2009. [8] E. J. Candès and T. Tao. The power of convex relaxation: near-optimal matrix completion. IEEE Trans. Information Theory, 56(5):2053 2080, 2010. [9] I. Chien, C. Pan, and O. Milenkovic. Query k-means clustering and the double dixie cup problem. ar Xiv preprint ar Xiv:1806.05938, 2018. [10] T. M. Cover and J. A. Thomas. Elements of information theory, 2nd Ed. John Wiley & Sons, 2012. [11] H. S. M. Coxeter. Projective geometry. Springer Science & Business Media, 2003. [12] M. A. Davenport, Y. Plan, E. Van Den Berg, and M. Wootters. 1-bit matrix completion. Information and Inference: A Journal of the IMA, 3(3):189 223, 2014. [13] D. Firmani, B. Saha, and D. Srivastava. Online entity resolution using an oracle. PVLDB, 9(5):384 395, 2016. [14] R. H. Keshavan, A. Montanari, and S. Oh. Matrix completion from a few entries. IEEE transactions on information theory, 56(6):2980 2998, 2010. [15] T. Kim and J. Ghosh. Semi-supervised active clustering with weak oracles. ar Xiv preprint ar Xiv:1709.03202, 2017. [16] X. Mao, P. Sarkar, and D. Chakrabarti. Overlapping clustering models, and one (class) svm to bind them all. In Advances in Neural Information Processing Systems, pages 2126 2136, 2018. [17] A. Mazumdar and B. Saha. A Theoretical Analysis of First Heuristics of Crowdsourced Entity Resolution. The Thirty-First AAAI Conference on Artificial Intelligence (AAAI-17), 2017. [18] A. Mazumdar and B. Saha. Clustering with noisy queries. In Advances in Neural Information Processing Systems (NIPS) 31, 2017. [19] A. Mazumdar and B. Saha. Query complexity of clustering with side information. In Advances in Neural Information Processing Systems, pages 4682 4693, 2017. [20] R. Nugent and M. Meila. An overview of clustering applied to molecular biology. In Statistical methods in molecular biology, pages 369 404. Springer, 2010. [21] C. E. Tsourakakis, M. Mitzenmacher, K. G. Larsen, J. Błasiok, B. Lawson, P. Nakkiran, and V. Nakos. Predicting positive and negative links with noisy queries: Theory & practice. ar Xiv preprint ar Xiv:1709.07308, 2017. [22] N. Vesdapunt, K. Bellare, and N. Dalvi. Crowdsourcing algorithms for entity resolution. PVLDB, 7(12):1071 1082, 2014. [23] R. K. Vinayak and B. Hassibi. Crowdsourced clustering: Querying edges vs triangles. In Advances in Neural Information Processing Systems, pages 1316 1324, 2016. [24] J. Wang, T. Kraska, M. J. Franklin, and J. Feng. Crowder: Crowdsourcing entity resolution. PVLDB, 5(11):1483 1494, 2012. [25] J. Wang, G. Li, T. Kraska, M. J. Franklin, and J. Feng. Leveraging transitive relations for crowdsourced joins. In SIGMOD Conference, pages 229 240, 2013. [26] O. Zamir and O. Etzioni. Web document clustering: A feasibility demonstration. In Proceedings of the 21st annual international ACM SIGIR conference on Research and development in information retrieval, pages 46 54. ACM, 1998. [27] J. Zou, K. Chaudhuri, and A. T. Kalai. Crowdsourcing feature discovery via adaptively chosen comparisons. In Proceedings, The Third AAAI Conference on Human Computation and Crowdsourcing (HCOMP-15), November 2015.