# interactive_structure_learning_with_structural_querybycommittee__914f377d.pdf Interactive Structure Learning with Structural Query-by-Committee Christopher Tosh Columbia University c.tosh@columbia.edu Sanjoy Dasgupta UC San Diego dasgupta@cs.ucsd.edu In this work, we introduce interactive structure learning, a framework that unifies many different interactive learning tasks. We present a generalization of the queryby-committee active learning algorithm for this setting, and we study its consistency and rate of convergence, both theoretically and empirically, with and without noise. 1 Introduction We introduce interactive structure learning, an abstract problem that encompasses many interactive learning tasks that have traditionally been studied in isolation, including active learning of binary classifiers, interactive clustering, interactive embedding, and active learning of structured output predictors. These problems include variants of both supervised and unsupervised tasks, and allow many different types of feedback, from binary labels to must-link/cannot-link constraints to similarity assessments to structured outputs. Despite these surface differences, they conform to a common template that allows them to be fruitfully unified. In interactive structure learning, there is a space of items X for instance, an input space on which a classifier is to be learned, or points to cluster, or points to embed in a metric space and the goal is to learn a structure on X, chosen from a family G. This set G could consist, for example, of all linear classifiers on X, or all hierarchical clusterings of X, or all knowledge graphs on X. There is a target structure g 2 G and the hope is to get close to this target. This is achieved by combining a loss function or prior on G with interactive feedback from an expert. We allow this interaction to be fairly general. In most interactive learning work, the dominant paradigm has been question-answering: the learner asks a question (like what is the label of this point x? ) and the expert provides the answer. We allow a more flexible protocol in which the learner provides a constant-sized snapshot of its current structure and asks whether it is correct ( does the clustering, restricted to these ten points, look right? ). If the snapshot is correct, the expert accepts it; otherwise, the expert fixes some part of it. This type of feedback, first studied in generality by [15], can be called partial correction. It is a strict generalization of question-answering, and as we explain in more detail below, it allows more intuitive interactions in many scenarios. In Section 3, we present structural query-by-committee, a simple algorithm that can be used for any instance of interactive structure learning. It is a generalization of the well-known query-bycommittee (QBC) algorithm [33, 16], and operates, roughly, by maintaining a posterior distribution over structures and soliciting feedback on snapshots on which there is high uncertainty. We also introduce an adaptation of the algorithm that allows convex loss functions to handle the noise. This helps computational complexity in some practical settings, most notably when G consists of linear functions, and also makes it possible to efficiently kernelize structural QBC. In Section 4, we show that structural QBC is guaranteed to converge to the target g , even when the expert s feedback is noisy. In the appendix, we give rates of convergence in terms of a shrinkage 32nd Conference on Neural Information Processing Systems (Neur IPS 2018), Montréal, Canada. coefficient, present experiments on a variety of interactive learning tasks, and give an overview of related work. 2 Interactive structure learning The space of possible interactive learning schemes is large and mostly unexplored. We can get a sense of its diversity from a few examples. In active learning [32], a machine is given a pool of unlabeled data and adaptively queries the labels of certain data points. By focusing on informative points, the machine may learn a good classifier using fewer labels than would be needed in a passive setting. Sometimes, the labels are complex structured objects, such as parse trees for sentences or segmentations of images. In such cases, providing an entire label is time-consuming and it is easier if the machine simply suggests a label (such as a tree) and lets the expert either accept it or correct some particularly glaring fault in it. This is interaction with partial correction. It is more general than the question-answering usually assumed in active learning, and more convenient in many settings. Interaction has also been used to augment unsupervised learning. Despite great improvements in algorithms for clustering, topic modeling, and so on, the outputs of these procedures are rarely perfectly aligned with the user s needs. The problem is one of underspecification: there are many legitimate ways to organize complex high-dimensional data, and no algorithm can magically guess which a user has in mind. However, a modest amount of interaction may help overcome this issue. For instance, the user can iteratively provide must-link and cannot-link constraints [37] to edit a flat clustering, or triplet constraints to edit a hierarchy [36]. These are just a few examples of interactive learning that have been investigated. The true scope of the settings in which interaction can be integrated is immense, ranging from structured output prediction to metric learning and beyond. In what follows, we aim to provide a unifying framework to address this profusion of learning problems. 2.1 The space of structures Let X be a set of data points. This could be a pool of unlabeled data to be used for active learning, or a set of points to be clustered, or an entire instance space on which a metric will be learned. We wish to learn a structure on X, chosen from a class G. This could, for instance, be the set of all labelings of X consistent with a function class F of classifiers (binary, multiclass, or with complex structured labels), or the set of all partitions of X, or the set of all metrics on X. Of these, there is some target g 2 G that we wish to attain. Although interaction will help choose a structure, it is unreasonable to expect that interaction alone could be an adequate basis for this choice. For instance, pinpointing a particular clustering over n points requires (n) must-link/cannot-link constraints, which is an excessive amount of interaction when n is large. To bridge this gap, we need a prior or a loss function over structures. For instance, if G consists of flat k-clusterings, then we may prefer clusterings with low k-means cost. If G consists of linear separators, then we may prefer functions with small norm kgk. In the absence of interaction, the machine would simply pick the structure that optimizes the prior or cost function. In this paper, we assume that this preference is encoded as a prior distribution over G. We emphasize that although we have adopted a Bayesian formulation, there is no assumption that the target structure g is actually drawn from the prior. 2.2 Feedback We consider schemes in which each individual round of interaction is not expected to take too long. This means, for instance, that the expert cannot be shown an entire clustering, of unrestricted size, and asked to comment upon it. Instead, he or she can only be given a small snapshot of the clustering, such as its restriction to 10 elements. The feedback on this snapshot will be either be to accept it, or to provide some constraint that fixes part of it. In order for this approach to work, it is essential that structures be locally checkable: that is, g corresponds to the target g if and only if every snapshot of g is satisfactory. When g is a clustering, for instance, the snapshots could be restrictions of g to subsets S X of some fixed size s. Technically, it is enough to take s = 2, which corresponds to asking the user questions of the form Do you agree with having zebra and giraffe in the same cluster? From the viewpoint of human-computer interaction, it might be preferable to use larger subsets (like s = 5 or s = 10), with questions such as Do you agree with the clustering {zebra, giraffe, dolphin}, {whale, seal}? Larger substructures provide more context and are more likely to contain glaring faults that the user can easily fix (dolphin and whale must go together). In general, we can only expect the user to provide partial feedback in these cases, rather than fully correcting the substructure. 2.3 Snapshots Perhaps the simplest type of snapshot of a structure g is the restriction of g to a small number of points. We start by discussing this case, and later present a generalization. 2.3.1 Projections For any g 2 G and any subset S X of size s = O(1), let g|S be a suitable notion of the restriction of g to S, which we will sometimes call the projection of g onto S. For instance: G is a set of classifiers on X. We can take s = 1 and let g|x be (x, g(x)) for any x 2 X. G is a set of partitions (flat clusterings) of X. For a set S X of size s 2, let g|S be the induced partition on just the points S. As discussed earlier, it will often be helpful to pick projections of size larger than the minimal possible s. For clusterings, for instance, any s 2 satisfies local checkability, but human feedback might be more effective when s = 10 than when s = 2. Thus, in general, the queries made to the expert will consist of snapshots (projections of size s = 10, say) that can in turn be decomposed further into atomic units (projections of size 2). 2.3.2 Atomic decompositions of structures Now we generalize the notion of projection to other types of snapshots and their atomic units. We will take a functional view of the space of structures G, in which each structure g is specified by its answers to a set of atomic questions A. For instance, if G is the set of partitions of X, then A = , with g({x, x0}) = 1 if g places x, x0 in the same cluster and 0 otherwise. The queries made during interaction can, in general, be composed of multiple atomic units, and feedback will be received on at least one of these atoms. Formally, let Q be the space of queries. In the partitioning example, this might be . The relationship between Q and A is captured by the following requirements: Each q 2 Q can be decomposed as a set of atomic questions A(q) A, and we write g(q) = {(a, g(a)) : a 2 A(q)}. In the partitioning example, A(q) is the set of all pairs in q. The user accepts g(q) if and only if g satisfactorily answers every atomic question in q, that is, if and only if g(a) = g (a) for all a 2 A(q). 2.4 Summary of framework To summarize, interactive structure learning has two key components: A reduction to multiclass classifier learning. We view each structure g 2 G as a function on atomic questions A. Thus, learning a good structure is equivalent to picking one whose labels g(a) are correct. Feedback by partial correction. For practical reasons we consider broad queries, from a set Q, where each query can be decomposed into atomic questions, allowing for partial corrections. This decomposition is given by the function A : Q ! 2A. Algorithm 1 STRUCTURAL QBC Input: Distribution1 over query space Q and initial prior distribution o over G Output: Posterior distribution t over G for t = 1, 2, . . . do Draw gt t 1 while Next query qt has not been chosen do Draw q and g, g0 t 1 With probability d(g, g0; q): take qt = q end while Show user qt and gt(qt) and receive feedback in form of pairs (at, yt) Update posterior: t(g) / t 1(g) exp( β 1(g(at) 6= yt)) end for The reduction to multiclass classification immediately suggests algorithms that can be used in the interactive setting. We are particular interested in adaptive querying, with the aim of finding a good structure with minimal interaction. Of the many schemes available for binary classifiers, one that appears to work well in practice and has good statistical properties is query-by-committee [33, 16]. It is thus a natural candidate to generalize to the broader problem of structure learning. 3 Structural QBC Query-by-committee, as originally analyzed by [16], is an active learning algorithm for binary classification in the noiseless setting. It uses a prior probability distribution over its classifiers and keeps track of the current version space, i.e. the classifiers consistent with the labeled data seen so far. At any given time, the next query is chosen as follows: Pick x 2 X at random Pick classifiers h, h0 at random from restricted to the current version space If h(x) 6= h0(x): halt with x as the query In our setting, the feedback at time t is the answer yt to some atomic question at 2 A, and we can define the resulting version space to be {g 2 G : g(at0) = yt0 for all t0 t}. The immediate generalization of QBC would involve picking a query q 2 Q at random (or more generally, drawn from some query distribution ), and then choosing it if g, g0 sampled from restricted to our version space happen to disagree on it. But this is unlikely to work well, because the answers to queries are no longer binary labels but mini-structures. As a result, g, g0 are likely to disagree on minor details even when the version space is quite small, leading to excessive querying. To address this, we will use a more refined notion of the difference between g(q) and g0(q): d(g, g0; q) = 1 |A(q)| 1[g(a) 6= g0(a)]. In words, this is the fraction of atomic subquestions of q on which g and g0 disagree. It is a value between 0 and 1, where higher values mean that g(q) differs significantly from g0(q). Then we will query q with probability d(g, g0; q). 3.1 Accommodating noisy feedback We are interested in the noisy setting, where the user s feedback may occasionally be inconsistent with the target structure. In this case, the notion of a version space is less clear-cut. Our modification is very simple: the feedback at time t, say (at, yt), causes the posterior to be updated as follows: t(g) / t 1(g) exp( β 1[g(at) 6= yt]). (1) 1In the setting where Q is finite, a reasonable choice of would be uniform over Q. Here β > 0 is a constant that controls how aggressively errors are punished. In the noiseless setting, we can take β = 1 and recover the original QBC update. Even with noise, however, this posterior update still enjoys nice theoretical properties. The full algorithm is shown in Algorithm 1. 3.2 Uncertainty and informative queries What kinds of queries will structural QBC make? To answer this, we first quantify the uncertainty in the current posterior about a particular query or atom. Define the uncertainty of atom a 2 A under distribution distribution as u(a; ) = Prg,g0 (g(a) 6= g0(a)) and u(q; ) as the average uncertainty of its atoms A(q). These values lie in the range [0, 1]. The probability that a particular query q 2 Q is chosen in round t by structural QBC is proportional to (q)u(q; t 1). Thus, queries with higher uncertainty under the current posterior are more likely to be chosen. As the following lemma demonstrates, getting feedback on uncertain atoms eliminates, or down-weights in the case of noisy feedback, many structures inconsistent with g . Lemma 1. For any distribution over G, we have ({g : g(a) 6= y}) u(a; )/2. The proof of Lemma 1 is deferred to the appendix. This gives some intuition for the query selection criterion of structural QBC, and will later be used in the proof of consistency. 3.3 General loss functions The update rule for structural QBC, equation (1), results in a posterior of the form t(g) / (g) exp( β #(mistakes made by g)), which may be difficult to sample from. Thus, we consider a broader class of updates, t(g) / t 1(g) exp( β (g(at), yt)), (2) where ( , ) is a general loss function. In the special case where G consists of linear functions and is convex, t will be a log-concave distribution, which allows for efficient sampling [28]. We will show that this update also enjoys nice theoretical properties, albeit under different noise conditions. To formally specify this setting, let Y be the space of answers to atomic questions A, and suppose that structures in G generate values in some prediction space Z Rd. That is, each g 2 G is a function g : A ! Z, and any output z 2 Z gets translated to some prediction in Y. The loss associated with predicting z when the true answer is y is denoted (z, y). Here are some examples: 0 1 loss. Z = Y and (z, y) = 1(y 6= z). Squared loss. Y = { 1, 1}, Z = [ B, B], and (z, y) = (y z)2. Logistic loss. Y = { 1, 1}, Z = [ B, B] for some B > 0, and (z, y) = ln(1 + e yz). When moving from a discrete to a continuous prediction space, it becomes very possible that the predictions, on a particular atom, of two randomly chosen structures will be close but not perfectly aligned. Thus, instead of checking strict equality of these predictions, we need to modify our querying strategy to take into account the distance between them. To this end, we will use the normalized average squared Euclidean distance: d2(g, g0; q) = 1 |A(q)| kg(a) g0(a)k2 where D = maxa2A maxg,g02G kg(a) g0(a)k2. Note that d2(g, g0; q) is a value between 0 and 1. We treat it as a probability, in exactly the same way we used d(g, g0; q) in the 0-1 loss setting. In the 0-1 loss setting, structural QBC chooses queries proportional to their uncertainty. What queries will structural QBC make in the general loss setting? Define the variance of a 2 A under as var(a; ) = 1 (g) (g0) kg(a) g0(a)k2 and var(q; ) as the average variance of its atoms A(q). Then the probability that structural QBC chooses q 2 Q at step t is proportional to (q)var(q; t 1) in the general loss setting. Algorithm 2 ROBUST QUERY SELECTION Input: Fixed set of queries q1, . . . , qm 2 Q, current distribution over G Output: Query qi Initial shrinkage estimate: buo = 1/2 for t = 0, 1, 2, . . . do Draw g1, g0 1, . . . , gnt, g0 nt If there exists qj such that 1 nt i=1 d(gi, g0 i; qj) but then we halt and query qj Otherwise, let but+1 = but/2. end for 3.4 Kernelizing structural QBC Consider the special case where G consists of linear functions, i.e. G = {gw(x) = hx, wi : w 2 Rd}. As mentioned above, when our loss function is convex, the posteriors we encounter are log-concave, and thus efficiently samplable. But what if we want a more expressive class than linear functions? To address this, we will resort to kernels. Gilad-Bachrach et al. [17] investigated the use of kernels in QBC. In particular, they observed that QBC does not actually need samples from the prior restricted to the current version space. Rather, given a candidate query x, it is enough to be able to sample from the distribution the posterior induces over the labelings of x. Although their work was in the realizable binary setting, this observation still applies to our setting. Given a feature mapping φ : X ! Rd, our posterior update becomes t(gw) / t 1(gw) exp ( β (hφ(xt), wi, yt)) . As the following lemma shows, when ( , ) is the squared-loss and our prior is Gaussian, the predictions of the posterior have a univariate normal distribution. Lemma 2. Suppose = N(0, σ2 o Id), ( , ) is the squared-loss, and we have observed (x1, y1), , (xt, yt). If gw t, then hw, φ(x)i N(µ, σ2) where oβ T (It o K) y and σ2 = σ2 φ(x)T φ(x) T o for Kij = hφ(xi), φ(xj)i, i = hφ(xi), φ(x)i, and o = 1 2σ2β It + K The proof is deferred to the appendix. The important observation here is that all the quantities involving the feature mapping in Lemma 2 are inner products. Thus we never need to explicitly construct any feature vectors. 3.5 Reducing the randomness in structural QBC It is easy to see that the query selection procedure of structural QBC is a rejection sampler where each query q is chosen with probability proportional to (q)u(q; t) (in the case of the 0-1 loss) or (q)var(q; t) (for general losses). However, it is possible for the rejection rate to be quite high, even when there are many queries that have much higher uncertainty or variance than the rest. To circumvent this issue, we introduce a robust version of structural QBC, wherein many candidate queries are sampled, and the query that has the highest uncertainty or variance is chosen. In the 0-1 loss case, we can estimate the uncertainty of a candidate query q by drawing many pairs g1, g0 1, . . . , gn, g0 n t and using the unbiased estimator bu(q; t) := 1 i=1 d(gi, g0 Unfortunately, the number of structures we need to sample in order to identify the most uncertain query depends on its uncertainty, which we do not know a priori. To circumvent this difficulty, we can use the halving procedure shown in Algorithm 2. If the appropriate number of structures are sampled at each round t, on the order of O((1/u2 t) log(m log(1/uo))) for some crude lower bound uo on the highest uncertainty, then with high probability this procedure terminates with a candidate query whose uncertainty is within a constant factor of the highest uncertainty [35]. 4 Consistency of structural QBC In this section, we look at a typical setting in which there is a finite but possibly very large pool of candidate questions Q, and thus the space of structures G is effectively finite. Let g 2 G be the target structure, as before. Our goal in this setting is to demonstrate the consistency of structural QBC, meaning that limt!1 t(g ) = 1 almost surely. To do so, we formalize our setting. Note that the random outcomes during time step t of structural QBC consist of the query qt, the atomic question at that the expert chooses to answer (pick one at random if the expert answers several of them), and the response yt to at. Let Ft denote the sigma-field of all outcomes up to, and including, time t. 4.1 Consistency under 0-1 loss In order to prove consistency, we will have to make some assumptions about the feedback we receive from a user. For query q 2 Q and atomic question a 2 A(q), let (y|a, q) denote the conditional probability that the user answers y to atomic question a, in the context of query q. Our first assumption is that the single most likely answer is g (a). Assumption 1. There exists 0 < λ 1 such that (g (a)|a, q) (y|a, q) λ for all q 2 Q and a 2 A(q) and all y 6= g (a). (We will use the convention λ = 1 for the noiseless setting.) In the learning literature, Assumption 1 is known as Massart s bounded noise condition [2]. The following lemma, whose proof is deferred to the appendix, demonstrates that under Assumption 1, the posterior probability of g increases in expectation with each query, as long as the β parameter of the update rule in equation (1) is small enough relative to λ. Lemma 3. Fix any t and suppose the expert provides an answer to atomic question at 2 A(qt) at time t. Let γt = t 1({g 2 G : g(at) = g (at)}). Define t by: ))))Ft 1, qt, at = (1 t) 1 t 1(g ), Under Assumption 1, t can be lower-bounded as follows: (a) If λ = 1 (noiseless setting), t (1 γt)(1 e β). (b) For λ 2 (0, 1), if β λ/2, then t βλ(1 γt)/2. To understand the requirement β = O(λ), consider an atomic question on which there are just two possible labels, 1 and 2, and the expert chooses these with probabilities p1 and p2, respectively. If the correct answer according to g is 1, then p1 p2 + λ under Assumption 1. Let G2 denote structures that answer 2. With probability p1, the expert answers 1, and the posterior mass of G2 is effectively multiplied by e β. With probability p2, the expert answers 2, and the posterior mass of G2 is effectively multiplied by eβ. The second outcome is clearly undesirable. In order for it to be counteracted, in expectation, by the first, β must be small relative to p1/p2. The condition β λ/2 ensures this. Lemma 3 does not, in itself, imply consistency. It is quite possible for 1/ t(g ) to keep shrinking but not converge to 1. Imagine, for instance, that the input space has two parts to it, and we keep improving on one of them but not the other. What we need is, first, to ensure that the queries qt capture some portion of the uncertainty in the current posterior, and second, that the user chooses atoms that are at least slightly informative. The first condition is assured by the SQBC querying strategy. For the second, we need an assumption. Assumption 2. There is some minimum probability po > 0 for which the following holds. If the user is presented with a query q and a structure g 2 G such that g(q) 6= g (q), then with probability at least po the user will provide feedback on some a 2 A(q) such that g(a) 6= g (a). Assumption 2 is one way of avoiding scenarios in which a user never provides feedback on a particular atom a. In such a pathological case, we might not be able to recover g (a), and thus our posterior will always put some probability mass on structures that disagree with g on a. The following lemma gives lower bounds on 1 γt under Assumption 2. Lemma 4. Suppose that G is finite and the user s feedback obeys Assumption 2. Then there exists a constant c > 0 such that for every round t E[1 γt | Ft 1] c t 1(g )2(1 t 1(g ))2 where γt = t 1({g 2 G : g(at) = g (at)}) and at is the atom the user provides feedback on. Together, Lemmas 3 and 4 show that the sequence 1 t(g ) is a positive supermartingale that decreases in expectation at each round by an amount that depends on t(g ). The following lemma tells us exactly when such stochastic processes can be guaranteed to converge. Lemma 5. Let f : [0, 1] ! R 0 be a continuous function such that f(1) = 0 and f(x) > 0 for all x 2 (0, 1). If 1 t 1(g ) f( t 1(g )) for each t 2 N, then t(g ) ! 1 almost surely. As a corollary, we see that structural QBC is consistent. Theorem 6. Suppose that G is finite, and Assumptions 1 and 2 hold. If (g ) > 0, then t(g ) ! 1 almost surely under structural QBC s query strategy. We provide a proof of Theorem 6 in the appendix, where we also provide rates of convergence. 4.2 Consistency under general losses We now turn to analyzing structural QBC with general losses. As before, we will need to make some assumptions. The first is that the loss function is well-behaved. Assumption 3. The loss function is bounded, 0 (z, y) B, and Lipschitz in its first argument, i.e. (z, y) (z0, y) Ckz z0k, for some constants B, C > 0. It is easily checked that this assumption holds for the three loss functions we mentioned earlier. In the case of 0-1 loss, we assumed that for any atomic question a, the correct answer g (a) would be given with higher probability than any incorrect answer. We now formulate an analogous assumption for the case of more general loss functions. Recall that ( |a) is the conditional probability distribution over the user s answers to a 2 A (we can also allow to also depend upon the context q, as we did before; here we drop the dependence for notational convenience). The expected loss incurred by z 2 Z on this atom is thus (y|a) (z, y). We will require that for any atomic question a, this expected loss is minimized when z = g (a), and predicting any other z results in expected loss that grows with the distance between z and g (a). Assumption 4. There exists a constant λ > 0 such that L(z, a) L(g (a), a) λkz g (a)k2 for any atomic question a 2 A and any z 2 Z. Let s look at some concrete settings: 0 1 loss with Y = Z = {0, 1}. Assumption 4 is equivalent to Assumption 1. Squared loss with Y = { 1, 1} and Z R. Assumption 4 is satisfied when g (a) = E[y|a] Logistic loss with Y = { 1, 1} and Z = [ B, B]. For a 2 A, let p = (1|a). Assump- tion 4 is satisfied when g (a) = ln p 1 p and λ = 2e2B (1+e B)4 . From these examples, it is clear that requiring g (a) to be the minimizer of L(z, a) is plausible if Z is a discrete space but much less so if Z is continuous. In general, we can only hope that this holds approximately. With this caveat in mind, we stick with Assumption 4 as a useful but idealized mathematical abstraction. With these assumptions in place, the following theorem guarantees the consistency of structural QBC under general losses. Its proof is deferred to the appendix. Theorem 7. Suppose we are in the general loss setting, G is finite, and the user s feedback satisfies Assumptions 2, 3, and 4. If (g ) > 0, then t(g ) ! 1 almost surely. 5 Conclusion In this work, we introduced interactive structure learning, a generic framework for learning structures under partial correction feedback. This framework can be applied to any structure learning problem in which structures are in one-to-one correspondence with their answers to atomic questions. Thus, interactive structure learning may be viewed as a generalization of active learning, interactive clustering with pairwise constraints, interactive hierarchical clustering with triplet constraints, and interactive ordinal embeddings with quadruplet constraints. On the algorithmic side, we introduced structural QBC, a generalization of the classical QBC algorithm to the interactive structure learning setting. We demonstrated that this algorithm is consistent, even in the presence of noise, provided that we can sample from a certain natural posterior. In the appendix, we also provided rates of convergence. Because this posterior is often intractable to sample from, we also considered an alternative posterior based on convex loss functions that sometimes allows for efficient sampling. We showed that structural QBC remains consistent in this setting, albeit under different noise conditions. In the appendix, we provide experiments on both interactive clustering and active learning tasks. On the interactive clustering side, these experiments demonstrate that even when the prior distribution places relatively low mass on the target clustering, structural QBC is capable of recovering a lowerror clustering with relatively few rounds of interaction. In contrast, these experiments also show that random corrections are not quite as useful. On the active learning side, there are experiments demonstrating the good empirical performance of structural QBC using linear classifiers with the squared-loss posterior update, with and without kernelization. Acknowledgments The authors are grateful to the reviewers for their feedback and to the NSF for support under grant CCF-1813160. Part of this work was done at the Simons Institute for Theoretical Computer Science, Berkeley, during the Foundations of Machine Learning program. CT also thanks Stefanos Poulis and Sharad Vikram for helpful discussions and feedback. [1] H. Ashtiani, S. Kushagra, and S. Ben-David. Clustering with same-cluster queries. In Advances in Neural Information Processing Systems, pages 3216 3224, 2016. [2] P. Awasthi, M.-F. Balcan, N. Haghtalab, and R. Urner. Efficient learning of linear separators under bounded noise. In Proceedings of the 28th Annual Conference on Learning Theory, pages 167 190, 2015. [3] P. Awasthi, M.-F. Balcan, and K. Voevodski. Local algorithms for interactive clustering. In Proceedings of the 31st International Conference on Machine Learning, 2014. [4] P. Awasthi and R.B. Zadeh. Supervised clustering. In Advances in Neural Information Process- ing Systems, 2010. [5] K. Azuma. Weighted sums of certain dependent random variables. Tohoku Mathematical Journal, Second Series, 19(3):357 367, 1967. [6] M.-F. Balcan and A. Blum. Clustering with interactive feedback. In Algorithmic Learning Theory (volume 5254 of the series Lecture Notes in Computer Science), pages 316 328, 2008. [7] A. Beygelzimer, S. Dasgupta, and J. Langford. Importance weighted active learning. In Proceedings of the 26th International Conference on Machine Learning, 2009. [8] R. Castro and R. Nowak. Minimax bounds for active learning. IEEE Transactions on Information Theory, 54(5):2339 2353, 2008. [9] D. Cohn, L. Atlas, and R. Ladner. Improving generalization with active learning. Machine Learning, 15(2):201 221, 1994. [10] G. Dasarathy, R. Nowak, and X. Zhu. S2: An efficient graph based active learning algorithm with application to nonparametric classification. In 28th Annual Conference on Learning Theory, pages 503 522, 2015. [11] S. Dasgupta. Analysis of a greedy active learning strategy. In Advances in Neural Information Processing Systems, 2004. [12] S. Dasgupta. Coarse sample complexity bounds for active learning. In Advances in Neural Information Processing Systems, 2005. [13] S. Dasgupta and D.J. Hsu. Hierarchical sampling for active learning. In Proceedings of the 25th International Conference on Machine Learning, 2008. [14] S. Dasgupta, D.J. Hsu, and C. Monteleoni. A general agnostic active learning algorithm. In Advances in Neural Information Processing Systems, 2007. [15] S. Dasgupta and M. Luby. Learning from partial correction. Ar Xiv e-prints, 2017. [16] Y. Freund, H. Seung, E. Shamir, and N. Tishby. Selective sampling using the query by committee algorithm. Machine Learning, 28(2):133 168, 1997. [17] R. Gilad-Bachrach, A. Navot, and N. Tishby. Query by committeee made real. In Advances in Neural Information Processing Systems, 2005. [18] A. Gonen, S. Sabato, and S. Shalev-Shwartz. Efficient active learning of halfspaces: an aggressive approach. Journal of Machine Learning Research, 14(1):2583 2615, 2013. [19] A. Guillory and J. Bilmes. Average-case active learning with costs. In Conference on Algorithmic Learning Theory, pages 141 155, 2009. [20] S. Hanneke. A bound on the label complexity of agnostic active learning. In Proceedings of the 25th International Conference on Machine Learning, 2007. [21] N. J. Higham. Accuracy and stability of numerical algorithms. SIAM, 2002. [22] W. Hoeffding. Probability inequalities for sums of bounded random variables. Journal of the American Statistical Association, 58(301):13 30, 1963. [23] T.-K. Huang, A. Agarwal, D.J. Hsu, J. Langford, and R.E. Schapire. Efficient and parsimonious agnostic active learning. In Advances in Neural Information Processing Systems, 2015. [24] D.M Kane, S. Lovett, S. Moran, and J. Zhang. Active classification with comparison queries. In IEEE Symposium on Foundations of Computer Science, pages 355 366, 2017. [25] S. Kpotufe, R. Urner, and S. Ben-David. Hierarchical label queries with data-dependent partitions. In Proceedings of the 28th Annual Conference on Learning Theory, 2015. [26] Y. Le Cun, L. Bottou, Y. Bengio, and P. Haffner. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278 2324, 1998. [27] M. Lichman. UCI machine learning repository, 2013. [28] L. Lovasz and S. Vempala. The geometry of logconcave functions and sampling algorithms. Random Structures and Algorithms, 30:307 358, 2007. [29] Nicolo N. Cesa-Bianchi, C. Gentile, and F. Vitale. Learning unknown graphs. In Conference on Algorithmic Learning Theory, pages 110 125, 2009. [30] R. Nowak. The geometry of generalized binary search. IEEE Transactions on Information Theory, 57(12):7893 7906, 2011. [31] S. Poulis and S. Dasgupta. Learning with feature feedback: from theory to practice. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, pages 1104 1113, 2017. [32] B. Settles. Active learning. Morgan Claypool, 2012. [33] H.S. Seung, M. Opper, and H. Sompolinsky. Query by committee. In Proceedings of the 5th Annual Workshop on Computational Learning Theory, pages 287 294, 1992. [34] C. Tosh and S. Dasgupta. Lower bounds for the gibbs sampler on mixtures of gaussians. In Thirty-First International Conference on Machine Learning, 2014. [35] C. Tosh and S. Dasgupta. Diameter-based active learning. In Proceedings of the 34th Interna- tional Conference on Machine Learning, pages 3444 3452, 2017. [36] S. Vikram and S. Dasgupta. Interactive Bayesian hierarchical clustering. In Proceedings of the 33rd International Conference on Machine Learning, 2016. [37] K. Wagstaff and C. Cardie. Clustering with instance-level constraints. In Proceedings of the 17th International Conference on Machine Learning, 2000. [38] Y. Xu, H. Zhang, K. Miller, A. Singh, and A. Dubrawski. Noise-tolerant interactive learning using pairwise comparisons. In Advances in Neural Information Processing Systems, pages 2431 2440, 2017. [39] X. Zhu, J. Lafferty, and Z. Ghahramani. Combining active learning and semi-supervised learning using gaussian fields and harmonic functions. In ICML Workshop on the Continuum from Labeled to Unlabeled Data, 2003.