# transductive_learning_is_compact__1b76a824.pdf Transductive Learning Is Compact Julian Asilis USC asilis@usc.edu Siddartha Devic USC devic@usc.edu Shaddin Dughmi USC shaddin@usc.edu Vatsal Sharan USC vsharan@usc.edu Shang-Hua Teng USC shanghua@usc.edu We demonstrate a compactness result holding broadly across supervised learning with a general class of loss functions: Any hypothesis class H is learnable with transductive sample complexity m precisely when all of its finite projections are learnable with sample complexity m. We prove that this exact form of compactness holds for realizable and agnostic learning with respect to any proper metric loss function (e.g., any norm on Rd) and any continuous loss on a compact space (e.g., cross-entropy, squared loss). For realizable learning with improper metric losses, we show that exact compactness of sample complexity can fail, and provide matching upper and lower bounds of a factor of 2 on the extent to which such sample complexities can differ. We conjecture that larger gaps are possible for the agnostic case. Furthermore, invoking the equivalence between sample complexities in the PAC and transductive models (up to lower order factors, in the realizable case) permits us to directly port our results to the PAC model, revealing an almost-exact form of compactness holding broadly in PAC learning. 1 Introduction Compactness results in mathematics describe the behavior by which, roughly speaking, an infinite system can be entirely understood by inspecting its finite subsystems: An infinite graph is k-colorable precisely when its finite subgraphs are all k-colorable [De Bruijn and Erdös, 1951], an infinite collection of compact sets in Rd has non-empty intersection precisely when the same is true of its finite subcollections, etc. In each case, compactness reveals a profound and striking structure, by which local understanding of a problem immediately yields global understanding. We demonstrate that supervised learning in the transductive model enjoys such structure. First, let us briefly review the transductive model, a close relative of the PAC model. In the realizable setting with a class of hypotheses H YX , it is defined by the following sequence of steps: 1. An adversary selects unlabeled data S = (x1, . . . , xn) X n and a hypothesis h H. 2. The unlabeled datapoints S are displayed to the learner. 3. One datapoint xi is selected uniformly at random from S. The remaining datapoints S i = (x1, . . . , xi 1, xi+1, . . . , xn) and their labels under h are displayed to the learner. 4. The learner is prompted to predict the label of xi, i.e., h(xi). The expected error incurred by the learner over the uniformly random choice of xi is its transductive error on this learning instance, from which one can easily define the transductive sample complexity of a learner and of a hypothesis class. 38th Conference on Neural Information Processing Systems (Neur IPS 2024). Notably, transductive learning, originally introduced by Vapnik and Chervonenkis [1974] and Vapnik [1982], is a fundamental approach to learning with deep theoretical connections to the PAC model. We study the transductive model as employed by the pioneering work of Haussler et al. [1994], who introduced the celebrated one-inclusion graph (OIG) to study transduction and used it to derive improved error bounds for VC classes. More recently, transductive learning and OIGs have been used to (among other work) establish the first characterizations of learnability for multiclass classification and realizable regression [Brukhim et al., 2022, Attias et al., 2023], to prove optimal PAC bounds across several learning settings [Aden-Ali et al., 2023b], and to understand regularization in multiclass learning [Asilis et al., 2024]. (See also Daniely and Shalev-Shwartz [2014], Alon et al. [2022], Montasser et al. [2022], Aden-Ali et al. [2023a].) The transductive model also naturally generalizes to the agnostic setting, much like PAC learning, as articulated by Asilis et al. [2024]. 1.1 Contributions Our results involve comparing a hypothesis class H to its finite projections. Formally, for a hypothesis class H YX and any finite collection of unlabeled data S X, we refer to the finite subsets of H|S as finite projections of H. Note that H is being made finite at two levels: first by restricting its functions to a finite region S X of the domain, and second by passing to a finite subset of H|S. Thus, any finite projection of H, e.g. F H|S, is necessarily a finite set of behaviors, |F| < , regardless of whether H|S in its totality is infinite (as may easily occur if Y is infinite). As our cornerstone result, we demonstrate in Theorem 3.6 that for the case of supervised learning with a large class of proper1 metric loss functions (including any norm on Rd or its closed subsets; see Definition 3.2) a class H can be learned with transductive sample complexity m precisely when the same is true of all its finite projections. In fact, in Theorem 3.7 we extend our results to arbitrary continuous losses on compact metric spaces, e.g., cross-entropy loss on finite-dimensional probability spaces and squared ℓ2 loss on compact subsets of Rd. For learning over arbitrary label spaces, we demonstrate in Theorems 3.8 and 3.9 that compactness fails: for realizable learning with metric losses, we provide matching upper and lower bounds of a factor of 2 on the extent to which such transductive sample complexities can differ. Our lower bound transfers directly to transductive learning in the agnostic case, for which we conjecture that larger gaps in sample complexity are possible. We stress that our compactness results are exact in the transductive model, avoiding dilution by asymptotics or even by constants. In addition, there is a growing body of work relating sample complexities in the transductive and PAC models, by which our results directly transfer in a black-box manner [Asilis et al., 2024, Aden-Ali et al., 2023b, Dughmi et al., 2024]. Notably, for realizable learning with any bounded loss, PAC sample complexities differ from their transductive counterparts by at most a logarithmic factor in δ, the confidence parameter. Combined with our results, this reveals an almost-exact form of compactness for realizable PAC learning, as we describe in Section 3.4.2 Our results hold for improper learners, i.e., learners that are permitted to emit a predictor outside the underlying class H. Curiously, compactness of sample complexity can be seen to fail strongly when one requires that learners be proper, using the work of Ben-David et al. [2019]. This demonstrates a structural difference between proper and improper learning; see Appendix B for further detail. Our compactness results are underpinned by a generalization of the classic marriage theorems for bipartite graphs which may be of independent mathematical interest. The original marriage theorem, due to Philip Hall [Hall, 1935], articulates a necessary and sufficient condition for the existence of a perfect matching from one side of a finite bipartite graph to the other. Subsequently, Marshall Hall [Hall Jr, 1948] extended the same characterization, referencing only finite subgraphs, to infinite graphs of arbitrary cardinality, provided the side to be matched has finite degrees the characterization being false otherwise, as can be seen by a simple countable example. This characterization therefore serves as a compactness result for matching on such infinite graphs. The proof of M. Hall features an involved analysis of the lattice of blocking sets , and invokes the axiom of choice through Zorn s lemma. Simpler proofs have since been discovered: a topological proof by Halmos and Vaughan 1We warn that we will shortly be overloading the term proper , as we discuss proper metric spaces and proper functions between metric spaces. We also note that our notion of properness is unrelated to losses which incentivize predicting the true probability, from e.g. Blasiok et al. [2023]. (And unrelated to proper vs. improper learners; we consider improper learners throughout the paper, which can emit predictors outside the class H.) 2Note too that any future improvements to the connections between the PAC and transductive models, whether in the realizable or agnostic settings, will be automatically inherited by our results in a black-box manner. [1950] which invokes the axiom of choice through Tychonoff s theorem, and an algebraic proof by Rado [1967] which also uses Zorn s lemma. At the heart of our paper is a compactness result (Theorem 3.3) for a variable-assignment problem which generalizes both supervised learning and bipartite matching: one side of a bipartite graph indexes infinitely many variables, the other indexes infinitely many functions that depend on finitely many variables each, and the goal is to assign all the variables in a manner that maintains all functions below a target value. Our proof draws inspiration from all three of the aforementioned proofs of M. Hall s theorem, and goes through Zorn s lemma. 1.2 Related Work The transductive approach to learning dates to the work of Vapnik and Chervonenkis [1974] and Vapnik [1982], and has inspired a breadth of recent advances across regression, classification, and various other learning regimes; see our introduction for a brief overview. Regarding transductive sample complexities, Hanneke et al. [2023] recently demonstrated a trichotomy result for optimal transductive error rates in the online setting of Ben-David et al. [1997]. In contrast, we focus on the classical (batch) setting, as described in Section 2.2. Perhaps most related to the present work is Attias et al. [2023], which introduces the γ-OIG dimension and demonstrates that it characterizes learnability for supervised learning problems with pseudometric losses. Notably, this is the first general dimension characterizing learnability across essentially the entirety of supervised learning. The γ-OIG dimension itself establishes a qualitative form of compactness as it is defined using only the finite projections of a class but we note that it has not been shown to tightly characterize the sample complexity of learning. Furthermore, it is analyzed only for realizable learning, which is in general not equivalent to agnostic learning (e.g., for regression). Our work, in contrast, establishes exact compactness for the sample complexity of transductive learning for both the realizable and agnostic settings, with respect to a general class of loss functions. Moreover, in Appendix B we extend our results to certain cases of distribution-family learning, including realizable learning of partial concept classes. 2 Preliminaries 2.1 Notation For a natural number n N, [n] denotes the set {1, . . . , n}. For a predicate P, [P] denotes the Iverson bracket of P, i.e., [P] = 1 when P is true and 0 otherwise. When Z is a set, Z<ω denotes the set of all finite sequences in Z, i.e., Z<ω = S i=1 Zi. For a tuple S = (z1, . . . , zn), we use S i to denote S with its ith entry removed, i.e., S i = (z1, . . . , zi 1, zi+1, . . . , zn). 2.2 Transductive Learning Let us recall the standard toolkit of supervised learning. A learning problem is determined by a domain X, label space Y, and hypothesis class H YX . The elements of H are functions X Y; such functions are referred to as hypotheses or predictors. Learning also requires a loss function ℓ (or d) from Y Y to R 0, which often endows Y with the structure of a metric space. Throughout the paper, we permit X to be arbitrary. A labeled datapoint is a pair (x, y) X Y and an unlabeled datapoint is an element x X. A training set, or training sample, is a tuple of labeled datapoints S (X Y)<ω. A learner is a function from training sets to predictors, i.e., A : (X Y)<ω YX . Definition 2.1. Realizable transductive learning is defined as follows: An adversary selects S = (xi)i [n] X n and a hypothesis h H. The unlabeled datapoints S are displayed to the learner. Then one datapoint xi is selected uniformly at random from S, and the remaining datapoints and their labels under h are displayed to the learner. Lastly, the learner is prompted to predict the label of xi, i.e., h(xi). We refer to the information of (S, h) as in Definition 2.1 as an instance of transductive learning, and to xi as the (randomly selected) test datapoint and S i the (randomly selected) training datapoints. The transductive error incurred by a learner A on an instance (S, h) is its average error over the uniformly random choice of test datapoint, i.e., LTrans S,h (A) = 1 i [n] ℓ A(S i, h)(xi), h(xi) , where A(S i, h) denotes the output of A on the sample (xj, h(xj))xj S i. Having defined transductive error, it is natural to define error rates and sample complexity. Definition 2.2. The transductive error rate of a learner A for H is the function ξA,H : N R defined by ξA,H(n) = sup S X n, h H LTrans S,h (A). The transductive sample complexity of a learner A for H is the function m Trans,A(ϵ) = min{m N : ξA,H(m ) ϵ, m m}. Definition 2.3. The transductive error rate of a class H is the minimal error rate attained by any of its learners, i.e., ξH(n) = inf A ξA,H(n). The transductive sample complexity m Trans,H : R>0 N of H is the function mapping ϵ to the minimal m for which ξH(m ) ϵ for all m m. That is, m Trans,H(ϵ) = min{m N : ξH(m ) ϵ, m m}. We say that H is learnable in the realizable case with transductive sample function m when m Trans,H(ϵ) m(ϵ) for all ϵ. Informally, agnostic transductive learning is the analogue in which the adversary is permitted to label the data in S arbitrarily, and in which the learner need only compete with the best hypothesis in H. We defer the formal definition to Section 3.3. 3 Compactness of Learning We present the central result of the paper in this section: the transductive sample complexity of learning is a compact property of a hypothesis class. In Section 3.1 we study compactness of realizable supervised learning over proper loss functions, and demonstrate a strong compactness result: a class H is learnable with transductive sample complexity m if and only if all its finite projections are learnable with the same complexity. In Section 3.2 we examine the case of realizable supervised learning over improper loss functions and prove a negative result: the previous compactness result no longer holds in this more general setting. Nevertheless, we demonstrate an approximate form of compactness, up to a factor of 2, for (improper) metric losses. Moreover, we show exact compactness for the special case of the (improper) 0-1 loss function, i.e., multiclass classification over arbitrary, possibly infinite label sets. Notably, this recovers M. Hall s classic matching theorem for infinite graphs [Hall Jr, 1948] as a corollary to our central result. In Section 3.3 we examine analogues of our results for agnostic learning, and in Section 3.4 we transfer our results to the PAC model via standard equivalences, obtaining approximate compactness of sample complexities. Due to space constraints, we defer an extension of our results to distribution-family PAC learning to Appendix B. 3.1 Realizable Learning With Proper Loss Functions We first consider the case of loss functions ℓ: Y Y R 0 defined on a proper metric space Y. Definition 3.1. A metric space is proper if its closed and bounded subsets are all compact. A related notion is that of a proper map between metric spaces. Definition 3.2. A function f : X Y between metric spaces is proper if it reflects compact sets, i.e., f 1(U) X is compact when U Y is compact. We remark that proper spaces are sometimes referred to as Heine-Borel spaces, and that their examples include Rd endowed with any norm, all closed subsets of Rd (under the same norms), and all finite sets endowed with arbitrary metrics. Further discussion of proper metric spaces is provided in Appendix A. The central technical result of this subsection is a compactness property concerning assignments of variables to metric spaces that maintain a family of functions below a target value ϵ. Theorem 3.3. Let L be a collection of variables, with each variable ℓ L taking values in a metric space Mℓ. Let R be a collection of proper functions, each of which depends upon finitely many variables in L and has codomain R 0. Then the following conditions are equivalent for any ϵ > 0. 1. There exists an assignment of all variables in L which keeps the output of each function r R no greater than ϵ. 2. For each finite subset R of R, there exists an assignment of all variables in L which keeps the output of each function r R no greater than ϵ. Proof. (1.) = (2.) is immediate. Before arguing the reverse direction, some terminology: a partial assignment of variables is an assignment of variables for a subset of L. A partial assignment is said to be completable with respect to R R if its unassigned variables can all be assigned so that all functions r R are kept below ϵ. A partial assignment is finitely completable if it is completable with respect to all finite subsets of R. This is a pointwise condition: the completions are permitted to vary across R s subsets. Lemma 3.4. Given a finitely completable partial assignment with an unassigned variable, one such variable can be assigned while preserving finite completability. Proof. Fix any unassigned variable ℓ L; we will assign it while preserving finite completability. For each set R R, let NR (ℓ) consist of those assignments of ℓthat preserve completability with respect to R . By the assumption of finite completability, we have that NR (ℓ) is non-empty for all finite R . We claim furthermore that NR (ℓ) is compact for finite R . To see why, let |R | = k and let ℓ1, . . . , ℓm be the variables in L upon which the functions in R depend. Suppose without loss of generality that ℓ= ℓ1 and that nodes ℓi+1, . . . , ℓm have already been assigned. Consider the function f R : Qi j=1 Mℓj Rk mapping assignments of the ℓ1, . . . , ℓi to the outputs they induce on the functions in R . (Notably, this includes the assignments already made for ℓi+1, . . . , ℓm.) As the functions in R are proper, including when fixing some of their inputs, f R is as well. Thus f 1 R ([0, ϵ]k) is compact, as is its projection onto its first coordinate. That set is precisely NR (ℓ), demonstrating our intermediate claim. We thus have a family of compact, non-empty sets I = {NR (ℓ) : R R, |R | < }. Note that finite intersections of elements of I are non-empty, as Tj i=1 NRi(ℓ) NSj i=1 Ri(ℓ) = . In metric spaces, an infinite family of compact sets has non-empty intersection if and only if the same holds for its finite intersections. Thus, by compactness of each element of I, the intersection across all of I is non-empty. That is, there exists an assignment for ℓwhich is completable with respect to all finite subsets of R. The claim follows. We now complete the argument using Zorn s lemma. Let P be the poset whose elements are finitely completable assignments, where ϕ1 ϕ2 if ϕ2 agrees with all assignments made by ϕ1 and perhaps assigns additional variables. Note first that chains in P have upper bounds. In particular, let C P be a chain and define ϕC to be the union of assignments in C, i.e., ϕC leaves ℓunassigned if all ϕ C leave ℓunassigned, otherwise assigns ℓto the unique element used by assignments in C. Clearly ϕC serves as an upper bound of C, provided that ϕC P. To see that ϕC P, fix a finite set R R. R is incident to a finite collection of nodes in L, say ℓ1, . . . , ℓm. Suppose ℓ1, . . . , ℓi are those which are assigned by ϕC, and let ϕ1, . . . , ϕi C be assignments which assign (i.e., do not leave free) the respective nodes ℓ1, . . . , ℓi. Then, as C is totally ordered, it must be that one of ϕ1, . . . , ϕi assigns all of the variables ℓ1, . . . , ℓi. That is, there exists ϕj C which agrees with ϕC in its action on ℓ1, . . . , ℓm. As ϕj P, it must be that ϕj is completable with respect to R . Then ϕC is also completable with respect to R , as R depends only upon ℓ1, . . . , ℓm. Thus, invoking Zorn s lemma, P has a maximal element ϕmax. By Lemma 3.4, it must be that ϕmax does not leave a single variable unassigned, otherwise it could be augmented with an additional assignment. There thus exists a total assignment that is finitely completable. As it has no free variables, it must indeed be maintaining all functions in R below ϵ. The claim follows. Remark 3.5. A corollary to Theorem 3.3 is that the same claim holds when the target values ϵ vary over the functions r R, as translations and scalings of proper functions Z R are proper. Figure 1: Depiction of variables L and functions R which model transductive learning, for a sequence of unlabeled datapoints |S| = 3 such that H|S contains the behaviors (0, 0, 0), (1, 0, 0), and (0, 1, 0). Arrows denote functional dependence, i.e., each r R depends upon its incident variables. Theorem 3.6. Let X be an arbitrary domain, Y a label set, and d a loss function such that (Y, d) is a proper metric space. Then the following are equivalent for any H YX and m: R>0 N: 1. H is learnable in the realizable case with transductive sample function m. 2. For any finite X X and finite H H|X, H is learnable in the realizable case with transductive sample function m. Proof. (1.) = (2.) is immediate. For the reverse direction, fix an ϵ > 0 and set n = m(ϵ). Then fix a sequence of unlabeled datapoints S X n. It suffices to demonstrate that a transductive learner for H on instances of the form {(S, h)}h H can be designed which attains error ϵ. We will capture the task of transductively learning H on such instances by way of a certain collection R of functions and L of variables. Each variable ℓ L will be permitted to take values in Y, while each function r R depends upon exactly n variables in L and outputs values in R 0. More precisely, let R = H|S and L = S S S,|S |=n 1 H|S . These serve merely as representations for the functions in R and variables in L, not their true definitions (which will be established shortly). Note now that by suppressing the unlabeled datapoints of S, we can equivalently represent elements of R as sequences in Yn and elements of L as sequences in (Y {?})n. In this view, each element of L is precisely an element of R which had exactly one entry replaced with a ? . See Figure 1. Now, to model transductive learning, fix an element r R represented by (y1, . . . , yn) Yn. Then we will define r to be a function depending upon the variables ℓ1, . . . , ℓn L, where ℓi = (y1, . . . , yi 1, ?, yi+1, . . . , yn). Given assignments for each of the variables ℓ1, . . . , ℓn as values in Y semantically, completions of their ? entries the node r then outputs the value 1 n Pn i=1 d(yi, ℓi). The two crucial observations are as follows: an assignment of each ℓ L corresponds precisely to the action of a learner responding to a query at test time, and the output of node r equals the error of a learner when r is the ground truth. Thus, it remains to show that the variables in L can all be assigned so as to keep the outputs of the functions in R less than ϵ. The condition (2.) grants us that this is true for each finite collection of functions R R. Now note that the functions r R are proper, as each such r is continuous and reflects bounded sets, and as Y itself is proper. Invoke Theorem 3.3 to complete the proof. Theorem 3.6 establishes an exact compactness in learning with respect to a flexible class of metric loss functions. One may note, however, that some non-metric losses are of central importance to machine learning, including the squared error on compact subsets of R (which violates the triangle inequality) and the cross-entropy loss for finite-dimensional distributions (which is not symmetric). We now provide a modified form of Theorem 3.6 which captures these loss functions, in which the loss function ℓY is permitted to differ from the underlying metric d on Y. (E.g., such that d is the usual Euclidean norm on a compact subset Y of Rd, and ℓY is any continuous loss function.) Theorem 3.7. Let X be an arbitrary domain, (Y, d) a compact metric space, and H YX a hypothesis class. Let ℓY : Y Y R 0 be a loss function employed for learning that is continuous with respect to the metric d. Then the following are equivalent for any m: R>0 N: 1. H is learnable in the realizable case with transductive sample function m. 2. For any finite X X and finite H H|X, H is learnable in the realizable case with transductive sample function m. Proof. We adopt precisely the perspective of Theorem 3.6, seeing transductive learning modeled as a variable assignment problem with the same variables L and functions R. To invoke Theorem 3.3, it remains only to show that the functions r R are proper. First note a continuous function from a compact space to R is automatically proper, as closed subsets of compact sets are compact. Now recall that each r R is a sum of scaled copies of ℓY with one input fixed. As each such function is continuous, r itself is continuous and thus proper. 3.2 Realizable Learning With Improper Loss Functions It is natural to ask whether the requirement that Y be a proper metric space is essential to Theorem 3.6 or merely an artifact of the proof. We now demonstrate the former: for arbitrary metric losses, the error rate of learning H can exceed that of all its finite projections by a factor of 2. Recall that ξH : N R 0 denotes the transductive error rate of learning a class H, i.e., ξH(n) denotes the error incurred by an optimal learner for H on (worst-case) samples of size n. Theorem 3.8. There exists a hypothesis class H YX , metric loss function d on Y, and n N such that for any finite X X and finite H H|X, ξH(n) 2 ξH (n). Let us describe the main idea of Theorem 3.8, whose proof is deferred to Appendix C.1. The crucial step lies in the creation of the label space Y = R S, where R is an infinite set whose points are all distance 2 apart, and S is an infinite set whose elements are indexed by the finite subsets of R, e.g., as in s R S for finite R R. For all such R , define s R to be distance 1 from the elements of R , distance 2 from the other elements of R, and distance 1 from all other points in S. Then Y indeed forms a metric space, and it is straightforward to see that, for instance, the class of all functions from a one-element set to Y is more difficult to learn than its finite projections (equivalently, finite subsets). We now prove a matching upper bound to Theorem 3.8, demonstrating that a factor of 2 is the greatest possible gap between the error rate of H and its projections when the loss function is a metric. Theorem 3.9. Let Y be a label set with a metric loss function and H YX a hypothesis class. Fix ξ : N R 0, and suppose that for any finite X X and finite H H|X, H has transductive error rate ξH ξ. Then H has transductive error rate ξH 2 ξ. The proof of Theorem 3.9 is deferred to Appendix C.2, but let us briefly sketch the main idea. Fix n N, S X n, and set ϵ = ξ(n). Consider again the collection of functions R = H|S and variables L = S S S,|S |=n 1 H|S , as described in the proof of Theorem 3.6. By the premise of the theorem, for any finite subset R R, there exists an assignment of variables L Y which maintains all functions in R below ϵ. Each such assignment induces an apportionment of error to each function in R , i.e., a vector of length n with positive entries summing to ϵ. For r R depending upon variables ℓ1, . . . , ℓn, this apportionment tracks the contribution of each ℓi to the output of r. The central technical step of the proof is to demonstrate that one can assign apportionments to each node r R such that any finite subset of the apportionments can be satisfied by an assignment of variables L Y. Then let ℓ= (y1, . . . , yi 1, ?, yi+1, . . . , yn) be a variable. We assign ℓto the value ˆy such that the function (y1, . . . , yi, ˆy, yi+1, . . . , yn) has minimal budget apportioned to ℓ, among all such ˆy. From an invocation of the triangle inequality, this learner at most doubles the output of any r R. Recall from Section 3.1 that proper metric spaces are sufficiently expressive to describe many of the most frequently studied label spaces, including Rd (equipped with any norm) and its closed subsets. What, then, is a typical example of a label space which fails to be proper? Perhaps the most natural example is multiclass classification over infinite label sets, i.e., Y equipped with the discrete metric ℓ0 1(y, y ) = [y = y ]. We will now demonstrate, however, that the particular structure of multiclass classification can be exploited to recover an exact compactness result in the style of Theorem 3.6. Notably, we do so by invoking M. Hall s classic matching theorem for infinite graphs, which for good measure we show to be a special case of our Theorem 3.3. Definition 3.10. Let G = (L R, E) be a bipartite graph. An R-matching is a set E E of disjoint edges which covers R. A graph with an R-matching is said to be R-matchable. Definition 3.11. A bipartite graph G = (L R, E) is finitely R-matchable if for each finite subset R of R, there exists a set E E of disjoint edges which covers R . M. Hall s theorem states that an infinite bipartite graph G is R-matchable if and only if it is finitely R-matchable, provided that all nodes in R have finite degree. Before proving M. Hall s theorem by way of Theorem 3.3, we establish an intermediate lemma. Lemma 3.12. Let G = (L R, E) be a bipartite graph such that all nodes r R have finite degree and G is finitely R-matchable. Then there exists a collection of edges E E such that G = (L R, E ) is finitely R-matchable and all nodes in G have finite degree. The proof of Lemma 3.12 is deferred to Appendix C.3, but its intuition is fairly simple: by P. Hall s theorem, G is finitely R-matchable precisely when Hall s condition holds, i.e., |N(R )| |R | for all finite R R [Hall, 1935]. Thus any ℓ L which is not incident to a Hall blocking set can be removed from G while preserving Hall s condition and finite R-matchability. Proceeding in this way, nodes can be removed until each remaining ℓ L is contained in a Hall blocking set R ℓ. At this point, ℓ s incident edges can be safely restricted to those which are incident with R ℓ, a finite set. We now prove M. Hall s theorem as a consequence of our Theorem 3.3. Theorem 3.13 (Hall Jr [1948]). Let G = (L R, E) be a bipartite graph in which all nodes r R have finite degree. Then G has an R-matching if and only if it is finitely R-matchable. Proof. The forward direction is clear. For the reverse, suppose G is finitely R-matchable. Then we may assume as a consequence of Lemma 3.12 that the nodes in L have finite degree as well. Let us think of each node ℓ L as a variable residing in the discrete metric space on its neighbors. We will also think of each node r R as a function of its neighbors, which outputs the number of neighbors that have not been assigned to r itself. Note that the discrete metric space on finitely many elements is proper, and furthermore that any function out of such a space is automatically proper. Then invoke Theorem 3.3 with ϵ(r) = 1 1 deg(r) to complete the proof. (See Remark 3.5.) Corollary 3.14. Let H YX be a classification problem, i.e., employing the 0-1 loss function. Then the following are equivalent for any m: R>0 N: 1. H is learnable in the realizable case with transductive sample function m. 2. For any finite X X and finite H H|X, H is learnable in the realizable case with transductive sample function m. Proof. Certainly (1.) = (2.). Then suppose (2.) and fix S X n. Now consider the bipartite graph G = (L R, E) with R = H|S, L = S S S,|S |=n 1 H|S , and where edges in E connect functions agreeing on common inputs. Then a learner for instances of the form {(S, h)}h H amounts precisely to a choice of incident node (equivalently, edge) for each ℓ L. Furthermore, such a learner attains error ϵ precisely when its selected edges contribute indegree at least d = n (1 ϵ) to each node in R. Using a splitting argument (i.e., creating d copies of each node in R), this is equivalent to asking for an R-perfect matching in a graph which, by (2.), is finitely R-matchable. Note that each node in R has degree n < and appeal to Theorem 3.13 to complete the proof. 3.3 Agnostic Learning Our discussion thus far has restricted attention to realizable learning: what can be said of the agnostic case? In short, all results from Sections 3.1 and 3.2 can be claimed for agnostic learning (with nearly identical proofs), with the exception of Theorem 3.9. To begin, let us briefly review transductive learning in the agnostic case. See Asilis et al. [2024] or Dughmi et al. [2024] for further detail. Definition 3.15. The setting of transductive learning in the agnostic case is defined as follows: 1. An adversary selects a collection of n labeled datapoints S (X Y)<ω. 2. The unlabeled datapoints in S are all revealed to the learner. 3. One labeled datapoint (xi, yi) is selected uniformly at random from S. The remaining labeled datapoints S i are displayed to the learner. 4. The learner is prompted to predict the label of xi. Notably, transductive learning in the agnostic case differs from the realizable case in that the adversary is no longer restricted to label the datapoints in S using a hypothesis H. To compensate for the increased difficulty, and in accordance with the PAC definition of agnostic learning, a learner is only judged relative to best-in-class performance across H. Formally, LTrans S (A) = 1 i [n] ℓ(A(S i)(xi), yi) inf h H 1 n i [n] ℓ(h(xi), yi). Furthermore, one can use nearly identical reasoning as in the proofs of Theorems 3.6 and 3.7 to see that agnostic transductive learning is described by a system of variables L and functions R. In particular, set R = Yn and let L (Y {?})n contain all sequences with exactly one ? . Then a function r = (y1, . . . , yn) depends upon the variables {ℓi}i [n], where ℓi = (y1, . . . , yi 1, ?, yi+1, . . . , yn) and r(ℓ1, . . . , ℓn) = 1 i [n] d Y(yi, ℓi) infh H 1 i [n] ℓ(h(xi), yi). Now, as in the realizable case, a learner A corresponds precisely to an assignment of each variable ℓ L to a value in Y, and A incurs agnostic transductive error at most ϵ if and only if the outputs of all nodes in R are maintained below ϵ. Under the conditions of Theorems 3.6 or 3.7, exact compactness of sample complexity thus comes as an immediate consequence of Theorem 3.3 and our preceding discussion. Furthermore, when Y bears a discrete metric, learning reduces to an assignment problem in graphs, and exact compactness follows from a straightforward splitting argument applied to M. Hall s matching theorem (as in Corollary 3.14). We thus have the following theorem. Theorem 3.16. Let Y and the loss function satisfy the conditions of Theorem 3.6, Theorem 3.7, or Corollary 3.14. Then the following conditions are equivalent for any H YX and m: R>0 N: 1. H is learnable in the agnostic case with transductive sample function m. 2. For any finite X X and finite H H|X, H is learnable in the agnostic case with transductive sample function m. Regarding improper metric losses, note that our lower bound from Theorem 3.8 transfers directly to the agnostic case, as it established for a hypothesis class for which agnostic learning is precisely as difficult as realizable learning. We conjecture that larger differences in such error rates perhaps of arbitrarily large ratio are possible for the agnostic case. 3.4 PAC Learning Though our results have thus far been phrased in the language of transductive learning, we now demonstrate that they may be easily extended (in an approximate manner) to Valiant s celebrated PAC model [Valiant, 1984]. The PAC model makes use of probability measures D over X Y, for which the true error incurred by a predictor h is defined as LD(h) = E(x,y) D ℓ(h(x), y). Definition 3.17. Let D be a collection of probability measures over X Y and H YX a hypothesis class. A learner A is a PAC learner for H with respect to D if there exists a sample function m : (0, 1)2 N such that the following holds: for any D D and ϵ, δ (0, 1)2, a D-i.i.d. sample S with |S| m(ϵ, δ) is such that, with probability at least 1 δ over the choice of S, LD(A(S)) inf H LD(h) + ϵ. Agnostic PAC learning refers to the case in which D consists of all measures over X Y, and realizable PAC learning to the case in which D = {D : min H LD(h) = 0}. Definition 3.18. The sample complexity of a learner A with respect to a hypothesis class H, m PAC,A : (0, 1)2 N, is the minimal sample function it attains as a learner for H. The sample complexity of a class H is the pointwise minimal sample complexity attained by any of its learners, i.e., m PAC,H(ϵ, δ) = min A m PAC,A(ϵ, δ). As previously mentioned, transductive learning bears a close connection to PAC learning: see Asilis et al. [2024] and Dughmi et al. [2024] for further detail on their approximate equivalence. Lemma 3.19 (Asilis et al. [2024, Proposition 3.6]). Let X be a domain, Y a label set, and H YX a hypothesis class. Fix a loss function taking values in [0, 1]. Then the following inequality holds for all ϵ, δ (0, 1) and the constant e 2.718: m Trans,H(e (ϵ + δ)) m PAC,H(ϵ, δ) O m Trans,H(ϵ/2) log(1/δ) . We now follow through on porting our results from the transductive model to the PAC model. The following is an immediate consequence of applying Lemma 3.19 to Theorems 3.9 and 3.7. Corollary 3.20. Let X be a domain, Y a label set, and H YX a hypothesis class. Suppose that the loss function ℓis bounded and satisfies either of the following conditions: ℓis a metric on Y, or (Y, d) is a compact metric space and ℓis continuous with respect to this topology. Then if all finite projections of H are learnable with realizable PAC sample function m : (0, 1)2 N, H is learnable with sample complexity O m( ϵ 4e) log(1/δ) . Let us mention briefly that the connection between transductive learning and PAC learning may not be as tight in the agnostic case as in the realizable case. Through a straightforward use of Markov s inequality and a repetition argument, one can show that agnostic PAC sample complexities exceed transductive by at most a factor of 1/ϵ, but this is an unimpressive bound. 4 Conclusion In this work, we studied the conditions under which the sample complexity of learning a class H can be detected by examining its finite projections. Notably, we established exact compactness results for transductive learning with a broad class of proper or continuous loss functions, across both realizable and agnostic learning. Using bounds relating the transductive and PAC models, we were able to transfer many of our results (in an approximate form) to realizable PAC learning. We leave as an open problem whether compactness of agnostic transductive sample complexities can fail by more than a factor of 2 for arbitrary (improper) metric losses. Additional future work includes better understanding the relationship between the transductive and PAC models in the agnostic case, and examining compactness for loss functions which do not satisfy any of our properness, metric, or continuity conditions (though they may be of somewhat limited interest in learning theory). It would also be of interest to study the compactness of error rates in settings other than supervised learning, such as online or unsupervised learning. Acknowledgments and Disclosure of Funding Julian Asilis was supported by the Simons Foundation and by the National Science Foundation Graduate Research Fellowship Program under Grant No. DGE-1842487. Siddartha Devic was supported by the Department of Defense through the National Defense Science & Engineering Graduate (NDSEG) Fellowship Program. Shaddin Dughmi was supported by NSF Grant CCF2009060. Vatsal Sharan was supported by NSF CAREER Award CCF-2239265 and an Amazon Research Award. Shang-Hua Teng was supported in part by the Simons Investigator Award from the Simons Foundation. Any opinions, findings, conclusions, or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of any of the sponsors such as the NSF. Ishaq Aden-Ali, Yeshwanth Cherapanamjeri, Abhishek Shetty, and Nikita Zhivotovskiy. The oneinclusion graph algorithm is not always optimal. In The Thirty Sixth Annual Conference on Learning Theory, pages 72 88. PMLR, 2023a. 2 Ishaq Aden-Ali, Yeshwanth Cherapanamjeri, Abhishek Shetty, and Nikita Zhivotovskiy. Optimal pac bounds without uniform convergence. In 2023 IEEE 64th Annual Symposium on Foundations of Computer Science (FOCS), pages 1203 1223. IEEE Computer Society, 2023b. 2 Noga Alon, Steve Hanneke, Ron Holzman, and Shay Moran. A theory of pac learnability of partial concept classes. In 2021 IEEE 62nd Annual Symposium on Foundations of Computer Science (FOCS), pages 658 671. IEEE, 2022. 2, 14 Julian Asilis, Siddartha Devic, Shaddin Dughmi, Vatsal Sharan, and Shang-Hua Teng. Regularization and optimal multiclass learning. In The Thirty Seventh Annual Conference on Learning Theory. PMLR, 2024. 2, 9, 10, 14 Idan Attias, Steve Hanneke, Alkis Kalavasis, Amin Karbasi, and Grigoris Velegkas. Optimal learners for realizable regression: PAC learning and online learning. In Thirty-seventh Conference on Neural Information Processing Systems, 2023. 2, 3 Shai Ben-David, Eyal Kushilevitz, and Yishay Mansour. Online learning versus offline learning. Machine Learning, 29:45 63, 1997. 3 Shai Ben-David, Pavel Hrubeš, Shay Moran, Amir Shpilka, and Amir Yehudayoff. Learnability can be undecidable. Nature Machine Intelligence, 1(1):44 48, 2019. 2, 14, 15 Gyora M Benedek and Alon Itai. Learnability with respect to fixed distributions. Theoretical Computer Science, 86(2):377 389, 1991. 13 Jaroslaw Blasiok, Parikshit Gopalan, Lunjia Hu, and Preetum Nakkiran. When does optimizing a proper loss yield calibration? Advances in Neural Information Processing Systems, 36, 2023. 2 Nataly Brukhim, Daniel Carmon, Irit Dinur, Shay Moran, and Amir Yehudayoff. A characterization of multiclass learnability. In 2022 IEEE 63rd Annual Symposium on Foundations of Computer Science (FOCS), pages 943 955. IEEE, 2022. 2 Amit Daniely and Shai Shalev-Shwartz. Optimal learners for multiclass problems. In Conference on Learning Theory, pages 287 316. PMLR, 2014. 2, 14 NG De Bruijn and P Erdös. A colour problem for infinite graphs and a problem in the theory of relations. Indigationes Mathematicae, 13:371 373, 1951. 1 Shaddin Dughmi, Yusuf Kalayci, and Grayson York. Is transductive learning equivalent to pac learning? ar Xiv preprint ar Xiv:2405.05190, 2024. 2, 9, 10 P. Hall. On representatives of subsets. Journal of the London Mathematical Society, s1-10(1):26 30, 1935. doi: https://doi.org/10.1112/jlms/s1-10.37.26. 2, 8, 18 Marshall Hall Jr. Distinct representatives of subsets. Bulletin of the American Mathematical Society, 54(10):922 926, 1948. 2, 4, 8 Paul Halmos and Herbert Vaughan. The marriage problem. American Journal of Mathematics, 72(1): 214 215, 1950. 2 Steve Hanneke, Shay Moran, and Jonathan Shafer. A trichotomy for transductive online learning. Advances in Neural Information Processing Systems, 36, 2023. 3 David Haussler, Nick Littlestone, and Manfred K Warmuth. Predicting {0, 1}-functions on randomly drawn points. Information and Computation, 115(2):248 292, 1994. 2, 14 Tosca Lechner and Shai Ben-David. Impossibility of characterizing distribution learning a simple solution to a long-standing problem. ar Xiv preprint ar Xiv:2304.08712, 2023. 13 Omar Montasser, Steve Hanneke, and Nati Srebro. Adversarially robust learning: A generic minimax optimal learner and characterization. Advances in Neural Information Processing Systems, 35: 37458 37470, 2022. 2 Richard Rado. Note on the transfinite case of hall s theorem on representatives. J. London Math. Soc, 42:321 324, 1967. 3 Leslie G Valiant. A theory of the learnable. Communications of the ACM, 27(11):1134 1142, 1984. Vladimir Vapnik. Estimation of dependences based on empirical data: Springer series in statistics (springer series in statistics), 1982. 2, 3 Vladimir Vapnik and Alexey Chervonenkis. Theory of pattern recognition, 1974. 2, 3 Robert Williamson and Ludvik Janos. Constructing metrics with the heine-borel property. Proceedings of the American Mathematical Society, 100(3):567 573, 1987. 13 A Proper metric spaces Several of our results concern proper metric spaces. Let us expand briefly upon this condition, and present an equivalent definition. Definition A.1. A metric space (Y, d) is proper if either of the following equivalent conditions hold: 1. For all Y Y, if Y is closed and bounded then it is compact. 2. For any y Y and r > 0, the closed ball Br(y) = {y Y : d(y, y ) r} is compact. Note that the conditions are indeed equivalent. That (1.) implies (2.) is immediate. Supposing (2.), note that any closed and bounded subset Y is a closed subset of some closed ball, and thus compact. We now discuss various sufficient conditions in order for a metric space to be proper. Lemma A.2. Let (Y, d) be a metric space. Any of the following conditions suffice to ensure that (Y, d) be a proper metric space. 1. Y is compact. 2. Y is finite. 3. Y is a closed subset of a proper metric space. Proof. If Y is compact, then its closed subsets are all compact. If Y is finite, then it is compact. If Y is a closed subset of a proper metric space Y , then its closed and bounded subsets are compact in Y and thus compact in Y. Recall now that Rd endowed with the usual Euclidean norm is a proper metric space, owing to the Heine-Borel theorem. Invoking the equivalence of all norms on Rd, it follows that Rd endowed with any norm enjoys the structure of a proper metric space. Corollary A.3. The following classes of metric spaces are proper: 1. All finite metric spaces. 2. All compact metric spaces. 3. Rd, with any norm. 4. Any closed subset of Rd, with any norm. Regarding necessary conditions for properness, note that all proper metric spaces are complete. Thus subsets of Rd which are not closed will not be proper, e.g., Q R1. See, e.g., Williamson and Janos [1987] for additional discussion and properties of proper metric spaces, which are sometimes referred to as Heine-Borel metric spaces. B Distribution-family Learning The analysis of PAC learning with respect to more flexible distribution classes than the realizable and agnostic cases falls largely under the purview of distribution-family learning [Benedek and Itai, 1991]. Formally, a problem in distribution-family learning of a class H YX is defined by a family of distributions D over X, such that unlabeled datapoints are drawn from a distribution D D and labeled by a hypothesis h H (in the realizable case) or arbitrarily (in the agnostic case). Notably, distribution-family learning has infamously resisted any characterization of learnability, combinatorial or otherwise, for the 40 years since its inception. In fact, there is some evidence to suggest that no such characterization may exist [Lechner and Ben-David, 2023]. Furthermore, it is a setting in which uniform convergence fails to characterize learning, rendering ineffective many of the standard and most celebrated techniques of learning theory. Nevertheless, we now demonstrate that compactness sheds light on the problem of distribution-family learning, at least for the case of well-behaved distribution classes. Definition B.1. A family of distributions D over a set Z is well-behaved if whenever S = (z1, . . . , zn) lies in the support of some D D, then Unif(S), the uniform distribution over S, lies in D as well. Definition B.1 is sufficiently flexible that we may apply it to distribution-family learning with Z = X or to PAC learning over arbitrary distribution classes D with Z = X Y. Though it may appear overly restrictive at first glance, note that well-behavedness is satisfied not only for ordinary PAC learning in the agnostic and realizable cases, but also for learning of partial concept classes in the realizable case [Alon et al., 2022] and for the EMX learning of Ben-David et al. [2019]. In particular, though EMX learning is not presented as a supervised learning problem in Ben-David et al. [2019], it can be seen as a binary classification problem over a domain X for which H consists of those functions outputting finitely many 1 s and D contains all realizable, discrete distributions placing all Y-mass on the label 1. Crucially, well-behavedness permits us to study PAC learning by way of transductive error. Proposition B.2. Let H YX be a hypothesis class and D a well-behaved family of distributions which are realizable (i.e., inf H LD(h) = 0 D D). Fix a loss function taking values in [0, 1]. Then the following inequality holds for all ϵ, δ (0, 1) and the constant e 2.718: m Trans,H(e (ϵ + δ)) m PAC,H(ϵ, δ) O m Trans,H(ϵ/2) log(1/δ) . Proof. The proof is nearly identical to that of [Asilis et al., 2024, Proposition 3.6]. In particular, let m Exp,H denote the sample complexity of learning H in the expected error regime (i.e., m Exp,H(ϵ) equals the number of datapoints needed to incur expected error at most ϵ). Then we have m Exp,H(ϵ + δ) m PAC,H(ϵ, δ) O m Exp,H(ϵ/2) log(1/δ) . The first inequality follows immediately from the fact that the loss function is bounded above by 1. The second inequality follows from a repetition argument, i.e., a learner attaining expected error ϵ/2 on samples of size n can be boosted to attain expected error ϵ with probability 1 δ by using an additional factor of O(log(1/δ)) many samples, as described in [Daniely and Shalev-Shwartz, 2014]. We now show that m Exp,H(ϵ) and m Trans,H(ϵ) are essentially equivalent, i.e., that m Exp,H(ϵ) m Trans,H(ϵ) m Exp,H(ϵ/e). The first inequality follows inequality from a standard leave-one-out argument of Haussler et al. [1994]. For the second inequality, let A be an optimal learner in the expected error paradigm, and note that a transductive learning instance S = (xi, yi)i [n] can be solved as follows: Given training set S j and test point xj, return A(T)(xj), where T is a sample of points drawn independently and uniformly at random from S j. The crucial observation is that T mimics an i.i.d. sample drawn from S, as such a sample will happen to avoid xi with probability 1/e. Thus the transductive error of our described learner can exceed that of A by at most a factor of e. Note that Proposition B.2 holds for the natural definition of transductive learning with respect to D, i.e., in which the adversary must select a sequence of unlabeled datapoints S X n which lie in the support of some D D. It is now immediate from the proof of Theorem 3.6 that distribution-family learning is a setting in which the transductive sample complexity of learning is H equals the sample complexity of learning its most challenging finite projections. In the following, we let D|X D denote the distributions of D which place full measure on X X. Theorem B.3. Let X be an arbitrary domain, Y a proper metric space, and H YX a hypothesis class. Let D be a family of well-behaved, realizable distributions. Then the following are equivalent for any m : R>0 N: 1. H is learnable with respect to D with transductive sample function m. 2. For any finite X X and finite H H|X, H is learnable with respect to D|X with transductive sample function m. As in Section 3.4, Proposition B.2 applied to Theorem B.3 immediately yields an almost-exact form of compactness for distribution-family PAC learning with realizable and well-behaved distribution classes. This demonstrates that the learnability of even problems as exotic as EMX learning can be detected by examining all their finite projections, provided that no restrictions are placed upon learners. In Ben-David et al. [2019], however, learners were required to only emit hypotheses in H. Our work demonstrates that the nature of their undecidability result in which the learnability of a class H is determined entirely by its cardinality, despite all its projections being easily learned could otherwise not appear in supervised learning with metric losses. C Omitted proofs C.1 Proof of Theorem 3.8 Proof. Set X = N and let Y be the metric space defined as follows. Y = R S, where R is an infinite set whose points are all distance 2 apart. S is an infinite set whose points are indexed by finite subsets of R, e.g., as in s R S for finite R R. For all such R , define s R to be distance 1 from the elements of R , distance 2 from the other elements of R, and distance 1 from all other points in S. Note that Y indeed forms a metric space as its distance function is positive-definite, symmetric, and only uses the non-zero values of 1 and 2. (In particular, the triangle inequality is satisfied as an automatic consequence of the latter fact.) Now fix an r R and k N. Define H YX to consist of all those functions which output r on all inputs x > k. Notably, any h H may take arbitrary values on x k. Let us first analyze the sample complexity of learning a finite projection of H. Fix finite X X and finite H H|X. Then, as X and H are each finite, the images of all the H are contained in a finite set Y Y. Let Y decompose as Y = R S with R R, S S. Then the following learner attains error k n on instances of size n: A : (X Y)<ω YX S 7 x 7 s R x k, r x > k. In particular, for any x k and h H , d(A(S)(x), h(x)) 1, while for any x > k, A(S) emits the correct prediction by definition of H. Furthermore, we may assume without loss of generality that transductive learning instances do not contain repeated datapoints, as these only lessen the difficulty of learning. Thus any sample S has at most k unlabeled datapoints in [k], and the previous analysis demonstrates that H can be learned with error k n when |S| = n. On the other hand, for the case of learning H itself, there exist n such that the worst-case error incurred by any learner on |S| = n is at least 2k n . In particular, take n = 1. As Y has radius 2 and furthermore for any y Y there exists y Y with d(y, y ) = 2 transductively learning H with n = 1 is guaranteed to incur an error of at least 2 in the worst case. (That is, by taking S = {x} with x k.) Thus, for n = 1 and any finite projection H we have ξH (n) 1 and ξH(n) 2, and the claim follows. C.2 Proof of Theorem 3.9 Before commencing with the proof, let us establish some terminology and a supporting lemma. Fix a hypothesis class H along with S X n and ϵ > 0. Recall the collection of functions R = H|S and variables L = S S S,|S |=n 1 H|S which capture the structure of transductive learning on instances of the form {(S, h)}h H, as described in the proof of Theorem 3.6. In this setting, for any number α > 0, we will let α-apportionment refer to a vector ν = (x1, . . . , xn) with xi 0 and P xi = α. We may suppress α and refer simply to apportionments. Given an α-apportionment ν for a node r R, we will say that an assignment of variables in L satisfies this apportionment if the output of r as a function decomposes according to ν. More explicitly, recall that r = (y1, . . . , yn) depends upon the variables {ℓi}i [n] L, where ℓi = (y1, . . . , yi 1, ?, yi+1, . . . , yn) and r(ℓ1, . . . , ℓn) = 1 i [n] d(yi, ℓi). Then an assignment of variables {ℓi}i [n] Y satisfies ν = (x1, . . . , xn) if 1 nd(yi, ℓi) xi for all i [n]. Intuitively, ν tracks the manner in which r produces its output. Similarly, given apportionments for various nodes in R, we say that a given assignment of variables satisfies the apportionments if it satisfies each of them at once. With a slight abuse of terminology, we say that a node r R without an apportionment is satisfied by an assignment of variables L Y if its output is at most ϵ under the assignment. There should be no risk of confusion, as it will always be clear whether a given node r R is endowed with an apportionment or not. In a similar fashion to Theorem 3.6, we say a partial assignment of apportionments is an assignment of apportionments to a subset of R. An assignment which happens to be total is referred to as a total assignment of apportionments. A partial assignment ϕ is satisfiable with respect to R R if there exists an assignment of variables L Y such that all nodes in R are satisfied. (I.e., have their apportionments satisfied if equipped with one, and are otherwise simply maintained below ϵ.) We say ϕ is finitely satisfiable if it is satisfiable with respect to all finite subsets of R.3 Lemma C.1. Let X be a domain, Y a metric space, and H Y a hypothesis class. Fix S X n and the corresponding collections of functions R = H|S and variables L = S S S,|S |=n 1 H|S . Let δ > 0. If R is finitely satisfiable, when none of its elements are endowed with apportionments, then there exists a total assignment of (ϵ + δ)-apportionments to R which is finitely satisfiable. Proof. We appeal to Zorn s lemma. The argument relies crucially upon the fact that any partial assignment of (ϵ + δ)-apportionments which is finitely satisfiable can be augmented by assigning an additional apportionment to an unassigned function in R. Lemma C.2. Let ϕ be a partial assignment of (ϵ+δ)-apportionments to R which is finitely satisfiable and leaves a function in R unassigned. Then one such unassigned variable can receive an (ϵ + δ)-apportionment while preserving finite satisfiability. Proof. Fix an unassigned variable r R, along with a finite collection of nodes R R. Let R = R0 R1, where the nodes in R0 are unassigned by ϕ (i.e., need only be maintained below ϵ), and those in R1 are assigned by ϕ (i.e., need be maintained below ϵ + δ and furthermore have their apportionments respected). Let A(R ) denote the collection of variable assignments L Y which satisfy all nodes in R along with r. By the supposition that ϕ is finitely satisfiable, A(R ) is non-empty. Now let Φ(R ) denote all ϵ-apportionments for r that are satisfied by an assignment in A(R ). Informally, these are the ϵ-apportionments with which we could endow r, if we only needed to consider the nodes R . Then let cl(Φ(R )) denote the closure of Φ(R ) in Rn, and consider the family of sets I = {cl(Φ(R )) : R R, |R | < }. Each set I I is compact, as it is closed and bounded. Furthermore, finite intersections of such sets are non-empty, as i=1 cl(Φ(Ri)) i=1 Φ(Ri) Φ( i=1 Ri) = . Then there exists an ϵ-apportionment ν T cl(Φ(R )), where the intersection ranges over all finite subsets of R. As we took the closures of the sets Φ(R ), this does not suffice to guarantee us that T Φ(R ) is non-empty. Let us now increase each of the entries of ν by an arbitrarily small amount, say δ/n, and call the resulting (ϵ + δ)-apportionment ν . For any finite R R, recall that ν cl(Φ(R )), meaning there exist assignments satisfying R and r which induce ϵ-apportionments on r of arbitrarily small proximity to ν. As ν strictly exceeds ν in each coordinate, then there exists an assignment satisfying R which also satisfies ν . And R was chosen arbitrarily, so the claim follows. 3Notably, this is a pointwise condition, not a uniform one. The satisfying assignments are permitted to vary across the finite subsets of R. Consider now the poset P whose elements are finitely satisfiable partial assignments of (ϵ + δ)- apportionments to R. The partial ordering is such that ϕ1 ϕ2 if ϕ2 agrees with all assignments of apportionments made by ϕ1 and perhaps assigns additional apportionments. It is straightforward to see that chains in P have upper bounds: let C P be a chain and define ϕC to be the union" of assignments in C, i.e., ϕC leaves r R unassigned if all ϕ C leave r unassigned, otherwise it assigns r to the unique apportionment used by the assignments in C. Certainly ϕC serves as an upper bound for C, provided that ϕC P. To see that ϕC P, fix a finite set R R. Suppose S R is the collection of nodes in R which receive (ϵ + δ)-apportionments from ϕC. Each s S receives an apportionment from a ϕs C. Then, as C is a chain and {ϕs}s S C is a finite set, there exists an s S with ϕs = max{ϕs}s S. By the definition of the partial order with which we endowed P, ϕs then agrees exactly with ϕC when restricted to the set R . As ϕs P, R is satisfiable with respect to ϕs and thus satisfiable with respect to ϕC. As R was selected arbitrarily, we have that ϕC is finitely satisfiable, meaning ϕC P and chains indeed have upper bounds. Then, invoking Zorn s lemma, P contains a maximal element ϕmax. By Lemma C.2, it must be that ϕmax does not leave any function in R unassigned, otherwise it could be augmented with an additional apportionment, contradicting maximality. Thus there exists a total assignment of (ϵ + δ)-apportionments which is finitely satisfiable, completing the argument. We are now equipped to prove Theorem 3.9 itself. Proof. Let H be as in the theorem statement, fix an n N, and let ϵ = ξ(n). We will exhibit a learner A for H attaining error at most 2ϵ + δ on samples of size n, for arbitrarily small δ > 0. To this end, fix one such δ > 0 and an S X n, and recall the system of functions R and variables L which capture learning on transductive instances of the form {(S, h)}h H. That is, we represent the functions in R as R = H|S and the variables in L as L = S S,|S |=n 1H|S . As described in the proof of Theorem 3.6, we may suppress the unlabeled datapoints in the definitions of R and S, and represent each r R as an element of Yn and each ℓ L as an element of (Y {?})n with exactly one ? . Recall too that each function r R, represented by (y1, . . . , yn) Yn, depends upon the variables ℓ1, . . . , ℓn L, where ℓi = (y1, . . . , yi 1, ?, yi+1, . . . , yn). Upon assigning each such variable ℓi to an element of Y (semantically, a completion of its ?" entry corresponding to a query at test time) the function r outputs the value i=1 d Y(yi, ℓi). The central observation is that defining a learner A for H which attains error 2ϵ + δ amounts precisely to assigning each variable in L to a value in Y such that the functions in R are all maintained below 2ϵ + δ. By the premise of the theorem, this collection of functions and variables is finitely satisfiable. That is, for each finite R R, there exists an assignment of all variables L Y such that all functions in R are maintained below ϵ.4 Then, by Lemma C.1, there exists an assignment of (ϵ + δ/3)-apportionments to each function in R which is finitely satisfiable. Now fix a node ℓ L: we will demonstrate how to assign it to a value of Y. Note that ℓinfluences a potentially infinite collection of functions {ri}i I R. Each such function has an apportionment of error for ℓ. Call these values {λi}i I. Recall that each node ri for i I computes its error incurred on ℓrelative to a label yi Y. Now choose an index N I such that λN inf I λi δ/3. We then set ℓ= y N. In order to analyze this assignment, fix an arbitrary i I and consider the set R = {ri, r N}. By finite satisfiability of our system of (ϵ + δ/3)-apportionments, there exists an assignment of variables 4Strictly speaking, the definition of error rate involves an infimum, meaning we are only guaranteed that the functions in R can be maintained arbitrarily close to ϵ. It is straightforward to see that this suffices for our purposes, however, as Lemma C.1 results in the addition of an arbitrarily small term to ϵ anyway. L Y satisfying the functions in R . Let y be the value of variable ℓin this assignment. We have: d Y(ℓ, yi) = d Y(y N, yi) d Y(y N, y ) + d(y , yi) λN + λi inf I λi + δ Now assign all variables in L in this manner, and consider an arbitrary node r R with error apportionment λ1, . . . , λn for each of the variables upon which it depends. Then, using the above analysis, r evaluates to at most 3 = 2 ϵ + δ, completing the argument. C.3 Proof of Lemma 3.12 Proof. Fix a bipartite graph G = (L R, E) such that all nodes in R have finite degree and G is finitely R-matchable. We will demonstrate the existence of a subgraph G = (L R, E ) of G such that G is finitely R-matchable. First note that G is finitely R-matchable if and only if P. Hall s condition holds for all finite subsets R of R. That is, if and only if |N(R )| |R |, where N(R ) denotes the set of neighbors R has in L, by Hall [1935]. We refer to a finite subset R of R as a blocking set if |N(R )| = |R |. We say a node ℓ L is contained in a blocking set if there exists a blocking set R such that ℓ N(R ). Now consider all nodes in L; if there exists a node ℓ L such that ℓis not in any blocking set, then it can be removed from L while preserving Hall s condition in the graph (i.e., while preserving finite R-matchability). Repeatedly removing nodes from L in this way and applying Zorn s lemma, we arrive at a collection of nodes L L such that Hall s condition is preserved and each node in L is contained in a blocking set. Call the resulting graph G . Now fix a node ℓ L and pick a blocking set T R containing ℓ L . We remove all edges incident to ℓwhich are not incident to R . Let us demonstrate that the remaining graph G remains finitely R-matchable, i.e., satisfies Hall s condition. Suppose not, so that there exists a finite set S R violating Hall s condition. Then, as G satisfies Hall s condition, it must be that ℓis incident to S in G but not in G . Furthermore, it must be that S is a blocking set in G , as was R. Then consider the set S T in G . We have: |N(S T)| = |N(S) N(T)| |N(S)| + |N(T)| |N(T) N(S)| |N(S)| + |N(T)| |N(T S)| |{ℓ}| |S| + |T| |T S| 1 < |S T| producing contradiction with Hall s condition for G . Note that the third line makes use of the fact that |N(T) N(S)| |N(T S)| {ℓ}, as clearly |N(T) N(S)| |N(T S)| and furthermore ℓ N(T) N(S) \ N(T S) owing to the fact that ℓis incident to both T and S but not due to a node in T S (otherwise ℓwould have remained incident to S in G ). We are thus permitted to perform the operation on any single node of L to make its degree finite while preserving Hall s condition. As any failure of Hall s condition can be detected by way of finitely many nodes in L , it follows that we can do so for all nodes of L in concert. The resulting graph is a subgraph of G which is finitely R-matchable and for which all nodes have finite degree, as desired. Neur IPS Paper Checklist Question: Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? Answer: [Yes] Justification: All claims made in the abstract and introduction are formally proven within the body of the paper (including the appendix). Guidelines: The answer NA means that the abstract and introduction do not include the claims made in the paper. The abstract and/or introduction should clearly state the claims made, including the contributions made in the paper and important assumptions and limitations. A No or NA answer to this question will not be perceived well by the reviewers. The claims made should match theoretical and experimental results, and reflect how much the results can be expected to generalize to other settings. It is fine to include aspirational goals as motivation as long as it is clear that these goals are not attained by the paper. 2. Limitations Question: Does the paper discuss the limitations of the work performed by the authors? Answer: [Yes] Justification: All our results are theoretical, and stated with their full set of required assumptions. Absent such assumptions, we make it clear that our results do not apply. (E.g., by conjecturing that larger gaps between sample complexities may be possible in the agnostic case.) Guidelines: The answer NA means that the paper has no limitation while the answer No means that the paper has limitations, but those are not discussed in the paper. The authors are encouraged to create a separate "Limitations" section in their paper. The paper should point out any strong assumptions and how robust the results are to violations of these assumptions (e.g., independence assumptions, noiseless settings, model well-specification, asymptotic approximations only holding locally). The authors should reflect on how these assumptions might be violated in practice and what the implications would be. The authors should reflect on the scope of the claims made, e.g., if the approach was only tested on a few datasets or with a few runs. In general, empirical results often depend on implicit assumptions, which should be articulated. The authors should reflect on the factors that influence the performance of the approach. For example, a facial recognition algorithm may perform poorly when image resolution is low or images are taken in low lighting. Or a speech-to-text system might not be used reliably to provide closed captions for online lectures because it fails to handle technical jargon. The authors should discuss the computational efficiency of the proposed algorithms and how they scale with dataset size. If applicable, the authors should discuss possible limitations of their approach to address problems of privacy and fairness. While the authors might fear that complete honesty about limitations might be used by reviewers as grounds for rejection, a worse outcome might be that reviewers discover limitations that aren t acknowledged in the paper. The authors should use their best judgment and recognize that individual actions in favor of transparency play an important role in developing norms that preserve the integrity of the community. Reviewers will be specifically instructed to not penalize honesty concerning limitations. 3. Theory Assumptions and Proofs Question: For each theoretical result, does the paper provide the full set of assumptions and a complete (and correct) proof? Answer: [Yes] Justification: All our results are stated with their full set of assumptions and equipped with a complete proof (in some cases, deferred to the appendix). Guidelines: The answer NA means that the paper does not include theoretical results. All the theorems, formulas, and proofs in the paper should be numbered and crossreferenced. All assumptions should be clearly stated or referenced in the statement of any theorems. The proofs can either appear in the main paper or the supplemental material, but if they appear in the supplemental material, the authors are encouraged to provide a short proof sketch to provide intuition. Inversely, any informal proof provided in the core of the paper should be complemented by formal proofs provided in appendix or supplemental material. Theorems and Lemmas that the proof relies upon should be properly referenced. 4. Experimental Result Reproducibility Question: Does the paper fully disclose all the information needed to reproduce the main experimental results of the paper to the extent that it affects the main claims and/or conclusions of the paper (regardless of whether the code and data are provided or not)? Answer: [NA] Justification: The paper does not include any experiments. Guidelines: The answer NA means that the paper does not include experiments. If the paper includes experiments, a No answer to this question will not be perceived well by the reviewers: Making the paper reproducible is important, regardless of whether the code and data are provided or not. If the contribution is a dataset and/or model, the authors should describe the steps taken to make their results reproducible or verifiable. Depending on the contribution, reproducibility can be accomplished in various ways. For example, if the contribution is a novel architecture, describing the architecture fully might suffice, or if the contribution is a specific model and empirical evaluation, it may be necessary to either make it possible for others to replicate the model with the same dataset, or provide access to the model. In general. releasing code and data is often one good way to accomplish this, but reproducibility can also be provided via detailed instructions for how to replicate the results, access to a hosted model (e.g., in the case of a large language model), releasing of a model checkpoint, or other means that are appropriate to the research performed. While Neur IPS does not require releasing code, the conference does require all submissions to provide some reasonable avenue for reproducibility, which may depend on the nature of the contribution. For example (a) If the contribution is primarily a new algorithm, the paper should make it clear how to reproduce that algorithm. (b) If the contribution is primarily a new model architecture, the paper should describe the architecture clearly and fully. (c) If the contribution is a new model (e.g., a large language model), then there should either be a way to access this model for reproducing the results or a way to reproduce the model (e.g., with an open-source dataset or instructions for how to construct the dataset). (d) We recognize that reproducibility may be tricky in some cases, in which case authors are welcome to describe the particular way they provide for reproducibility. In the case of closed-source models, it may be that access to the model is limited in some way (e.g., to registered users), but it should be possible for other researchers to have some path to reproducing or verifying the results. 5. Open access to data and code Question: Does the paper provide open access to the data and code, with sufficient instructions to faithfully reproduce the main experimental results, as described in supplemental material? Answer: [NA] Justification: The paper does not include any experiments requiring code. Guidelines: The answer NA means that paper does not include experiments requiring code. Please see the Neur IPS code and data submission guidelines (https://nips.cc/ public/guides/Code Submission Policy) for more details. While we encourage the release of code and data, we understand that this might not be possible, so No is an acceptable answer. Papers cannot be rejected simply for not including code, unless this is central to the contribution (e.g., for a new open-source benchmark). The instructions should contain the exact command and environment needed to run to reproduce the results. See the Neur IPS code and data submission guidelines (https: //nips.cc/public/guides/Code Submission Policy) for more details. The authors should provide instructions on data access and preparation, including how to access the raw data, preprocessed data, intermediate data, and generated data, etc. The authors should provide scripts to reproduce all experimental results for the new proposed method and baselines. If only a subset of experiments are reproducible, they should state which ones are omitted from the script and why. At submission time, to preserve anonymity, the authors should release anonymized versions (if applicable). Providing as much information as possible in supplemental material (appended to the paper) is recommended, but including URLs to data and code is permitted. 6. Experimental Setting/Details Question: Does the paper specify all the training and test details (e.g., data splits, hyperparameters, how they were chosen, type of optimizer, etc.) necessary to understand the results? Answer: [NA] Justification: The paper does not include any experiments. Guidelines: The answer NA means that the paper does not include experiments. The experimental setting should be presented in the core of the paper to a level of detail that is necessary to appreciate the results and make sense of them. The full details can be provided either with the code, in appendix, or as supplemental material. 7. Experiment Statistical Significance Question: Does the paper report error bars suitably and correctly defined or other appropriate information about the statistical significance of the experiments? Answer: [NA] Justification: The paper does not include any experiments. Guidelines: The answer NA means that the paper does not include experiments. The authors should answer "Yes" if the results are accompanied by error bars, confidence intervals, or statistical significance tests, at least for the experiments that support the main claims of the paper. The factors of variability that the error bars are capturing should be clearly stated (for example, train/test split, initialization, random drawing of some parameter, or overall run with given experimental conditions). The method for calculating the error bars should be explained (closed form formula, call to a library function, bootstrap, etc.) The assumptions made should be given (e.g., Normally distributed errors). It should be clear whether the error bar is the standard deviation or the standard error of the mean. It is OK to report 1-sigma error bars, but one should state it. The authors should preferably report a 2-sigma error bar than state that they have a 96% CI, if the hypothesis of Normality of errors is not verified. For asymmetric distributions, the authors should be careful not to show in tables or figures symmetric error bars that would yield results that are out of range (e.g. negative error rates). If error bars are reported in tables or plots, The authors should explain in the text how they were calculated and reference the corresponding figures or tables in the text. 8. Experiments Compute Resources Question: For each experiment, does the paper provide sufficient information on the computer resources (type of compute workers, memory, time of execution) needed to reproduce the experiments? Answer: [NA] Justification: The paper does not include any experiments. Guidelines: The answer NA means that the paper does not include experiments. The paper should indicate the type of compute workers CPU or GPU, internal cluster, or cloud provider, including relevant memory and storage. The paper should provide the amount of compute required for each of the individual experimental runs as well as estimate the total compute. The paper should disclose whether the full research project required more compute than the experiments reported in the paper (e.g., preliminary or failed experiments that didn t make it into the paper). 9. Code Of Ethics Question: Does the research conducted in the paper conform, in every respect, with the Neur IPS Code of Ethics https://neurips.cc/public/Ethics Guidelines? Answer: [Yes] Justification: Our research does not involve datasets or human participants. Furthermore, the research does not directly contribute to the development of new models or technology. Guidelines: The answer NA means that the authors have not reviewed the Neur IPS Code of Ethics. If the authors answer No, they should explain the special circumstances that require a deviation from the Code of Ethics. The authors should make sure to preserve anonymity (e.g., if there is a special consideration due to laws or regulations in their jurisdiction). 10. Broader Impacts Question: Does the paper discuss both potential positive societal impacts and negative societal impacts of the work performed? Answer: [NA] Justification: There is no societal impact of the work performed, as it is purely theoretical in nature. Our research does not directly contribute to the development of new models or technology. Guidelines: The answer NA means that there is no societal impact of the work performed. If the authors answer NA or No, they should explain why their work has no societal impact or why the paper does not address societal impact. Examples of negative societal impacts include potential malicious or unintended uses (e.g., disinformation, generating fake profiles, surveillance), fairness considerations (e.g., deployment of technologies that could make decisions that unfairly impact specific groups), privacy considerations, and security considerations. The conference expects that many papers will be foundational research and not tied to particular applications, let alone deployments. However, if there is a direct path to any negative applications, the authors should point it out. For example, it is legitimate to point out that an improvement in the quality of generative models could be used to generate deepfakes for disinformation. On the other hand, it is not needed to point out that a generic algorithm for optimizing neural networks could enable people to train models that generate Deepfakes faster. The authors should consider possible harms that could arise when the technology is being used as intended and functioning correctly, harms that could arise when the technology is being used as intended but gives incorrect results, and harms following from (intentional or unintentional) misuse of the technology. If there are negative societal impacts, the authors could also discuss possible mitigation strategies (e.g., gated release of models, providing defenses in addition to attacks, mechanisms for monitoring misuse, mechanisms to monitor how a system learns from feedback over time, improving the efficiency and accessibility of ML). 11. Safeguards Question: Does the paper describe safeguards that have been put in place for responsible release of data or models that have a high risk for misuse (e.g., pretrained language models, image generators, or scraped datasets)? Answer: [NA] Justification: The paper does not release, or directly contribute to the release of, any data or models. Guidelines: The answer NA means that the paper poses no such risks. Released models that have a high risk for misuse or dual-use should be released with necessary safeguards to allow for controlled use of the model, for example by requiring that users adhere to usage guidelines or restrictions to access the model or implementing safety filters. Datasets that have been scraped from the Internet could pose safety risks. The authors should describe how they avoided releasing unsafe images. We recognize that providing effective safeguards is challenging, and many papers do not require this, but we encourage authors to take this into account and make a best faith effort. 12. Licenses for existing assets Question: Are the creators or original owners of assets (e.g., code, data, models), used in the paper, properly credited and are the license and terms of use explicitly mentioned and properly respected? Answer: [NA] Justification: The paper does not use any existing assets. Guidelines: The answer NA means that the paper does not use existing assets. The authors should cite the original paper that produced the code package or dataset. The authors should state which version of the asset is used and, if possible, include a URL. The name of the license (e.g., CC-BY 4.0) should be included for each asset. For scraped data from a particular source (e.g., website), the copyright and terms of service of that source should be provided. If assets are released, the license, copyright information, and terms of use in the package should be provided. For popular datasets, paperswithcode.com/datasets has curated licenses for some datasets. Their licensing guide can help determine the license of a dataset. For existing datasets that are re-packaged, both the original license and the license of the derived asset (if it has changed) should be provided. If this information is not available online, the authors are encouraged to reach out to the asset s creators. 13. New Assets Question: Are new assets introduced in the paper well documented and is the documentation provided alongside the assets? Answer: [NA] Justification: The paper does not release any new assets. Guidelines: The answer NA means that the paper does not release new assets. Researchers should communicate the details of the dataset/code/model as part of their submissions via structured templates. This includes details about training, license, limitations, etc. The paper should discuss whether and how consent was obtained from people whose asset is used. At submission time, remember to anonymize your assets (if applicable). You can either create an anonymized URL or include an anonymized zip file. 14. Crowdsourcing and Research with Human Subjects Question: For crowdsourcing experiments and research with human subjects, does the paper include the full text of instructions given to participants and screenshots, if applicable, as well as details about compensation (if any)? Answer: [NA] Justification: Our work does not include the involvement of any human participants. Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Including this information in the supplemental material is fine, but if the main contribution of the paper involves human subjects, then as much detail as possible should be included in the main paper. According to the Neur IPS Code of Ethics, workers involved in data collection, curation, or other labor should be paid at least the minimum wage in the country of the data collector. 15. Institutional Review Board (IRB) Approvals or Equivalent for Research with Human Subjects Question: Does the paper describe potential risks incurred by study participants, whether such risks were disclosed to the subjects, and whether Institutional Review Board (IRB) approvals (or an equivalent approval/review based on the requirements of your country or institution) were obtained? Answer: [NA] Justification: Our work does not include the involvement of any human participants. Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Depending on the country in which research is conducted, IRB approval (or equivalent) may be required for any human subjects research. If you obtained IRB approval, you should clearly state this in the paper. We recognize that the procedures for this may vary significantly between institutions and locations, and we expect authors to adhere to the Neur IPS Code of Ethics and the guidelines for their institution. For initial submissions, do not include any information that would break anonymity (if applicable), such as the institution conducting the review.