# label_embedding_via_lowcoherence_matrices__03575efa.pdf Published in Transactions on Machine Learning Research (09/2025) Label Embedding via Low-Coherence Matrices Jianxin Zhang Electrical Engineering and Computer Science University of Michigan Ann Arbor, MI 48109 jianxinz@umich.edu Clayton Scott Electrical Engineering and Computer Science University of Michigan Ann Arbor, MI 48109 clayscot@umich.edu Reviewed on Open Review: https://openreview.net/forum?id=vrc WXcr4On Label embedding is a framework for multiclass classification problems where each label is represented by a distinct vector of some fixed dimension, and training involves matching model output to the vector representing the correct label. While label embedding has been successfully applied in extreme classification and zero-shot learning, and offers both computational and statistical advantages, its theoretical foundations remain poorly understood. This work presents an analysis of label embedding in the context of extreme multiclass classification, where the number of classes C is very large. We present an excess risk bound that reveals a trade-off between computational and statistical efficiency, quantified via the coherence of the embedding matrix. We further show that under the Massart noise condition, the statistical penalty for label embedding vanishes with sufficiently low coherence. Our analysis supports an algorithm that is simple, scalable, and easily parallelizable, and experimental results demonstrate its superiority in large-scale applications. 1 Introduction In classification, the goal is to learn from feature-label pairs {(xi, yi)}N i=1 a classifier h : X Y that accurately maps a feature vector x in the feature space X to its label y in the label space Y. For implementation purposes, labels are typically represented using the one-hot encoding which, for a C-class classification problem, represents the i-th class by the i-th standard basis vector in RC. Unfortunately, the one-hot encoding of labels is limited in the context of extreme classification, which refers to multiclass and multilabel classification problems involving thousands of classes or more (Wei et al., 2022). Extreme classification has emerged as an essential research area in machine learning, owing to an increasing number of real-world applications involving massive numbers of classes, such as image recognition (Zhou et al., 2014), natural language processing (Le and Mikolov, 2014; Jernite et al., 2017), and recommendation systems (Bhatia et al., 2015; Chang et al., 2020). Classification methods based on the one-hot encoding often struggle to scale effectively in these scenarios due to the high computational cost and memory requirements associated with handling large label spaces. To overcome this challenge, the label embedding framework represents each label by a vector of fixed dimension, typically much lower than C, and learns a function that maps a feature vector to the vector representing its label. At inference time, the label of a test data point is assigned to match the nearest label representative in the embedding space. By employing a lower-dimensional label space, label embedding overcomes the Published in Transactions on Machine Learning Research (09/2025) computational challenge posed by large label space. At the same time, there is a growing need for efficient and scalable algorithms that can tackle extreme classification problems without compromising on performance (Prabhu and Varma, 2014; Prabhu et al., 2018; Deng et al., 2018). Successful applications of label embedding to extreme classification include Yu et al. (2014); Jain et al. (2019); Bhatia et al. (2015); Guo et al. (2019); Evron et al. (2018); Hsu et al. (2009). Furthermore, Rodríguez et al. (2018) argues that label embedding can accelerate the convergence rate and better capture latent relationships between categories. Despite its widespread use, the theoretical basis of label embedding has not been thoroughly explored. This paper presents a new excess risk bound that provides insight into how label embedding algorithms work, and then exploits this insight to identify a simple implementation of label embedding with excellent performance. Specifically, our bound establishes a trade-off between computational efficiency and classification accuracy, quantified in terms of the coherence of the label embedding matrix. Our theory applies to label-embedding algorithms described by Meta-Algorithm 1, including various types of embeddings: dataindependent embeddings (Weston et al., 2002; Hsu et al., 2009), those anchored in auxiliary information (Akata et al., 2013), and embeddings co-trained with models (Weston et al., 2010). Furthermore, under the multiclass noise condition of Massart and Nédélec (2006), the statistical penalty associated with a positive matrix coherence, which results from reducing the dimension of the label space, disappears. Meta-Algorithm 1 is not our contribution, but rather a unification of several existing approaches. However, by selecting a label embedding matrix to optimize the tradeoff described by our bound, namely, by choosing a low-coherence embedding matrix, we present instantiations of this meta-algorithm that outperform competing approaches across several benchmark datasets. 2 Related work While label embedding has also been successfully applied to zero-shot learning (Wang et al., 2019; Akata et al., 2013), we focus here on extreme classification, together with related theoretical contributions. 2.1 Extreme classification Besides label embedding, existing methods for extreme multiclass classification can be grouped into three main categories: label hierarchy, one-vs-all methods, and other methods. Label Embedding. LEML (Yu et al., 2014) leverages a low-rank assumption on linear models and effectively constrains the output space of models to a low-dimensional space. SLEEC (Bhatia et al., 2015) is a local embedding framework that preserves the distance between label vectors. Guo et al. (2019) point out that low-dimensional embedding-based models could suffer from significant overfitting. Their theoretical insights inspire a novel regularization technique to alleviate overfitting in such models. WLSTS (Evron et al., 2018) is an extreme multiclass classification framework based on error correcting output coding, which embeds labels with codes induced by graphs. Hsu et al. (2009) use column vectors from a matrix with the restricted isometry property (RIP) to represent labels. Their analysis is primarily tailored to multilabel classification. They deduce bounds for the conditional ℓ2-error, which measures the squared 2-norm difference between the prediction and the label vector a metric that is not a standard measure of classification error. In contrast, our work analyzes the standard classification error. Label Hierarchy. Numerous methods such as Parabel (Prabhu et al., 2018), Bonsai (Khandagale et al., 2020), Attention XML (You et al., 2019), light XML (Jiang et al., 2021), XR-Transformer (Zhang et al., 2021), X-Transformer (Wei et al., 2019), XR-Linear (Yu et al., 2022), and ELIAS (Gupta et al., 2022) partition the label spaces into clusters. This is typically achieved by performing k-means clustering on the feature space. The training process involves training a cluster-level model to assign a cluster to a feature vector, followed by training a label-level model to assign labels within the cluster. One-vs-all methods. One-vs-all (OVA) algorithms address extreme classification problems with C labels by modeling them as C independent binary classification problems. For each label, a classifier is trained to predict its presence. Di SMEC (Babbar and Schölkopf, 2017) introduces a large-scale distributed framework to train linear OVA models, albeit at an expensive computational cost. Pro XML (Babbar and Schölkopf, 2019) Published in Transactions on Machine Learning Research (09/2025) Meta-Algorithm 1 Label Embedding. 1: Input: dataset D = {(xi, yi)}N i=1, embedding matrix G, multi-output regression algorithm A, the decoder function βG from the embedding space to the label space. 2: Form the regression dataset Dr = {(xi, gyi)}N i=1. 3: Train a regression model f with A on Dr. 4: Return: βG f. mitigates the impact of data scarcity with adversarial perturbations. SLICE (Jain et al., 2019) accelerates negative sampling based on a generative model approximation. PD-Sparse (Yen et al., 2016) and PPD-Sparse (Yen et al., 2017a) propose optimization algorithms to exploit a sparsity assumption on labels and feature vectors. Other methods. Beyond the above categories, Deep XML (Dahiya et al., 2021) uses a negative sampling procedure that shortlists O(log C) relevant labels during training and prediction. VM (Choromanska and Langford, 2015) constructs trees with O(log C) depth that have leaves with low label entropy. Based on the standard random forest training algorithm, Fast XML (Prabhu and Varma, 2014) proposes to directly optimize the Discounted Cumulative Gain to reduce the training cost. Annex ML (Tagami, 2017) constructs a k-nearest neighbor graph of the label vectors and attempts to reproduce the graph structure in a lower-dimension feature space. 2.2 Excess risk bounds Our theoretical contributions are expressed as excess risk bounds, which quantify how the excess risk associated to a surrogate loss relates to the excess risk for the 0-1 loss. Excess risk bounds for classification were developed by Zhang (2004); Bartlett et al. (2006); Steinwart (2007) and subsequently developed and extended by several authors. Ramaswamy and Agarwal (2012) shows that one needs to minimize a convex surrogate loss defined on at least a C 1 dimension space to achieve consistency for the standard C-class classification problem, i.e., any convex surrogate loss function operating on a dimension less than C 1 inevitably suffers an irreducible error. Complementing this, Ávila Pires et al. (2013) validates the consistency of simplex encoding (Mroueh et al., 2012), a variant of label embedding in a C 1 dimensional space, and introduces an excess risk bound. Previous excess risk bounds have been developed for consistent loss functions. In contrast, drawing from (Steinwart, 2007), we establish a novel excess risk bound for the label embedding framework, which admits an irreducible error and is inherently inconsistent. This error diminishes as the coherence of the embedding matrix decreases and ultimately vanishes under Massart s noise condition (Massart and Nédélec, 2006), leading to an excess risk bound of the conventional form. From a different perspective, Ramaswamy et al. (2018) put forth a novel surrogate loss function for multiclass classification with an abstain option. This abstain option enables the classifier to opt-out from making predictions at a certain cost. Remarkably, their proposed methods not only demonstrate consistency but also effectively reduce the multiclass problems to log C binary classification problems by encoding the classes with their binary representations. In particular, the region in X that causes the irreducible error in our excess risk bound is abstained from in Ramaswamy et al. (2018) to achieve lossless dimension reduction in the abstention setting. 3 Label embedding by low-coherence matrices Let X denote the feature space and Y = {1, . . . , C} denote the label space where C N. Let (X, Y ) be random variables in X Y, and let P be the probability measure that governs (X, Y ). We use PX to denote the marginal distribution of P on X. We first introduce the definitions of matrix coherence in Section 3.1, followed by the notations and problem statement in Section 3.2, and then present the associated algorithm in Section 3.3. The excess risk bound and Published in Transactions on Machine Learning Research (09/2025) Table 1: Frequently used symbols in Section 3 Symbol Description Symbol Description G Embedding matrix X Feature space Y Label space P Probability measure on X Y PX Marginal distribution of P on X L01 0-1 loss function ℓG Squared loss with embedding G RL,P Risk of L under P R L,P Bayes risk for L under P η(x) Class posterior probabilities d( ) Difference between top two posteriors λG Coherence of G βG Decoder from embedding to label LG LG(p, y) = L01(βG(p), y) its interpretation are presented in Section 3.4. Finally, we introduce the condition for lossless label embedding in Section 3.5. Frequently used notations appear in Table 1. 3.1 Matrix coherence Our theory relies on the notion of the coherence of a matrix A Cn C, which is the maximum magnitude of the dot products between distinct columns. Definition 1. Let {aj}C j=1 be the columns of the matrix A Cn C, where aj 2 = 1 for all j. The coherence of A is λ = max1 i =j C| ai, aj |. If n C, λ is 0 when the columns of A are orthonormal. When n < C, however, the coherence must be positive. Indeed, Welch (1974) showed that for A Cn C and n C, λ q C n n(C 1). There are a number of known constructions of low-coherence matrices when n < C. A primary class of examples is the random matrices with columns of unit norms that satisfy the Johnson-Lindenstrauss property (Johnson and Lindenstraus, 1984). For example, a Rademacher matrix has entries that are sampled i.i.d. from a uniform distribution on { 1 n, 1 n}. With high probability, a Rademacher random matrix of shape n C achieves a coherence λ q n for some constant c0 (Achlioptas, 2001). While random matrices can be easily obtained and have a low coherence in general, they require explicit storage (Nelson and Temlyakov, 2011) and can be outperformed in practical problems by some carefully crafted deterministic matrices (Naidu et al., 2016; Liu and Jia, 2020). Numerous deterministic constructions of low-coherence matrices have been proposed (Nelson and Temlyakov, 2011; Yu, 2011; Li et al., 2012; Xu, 2011; Naidu et al., 2016). In particular, Nelson and Temlyakov (2011) propose a deterministic construction that can achieve λ C 1 C, which avoids explicit storage of the matrix and can achieve a lower coherence in practice. There are also algorithms that directly optimize matrices for a smaller coherence (Wei et al., 2020; Abolghasemi et al., 2010; Obermeier and Martinez-Lorenzo, 2017; Li et al., 2013; Lu et al., 2018). 3.2 Problem statement To define the standard classification setting, denote the 0-1 loss L01 : Y Y R by L01(ˆy, y) = 1y =ˆy, where 1 is the indicator function. The risk of a classifier h is E[L01(h(X), Y )], and the goal of classification is to learn a classifier from training data whose risk is as close as possible to the Bayes risk minh H E[L01(h(X), Y )], where H = {measurable h : X Y}. We now describe an approach to classification based on label embedding, which represents labels as vectors in n < C dimensional complex space Cn. In particular, let G be an n C matrix with unit norm columns, called the embedding matrix, having coherence λG < 1. The columns of G are denoted by g1, g2, . . . , g C, and the column gi is used to embed the i-th label. Given an embedding matrix G, the original C-class classification problem may be reduced to a multi-output regression problem, where the classification instance (x, y) translates to the regression instance (x, gy). Given training data {(xi, yi)} for classification, we create training data {(xi, gyi)} for regression, and apply any algorithm for multi-output regression to learn a regression function f : X Cn. Published in Transactions on Machine Learning Research (09/2025) At test time, given a test point x, a label y is obtained by taking the nearest neighbor to f(x) among the columns of G. In particular, define the decoding function βG : Cn Y, βG(p) = min arg mini Y p gi 2 , where p represents the output of a regression model. (Since the arg min is potentially set-valued, the min breaks ties in favor of the label with the smallest index.) Then the label assigned to x is βG(f(x)). Thus, label embedding gives rise to classifiers of the form βG f, where f F = {all measurable f : X Cn}. Fortunately, according to the following result, no expressiveness is lost by considering classifiers of this form. Proposition 2. Recall the decoding function βG(p) = min arg mini Y p gi 2 , the regression models F = {all measurable f : X Cn}, and the classification models H = {all measurable h : X Y}. Then βG F = H. While the composition βG F preserves expressiveness, choosing a low-dimensional embedding in conjunction with a convex surrogate can introduce irreducible error (Ramaswamy and Agarwal, 2012). This and other results are proved in an appendix. It follows that minf F EP L01(βG(f(X)), Y ) is the Bayes risk for classification as defined earlier. This allows us to focus our attention on learning f : X Cn. Toward that end, we now formalize notions of loss and risk for the task of learning a multi-output function f for multiclass classification via label embedding. Definition 3. A loss function for label embedding is a function L : Cn Y R. Given such a loss function, define the L-risk of f with distribution P to be RL,P : F R, RL,P (f) := EP [L(f(X), Y )] and the L-Bayes risk to be R L,P := inff F RL,P (f). Using this notation, the target loss for learning with a fixed embedding matrix G is the loss function LG : Cn Y defined by LG(p, y) := L01(βG(p), y). By Prop. 2, the LG-risk of f is the usual classification risk of the associated classifier h = βG f, and the LG-Bayes risk is the Bayes risk of the original classification problem. The goal is to find a pair (f, G) that minimizes RLG,P (f), or in other words, a pair (f, G) that makes the excess target risk RLG,P R LG,P as small as possible. While the target loss LG defines the learning goal, it is not practical as a training objective because of its discrete nature. Therefore, for learning purposes, Hsu et al. (2009); Akata et al. (2013); Yu et al. (2014) suggest a surrogate loss, namely, the squared distance between f(x) and gy. More precisely, for a given embedding matrix G, define ℓG : Cn Y R as ℓG(p, y) := 1 2 p gy 2 2. This surrogate allows us to learn f by applying existing multi-output regression algorithms as we describe next. Thus, the label embedding learning problem is to learn f with small surrogate excess risk. Our subsequent analysis will connect RLG,P R LG,P to RℓG,P R ℓG,P . 3.3 Learning algorithms Learning algorithms for classification via label embedding, as described thus far, can be summarized by a conceptually simple meta-algorithm, depicted in Meta-Algorithm 1. This meta-algorithm should not be considered novel as its essential ingredients have been previously introduced Akata et al. (2013); Rodríguez et al. (2018), and several existing algorithms can be seen as instances (Hsu et al., 2009; Yu et al., 2014; Bhatia et al., 2015; Evron et al., 2018; Akata et al., 2013). The meta-algorithm takes as input a training dataset {(xi, yi)}N i=1, an embedding matrix G = [g1, g2, . . . , g C], and an algorithm A for multi-output regression. It forms the multi-output regression dataset {(xi, gyi)}N i=1, and applies A to produce a function f. The output is the classifier βG f. For example, the regression algorithm can be specified by selecting a model class F0 and a surrogate loss ℓG, and learning f by empirical risk minimization: min f F0 1 N i=1 ℓG(f(xi), yi). In our experiments we select F0 to be a neural network with n nodes in the output layer, and ℓG to be the squared error loss mentioned previously, the same surrogate analyzed in the next section. Published in Transactions on Machine Learning Research (09/2025) As a remark, we have treated G as a fixed input to the meta-algorithm, but it can also be trained jointly with f. Our analysis is independent of model training, and thus applies to this case as well. In the next section we present theory that supports selecting G with low coherence, which has not previously been examined in the label embedding literature. 3.4 Excess risk bound We present an excess risk bound, which relates the excess surrogate risk to the excess target risk. This bound justifies the use of the squared error surrogate, and also reveals a trade-off between the reduction in dimensionality (as reflected by λG) and the potential penalty in accuracy. To state the bound, define the class posterior η(x) = (η1(x), . . . , ηC(x)) where ηi(x) = PY |X=x(i). Define d(x) = maxi ηi(x) maxi/ arg maxj ηj(x) ηi(x), which is a measure of margin at x. We discuss this quantity further after the main result, which we now state. Theorem 4. Consider an embedding matrix G with unit norm columns g1, g2, . . . , g C and coherence λG = maxi =j| gi, gj |. Recall RLG,P and RℓG,P represent risks as defined in Definition 3, with R LG,P and R ℓG,P being the corresponding Bayes risks. Then for all f F, RLG,P (f) R LG,P inf r> 2λG 1 + λG PX (d(X) < r) (1 + λG)2 PX (d(X) < r)(RℓG,P (f) R ℓG,P ) (r(1 + λG) 2λG)2 RℓG,P (f) R ℓG,P As mentioned earlier, the goal of learning is to minimize the excess target risk RLG,P (f) R LG,P . The theorem shows that this goal can be achieved up to the first (irreducible) term by minimizing the excess surrogate risk RℓG,P (f) R ℓG,P . The excess surrogate risk can be driven to zero by any consistent algorithm for multi-output regression with squared error loss. We provide a formal definition of consistency and illustrate it with examples of both consistent and inconsistent loss functions in the appendix. The quantity d(x) can be viewed as a measure of noise (inherent in the joint distribution of (X, Y )) at a point x. While maxi ηi(x) represents the probability of the most likely label occurring, maxi/ arg maxj ηj(x) ηi(x) represents the probability of the second most likely label occurring. A large d(x) implies that arg maxi ηi(x) is, with high confidence, the correct prediction at x. In contrast, if d(x) is small, our confidence in predicting arg maxi ηi(x) is reduced, as the second most likely label has a similar probability of being correct. As pointed out by Ramaswamy and Agarwal (2012), any convex surrogate loss function operating on a dimension less than C 1 inevitably suffers an irreducible error. In the present setting, this irreducible error is manifested in the first term, which depends on the coherence λG of the embedding matrix. Given a classification problem with C classes, a larger embedding dimension n will lead to a smaller coherence λG, making d(x) > 2λG 1+λG on a larger region in X at the cost of increasing computational complexity. On the other hand, by choosing a smaller n, d(x) < 2λG 1+λG on a larger region in X, increasing the first term in Theorem 4. This interpretation highlights the balance between the benefits of dimensionality reduction and the potential impact on prediction accuracy, as a function of the coherence of the embedding matrix, λG, and the noisiness measure, d(x). 3.5 Improvement under low noise While Theorem 4 holds universally (for all distributions P), by considering a specific subset of distributions, we can derive a more conventional form of the excess risk bound. As a direct consequence of Theorem 4, under the multiclass extension of the Massart noise condition (Massart and Nédélec, 2006), which requires d(X) > c with probability 1 for some c, the first and second terms in Theorem 4 vanish. In this case, we Published in Transactions on Machine Learning Research (09/2025) recover a conventional excess risk bound, where RLG,P (f) R LG,P tends to 0 with RℓG,P (f) R ℓG,P . We now formalize this. Definition 5 (Multiclass Massart Noise Condition). The distribution P on X Y is said to satisfy the Multiclass Massart Noise Condition if and only if c > 0 such that PX (d(X) c) = 1. Corollary 6. Consider the same setup as in Theorem 4 and assume P satisfies the Multiclass Massart Noise Condition. If λG 0, ess inf d 2 ess inf d , then for all f F RLG,P (f) R LG,P 4 2λG ((1 + λG) ess inf d 2λG)2 RℓG,P (f) R ℓG,P , where ess inf d is the essential infimum of d, i.e., ess inf d = sup{a R : PX (d(X) < a) = 0}. For the special case where all labels are deterministic, we have ess inf d(x) = 1 for all x, leading to the simplified bound RLG,P (f) R LG,P 4 2λG (1 λG)2 (RℓG,P (f) R ℓG,P ). This observation implies that for deterministic labels, any embedding matrix with coherence less than 1 ensures consistency. Furthermore, a smaller coherence means faster convergence. 4 Experiments Table 2: Summary of the datasets used in the experiments. Ntrain is the number of training data points, Ntest the number of test data points, D the number of features, and C the number of classes. Dataset Ntrain Ntest D C LSHTC1 83805 5000 328282 12046 DMOZ 335068 38340 561127 11879 ODP 975936 493014 493014 103361 In this section, we present an experimental evaluation of our proposed method, LOCOLE (LOw COherence Label Embedding), for extreme multiclass classification. LOCOLE is an instance of Meta-Algorithm 1, as we explain below. 4.1 Experiment setup We conduct experiments on three large-scale datasets, DMOZ (Partalas et al., 2015), LSHTC1 (Partalas et al., 2015), and ODP (Bennett and Nguyen, 2009), which are extensively used for benchmarking extreme classification algorithms. The details of these datasets are provided in Table 2, with DMOZ and LSHTC1 available from (Yen et al., 2016)1, and ODP from (Medini et al., 2019). We apply LOCOLE where the multi-output regression algorithm is to train a multilayer perceptron with n output layer nodes using the surrogate loss ℓG. This is implemented using Py Torch, with a 2-layer fully connected neural network used for the LSHTC1 and DMOZ datasets and a 4-layer fully connected neural network for the ODP dataset. The hyperparameters are tuned on a held-out dataset. We experiment with the following types of embedding matrices: Rademacher: entries sampled i.i.d. from a uniform distribution on { 1 n, 1 n}. Gaussian: entries sampled i.i.d. from N(0, 1 n). Columns are normalized to have unit norm. C-Gaussian: the real and imaginary parts of each entry are sampled i.i.d. from N(0, 1 2n). Each column is normalized to have unit norm. 1https://people.csail.mit.edu/xrhuang/PDSparse/index.html Published in Transactions on Machine Learning Research (09/2025) Gaussian C-Gaussian Rademacher Nelson Cross Entropy Squared Loss Annex ML Parabel PD-Sparse PPD-sparse WLSTS 0 0.2 0.4 0.6 Figure 1: These plots reveal an inverse correlation between embedding matrix coherence and classification accuracy across different datasets, with coherence on the horizontal axis and accuracy on the vertical. Non-LOCOLE methods are plotted as horizontal lines because they do not depend on the embedding dimensionality. Nelson (Nelson and Temlyakov, 2011): a deterministic construction of low-coherence complex matrices in which n must be prime. If r is an integer, n > r is a prime number, and nr C, then the coherence of the constructed matrix is at most r 1 n . We choose r = 2 in experiments. We compare LOCOLE against the following state-of-the-art methods: PD-Sparse (Yen et al., 2016): an efficient solver designed to exploit the sparsity in extreme classification. PPD-Sparse (Yen et al., 2017a): a multi-process extension of PD-Sparse (Yen et al., 2016). Parabel (Prabhu et al., 2018): a tree-based method which builds a label-hierarchy. Annex ML (Tagami, 2017): a method which constructs a k-nearest neighbor graph of the label vectors and attempts to reproduce the graph structure from a lower-dimension feature space. WLSTS (Evron et al., 2018): a method based on error correcting output coding which embeds labels by codes induced by graphs. MLP (CE) Standard multilayer perceptron classifier with cross-entropy loss. MLP (SE): Standard multilayer perceptron classifier with squared error loss. For these methods, we use the hyperparameters suggested by their papers or accompanying code. While there are numerous label embedding algorithms (Hsu et al., 2009; Yu et al., 2014; Bhatia et al., 2015; Evron et al., 2018; Akata et al., 2013) which could be seen as specific subsets or instances of Meta-Algorithm 1, our comparisons will not include all of them. The approach in Akata et al. (2013) is not tailored for extreme classification and requires auxiliary information to construct the embedding matrices. On the other hand, methods like Yu et al. (2014); Bhatia et al. (2015); Hsu et al. (2009) have been outperformed significantly by the current state-of-the-art algorithms as shown in prior work (Yen et al., 2016; Bengio et al., 2010). Published in Transactions on Machine Learning Research (09/2025) Table 3: Accuracy on Various Datasets. RRM denotes Rademacher Random Matrices, GRM stands for Gaussian Random Matrices, and CGRM stands for Complex Gaussian Random Matrices. Methods without an embedding dimension ( -Dim ) are non-LOCOLE methods. We stop training when the training time reaches 50 hours and indicate this in the table. LSHTC1 DMOZ ODP Method(-Dim) Acc.(Mean Std) GRM 32 18.71 0.22% GRM 64 24.64 0.23% GRM 128 28.26 0.19% GRM 256 29.97 0.16% GRM 512 30.61 0.22% CGRM 32 24.77 0.31% CGRM 64 28.42 0.22% CGRM 128 29.97 0.21% CGRM 256 30.58 0.13% CGRM 512 30.98 0.14% RRM-32 18.66 0.20% RRM-64 24.77 0.17% RRM-128 28.30 0.27% RRM-256 30.06 0.21% RRM-512 30.66 0.22% Nelson-113 29.66 0.12% Nelson-127 29.71 0.13% Nelson-251 30.50 0.16% Nelson-509 31.08 0.15% MLP (CE) 26.30 0.36% MLP (SE) 28.03 0.17% Annex ML 29.06 0.35% Parabel 22.24 0.00% WLSTS 16.52 1.43% PDSparse 22.04 0.06% PPDSparse 22.60 0.11% Method(-Dim) Acc.(Mean Std) GRM 32 38.66 0.18% GRM 64 44.16 0.01% GRM 128 46.76 0.05% GRM 256 47.58 0.12% GRM 512 47.92 0.07% CGRM 32 44.11 0.04% CGRM 64 46.78 0.07% CGRM 128 47.51 0.09% CGRM 256 47.84 0.06% CGRM 512 47.90 0.08% RRM-32 38.41 0.17% RRM-64 44.15 0.05% RRM-128 46.60 0.04% RRM-256 47.67 0.09% RRM-512 47.82 0.06% Nelson-113 47.57 0.06% Nelson-127 47.71 0.04% Nelson-251 48.01 0.04% Nelson-509 48.02 0.06% MLP (CE) 46.71 0.06% MLP (SE) 38.38 0.14% Annex ML 39.82 0.14% Parabel 38.56 0.00% WLSTS 13.60 1.49% PDSparse 39.76 0.03% PPDSparse 39.30 0.07% Method(-Dim) Acc.(Mean Std) GRM 256 17.62 0.07% GRM 512 19.35 0.04% GRM 1024 20.77 0.03% GRM 2048 21.81 0.07% GRM 4096 22.20 0.04% CGRM 256 19.39 0.07% CGRM 512 20.74 0.08% CGRM 1024 21.78 0.04% CGRM 2048 22.14 0.06% CGRM 4096 21.90 0.07% RRM-256 17.63 0.04% RRM-512 19.36 0.04% RRM-1024 20.75 0.05% RRM-2048 21.81 0.06% RRM-4096 22.13 0.08% Nelson-331 20.14 0.06% Nelson-509 20.86 0.09% Nelson-1021 21.91 0.06% Nelson-2039 22.30 0.06% MLP (CE) 13.99 0.11% MLP (SE) 18.64 0.02% Annex ML 21.61 0.04% Parabel 17.09 0.00% WLSTS Train > 50 hrs PDSparse Train > 50 hrs PPDSparse 13.66 0.05% Table 4: Accuracy and training time across methods. See Section 4.3 for details. Dataset Metric Single Node Distributed PD-Sparse LOCOLE PPD-Sparse LOCOLE LSHTC1 Accuracy 22.04% 23.42% 22.60% 23.38% Training Time 230s 55s 135s 14s DMOZ Accuracy 39.76% 41.09% 39.30% 40.57% Training Time 829s 254s 656s 68s ODP Accuracy N/A 15.11% 13.66% 15.06% Training Time > 50 hrs 2045s 668s 350s While our theoretical framework is adaptable to any form of embedding, whether trained, fixed, or derived from auxiliary information, we focus on fixed embeddings in our empirical studies. This choice stems from the absence of a standardized or widely accepted methodology for jointly training the embedding function and the classifier. Although heuristic approaches exist, we are concerned that they may introduce confounding Published in Transactions on Machine Learning Research (09/2025) factors, complicating empirical interpretation. By centering on fixed embeddings, we ensure a controlled evaluation, minimizing such confounding factors and emphasizing the role of coherence of the embeddings. All neural network training is performed on a single NVIDIA A40 GPU with 48GB RAM. We explore different embedding dimensions and provide figures showing the relationship between the coherence of G and the accuracy. Full experimental details are presented in section B in the appendix. 4.2 Experimental results The experimental results in Table 3 highlight the superior performance of our proposed method across various datasets. In Table 3, the column Method (-Dim) denotes the method or embedding type along with its dimension, while the Acc. (Mean Std) column presents the mean and standard deviation of accuracy over 5 randomized repetitions. Methods without an embedding dimension (-Dim) are non-LOCOLE methods, which are unaffected by the embedding dimension. We highlight the best-performing method for each dataset. For each embedding method, there is a clear trend of increasing accuracy with larger embedding dimension. Ultimately, every embedding type outperforms all non-LOCOLE methods as the embedding dimension increases, with the Nelson embedding at dimension 509 achieving the highest accuracies across all datasets. We stop the training when the training time reaches 50 hours and indicate this in the table. We plot the accuracies as the coherence of the embedding matrix decreases in Figure 1 for the LSHTC1, DMOZ, and ODP datasets. Alongside, we include several baselines for comparison. Figure 1 demonstrates a negative correlation between the coherence of the embedding matrix and the accuracy, confirming our theoretical analysis. We include plots of coherence vs precision and recall in Section B of the appendix, along with low-dimensional projections of the embedding vectors. 4.3 Computational advantage PD-Sparse (Yen et al., 2016) and PPD-Sparse (Yen et al., 2017b) are among the most computationally efficient methods for training for extreme classification, to the best of our knowledge. PD-Sparse and PPD-Sparse both efficiently fit a linear model for classification with a multiclass hinge loss and elastic net regularization, which is regularization with both ℓ1 and ℓ2 penalties. For comparison, we apply LOCOLE using the Rademacher embedding to elastic net-regularized (Zou and Hastie, 2005) linear regression with ℓG loss. To compare the computational efficiency, we set the embedding dimension to n = 360. In our distributed implementation (multi-output linear regression can be trivially parallelized by solving multiple scalar-output linear regressions), each node independently solves a subset of elastic net linear regressions with scalar output, effectively spreading out the computation. In Table 4, we compare LOCOLE with Rademacher embedding with PD-sparse and PPD-sparse. LOCOLE clearly outperforms PD-sparse and PPD-sparse in both runtime and accuracy. We train the PD-Sparse method and Single-Node LOCOLE on Intel Xeon Gold 6154 processors, equipped with 36 cores and 180GB of memory. The distributed LOCOLE and the PPD-Sparse method also implemented in a distributed fashion are trained across 10 CPU nodes, harnessing 360 cores and 1.8TB of memory in total. 5 Conclusion and future work We provide a theoretical analysis for label embedding methods in the context of extreme multiclass classification. Our analysis confers a deeper understanding of the tradeoffs between dimensionality reduction and accuracy. We derive an excess risk bound that quantifies this tradeoff in terms of the coherence of the embedding matrix, and show that the statistical penalty for label embedding vanishes under the multiclass Massart condition. Through extensive experiments, we demonstrated that label embedding with low-coherence matrices outperforms existing techniques in both accuracy and runtime. While our analysis focuses on excess risk, the reduction of classification to regression means that existing generalization error bounds (for multi-output regression with squared error loss) can be applied to analyze the generalization error in our context. For example, Reeve and Kabán (2020) show that the generalization Published in Transactions on Machine Learning Research (09/2025) error grows with the dimension of the output space. This suggests that smaller embedding dimension leads to tighter control of the generalization error. Building on out theoretical framework, future work may consider extensions to multilabel classification, online learning, zero-shot learning, and learning with rejection. 6 Limitations Lossless label embedding relies on Massart s noise condition. Although Massart s condition is a well-recognized assumption in learning theory, it is important to note that it remains a theoretical construct that cannot be directly verified in most practical scenarios. This assumption facilitates theoretical analysis but may not always reflect real-world data distributions. Published in Transactions on Machine Learning Research (09/2025) Abolghasemi, V., Ferdowsi, S., Makkiabadi, B., and Sanei, S. (2010). On optimization of the measurement matrix for compressive sensing. In 2010 18th European Signal Processing Conference, pages 427 431. Achlioptas, D. (2001). Database-friendly random projections. In Proceedings of the Twentieth ACM SIGMODSIGACT-SIGART Symposium on Principles of Database Systems, PODS 01, page 274 281, New York, NY, USA. Association for Computing Machinery. Akata, Z., Perronnin, F., Harchaoui, Z., and Schmid, C. (2013). Label-embedding for attribute-based classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Babbar, R. and Schölkopf, B. (2017). Di SMEC: Distributed Sparse Machines for Extreme Multi-Label Classification. In Proceedings of the Tenth ACM International Conference on Web Search and Data Mining, WSDM 17, page 721 729, New York, NY, USA. Association for Computing Machinery. Babbar, R. and Schölkopf, B. (2019). Data scarcity, robustness and extreme multi-label classification. Machine Learning, 108. Bartlett, P. L., Jordan, M. I., and Mc Auliffe, J. D. (2006). Convexity, classification, and risk bounds. Journal of the American Statistical Association, 101(473):138 156. Bengio, S., Weston, J., and Grangier, D. (2010). Label embedding trees for large multi-class tasks. In Lafferty, J., Williams, C., Shawe-Taylor, J., Zemel, R., and Culotta, A., editors, Advances in Neural Information Processing Systems, volume 23. Curran Associates, Inc. Bennett, P. N. and Nguyen, N. (2009). Refined experts: improving classification in large taxonomies. In Proceedings of the 32nd International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 09, page 11 18, New York, NY, USA. Association for Computing Machinery. Bhatia, K., Jain, H., Kar, P., Varma, M., and Jain, P. (2015). Sparse Local Embeddings for Extreme Multi-label Classification. In Cortes, C., Lawrence, N., Lee, D., Sugiyama, M., and Garnett, R., editors, Advances in Neural Information Processing Systems, volume 28. Curran Associates, Inc. Chang, W.-C., Yu, H.-F., Zhong, K., Yang, Y., and Dhillon, I. S. (2020). Taming pretrained transformers for extreme multi-label text classification. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pages 3163 3171. Choromanska, A. E. and Langford, J. (2015). Logarithmic Time Online Multiclass prediction. In Cortes, C., Lawrence, N., Lee, D., Sugiyama, M., and Garnett, R., editors, Advances in Neural Information Processing Systems, volume 28. Curran Associates, Inc. Dahiya, K., Saini, D., Mittal, A., Dave, K., Jain, H., Agarwal, S., and Varma, M. (2021). Deep XML: A deep extreme multi-Label learning framework applied to short text documents. In ACM International Conference on Web Search and Data Mining. Deng, H., Han, J., Zhao, Z., and Liu, H. (2018). Dual principal component pursuit: Improved analysis and efficient algorithms. In Advances in Neural Information Processing Systems, pages 1514 1524. Evron, I., Moroshko, E., and Crammer, K. (2018). Efficient Loss-Based Decoding on Graphs for Extreme Classification. In Bengio, S., Wallach, H., Larochelle, H., Grauman, K., Cesa-Bianchi, N., and Garnett, R., editors, Advances in Neural Information Processing Systems, volume 31. Curran Associates, Inc. Guo, C., Mousavi, A., Wu, X., Holtmann-Rice, D. N., Kale, S., Reddi, S., and Kumar, S. (2019). Breaking the Glass Ceiling for Embedding-Based Classifiers for Large Output Spaces. In Wallach, H., Larochelle, H., Beygelzimer, A., d'Alché-Buc, F., Fox, E., and Garnett, R., editors, Advances in Neural Information Processing Systems, volume 32. Curran Associates, Inc. Published in Transactions on Machine Learning Research (09/2025) Gupta, N., Chen, P., Yu, H.-F., Hsieh, C.-J., and Dhillon, I. (2022). ELIAS: End-to-End Learning to Index and Search in Large Output Spaces. In Koyejo, S., Mohamed, S., Agarwal, A., Belgrave, D., Cho, K., and Oh, A., editors, Advances in Neural Information Processing Systems, volume 35, pages 19798 19809. Curran Associates, Inc. Hsu, D., Kakade, S. M., Langford, J., and Zhang, T. (2009). Multi-Label Prediction via Compressed Sensing. In Proceedings of the 22nd International Conference on Neural Information Processing Systems, NIPS 09, page 772 780, Red Hook, NY, USA. Curran Associates Inc. Jain, H., Balasubramanian, V., Chunduri, B., and Varma, M. (2019). Slice: Scalable Linear Extreme Classifiers trained on 100 Million Labels for Related Searches. In WSDM 19, February 11 15, 2019, Melbourne, VIC, Australia. ACM. Best Paper Award at WSDM 19. Jernite, Y., Choromanska, A., and Sontag, D. (2017). Simultaneous Learning of Trees and Representations for Extreme Classification and Density Estimation. In Precup, D. and Teh, Y. W., editors, Proceedings of the 34th International Conference on Machine Learning, volume 70 of Proceedings of Machine Learning Research, pages 1665 1674. PMLR. Jiang, T., Wang, D., Sun, L., Yang, H., Zhao, Z., and Zhuang, F. (2021). Light XML: Transformer with Dynamic Negative Sampling for High-Performance Extreme Multi-label Text Classification. Proceedings of the AAAI Conference on Artificial Intelligence, 35(9):7987 7994. Johnson, W. B. and Lindenstraus, J. (1984). Extensions of Lipschitz mappings into Hilbert space. Contemporary mathematics, 26:189 206. Khandagale, S., Xiao, H., and Babbar, R. (2020). Bonsai: Diverse and Shallow Trees for Extreme Multi-Label Classification. Mach. Learn., 109(11):2099 2119. Le, Q. and Mikolov, T. (2014). Distributed representations of sentences and documents. In International conference on machine learning, pages 1188 1196. PMLR. Li, G., Zhu, Z., Yang, D., Chang, L., and Bai, H. (2013). On projection matrix optimization for compressive sensing systems. IEEE Transactions on Signal Processing, 61(11):2887 2898. Li, S., Gao, F., Ge, G., and Zhang, S. (2012). Deterministic construction of compressed sensing matrices via algebraic curves. IEEE Transactions on Information Theory, 58(8):5035 5041. Liu, X. and Jia, L. (2020). Deterministic construction of compressed sensing matrices via vector spaces over finite fields. IEEE Access, 8:203301 203308. Lu, C., Li, H., and Lin, Z. (2018). Optimized projections for compressed sensing via direct mutual coherence minimization. Signal Processing, 151:45 55. Massart, P. and Nédélec, E. (2006). Risk Bounds for Statistical Learning. The Annals of Statistics, 34(5):2326 2366. Medini, T. K. R., Huang, Q., Wang, Y., Mohan, V., and Shrivastava, A. (2019). Extreme Classification in Log Memory using Count-Min Sketch: A Case Study of Amazon Search with 50M Products. In Wallach, H., Larochelle, H., Beygelzimer, A., d'Alché-Buc, F., Fox, E., and Garnett, R., editors, Advances in Neural Information Processing Systems, volume 32. Curran Associates, Inc. Mroueh, Y., Poggio, T., Rosasco, L., and Slotine, J.-j. (2012). Multiclass learning with simplex coding. In Pereira, F., Burges, C., Bottou, L., and Weinberger, K., editors, Advances in Neural Information Processing Systems, volume 25. Curran Associates, Inc. Naidu, R. R., Jampana, P., and Sastry, C. S. (2016). Deterministic compressed sensing matrices: Construction via euler squares and applications. IEEE Transactions on Signal Processing, 64(14):3566 3575. Nelson, J. and Temlyakov, V. (2011). On the size of incoherent systems. Journal of Approximation Theory, 163(9):1238 1245. Published in Transactions on Machine Learning Research (09/2025) Obermeier, R. and Martinez-Lorenzo, J. A. (2017). Sensing matrix design via mutual coherence minimization for electromagnetic compressive imaging applications. IEEE Transactions on Computational Imaging, 3(2):217 229. Partalas, I., Kosmopoulos, A., Baskiotis, N., Artieres, T., Paliouras, G., Gaussier, E., Androutsopoulos, I., Amini, M.-R., and Galinari, P. (2015). Lshtc: A benchmark for large-scale text classification. Prabhu, Y., Kag, A., Harsola, S., Agrawal, R., and Varma, M. (2018). Parabel: Partitioned Label Trees for Extreme Classification with Application to Dynamic Search Advertising. In ACM International WWW Conference, pages 993 1002. International World Wide Web Conferences Steering Committee, International World Wide Web Conferences Steering Committee. Prabhu, Y. and Varma, M. (2014). Fast XML: A Fast, Accurate and Stable Tree-Classifier for Extreme Multi-Label Learning. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD 14, page 263 272, New York, NY, USA. Association for Computing Machinery. Ramaswamy, H. G. and Agarwal, S. (2012). Classification calibration dimension for general multiclass losses. In Pereira, F., Burges, C., Bottou, L., and Weinberger, K., editors, Advances in Neural Information Processing Systems, volume 25. Curran Associates, Inc. Ramaswamy, H. G., Tewari, A., and Agarwal, S. (2018). Consistent algorithms for multiclass classification with an abstain option. Electronic Journal of Statistics, 12(1):530 554. Reeve, H. W. J. and Kabán, A. (2020). Optimistic bounds for multi-output prediction. In Proceedings of the 37th International Conference on Machine Learning, ICML 20. JMLR.org. Rodríguez, P., Bautista, M. A., Gonzàlez, J., and Escalera, S. (2018). Beyond one-hot encoding: Lower dimensional target embedding. Image and Vision Computing, 75:21 31. Steinwart, I. (2007). How to compare different loss functions and their risks. Constructive Approximation, 26:225 287. Tagami, Y. (2017). Annex ML: Approximate Nearest Neighbor Search for Extreme Multi-Label Classification. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD 17, page 455 464, New York, NY, USA. Association for Computing Machinery. Tewari, A. and Bartlett, P. (2005). On the validity of pairwise surrogate losses in kernel methods. In Advances in Neural Information Processing Systems 18, pages 1217 1224. Tewari, A. and Bartlett, P. L. (2007). On the consistency of multiclass classification methods. The Annals of Statistics, 35(1):605 636. Wang, W., Zheng, V. W., Yu, H., and Miao, C. (2019). A survey of zero-shot learning: Settings, methods, and applications. ACM Trans. Intell. Syst. Technol., 10(2). Wei, T., Mao, Z., Shi, J.-X., Li, Y.-F., and Zhang, M.-L. (2022). A survey on extreme multi-label learning. Wei, T., Tu, W.-W., and Li, Y.-F. (2019). Learning for Tail Label Data: A Label-Specific Feature Approach. In Proceedings of the 28th International Joint Conference on Artificial Intelligence, IJCAI 19, page 3842 3848. AAAI Press. Wei, Z., Zhang, J., Xu, Z., and Liu, Y. (2020). Measurement matrix optimization via mutual coherence minimization for compressively sensed signals reconstruction. Mathematical Problems in Engineering, 2020:7979606. Welch, L. (1974). Lower bounds on the maximum cross correlation of signals (corresp.). IEEE Transactions on Information Theory, 20(3):397 399. Published in Transactions on Machine Learning Research (09/2025) Weston, J., Bengio, S., and Usunier, N. (2010). Large scale image annotation: learning to rank with joint word-image embeddings. Machine Learning, 81(1):21 35. Weston, J., Chapelle, O., Elisseeff, A., Schölkopf, B., and Vapnik, V. (2002). Kernel dependency estimation. In Proceedings of the 15th International Conference on Neural Information Processing Systems, NIPS 02, page 897 904, Cambridge, MA, USA. MIT Press. Xu, Z. (2011). Deterministic sampling of sparse trigonometric polynomials. Journal of Complexity, 27(2):133 140. Yen, I. E., Huang, X., Dai, W., Ravikumar, P., Dhillon, I., and Xing, E. (2017a). PPDsparse: A Parallel Primal Dual Sparse Method for Extreme Classification. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD 17, page 545 553, New York, NY, USA. Association for Computing Machinery. Yen, I. E., Huang, X., Dai, W., Ravikumar, P., Dhillon, I., and Xing, E. (2017b). PPDsparse: A Parallel Primal Dual Sparse Method for Extreme Classification. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD 17, page 545 553, New York, NY, USA. Association for Computing Machinery. Yen, I. E.-H., Huang, X., Ravikumar, P., Zhong, K., and Dhillon, I. (2016). PD-Sparse : A Primal and Dual Sparse Approach to Extreme Multiclass and Multilabel Classification. In Balcan, M. F. and Weinberger, K. Q., editors, Proceedings of The 33rd International Conference on Machine Learning, volume 48 of Proceedings of Machine Learning Research, pages 3069 3077, New York, New York, USA. PMLR. You, R., Zhang, Z., Wang, Z., Dai, S., Mamitsuka, H., and Zhu, S. (2019). Attentionxml: Label tree-based attention-aware deep model for high-performance extreme multi-label text classification. Advances in Neural Information Processing Systems, 32. Yu, H.-F., Jain, P., Kar, P., and Dhillon, I. S. (2014). Large-scale multi-label learning with missing labels. In Proceedings of the 31st International Conference on International Conference on Machine Learning - Volume 32, ICML 14, page I 593 I 601. JMLR.org. Yu, H.-F., Zhong, K., Zhang, J., Chang, W.-C., and Dhillon, I. S. (2022). PECOS: Prediction for enormous and correlated output spaces. Journal of Machine Learning Research. Yu, N. Y. (2011). Deterministic compressed sensing matrices from multiplicative character sequences. In 2011 45th Annual Conference on Information Sciences and Systems, pages 1 5. Zhang, J., Chang, W.-C., Yu, H.-F., and Dhillon, I. (2021). Fast Multi-Resolution Transformer Fine-tuning for Extreme Multi-label Text Classification. In Ranzato, M., Beygelzimer, A., Dauphin, Y., Liang, P., and Vaughan, J. W., editors, Advances in Neural Information Processing Systems, volume 34, pages 7267 7280. Curran Associates, Inc. Zhang, T. (2004). Statistical behavior and consistency of classification methods based on convex risk minimization. The Annals of Statistics, 32(1):56 85. Zhou, D.-X., Tang, J., and Yang, Y. (2014). Large scale multi-label learning with missing labels. In International Conference on Machine Learning, pages 593 601. PMLR. Zou, H. and Hastie, T. (2005). Regularization and Variable Selection Via the Elastic Net. Journal of the Royal Statistical Society Series B: Statistical Methodology, 67(2):301 320. Ávila Pires, B., Szepesvari, C., and Ghavamzadeh, M. (2013). Cost-sensitive multiclass classification risk bounds. In Dasgupta, S. and Mc Allester, D., editors, Proceedings of the 30th International Conference on Machine Learning, volume 28 of Proceedings of Machine Learning Research, pages 1391 1399, Atlanta, Georgia, USA. PMLR. Published in Transactions on Machine Learning Research (09/2025) In this section, we present the proofs of the results in the main paper. A.1 Proof of proposition 2 Proof. This stems from the following facts: (i) for any given f F, the function βG f is also measurable, i.e., βG F H, (ii) for all h H, the function f(x) = gh(x) ensures that βG f = h, and (iii) f(x) = gh(x) is measurable as it only attains finite number of values. So (ii) and (iii) imply H βG F. A.2 Proof of the main theorem Recall that βG(p) = min arg mini Y p gi 2 . Lemma 7. p Cn, x X, C1,x(p) = maxi ηi(x) ηβG(p)(x). . Proof. Recall that LG(p, y) = L01(βG(p), y) = 1{min{arg mini Y p gi 2} =y}. The result follows from the observation that CLG,x(p) = 1 ηβG(p)(x) and C LG,x = 1 maxi ηi(x). Lemma 8. C2,x is strictly convex and minimized at p x = Gη(x) = PC i=1 ηi(x)gi. CℓG,x(p) = Ey PY |X=xℓ2(p, gy) i=1 ηi(x) p gi 2 2 i=1 ηi(x) p gi, p gi i=1 ηi(x)gi, p i=1 ηi(x) gi 2 2 i=1 ηi(x)gi i=1 ηi(x) gi 2 2 1 i=1 ηi(x)gi So C2,x(p) is strictly convex and minimized at p x = PC i=1 ηi(x)gi = Gη(x). We ll use p x to denote Gη(x) henceforward. Lemma 9. Let V be a normed space with norm V . Let u : V R be a strictly convex function. Let u be minimized at x with u(x ) = 0. x V , δ0 > 0, if u(x) < δ := infx: x x V =δ0 u(x), then x x V < δ0. Proof. We first confirm the fact that t (0, 1) and q V {0}, u(x + tq) < u(x + q). u(x + tq) = u((1 t)x + t(x + q)) < (1 t)u(x ) + tu(x + q) = tu(x + q) < u(x + q). Assume the opposite: For some u V , δ0 > 0, u(x) < δ = infx: x x V =δ0 u(x) and x x V δ0. Then u(x) = u(x + (x x )) u(x + δ0 x x V (x x )) δ, which results in a contradiction. Published in Transactions on Machine Learning Research (09/2025) Lemma 10. Fix x X. δ0 > 0 p Cn, C2,x(p) < 1 2δ2 0 = p p x < δ0. Proof. Applying Lemma 9 with u = C2,x, V = Cn, V = 2, and x = p x, we have p Cn, C2,x(p) < infs: s p x =δ0 C2,x(s) = p p x 2 < δ0. Furthermore, inf s: s p x =δ0 C2,x(s) = inf s: s p x =δ0 2 s, s Re p x, s 1 2 p x, p x Re p x, p x = inf s: s p x =δ0 1 2 s p x 2 2 To facilitate our proofs, we denote λj,k = gj, gk , λRe j,k = Reλj,k, and λG = maxj =k|λj,k|. Lemma 11. x X, j, k [C], p Cn, (1 + λG)(ηk(x) ηj(x)) 2λG 2 2λG > p x p 2 = p gj 2 > p gk 2. Proof. We first consider the general case when p = p x. Write p = p x + δ0v where v = p p x p p x 2 and δ0 = p p x 2. Recall that p x = Gη(x). Note the inequality immediately implies ηk(x) > ηj(x). (1 + λG)(ηk(x) ηj(x)) 2λG = (1 λG)(ηk(x) ηj(x)) 2λG(1 (ηk(x) ηj(x))) > δ0 p = (1 λG)(ηk(x) ηj(x)) 2λG(1 ηk(x) ηj(x)) > δ0 p 1 λG(ηk(x) ηj(x)) 2λG 1 λG (1 ηk(x) ηj(x)) > 1 λRe j,k(ηk(x) ηj(x)) + 1 q i =j,k (λRe i,k λRe i,j )ηi(x) > = (1 λRe j,k)(ηk(x) ηj(x)) + X i =j,k (λRe i,k λRe i,j )ηi(x) > δ0 q = Re p , gk gj > δ0 gj gk 2 (2) = Re p , gk gj > Re δ0v, gj gk (3) = Re p + δ0v, gk > Re p + δ0v, gj = p gj 2 2 > p gk 2 2 (4) Inequality (1) follows the fact that i, i , λRe i,i λG. Inequality (2) follows from p x = PC i=1 ηi(x)gi and gj gk 2 = q 2 2λRe j,k. Inequality (3) is implied by the Cauchy-Schwarz inequality. In the last inequality (4), we use the fact that gj 2 = gk 2 = 1. Published in Transactions on Machine Learning Research (09/2025) Now let p = p x. Let (1+λG)(ηk(x) ηj(x)) 2λG p gj 2 2 p gk 2 2 =2Re p x, gk gj =2(1 λRe j,k)(ηk(x) ηj(x)) + 2 X i =j,k (λRe i,k λRe i,j )ηi(x) 2(1 λG)(ηk(x) ηj(x)) 4λG(1 ηk(x) ηj(x)) 2(1 λG)(ηk(x) ηj(x)) 4λG(1 (ηk(x) ηj(x))) =2(1 + λG)(ηk(x) ηj(x)) 4λG > 0 Lemma 12. x X, r > 2λG 1+λG , and p Cn, C2,x(p) < ((1+λG)r 2λG) 2 4 2λG = C1,x(p) < r. Proof. By Lemma 10, C2,x(p) < ((1+λG)r 2λG) 2 4 2λG = p p x 2 < (1+λG)r 2λG 2 λG . Fix x X. Recall that βG(p) = min arg mini Y p gi 2 . We claim p p x < (1 + λG)r 2λG 2 λG = max i ηi(x) ηβG(p)(x) < r. Assume p p x < (1+λG)r 2λG 2 λG and maxi ηi(x) ηβG(p)(x) r. By Lemma 11, p gβG(p) > p gmin{arg maxi ηi(x)} , contradicting the definition of βG(p). Hence, C1,x(p) = maxi ηi(x) ηβG(p)(x) < r. Now we re ready to prove Theorem 4. Proof of Theorem 4. RLG,P (f) R LG,P = Z X C1,x(f(x)) x:d(x) 2λG 1+λG , and p Cn, C1,x(p) r = C2,x(p) (1 + λG)r 2λG 2 4 2λG . (5) C1,x(p) > 2λG 1 + λG = C2,x(p) (1 + λG)C1,x(p) 2λG 2 = C1,x(p) 2λG 1 + λG + 1 1 + λG (4 2λG)C2,x(p). Note the last inequality actually holds for all p Cn, that is, it holds even when C1,x(p) 2λG 1+λG . Then, Published in Transactions on Machine Learning Research (09/2025) x:d(x) 0, C1,x(p) = maxi ηi(x) ηβG(p)(x) maxi ηi(x) maxi/ arg maxj ηj(x) ηi(x) = d(x). By (5), if d(x) r and C1,x(p) > 0, then C2,x(p) ((1+λG)r 2λG) 2 4 2λG . As C1,x(p) [0, 1], d(x) r and C1,x(p) > 0 = C2,x(p) ((1+λG)r 2λG) 2 4 2λG C1,x(p). It is trivial that C2,x(p) ((1+λG)r 2λG) 2 4 2λG C1,x(p) also holds when C1,x(p) = 0. Thus, x X and p Cn, d(x) r = C2,x(p) ((1+λG)r 2λG) 2 4 2λG C1,x(p) = C1,x(p) ((1+λG)r 2λG)2 C2,x(p). Therefore, x:d(x) r C1,x(f(x)) Z ((1 + λG)r 2λG)2 C2,x(f(x)) ((1 + λG)r 2λG)2 X C2,x(f(x)) ((1 + λG)r 2λG)2 RℓG,P (f) R ℓG,P . B Experiment details In this section, we provide the details of our experiments. B.1 Neural networks In this section, we provide details on the architectures and hyperparameter choices for the neural networks used in our experiments. The architectures and hyperparameters are selected by trial-and-error on a heldout dataset. B.1.1 LSHTC1 The proposed embedding strategy adopts a 2-layer neural network architecture, employing a hidden layer of 4096 neurons with Re LU activation. The output of the neural network is normalized to have a Euclidean norm of 1. An Adamax optimizer with a learning rate of 0.001 is utilized together with a batch size of 128 for training. The model is trained for a total of 5 epochs. In order to effectively manage the learning rate, a scheduler is deployed, which scales down the learning rate by a factor of 0.1 at the second epoch. Published in Transactions on Machine Learning Research (09/2025) Our cross-entropy baseline retains a similar network architecture to the embedding strategy, with an adjustment in the output layer to reflect the number of classes. Here, the learning rate is 0.01 and the batch size is 128 for training. The model undergoes training for a total of 5 epochs, with a scheduler set to decrease the learning rate after the third epoch. Finally, the squared loss baseline follows the architecture of our cross-entropy baseline, with the learning rate and batch size mirroring the parameters of the embedding strategy. As with the embedding strategy, the output is normalized. The model is trained for a total of 5 epochs, and the learning rate is scheduled to decrease after the third epoch. For the DMOZ dataset, our proposed label embedding strategy employed a 2-layer neural network with a hidden layer of 2500 neurons activated by the Re LU function. The output of the network is normalized to have a norm of 1. We trained the model using the Adamax optimizer with a learning rate of 0.001 and a batch size of 256. The model was trained for 5 epochs, and the learning rate was scheduled to decrease at the second and fourth epochs by a factor of 0.1. For the cross-entropy loss baseline, we used the same network architecture with an adjustment in the output layer to match the number of classes. The learning rate was 0.01 and the batch size was 256. The model underwent training for a total of 5 epochs, with the learning rate decreasing after the third epoch as determined by the scheduler. Lastly, the squared loss baseline utilized the same architecture, learning rate, and batch size as the proposed label embedding strategy. Similarly, the model s output was normalized. The model was trained for 5 epochs, with the learning rate scheduled to decrease after the third epoch. For the ODP dataset, the experiments utilized a neural network model composed of 4 layers. The size of the hidden layers progressively increased from 210 to 214, then decreased to 213. Each of these layers employed the Re LU activation function and was followed by batch normalization to promote faster, more stable convergence. The final layer output size corresponded with the embedding dimension for the label embedding strategy and the number of classes for the cross-entropy and squared loss baselines. In the label embedding framework, the output was normalized to yield a norm of 1. This model was trained using the Adamax optimizer, a learning rate of 0.001, and a batch size of 2048. The training spanned 20 epochs, with a learning rate decrease scheduled after the 10th epoch by a factor of 0.1. For the cross-entropy loss baseline, the same network architecture was preserved, with an adjustment to the penultimate layer, reduced by half, and the final output layer resized to match the number of classes. This slight modification in the penultimate layer was necessary to accommodate the models within the 48GB GPU memory. Notably, the neural network output was normalized by dividing each output vector by its Euclidean norm before applying the softmax function, a non-standard operation that significantly enhanced performance. This model was trained using a learning rate of 0.01 over 20 epochs, following a similar learning rate schedule. Finally, the squared loss baseline used the same architecture as the cross-entropy baseline and the same learning rate and batch size as the label embedding model. Here, the output was also normalized. The model underwent training for 20 epochs, with a learning rate decrease scheduled after the 10th epoch. B.2 Elastic net We aim to solve min W RD n XW Y 2 fro + λ1 W 1,1 + λ2 W 2 fro, (7) where X RN D is the data matrix, the rows of Y RN n are embedding vectors. Published in Transactions on Machine Learning Research (09/2025) The problem (7) can be broken down into n independent real-output regression problems of the form min Wj RD XWj Yj 2 fro + λ1 Wj 1 + λ2 Wj 2 2, where Wj is the j-th column of W and Yj is the j-th column of Y . Consequently, We can distribute the n real-output regression problems across multiple cores. We develop a regression variant of the Fully-Corrective Block-Coordinate Frank-Wolfe (FC-BCFW) algorithm (Yen et al., 2016) and use it to solve the real-output regression problems. As the solver operates iteratively, we set it to run for a predefined number of iterations, denoted as Niter. The chosen hyperparameters are outlined in table 5. Dataset λ1 λ2 Niter LSHTC1 0.1 1 20 DMOZ 0.01 0.01 20 ODP 0.01 0.1 40 Table 5: Hyperparameters for elastic net. B.3 Practical issues B.3.1 Choice of embedding dimension In practical settings, the choice of the embedding dimension n depends on available computational resources. Under a fixed computational budget, our theory recommends opting for an embedding with minimal coherence. When the type of embedding is fixed but computational resources are flexible, we advise treating the embedding dimension n as a tunable parameter. Specifically, it is beneficial to incrementally increase n, perhaps exponentially, and observe the performance. This process should continue until the improvement in performance plateaus or the additional computational cost becomes prohibitively high. B.3.2 Types of embedding matrices In our experiments, we used four types of embeddings. Random embeddings stand out for their simplicity and flexibility. They can be generated with a single line of code, and the embedding dimension can be any positive integer. However, their drawbacks include the need for explicit storage and potentially higher coherence compared to some deterministic matrices. On the other hand, deterministic matrices like those constructed using Nelson s method can achieve lower coherence, generally leading to better performance. They do not require explicit storage, as individual columns can be generated on demand. The trade-off is that they are more complex to implement, and the options for embedding dimensions are more limited (e.g., Nelson s construction requires prime numbers for the embedding dimension). B.4 Used assets We list the existing code used in our experiments. PD-Sparse (Yen et al., 2016): https://github.com/a061105/Extreme Multiclass (BSD-3-Clause license). PPD-Sparse (Yen et al., 2017a): https://github.com/a061105/Async PDSparse. Parabel (Prabhu et al., 2018): http://manikvarma.org/code/Parabel/download.html. Annex ML (Tagami, 2017): https://github.com/yahoojapan/Annex ML?tab=Apache-2. 0-1-ov-file(Apache-2.0 license). WLSTS(Evron et al., 2018): https://github.com/ievron/wltls/?tab=MIT-1-ov-file(MIT License). Published in Transactions on Machine Learning Research (09/2025) B.5 Visualization of Embeddings (a) t-SNE plots of 2,048 Rademacher embedding vectors in R128. (b) t-SNE plots of 2,048 Gaussian embedding vectors in R128. (c) t-SNE plots of 2,048 complex Guassian embedding vectors in R128. (d) t-SNE plots of 2,048 Nelson s embedding vectors in R127. Figure 2: t-SNE plots illustrating the distribution of vectors from four embedding types. To better understand the structural differences among various embedding types, we visualize 2,048 embedding vectors from each method using t-SNE, as shown in Figure 2. The embeddings include Rademacher, Gaussian, complex Gaussian, and Nelson s construction. The random embeddings all exhibit roughly isotropic spreads in the low-dimensional projection. In contrast, the Nelson embeddings display a more organized and geometrically constrained distribution. Published in Transactions on Machine Learning Research (09/2025) Macro Precision Rademacher Nelson Micro Precision Macro Recall Micro Recall Figure 3: Coherence vs metric score on LSHTC1 comparing Rademacher and Nelson embeddings across precision and recall (macro and micro). B.6 Precision and recall For a set of C classes, let TPc, FPc, FNc denote true positives, false positives, and false negatives for class c. Then Precisionc = TPc TPc + FPc , Recallc = TPc TPc + FNc . Macro-averages compute the unweighted mean across classes, Macro-Precision = 1 c=1 Precisionc, Macro-Recall = 1 c=1 Recallc, while micro-averages pool counts across classes before computing, Micro-Precision = PC c=1 TPc PC c=1(TPc + FPc) , Micro-Recall = PC c=1 TPc PC c=1(TPc + FNc) . To assess whether the benefits of low coherence extend beyond classification accuracy, we include additional plots showing four other performance metrics: macro-precision, macro-recall, micro-precision, and micro-recall. Figure 3 shows how these metrics vary with the coherence of the embedding matrix for LSHTC1 using Rademacher and Nelson constructions. Consistent with the accuracy trends reported in the main paper, we observe a clear inverse relationship between coherence and all measures. As coherence decreases, moving rightward along the reversed x-axis, all four scores consistently improve. C Extended discussion In this section, we provide supplementary insights and theoretical context that extend the main exposition. C.1 Consistent loss functions Let (X, Y ) P with Y {1, . . . , K} and a function f : X RK. For a surrogate ϕ : RK {1, . . . , K} R define the risk Rϕ(f) = E ϕ f(X), Y . We say ϕ is (Fisher) consistent w.r.t. the 0 1 loss ℓif every minimiser f arg minf Rϕ(f) induces a Bayes optimal classifier g (x) = arg maxk f k(x), i.e. inff Rϕ(f) = Rϕ(f ) infg Rℓ(g) = Rℓ(g ) (Zhang, 2004; Tewari and Bartlett, 2007; Steinwart, 2007). Canonical consistent surrogates include the mean-squared error ϕMSE(f, y) = ey f 2 2 and the multinomial logistic (cross-entropy) loss ϕlog(f, y) = log exp(fy) P k exp(fk). In contrast, margin-based hinge variants such as ϕhinge(f, y) = maxj =y 1 + + are known to be inconsistent in the multiclass setting (Tewari and Bartlett, 2005). Published in Transactions on Machine Learning Research (09/2025) C.2 Nelson s construction Let r 2 be a natural number and p > r be a prime number. Define an index list a := (a1, . . . , ar) where the ai s are integers in [1, n]. For each a, we define The n dimension embedding is constructed as n F(a, 1) , . . . , exp 2πi n F(a, n) T . Given n and r, each embedding vector is uniquely determined by a, yielding an embedding capacity of nr. The magnitude of the inner product between ga and ga is bounded by r 1 n when a = a (Nelson and Temlyakov, 2011). C.3 Dynamically growing label space In scenarios with dynamically growing label spaces, random embedding matrices present no significant challenges: as new labels arrive, we can simply sample their embeddings independently and identically distributed (i.i.d.). For deterministic constructions, the capacity depends on the construction algorithm of the embedding matrices. In Nelson s framework, the number of available low-coherence embeddings scales as nr, allowing expansion to a polynomially large number of labels while maintaining desirable geometric properties. However, once the (nr 1) cap is reached, the labels have to be rehashed .