# decodingbased_regression__6a187712.pdf Published in Transactions on Machine Learning Research (08/2025) Decoding-based Regression Xingyou Song , Dara Bahri Google Deep Mind Equal Contribution. Reviewed on Open Review: https://openreview.net/forum?id=av UQ8jguxg Language models have recently been shown capable of performing regression wherein numeric predictions are represented as decoded strings. In this work, we provide theoretical grounds for this capability and furthermore investigate the utility of causal sequence decoding models as numeric regression heads given any feature representation. We find that, despite being trained in the usual way - for next-token prediction via cross-entropy loss - decoder-based heads are as performant as standard pointwise heads when benchmarked over standard regression tasks, while being flexible enough to capture smooth numeric distributions, such as in the task of density estimation. 1 Introduction 1111 1001 0011 Regression Problem Figure 1: Given any feature representation ϕ(x), we attach a decoding-based head to output predictive distribution pθ(y|x). Despite being originally developed for the purpose of text generation and chat applications, large language models (LLMs) have recently been applied for new applications, particularly one of which is regression, and more broadly the prediction of numeric outcomes. Vacareanu et al. (2024) have shown service-based LLMs such as Chat GPT and Gemini capable of regression with performance comparable to that of traditional regression methods such as random forests, while Song et al. (2024); Akhauri et al. (2025) have shown smaller custom language models can be trained specifically on multiple regression tasks for transfer learning. For an input-output pair (x, y), where x is a feature vector and y is a real number, a regression model s performance is determined by two factors: how it processes x and how its output head represents y. While the mentioned works (Vacareanu et al., 2024; Song et al., 2024; Akhauri et al., 2025) can be seen as text-to-text regression where both x and y are represented as tokens, this combination is not necessarily required. Tang et al. (2024) investigate the isolated case where LLM embeddings of x are attached to feed-forward networks as regression heads, while Nguyen et al. (2024) investigate the case when these embeddings are eventually attached to Gaussian distribution heads. Both can be seen as particular instances when x is represented as text, while common regression heads are still used. However, there has not been work investigating the inverse situation, i.e. y is represented as text or structured tokens. One could do so by using decoding-based regression heads, where for example, tokens <1><.><2><3> can be decoded to represent 1.23, a technique used in several works Published in Transactions on Machine Learning Research (08/2025) training language models for specific numeric tasks, such as arithmetic (Nogueira et al., 2021), linear algebra (Charton, 2022), and symbolic regression (d Ascoli et al., 2022). In contrast to traditional deterministic or parametric distribution (e.g. Gaussian) regression heads, decodingbased heads may be much more flexible, as they can represent any numeric distribution approximately over R without the need for explicit normalization. However, due to their initial lack of inductive bias over numeric distances, numeric token representations and sequential dependencies may need to be learned using additional training data, and thus it is worth empirically investigating these trade-offs in isolated, controlled experiments. Our research provides valuable insights on using numbers as an output modality of token-based autoregressive decoders. Our specific contributions and findings are thus as follows: We formalize decoding-based regression, i.e. explicitly define tokenization schemes for numbers, establish training and inference procedures, discuss methods for pointwise estimation, and theoretically provide risk guarantees for density estimation under common assumptions. In experimental benchmarks, we find that properly tuned decoding-based regression heads are dataefficient and competitive with regular pointwise heads on tabular regression, yet are also expressive enough to perform against Gaussian mixture heads for density estimation. 2 Related Work and Motivation The idea of text-to-text regression is especially relevant as LLMs are currently being fine-tuned as Generative Reward Models (Mahan et al., 2024; Zhang et al., 2024), i.e. end-to-end scoring methods for reinforcement learning feedback (Ziegler et al., 2019; Bai et al., 2022) or LLM-as-a-Judge (Zheng et al., 2023). Such reward modeling methods can be simpler than other forms such as Bradley-Terry (Bradley & Terry, 1952) which requires appending additional prediction heads and custom losses. However, little analysis has been done in isolation on the theoretical and modeling capabilities of using text, or more generally tokens, to represent real numbers. Understandably, one could argue that regular supervised fine-tuning over numbers represented as strings is unprincipled, considering that there is no notion of numeric distance when using cross-entropy loss. However, we argue that token-based numeric modeling is actually natural, given observed phenomena and techniques proposed in recent works. Given a post-processed representation ϕ(x) Rd after x is sent through a task-specific encoder (MLP, CNN, etc.), we provide an overview of common regression heads pθ(y|ϕ(x)) with trainable parameters θ, which can be applied on top of ϕ(x) to return numeric outputs (additional techniques in Section 5). Pointwise Heads: By far, the most commonly used regression head is a learnable deterministic function with weights θ, typically a simple feed-forward network (often a single linear layer) mapping ϕ(x) to a single-point scalar, trained by minimizing a pointwise loss like mean squared error. To allow stability during training, the y-values must be normalized space, e.g. within [0, 1]. Parameteric Distribution Heads: In the case of probabilistic outputs, one may apply distributions with parametric forms. A common example is a Gaussian head, e.g. pθ(y|ϕ(x)) = N(µ, σ2) where µ, σ are deterministic learnable functions of ϕ(x). A more flexible variant is a finite mixture of Gaussians (Bishop, 1994), which can be extended to infinite mixtures (Rasmussen, 1999). Such mixture techniques can be more broadly seen within the realm of density estimation (Parzen, 1962; Rosenblatt, 1956) in which a complex distribution may be estimated using multiple simpler basis distributions. Histogram (Riemann) Distribution Heads: One such basis common in deep learning applications is the piece-wise constant basis, for learning histograms over a finite support set {y1, . . . , yn} R via softmax parametrization, i.e. pθ(yi|ϕ(x)) = Softmax(i)(ϕ(x)T θ) where θ Rn, which has led to strong results in value-based reinforcement learning (Imani & White, 2018; Bellemare et al., 2017) and tabular data (Hollmann et al., 2025; Chen et al., 2022). However, a drawback is that learning numeric distances between all of the bins requires more data as the size of the vocabulary increases. We term these as Riemann heads, following (Hollmann et al., 2025). Published in Transactions on Machine Learning Research (08/2025) While there have been works on ordinal regression to learn rankings among these bins, such as using rankconsistency (Cao et al., 2020) and soft labels / metric-awareness (Diaz & Marathe, 2019), we propose a much simpler way, by simply considering the histogram distribution as a special case of decoding a sequence of length 1. By extending the sequence length instead, there can be an exponential reduction in bin count e.g. 1000 (=103) bins can be expressed instead using 10 bins and 3 decoding steps. While this intuitive idea has been studied in extreme classification problems (Wydmuch et al., 2018), it has not been thoroughly examined for numeric regression, which is the focus of our work. Below, we introduce decoding-based regression heads. Note: To prevent confusion with the term decoder which is also a central component of generative models like Variational Autoencoders (VAEs) (Kingma & Welling, 2014), we stress a key distinction. While both VAE decoders and distributional regression heads map a feature vector to a probability distribution, their objectives differ: a VAE decoder models p(x|ϕ(x)) to reconstruct the input x from a latent ϕ(x), whereas a regression head models p(y|ϕ(x)) to predict a separate target y. Given this difference in the output space, we avoid referring to standard distributional heads (e.g., Gaussian) as decoders . 3 Decoding-Based Regression In this work, we define the decoder head as an autoregressive sequence model, such as a compact Transformer decoder. Given a vocabulary V, the Transformer takes the feature vector ϕ(x) as its initial context, and generates a discrete sequence of tokens (t1, . . . , t K) VK one token at a time. This sequence is a string representation of a number, and by modeling the probability pθ(t1, . . . , t K | ϕ(x)) = QK k=1 pθ(tk | ϕ(x), t1 . . . tk 1), the head implicitly defines a probability distribution over a discrete set of representable numbers. Below, we discuss natural token representations of numbers. 3.1 Numeric Token Representations Normalized Tokenization: If y is restricted to [0, 1] (via scale normalization for example), then in Section 3.3 we show any smooth density p(y|ϕ(x)) can be approximated with an increasing level of granularity as more tokens are used in the numeric representation, under some universality assumptions on pθ. This can be seen intuitively with a tree-based construction, i.e. for a base B, the vocabulary contains <0>, <1>, . . . , , and y is simply represented by its base-B expansion up to a length K. This setting aligns with standard data-science practices of also normalizing y-values according to training data or known bounds. Unnormalized Tokenization: However, there are cases in which we would like to use an unnormalized tokenization scheme. Such cases include multi-task regression (Song et al., 2024), in which different tasks may have varying y-scales, or express very wide y-ranges for which appropriately normalizing y-values for the correct balance between numeric stability and expressiveness would be very tedious. In this case, we may simply view normalized tokenizations as mantissas and then left-append sign and exponent tokens to form a base-B generalization of the common IEEE-754 floating point representation (IEEE, 2019). Given length parameters E and M, our tokenization is therefore . . . . . . where se, e1, e2, . . . , e E are the sign and base-B representation of the exponent and m1, m2, . . . , m M are the most significant base-B digits of the mantissa. E.g. if (B=10, E=3, M=4), then 10 222 1.23456789 will be represented as <+><-><2><2><2><1><2><3><4>. Signs , can have their own dedicated <->, <+> tokens or optionally reuse the <0>,<1> tokens from V; this made little difference in results. Note that the vocabulary can additionally contain special tokens for representing outputs not within a supported numeric range. For example, one can use a token to represent non-numbers, commonly used in cases where x may be an invalid input. We mention such cases for useful downstream applications, although the scope of this paper assumes y is always within R. Architecture: Any autoregressive model can be used, so long as it supports constrained token decoding to enforce proper sequences which represent a valid number. By default, we use a small Transformer (Vaswani et al., 2017) due to its strong autoregression capabilities, with the initial token embedding as ϕ(x). As we show in our experiment section, this Transformer size can be made minimal, with negligible contribution to parameter count in comparison to the encoder. Published in Transactions on Machine Learning Research (08/2025) Since the token space is finite while R is uncountable, this mapping is lossy (i.e. not invertible) and introduces a notion of rounding error. However, for large enough base B and sequence lengths (both normalized and unnormalized), practically any y-value will be within the expressible range and rounding errors will be minimal. The trade-off is that the vocabulary size and sequential dependencies between tokens will also increase, and learning better numeric representations may thus require more training data. While it s possible to first pretrain these numeric representations as in Hollmann et al. (2025) for the histogram distribution, we show that with proper hyperparameter tuning, the Transformer decoder can be used out-of-the-box as a randomly initialized regression head. 3.2 Pointwise Estimation In many cases, one may only be interested in a scalar quantity of interest M(pθ) of the model s distribution (e.g. its mean). If pθ matches the true distribution p perfectly, then for a given pointwise loss ℓ: R2 R the goal is then to select M(p) which minimizes Ey p( |x) [ℓ(M(p), y)]. It is well established that for common error functions like L2, L1, and L0, the optimal values are the mean, median, and mode of p( |x), respectively. Lukasik et al. (2024) also leverage this observation to enhance LLM decoding. To estimate these M(p), the mode can be approximated using e.g. beam search (Graves, 2012), but efficiently estimating other common general pointwise representatives M(p) from pure temperature samples is a broad topic - for example, one can efficiently approximate the true median from the Harrell-Davis estimator (Harrell & Davis, 1982), and more generally we refer the reader to Lehmann (1983) on statistical point estimators. Especially for unnormalized tokenization, additional care needs to be taken, since in practice, the model can have a miniscule but non-zero probability of decoding an arbitrarily large outlier, even if the underlying true distribution is bounded. Such outliers can easily sway non-robust estimators such as the sample mean, as observed in Song et al. (2024). This issue fundamentally comes from the fact that some tokens are more significant than others, prompting the use of alternative tokenizations based on coding theory which are robust to corruptions, which we show can be effective in our experiment section. Alternatively, decoding techniques from the LLM literature can also be used, e.g. top-k (Fan et al., 2018) or top-p (Holtzman et al., 2020), or even simply decreasing the temperature to increase model confidence and thereby filter out possible outliers. One can also avoid decoding altogether and use recently proposed RAFT (Lukasik et al., 2025; Chiang et al., 2025) which estimates M(p) using a query-based approach using a finite fixed evaluation set Y, e.g. for mean, Ey pθ[y] 1 N P y Y pθ(y ) y , although the choice of Y may be non-trivial to obtain an unbiased estimate, especially over unnormalized tokenizations. This may also defeat the purpose of using a decoding head, which offers several density estimation benefits, as we discuss below. Overall, the choice of method for computing pointwise representations we leave as a hyperparameter to be tuned depending on the application. 3.3 Density Estimation and Theory During training, to allow the full probabilistic modeling benefits of using a decoding head, we apply the standard cross-entropy loss over all sequence tokens. For a model pθ and target y = (t1, . . . , t K), the crossentropy loss (omitting x to simplify notation) will be: btk V 1(btk = tk) log pθ(btk|t1, . . . , tk 1) The expected loss over all y sampled from the true distribution is then Ey p [H(y, pθ)]. Given our tree-based tokenization and training loss, we provide formal guarantees for estimating onedimensional densities on [0, 1]. Note that densities with finite support can be shifted and rescaled to have support in [0, 1]. Define λk : [0, 1) {0, 1}k be the operation that returns the first k bits after the radix point in the (possibly infinite) binary representation of y. Concretely, if y = 0.b1b2b3b4... then λk(y) = (b1, . . . , bk). We abuse notation and interpret λk s output either as a sequence or as the real number it represents (Pk i=1 bi2 i) depending on the context. The analysis is presented using binary (base-2) repre- Published in Transactions on Machine Learning Research (08/2025) sentations (e.g. V = {0, 1}) for simplicity, but it holds for arbitrary bases. First, we provide an assumption on the learnability of our model and additional definitions: Definition 1 (K-bit universality). Let H(p, q) = Ey p log q(y) denote the cross-entropy between discrete distributions p and q. Note that H(p, p) is just the Shannon entropy of p. Call parametric model pθ K-bit universal if for all discrete distributions p on K-bit strings (equivalently, 2K categories), min θ H(p, pθ) = H(p, p). In other words, pθ is K-bit universal if it is flexible enough to fit any discrete distribution on 2K categories. Definition 2. Define pk θ as probability of the first k bits under pθ, marginalizing out the remaining bits. Concretely, pk θ((b1, . . . , bk)) = X {bk+1,...,b K} pθ((b1, . . . , b K)). Seen another way, pk θ is the distribution over k-bit strings that results from auto-regressive decoding pθ for exactly k steps. Definition 3. Let f : [0, 1] R be a density function. With {Y1, . . . , YN} as i.i.d. draws from f, define θ as the maximum likelihood estimator on the truncated sequence of K bits. Concretely, θ (Y1, . . . , YN) = argmin θ n=1 log pθ(λK(Yn)). Define risk as the mean integrated squared error between true density f and an estimator bf N(Y1, . . . , YN): R(f, bf N) = E Y1,...,YN f f(y) bf N(y) 2 dy. We now give our main result below, expressing the distributional fit in terms of bias and variance. The proof is deferred to Appendix B. Theorem 1. Assume our decoding-based regression model pθ : {0, 1}K 2K is K-bit universal, and f be any twice continuously differentiable density function. If the maximum likelihood estimator at k is f k N (y) = 2kpk θ (Y1,...,YN)(λk(y)) for y [0, 1], then the risk can be exactly computed: R f, f k N (y) = 2 2k 0 f (y)2dy | {z } Bias N |{z} V ariance + O(2 4k + 1/N) | {z } Negligible Note that this theorem is broad, applicable to both Riemann and decoding heads even if they perform inference at a lower token length k than the maximum length K used during training. For simplicity, let us assume that the maximal length is always used (i.e. k = K). Intuitively, this implies that one needs a higher resolution K to capture the curvature of f, but as the number of bins increases, more data points N are required to learn to separate these 2K bins. In Figure 2, for large N=16384, we show this trend holds empirically where there is an optimal K 5 which minimizes the error. When N is quite small (e.g. 1024) we see that the decoder head significantly deviates from the theoretical risk (for the better) when the number of bits is large (>9), while the Riemann head still fits it tightly. Recall that we required a universality assumption, which says that our model can learn any discrete distribution over K-bit strings perfectly. We can decompose this assumption further into two pieces: 1) that there exists θ in our model class that achieves the minimum cross-entropy (i.e. pθ = p in Definition 1), and 2) that our SGD-based training procedure is able to find it. An explanation for this phenomenon is that in this regime (low sample size and large number of bits, or equivalently, a large number of bins), the risk profile of Published in Transactions on Machine Learning Research (08/2025) 0.25 N=16384 Number of Bits (K) Theoretical Decoder Riemann Figure 2: Lower ( ) is better. Risk (theoretical and empirical) when varying K and N to fit a truncated N[0,1](0.5, 0.252) distribution using binary tokenization. Results averaged across 10 runs each. the classical Riemann estimator is dominated by the variance term. Few samples land in each bin and as a result the histogram-based density estimate for the bins is noisy. It is conceivable that a combination of the inductive bias of our model class and the implicit bias of our SGD training procedure makes the decoder less likely to fit noise; a concrete example would be that the model is biased to learn smooth distributions, and so when asked to fit the highly discontinuous empirical distribution arising from dropping few samples into a large number of bins, it refuses to, instead opting to learn a smooth approximation, and thereby driving down the variance term and hence the overall risk. This suggests the decoder head possesses implicit regularization properties which make it much more efficient with low training data. Figure 3: Visualization of fitting a truncated Gaussian distribution. Each level k of the binary tree represents the empirical fit using k bits, and each bin gets subdivided into two. Taking a closer look at the decoding mechanism, a crucial observation is that λk essentially discretizes the unit interval (and hence f as well) into bins {Bj}2k 1 j=0 , where Bj = [j2 k, (j + 1)2 k) so that P(x Bj) = R Bj f(y)dy. We can identify k-bit sequence y = 0.b1 . . . bk with the interval [y, y + 2 k]. With a single bit (K = 1) we learn a histogram estimator on two bins [0, 1/2) and [1/2, 1) representing 0 and 1. With Published in Transactions on Machine Learning Research (08/2025) two bits we refine our prediction using four bins: [0, 1/4), [1/4, 1/2), [1/2, 3/4), and [3/4, 1) representing (0, 0), (0, 1), (1, 0), (1, 1) respectively (because, for example (0, 1) means 0.012 = 1/4). We can interpret binary representations in terms of binary trees on 2K leaf nodes where nodes represent intervals (the root representing [0, 1)) and left and right children represent the left and right halves of the node s interval. Reading off bits tells us how to traverse this tree, where 0 and 1 mean traverse the left and right subtrees respectively. For example, to arrive at (0, 1, 1) = 0.0112 = 3/8 our traversal is: [0, 1) [0, 1/2) [1/4, 1/2) [3/8, 1/2). When trained on K-bit sequences, our decoding head pθ simultaneously learns K histogram estimators for f; 2kpk θ(λk(y)) is the k-th histogram estimator (over 2k bins). In other words, as we decode bits one-by-one auto-regressively, we are iteratively refining our prediction. Figure 3 shows this mechanism in detail in the case of fitting a truncated Gaussian distribution. There are alternatives to binary representations, for example p-adic expansions, or even the Stern Brocot tree which uses the mediant to determine the child-parent relationship. An interesting research question left for future work is whether these more exotic representations of real numbers are better suited for our sequence-based regression model than the standard representations. 4 Experiments Our main goals for experiments are to: Demonstrate decoder heads can be effective swap-in replacements to common pointwise regression heads. Establish the density estimation capabilities of the decoding-based head over any distribution over R. Ablate the effect of decoder head size and sequence-specific methods such as error correction on performance. To maintain fairness, all neural network methods have access to the same encoder ϕ(x), which is a large multi-layer perceptron (MLP) with Re LU activations, with hyperparameter sweeps over number of layers (up to 5) and hidden unit sizes (up to 2048). Furthermore, the decoder head uses only 1 layer and 32 units, making up for less than 10% of the total network parameter count, which minimizes its contribution to representation learning as a confounding factor. Furthermore, for distributional heads (e.g. decoder, Riemann), we sweep their specific settings (e.g. number of bins / tokenization) over reasonable values - additional details are found in Appendix C. For the vast majority of tabular regression problems, we found that the process of training and tuning only requires at most 20 minutes on a single Nvidia P100 GPU, making the decoder head relatively cheap to use. For comparisons, we use relative mean squared error within individual tasks and scale-invariant Kendall-Tau correlation for aggregate comparisons. 4.1 Curve Fitting Input Dimension Regression Head 5 10 15 20 Unnormalized Decoder 89.56 88.71 87.49 86.11 Normalized Decoder 89.40 88.54 86.90 86.02 Pointwise 89.08 88.25 88.06 86.78 Riemann 88.94 88.30 87.42 86.78 Table 1: Higher ( ) is better. Mean Kendall-Tau correlations over BBOB functions with ( 100K) training data. Individual function results can be seen in Appendix A.2. We begin by visually demonstrating the fundamental representation power of tokenization. In Figure 4, the unnormalized decoder head is able to successfully capture the shapes of various functions with which both the Riemann and pointwise head struggle. The issue with using the pointwise head stems from two main factors: (1) requiring y-normalization, which leads to numeric instabilities especially with functions with very high or unbounded y-ranges, and (2) struggling to model abrupt or high rates of change (i.e. large Lipschitz Published in Transactions on Machine Learning Research (08/2025) Weierstrass (Smooth) Unnormalized Decoder Normalized Decoder Riemann Pointwise Ground Truth Vertical Asymptote (Hyperbolic) Stepwise (Alternating) Weierstrass (Triangular) Vertical Asymptote (Tangent) Stepwise (Increasing) Figure 4: Visual fit to ground truth is better. Curve fitting plots for various 1D functions. Both models are trained over 100K (x, y) points, where x is uniformly sampled from a bounded range. Note: These results occur regardless of xy-scales, which are omitted for brevity. Riemann prediction for Vertical Asymptote (Tangent) went out of range. constants). In contrast, the unnormalized decoder head does not encounter these issues due to its ability to express a very high range of y-values, with the normalized decoder also performing decently. In Table 1, as a sanity check over higher-dimensional functions, synthetic continuous objectives from the Black-box Optimization Benchmarking (BBOB) suite (Elhara et al., 2019) can also be sufficiently fitted by both the unnormalized and normalized decoder heads just as well as the pointwise and Riemann heads. 4.2 Real-World Regression In Figure 5, over real-world Open ML (Vanschoren et al., 2013) regression tasks from Open ML-CTR23 (Fischer et al., 2023) and AMLB (Gijsbers et al., 2024), using the unnormalized decoder head is competitive to using a regular pointwise head given the same training data. In fact, in the majority of tasks, the decoder outperforms the pointwise head, and in a few cases, the gap can be quite significant (>0.3). Full results in Appendix A.3, Figure 13 also lead to the same conclusion for normalized decoder heads. 1 5 10 15 20 25 Task Index Kendall-Tau Correlation 1 5 10 15 20 25 30 Task Index Open ML-CTR23 Unnormalized Decoder Pointwise Figure 5: Higher ( ) is better. Kendall-Tau regression scores over AMLB and Open ML-CTR23 tasks using up to 10K maximum training points. Each bar averaged over 10 runs and bars from the same task (but different regressors) are stacked on top of each other and sorted by gap performance gap. Published in Transactions on Machine Learning Research (08/2025) In Figure 6, both decoding heads outperform the Riemann head in the vast majority of tasks as well, suggesting improved sample efficiency from minimizing vocabulary / bin sizes. 0.0 0.2 0.4 0.6 0.8 1.0 Riemann Kendall-Tau Score Decoder Kendall-Tau Score 0.0 0.2 0.4 0.6 0.8 1.0 Riemann Kendall-Tau Score Decoder Kendall-Tau Score Open ML-CTR23 Decoder vs. Riemann Performance Comparison Identity (No Difference) Unnormalized Decoder vs. Riemann Normalized Decoder vs. Riemann Figure 6: Upper left ( ) is better for decoder heads. Paired scatter plots for comparing both decoders against the Riemann head, over real-world regression tasks. In order to more rigorously validate the sample efficiency hypothesis, in Figure 7, we compare the use of all normalized heads (normalized decoder, Riemann histogram, and pointwise), when varying the amount of training data. We omit unnormalized decoder results, as it aggregates samples differently. We first observe the data inefficiency of using the histogram head on selected regression tasks - in certain cases, the histogram head plateaus, unable to even achieve the performance of the decoder head, regardless of the amount of training data. 102 103 104 Task ID: 233212 102 103 104 10 2 Task ID: 233215 102 103 104 7 10 2 Task ID: 359930 102 103 104 Task ID: 359938 102 103 104 6 10 2 Task ID: 359939 102 103 104 Task ID: 359948 102 103 104 Task ID: 359951 102 103 104 Task ID: 360945 Training Data Size Relative Mean Squared Error Normalized Decoder Riemann Pointwise Figure 7: Lower ( ) is better. Relative mean squared error (MSE) over selected AMLB tasks. Each method used a min-max linear scaling normalization on y-values. Full results in Appendix A.1, Figure 11. Published in Transactions on Machine Learning Research (08/2025) Furthermore, interesting observations can be made when comparing against the standard pointwise head as a baseline. In high data regimes ( 104 data points), there are cases in which it also plateaus earlier than the decoding head. In low data regimes ( 101 data points), one would expect the decoding head to struggle more as it needs to learn numeric token representations, but as it turns out, the pointwise head can perform worse due to numeric instabilities of its own. Due to undertraining, the pointwise head required appending a sigmoid activation to enforce the normalized output to be within [0,1] to avoid extremely high MSE errors. 4.3 Density Estimation In Figure 8, we further see the decoding head s ability to perform density estimation over various shapes. Given unbounded training data it is able to capture the overall distribution p(y|x) well, although there can be slight outlier noise as shown by lighter points. In Appendix A.6 we show that even baseline heads such as Mixture Density Networks (MDNs) (Bishop, 1994) and Riemann distributions also suffer from noisy outputs. While one can enforce the sampling to be tighter (e.g. lowering temperature) to remove noise, this tighter sampling can unfortunately also reduce expressivity. In general, we find that vanilla temperature sampling with temperature 1.0 is the best way to match p(y|x). Hollow Square Sample Ground Truth Figure 8: Fit to ground truth is better. Density estimation visualization over various shapes using an unnormalized decoder head with vanilla temperature sampling. Note that these results occur regardless of xy-scales, which are omitted for brevity. In Table 2, we display the negative log-likelihood (NLL) on a collection of representative real-world datasets from the UCI regression repository (Dua & Graff, 2017) (full results over 25 datasets in Appendix A.5). We see that MDN head performance has high variability, at times able to perform the best but also extremely poorly depending on the task. Meanwhile both decoding heads remain reliable overall (NLL<0.7 always). In comparison, the Riemann head consistently underperforms in every task. Dataset MDN UD ND R Airfoil 0.12 0.11 0.40 0.01 0.34 0.01 1.33 0.14 Bike 4.59 0.86 0.12 0.00 0.10 0.01 0.36 0.05 Elevators 0.30 0.43 0.15 0.00 0.13 0.00 1.12 0.02 Gas 0.68 0.25 0.02 0.01 0.02 0.00 0.20 0.09 Housing 0.22 0.13 0.41 0.03 0.38 0.03 1.56 0.21 Kin 40K 7.49 0.73 0.19 0.01 0.12 0.01 0.39 0.03 Pol 1.49 0.41 0.01 0.00 0.01 0.00 0.18 0.02 Protein 1.07 0.44 0.34 0.00 0.41 0.01 1.55 0.04 Pumadyn32nm 0.69 1.26 0.55 0.00 0.58 0.02 2.32 0.03 Wine 0.05 0.12 0.24 0.01 0.21 0.01 1.67 0.14 Yacht 0.21 0.10 0.39 0.02 0.23 0.05 1.29 0.38 Table 2: Lower ( ) is better. Avg. NLL ( Std Dev) of test examples on UCI datasets over 10 train-test splits. Abbreviations: (UD, ND) = (unnormalized, normalized) decoder heads respectively; R = Riemann. Published in Transactions on Machine Learning Research (08/2025) 4.4 Ablation: Role of Decoding Head Size We ablate the effect of the decoding head s size on performance. We first fix the tokenization for the normalized decoding head (B=10, K=4) and then sweep the number of layers, heads, and hidden units. In Figure 9, we observe that larger decoding heads do sometimes help, but only up to a certain point, at which overfitting can occur. This was also observed over regression over BBOB functions and with the unnormalized decoding head, but we omitted these results for brevity. 1 2 3 4 0.8 1 2 3 4 Num. Layers 1 2 3 4 1.4 2 4 6 8 Num. Heads 2 4 6 8 1.40 500 1000 0.7 500 1000 Num. Units Figure 9: Lower ( ) is better. NLL over UCI datasets, when varying different axis (layers, heads, units) from a fixed default of (3, 4, 128) respectively. 4.5 Ablation: Error Correction One can also improve regression behavior using techniques purely by modifying sequence representations. Inspired by the field of coding theory, we can use error correction, where we may simply have the decoding head repeat its output multiple times (t1, . . . , t K, t 1, . . . , t K, t 1, . . . , t K, . . .) during training, and at inference perform majority voting on each location k {1, . . . , K}. 101 102 103 10 3 Task ID: 233213 101 102 103 Task ID: 359934 101 102 103 Task ID: 359944 101 102 103 Task ID: 359945 Sampling Size Relative MSE Repeat Count Figure 10: Lower ( ) is better. Relative MSE over selected AMLB tasks, when varying output repetitions. In Figure 10, we focus on the unnormalized case when using mean aggregation, where performance can be significantly harmed from extreme outliers. We see that when using regular tokenization (repeat count=1), Published in Transactions on Machine Learning Research (08/2025) as more samples are drawn, the likelihood of drawing outliers increases the error. However, the error can be substantially decreased by training the decoding head to decode the same tokens repeatedly and allow better scaling with samples, although repeating too many times may make learning more difficult. Not all error correction techniques improve results however - in Appendix A.4, we briefly observe negative results applying other types of error correction, and we leave exploring the space of such methods for future work. 5 Discussion: Limitations and Extensions This work establishes the validity of training decoding heads with cross-entropy losses for regression. To minimize confounding factors, our designs remained very simple (e.g. using basic primitives such as softmax and vanilla attention mechanisms) yet principled (e.g. digit-by-digit tokenization and constrained sampling). We list some limitations of this work, along with more potential areas for exploration. Modern LLM Architectures: Many modern LLM architectures no longer use vanilla attention mechanisms, instead opting for sparsity or low-rank approximations. In addition, MLPs have typically been replaced by mixtures of experts for memory reduction. It is worth studying further how these changes affect the performance of numeric decoding. Tokenization and Sampling: Our digit-by-digit tokenization and constrained decoding can be considered ideal cases, which modern LLMs only approximately implement. The main sources of differences include vocabularies which do not use individual digit tokens, and sampling procedures which are Top-P or Top-K, which may accidentally choose invalid tokens. We hypothesize that these issues can decrease regression performance. Intermediate Language and Training: In many LLM use-cases involving numeric prediction problems, the decoder not only must return a tokenized number, but also other intermediate language tokens. Our work does not address these cases, and it is unclear how fine-tuning on these intermediate language tokens may affect training dynamics and numeric prediction performance. Furthermore, modern LLM training consists of reinforcement learning (RL), and it remains to be studied how reward-based training methods affect regression results. Multi-objective Regression: In this paper, we only studied the single-objective regression case. However, one may easily modify our paradigm to support multiple objectives p(y(1), . . . , y(M) | ϕ(x)), e.g. by decoding a concatenated sequence of those objectives. This has the particular benefit of modeling objectives autoregressively, which can be difficult to perform with classical techniques. The benefit of multi-objective density estimation is even more pronounced, due to the decoder s universal approximation abilities. Other Regression Architectures: We did not compare to significantly more complex methods such as stochastic networks, which use stochastic activations or weights. Early examples include Sigmoid Belief Nets (Neal, 1992; Tang & Salakhutdinov, 2013), which have not seen wide adoption due to their complex architectures and expectation-maximization training updates. Bayesian neural networks (Lampinen & Vehtari, 2001; Titterington, 2004; Goan & Fookes, 2020) can be seen as more modern stochastic networks, but still possess complex inference techniques, e.g. Markov Chain Monte Carlo (MCMC) or variational inference. Similarly, Energy-based models (Teh et al., 2003) can be used for image-based regression (Gustafsson et al., 2020; Liu et al., 2022) but still see limited use due MCMC required at inference time. 6 Conclusion and Future Work We thoroughly investigated the many benefits but also drawbacks of using decoding-based regression. We described a natural tokenization scheme for both normalized and unnormalized y-values, and theoretically established its risk minimization properties. Empirically, we showed that it can be competitive as, or even outperform traditional pointwise heads for regression tasks. Furthermore, it is also capable of density estimation over a variety of conditional distributions p(y|ϕ(x)), and can further outperform common baseline regression heads such as Gaussian mixtures and histogram distributions. We hope this work will also be a valuable reference for the language modeling community and that it provides a principled explanation for the use of supervised fine-tuning over numeric targets. Published in Transactions on Machine Learning Research (08/2025) Acknowledgements We would like to thank Yutian Chen for his valuable review of the work and Bangding (Jeffrey) Yang for technical help. We further thank Yash Akhauri, Aviral Kumar, Bryan Lewandowski, Michal Lukasik, Sagi Perel, David Smalling, and Subhashini Venugopalan for useful discussions, and Daniel Golovin and Denny Zhou for continuing support. Yash Akhauri, Bryan Lewandowski, Cheng-Hsi Lin, Adrian N. Reyes, Grant C. Forbes, Arissa Wongpanich, Bangding Yang, Mohamed S. Abdelfattah, Sagi Perel, and Xingyou Song. Performance prediction for large systems via text-to-text regression, 2025. Yuntao Bai, Saurav Kadavath, Sandipan Kundu, Amanda Askell, Jackson Kernion, Andy Jones, Anna Chen, Anna Goldie, Azalia Mirhoseini, Cameron Mc Kinnon, Carol Chen, Catherine Olsson, Christopher Olah, Danny Hernandez, Dawn Drain, Deep Ganguli, Dustin Li, Eli Tran-Johnson, Ethan Perez, Jamie Kerr, Jared Mueller, Jeffrey Ladish, Joshua Landau, Kamal Ndousse, Kamile Lukosiute, Liane Lovitt, Michael Sellitto, Nelson Elhage, Nicholas Schiefer, Noemí Mercado, Nova Das Sarma, Robert Lasenby, Robin Larson, Sam Ringer, Scott Johnston, Shauna Kravec, Sheer El Showk, Stanislav Fort, Tamera Lanham, Timothy Telleen-Lawton, Tom Conerly, Tom Henighan, Tristan Hume, Samuel R. Bowman, Zac Hatfield-Dodds, Ben Mann, Dario Amodei, Nicholas Joseph, Sam Mc Candlish, Tom Brown, and Jared Kaplan. Constitutional AI: harmlessness from AI feedback. Co RR, abs/2212.08073, 2022. doi: 10.48550/ARXIV.2212.08073. Marc G. Bellemare, Will Dabney, and Rémi Munos. A distributional perspective on reinforcement learning. In Doina Precup and Yee Whye Teh (eds.), Proceedings of the 34th International Conference on Machine Learning, ICML 2017, Sydney, NSW, Australia, 6-11 August 2017, volume 70 of Proceedings of Machine Learning Research, pp. 449 458. PMLR, 2017. Christopher M. Bishop. Mixture density networks. Technical report, Aston University, 1994. Ralph Allan Bradley and Milton E. Terry. Rank analysis of incomplete block designs: I. the method of paired comparisons. Biometrika, 39(3/4):324 345, 1952. ISSN 00063444, 14643510. Wenzhi Cao, Vahid Mirjalili, and Sebastian Raschka. Rank consistent ordinal regression for neural networks with application to age estimation. Pattern Recognit. Lett., 140:325 331, 2020. doi: 10.1016/J.PATREC. 2020.11.008. François Charton. Linear algebra with transformers. Trans. Mach. Learn. Res., 2022, 2022. Yutian Chen, Xingyou Song, Chansoo Lee, Zi Wang, Richard Zhang, David Dohan, Kazuya Kawakami, Greg Kochanski, Arnaud Doucet, Marc Aurelio Ranzato, Sagi Perel, and Nando de Freitas. Towards learning universal hyperparameter optimizers with transformers. In Sanmi Koyejo, S. Mohamed, A. Agarwal, Danielle Belgrave, K. Cho, and A. Oh (eds.), Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, Neur IPS 2022, New Orleans, LA, USA, November 28 - December 9, 2022, 2022. Cheng-Han Chiang, Hung-yi Lee, and Michal Lukasik. TRACT: regression-aware fine-tuning meets chainof-thought reasoning for llm-as-a-judge. Co RR, abs/2503.04381, 2025. doi: 10.48550/ARXIV.2503.04381. Stéphane d Ascoli, Pierre-Alexandre Kamienny, Guillaume Lample, and François Charton. Deep symbolic regression for recurrent sequences. Co RR, abs/2201.04600, 2022. Raul Diaz and Amit Marathe. Soft labels for ordinal regression. In IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, June 16-20, 2019, pp. 4738 4747. Computer Vision Foundation / IEEE, 2019. doi: 10.1109/CVPR.2019.00487. Dheeru Dua and Casey Graff. UCI machine learning repository, 2017. URL http://archive.ics.uci.edu/ ml. Published in Transactions on Machine Learning Research (08/2025) Ouassim Elhara, Konstantinos Varelas, Duc Nguyen, Tea Tusar, Dimo Brockhoff, Nikolaus Hansen, and Anne Auger. Coco: the large scale black-box optimization benchmarking (bbob-largescale) test suite. ar Xiv preprint ar Xiv:1903.06396, 2019. Angela Fan, Mike Lewis, and Yann N. Dauphin. Hierarchical neural story generation. In Iryna Gurevych and Yusuke Miyao (eds.), Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, ACL 2018, Melbourne, Australia, July 15-20, 2018, Volume 1: Long Papers, pp. 889 898. Association for Computational Linguistics, 2018. doi: 10.18653/V1/P18-1082. Sebastian Felix Fischer, Liana Harutyunyan Matthias Feurer, and Bernd Bischl. Open ML-CTR23 a curated tabular regression benchmarking suite. In Auto ML Conference 2023 (Workshop), 2023. Pieter Gijsbers, Marcos L. P. Bueno, Stefan Coors, Erin Le Dell, Sébastien Poirier, Janek Thomas, Bernd Bischl, and Joaquin Vanschoren. AMLB: an automl benchmark. J. Mach. Learn. Res., 25:101:1 101:65, 2024. Ethan Goan and Clinton Fookes. Bayesian Neural Networks: An Introduction and Survey. Springer International Publishing, Cham, 2020. ISBN 978-3-030-42553-1. doi: 10.1007/978-3-030-42553-1_3. Alex Graves. Sequence transduction with recurrent neural networks. Co RR, abs/1211.3711, 2012. Fredrik K. Gustafsson, Martin Danelljan, Goutam Bhat, and Thomas B. Schön. Energy-based models for deep probabilistic regression. In Andrea Vedaldi, Horst Bischof, Thomas Brox, and Jan-Michael Frahm (eds.), Computer Vision - ECCV 2020 - 16th European Conference, Glasgow, UK, August 23-28, 2020, Proceedings, Part XX, volume 12365 of Lecture Notes in Computer Science, pp. 325 343. Springer, 2020. doi: 10.1007/978-3-030-58565-5\_20. Frank E. Harrell and C. E. Davis. A new distribution-free quantile estimator. Biometrika, 69(3):635 640, 1982. ISSN 00063444, 14643510. Noah Hollmann, Samuel Müller, Lennart Purucker, Arjun Krishnakumar, Max Körfer, Shi Bin Hoo, Robin Tibor Schirrmeister, and Frank Hutter. Accurate predictions on small data with a tabular foundation model. Nat., 637(8044):319 326, 2025. doi: 10.1038/S41586-024-08328-6. Ari Holtzman, Jan Buys, Li Du, Maxwell Forbes, and Yejin Choi. The curious case of neural text degeneration. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020. Open Review.net, 2020. IEEE. Ieee standard for floating-point arithmetic. IEEE Std 754-2019 (Revision of IEEE 754-2008), pp. 1 84, 2019. doi: 10.1109/IEEESTD.2019.8766229. Ehsan Imani and Martha White. Improving regression performance with distributional losses. In Jennifer G. Dy and Andreas Krause (eds.), Proceedings of the 35th International Conference on Machine Learning, ICML 2018, Stockholmsmässan, Stockholm, Sweden, July 10-15, 2018, volume 80 of Proceedings of Machine Learning Research, pp. 2162 2171. PMLR, 2018. Diederik P. Kingma and Max Welling. Auto-encoding variational bayes. In Yoshua Bengio and Yann Le Cun (eds.), 2nd International Conference on Learning Representations, ICLR 2014, Banff, AB, Canada, April 14-16, 2014, Conference Track Proceedings, 2014. Jouko Lampinen and Aki Vehtari. Bayesian approach for neural networks review and case studies. Neural Networks, 14(3):257 274, 2001. doi: 10.1016/S0893-6080(00)00098-8. L.E. Lehmann. Theory of Point Estimation. A Wiley publication in mathematical statistics. Wiley, 1983. Xixi Liu, Che-Tsung Lin, and Christopher Zach. Energy-based models for deep probabilistic regression. In 2022 26th International Conference on Pattern Recognition (ICPR), pp. 2693 2699, 2022. doi: 10.1109/ ICPR56361.2022.9955636. Published in Transactions on Machine Learning Research (08/2025) Michal Lukasik, Harikrishna Narasimhan, Aditya Krishna Menon, Felix Yu, and Sanjiv Kumar. Regression aware inference with llms. In Yaser Al-Onaizan, Mohit Bansal, and Yun-Nung Chen (eds.), Findings of the Association for Computational Linguistics: EMNLP 2024, Miami, Florida, USA, November 12-16, 2024, pp. 13667 13678. Association for Computational Linguistics, 2024. Michal Lukasik, Zhao Meng, Harikrishna Narasimhan, Aditya Krishna Menon, Yin Wen Chang, Felix X. Yu, and Sanjiv Kumar. Better autoregressive regression with LLMs. In The Thirteenth International Conference on Learning Representations, ICLR 2025, Singapore, Singapore, April 24-28, 2025. Open Review.net, 2025. Dakota Mahan, Duy Phung, Rafael Rafailov, Chase Blagden, Nathan Lile, Louis Castricato, Jan-Philipp Fränken, Chelsea Finn, and Alon Albalak. Generative reward models. Co RR, abs/2410.12832, 2024. doi: 10.48550/ARXIV.2410.12832. Radford M. Neal. Connectionist learning of belief networks. Artif. Intell., 56(1):71 113, 1992. doi: 10.1016/ 0004-3702(92)90065-6. Tung Nguyen, Qiuyi Zhang, Bangding Yang, Chansoo Lee, Jorg Bornschein, Yingjie Miao, Sagi Perel, Yutian Chen, and Xingyou Song. Predicting from strings: Language model embeddings for bayesian optimization. Co RR, abs/2410.10190, 2024. doi: 10.48550/ARXIV.2410.10190. Rodrigo Frassetto Nogueira, Zhiying Jiang, and Jimmy Lin. Investigating the limitations of the transformers with simple arithmetic tasks. Co RR, abs/2102.13019, 2021. Emanuel Parzen. On Estimation of a Probability Density Function and Mode. The Annals of Mathematical Statistics, 33(3):1065 1076, 1962. doi: 10.1214/aoms/1177704472. Minghai Qin. Hamming-distance-based binary representation of numbers. In 2018 IEEE International Symposium on Information Theory (ISIT), pp. 2202 2205, 2018. doi: 10.1109/ISIT.2018.8437644. Carl Edward Rasmussen. The infinite gaussian mixture model. In Sara A. Solla, Todd K. Leen, and Klaus Robert Müller (eds.), Advances in Neural Information Processing Systems 12, [NIPS Conference, Denver, Colorado, USA, November 29 - December 4, 1999], pp. 554 560. The MIT Press, 1999. Murray Rosenblatt. Remarks on Some Nonparametric Estimates of a Density Function. The Annals of Mathematical Statistics, 27(3):832 837, 1956. doi: 10.1214/aoms/1177728190. Xingyou Song, Oscar Li, Chansoo Lee, Bangding Yang, Daiyi Peng, Sagi Perel, and Yutian Chen. Omnipred: Language models as universal regressors. Co RR, abs/2402.14547, 2024. Eric Tang, Bangding Yang, and Xingyou Song. Understanding LLM embeddings for regression. Co RR, abs/2411.14708, 2024. doi: 10.48550/ARXIV.2411.14708. Yichuan Tang and Ruslan Salakhutdinov. Learning stochastic feedforward neural networks. In Christopher J. C. Burges, Léon Bottou, Zoubin Ghahramani, and Kilian Q. Weinberger (eds.), Advances in Neural Information Processing Systems 26: 27th Annual Conference on Neural Information Processing Systems 2013. Proceedings of a meeting held December 5-8, 2013, Lake Tahoe, Nevada, United States, pp. 530 538, 2013. Yee Whye Teh, Max Welling, Simon Osindero, and Geoffrey E. Hinton. Energy-based models for sparse overcomplete representations. J. Mach. Learn. Res., 4:1235 1260, 2003. D. M. Titterington. Bayesian Methods for Neural Networks and Related Models. Statistical Science, 19(1): 128 139, 2004. doi: 10.1214/088342304000000099. Robert Vacareanu, Vlad-Andrei Negru, Vasile Suciu, and Mihai Surdeanu. From words to numbers: Your large language model is secretly A capable regressor when given in-context examples. Co RR, abs/2404.07544, 2024. Published in Transactions on Machine Learning Research (08/2025) Joaquin Vanschoren, Jan N. van Rijn, Bernd Bischl, and Luís Torgo. Openml: networked science in machine learning. SIGKDD Explor., 15(2):49 60, 2013. doi: 10.1145/2641190.2641198. Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. Attention is all you need. In Isabelle Guyon, Ulrike von Luxburg, Samy Bengio, Hanna M. Wallach, Rob Fergus, S. V. N. Vishwanathan, and Roman Garnett (eds.), Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4-9, 2017, Long Beach, CA, USA, pp. 5998 6008, 2017. Marek Wydmuch, Kalina Jasinska, Mikhail Kuznetsov, Róbert Busa-Fekete, and Krzysztof Dembczynski. A no-regret generalization of hierarchical softmax to extreme multi-label classification. In Samy Bengio, Hanna M. Wallach, Hugo Larochelle, Kristen Grauman, Nicolò Cesa-Bianchi, and Roman Garnett (eds.), Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, Neur IPS 2018, December 3-8, 2018, Montréal, Canada, pp. 6358 6368, 2018. Lunjun Zhang, Arian Hosseini, Hritik Bansal, Mehran Kazemi, Aviral Kumar, and Rishabh Agarwal. Generative verifiers: Reward modeling as next-token prediction. Co RR, abs/2408.15240, 2024. doi: 10.48550/ARXIV.2408.15240. Lianmin Zheng, Wei-Lin Chiang, Ying Sheng, Siyuan Zhuang, Zhanghao Wu, Yonghao Zhuang, Zi Lin, Zhuohan Li, Dacheng Li, Eric P. Xing, Hao Zhang, Joseph E. Gonzalez, and Ion Stoica. Judging llmas-a-judge with mt-bench and chatbot arena. In Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, Neur IPS 2023, New Orleans, LA, USA, December 10 - 16, 2023, 2023. Daniel M. Ziegler, Nisan Stiennon, Jeffrey Wu, Tom B. Brown, Alec Radford, Dario Amodei, Paul F. Christiano, and Geoffrey Irving. Fine-tuning language models from human preferences. Co RR, abs/1909.08593, 2019. Published in Transactions on Machine Learning Research (08/2025) A Additional Experiments A.1 Data Scaling: Extended For completeness, we display the plots over all tasks in AMLB (Gijsbers et al., 2024). We confirm the data-efficiency of the decoder head against the Riemann distribution head on nearly every regression task. Furthermore, we observe numerous cases where both distributional methods outperform the pointwise head, especially in low data regimes. 102 103 104 10 2 2 10 2 2.2 10 2 Task ID: 359944 102 103 104 Task ID: 233212 102 103 104 Task ID: 359937 102 103 104 Task ID: 359950 102 103 104 Task ID: 359938 102 103 104 Task ID: 233213 102 103 104 Task ID: 359942 102 103 104 Task ID: 233211 102 103 104 Task ID: 359936 102 103 104 4 10 2 Task ID: 359952 102 103 104 Task ID: 359951 102 103 104 Task ID: 359949 102 103 104 10 2 10 1 Task ID: 233215 102 103 104 Task ID: 360945 102 103 104 Task ID: 167210 102 103 104 4 10 2 Task ID: 359941 102 103 104 Task ID: 359946 102 103 104 7 10 2 Task ID: 359930 102 103 104 Task ID: 233214 102 103 104 Task ID: 359948 102 103 104 Task ID: 359931 102 103 104 Task ID: 359932 102 103 104 Task ID: 359933 102 103 104 Task ID: 359934 102 103 104 6 10 2 Task ID: 359939 102 103 104 Task ID: 359945 102 103 104 Task ID: 359935 102 103 104 Task ID: 317614 Training Data Size Relative Mean Squared Error Normalized Decoder Riemann Pointwise Figure 11: Lower ( ) is better. Regression performance as a function of training data scaling between using the normalized decoder vs. Reimannian distribution as regression heads. Each point was averaged over 10 training runs over random combinations of datapoints from the original AMLB task s training set. Published in Transactions on Machine Learning Research (08/2025) A.2 BBOB Curve Fitting: Extended In Figure 12, we compare the curve fitting properties of multiple regression heads. We see overall that the decoder head is competitive and has both pros and cons for specific function landscapes from the BBOB benchmark. 1 5 10 15 20 0.96 1 5 10 15 20 0.92 1 5 10 15 20 Bueche Rastrigin 1 5 10 15 20 Linear Slope 1 5 10 15 20 Attractive Sector 1 5 10 15 20 Step Ellipsoidal 1 5 10 15 20 Rosenbrock Rotated 1 5 10 15 20 Ellipsoidal 1 5 10 15 20 0.986 1 5 10 15 20 1 5 10 15 20 Griewank Rosenbrock 1 5 10 15 20 0.970 Sharp Ridge 1 5 10 15 20 Different Powers 1 5 10 15 20 0.2 Weierstrass 1 5 10 15 20 Schaffers F7 1 5 10 15 20 0.76 Schaffers F7Ill Conditioned 1 5 10 15 20 0.965 1 5 10 15 20 1 5 10 15 20 1 5 10 15 20 Gallagher101Me 1 5 10 15 20 1.00 Gallagher21Me 1 5 10 15 20 0.96 Negative Sphere Dimension (x) Kendall-Tau Correlation Normalized Decoder Unnormalized Decoder Riemann Pointwise Figure 12: Higher ( ) is better. Extended results from Table 1 in the main body. Regression performance as a function of input dimension over BBOB functions using Kendall-Tau correlation. Each point was averaged over 10 training runs, each with 100K training points (x, y) where each x is sampled uniformly from [ 5, 5] coordinate-wise. Note: Some functions such as Rosenbrock Rotated or Griewank Rosenbrock are undefined when dimension is 1, so we skip those points. Published in Transactions on Machine Learning Research (08/2025) A.3 Individual Open ML Kendall-Taus In Figure 13, we present full results on the AMLB and Open ML-CTR23 regression benchmarks, over all regression heads. We see that both the unnormalized and normalized decoder heads remain competitive throughout the benchmarks. 1 6 11 16 21 26 Task Index Kendall-Tau Correlation 1 6 11 16 21 26 31 Task Index Kendall-Tau Correlation Open ML-CTR23 Unnormalized Decoder Normalized Decoder Riemann Pointwise Figure 13: Higher ( ) is better. Extended results from Figure 5, but for all four regression heads on all tasks. Task IDs sorted by pointwise head performance. Published in Transactions on Machine Learning Research (08/2025) A.4 Alternative Tokenization Schemes: Hamming-Distance One possible criticism of the default tree-based tokenization in the normalized decoding case, is the vulnerability to small changes in the left-most significant tokens, which can cause large numeric changes in the actual number. Qin (2018) notes this and proposes an alternative Hamming Distance-based binary representation which is robust to bitwise edits, and upper bounds the possible distortion |y y| as a function of the edit distance between the Hamming representations of y and y. For example, if the binary length is 3, the representation for all integers {0, 1, . . . , 23} is {(000), (001), (010), (100), (011), (101), (110), (111)} which can also be used in the normalized case for {0/23, 1/23, . . . , 7/23} [0, 1]. In Figure 14, we show however, such a representation may not lead to better regression results, which we hypothesize is due to this representation being more difficult to learn. 101 102 103 Task ID: 233212 101 102 103 Task ID: 359950 101 102 103 Task ID: 233213 101 102 103 Task ID: 359951 101 102 103 Task ID: 359949 101 102 103 Task ID: 233215 101 102 103 Task ID: 167210 101 102 103 Task ID: 359932 101 102 103 Task ID: 359932 Sampling Size Relative Mean Squared Error Tree Hamming Figure 14: Lower ( ) is better. Regression performance while vary sampling size from y pθ( |x) using binary tree-based tokenization vs. Hamming representation on normalized decoder with mean aggregation. Each point was averaged over 10 training runs over random size-1000 combinations of the original AMLB task s training data points. Published in Transactions on Machine Learning Research (08/2025) A.5 Full UCI Density Estimation Results We obtained the datasets from https://github.com/treforevans/uci_datasets which preprocessed and scraped the data from the official UCI website at https://archive.ics.uci.edu/. Dataset Mixture Density Network Unnormalized Decoder Normalized Decoder Riemann Airfoil 0.12 0.11 0.40 0.01 0.34 0.01 1.33 0.14 Auto MPG 0.21 0.07 0.32 0.03 0.41 0.05 1.62 0.17 Autos 0.32 0.23 0.48 0.05 0.47 0.07 2.60 0.76 Bike 4.59 0.86 0.12 0.00 0.10 0.01 0.36 0.05 Breast Cancer 0.32 0.09 0.48 0.05 0.64 0.03 2.85 0.37 Challenger -0.29 0.66 0.14 0.14 0.06 0.08 0.87 0.77 Concrete 0.15 0.05 0.43 0.03 0.41 0.04 1.67 0.20 Elevators 0.30 0.43 0.15 0.00 0.13 0.00 1.12 0.02 Energy 0.40 0.14 0.17 0.03 0.16 0.05 0.38 0.20 Fertility -0.06 0.16 0.31 0.09 0.46 0.13 2.41 0.61 Gas 0.68 0.25 0.02 0.01 0.02 0.00 0.20 0.09 Housing 0.22 0.13 0.41 0.03 0.38 0.03 1.56 0.21 Kegg Directed 2.41 1.10 0.05 0.00 0.05 0.00 0.22 0.02 Kin 40K 7.49 0.73 0.19 0.01 0.12 0.01 0.39 0.03 Parkinsons 0.59 0.18 0.40 0.02 0.39 0.03 1.40 0.33 Pol 1.49 0.41 0.01 0.00 0.01 0.00 0.18 0.02 Protein 1.07 0.44 0.34 0.00 0.41 0.01 1.55 0.04 Pumadyn32nm 0.69 1.26 0.55 0.00 0.58 0.02 2.32 0.03 Slice 7.09 0.09 0.05 0.00 0.02 0.00 0.08 0.02 SML 1.31 0.59 0.21 0.01 0.11 0.01 0.35 0.03 Solar -1.40 0.29 0.04 0.01 0.04 0.01 0.61 0.12 Stock -0.15 0.15 0.27 0.04 0.32 0.04 1.63 0.46 Tami Electric 0.01 0.00 0.46 0.00 0.69 0.00 2.70 0.00 Wine 0.05 0.12 0.24 0.01 0.21 0.01 1.67 0.14 Yacht 0.21 0.10 0.39 0.02 0.23 0.05 1.29 0.38 Table 3: Lower ( ) is better. Avg. NLL ( Std Dev) of test examples on UCI datasets over 10 train-test splits. Published in Transactions on Machine Learning Research (08/2025) A.6 Density Estimation Visualization: Extended In Figure 15, we present further results on density estimation with various decoder sampling techniques (topk, top-p, low temperature) alongside MDN and Riemann baselines. We see that using vanilla temperature sampling for the decoder is optimal and unbiased for capturing the shapes of all problems. MDN (Mixtures=100) Hollow Square Decoder (Vanilla) Decoder (Temp=0.1) Decoder (Top P=0.9) Decoder (Top K=5) Riemann (Bins=1000) Sample Ground Truth Figure 15: Visualizing density estimation of p(y|x) on 1D problems. We used an unnormalized decoder with (B = 10, E = 1, M = 5). Note that these results occur regardless of xy-scales, which are omitted for brevity. Published in Transactions on Machine Learning Research (08/2025) B Extended Theory Proof of Theorem 1. Firstly, we observe that n=1 log pθ(λK(Yn)) = argmin θ H( ef K N , pθ), where ef k N is a discrete distribution that tracks the fraction of samples {Yn}N n=1 that fall within each of the 2k uniformly-spaced bins in [0, 1]. Formally, ef k N((b1, . . . , bk)) = 1 n=1 1(λk(Yn) = (b1, . . . , bk)). Conditioned on the samples {Yn}N n=1, ef k N is a distribution on k-bit strings, and so by the K-bit universality assumption, p K θ = ef K N . It follows that pk θ = ef k N k K since if two discrete distributions are equal so are any of their marginals. Then f k N (x) 2kpk θ (λk(x)) = 2k ef k N(λk(x)) lines up exactly as a 2k-bin histogram estimator for f, for all k K. Now, we can treat the problem as one of histogram estimation. Let s consider a fixed k. We first observe that the risk can be written as the sum of a squared bias term and a variance one. Specifically, R(f, f k N ) = Z 1 0 Bias(y)2dy + Z 1 0 Variance(y)dy, where Bias(y) = E[f k N (y)] f(y) and Variance(y) = V(f k N (y)) is the bias and variance of f k N (y) at fixed y respectively. Now, label bins {Bj}2k 1 j=0 , where Bj = [jε, (j + 1)ε) and ε = 2 k is the bin width. Let pj = R Bj f(z)dz be the true probability mass in bin Bj. With Nj as the number of samples in Bj, the expected value of the estimator for y Bj is E[f k N (y)] = E[Nj/(Nε)] = (Npj)/(Nε) = pj/ε. Assume the true density f is twice continuously differentiable on [0, 1] (i.e., f C2([0, 1])). This implies f, f , and f are bounded on [0, 1]. Let M1 = supy [0,1] |f (y)| and M2 = supy [0,1] |f (y)|. Bias Analysis: Let yj = (j + 1/2)ε be the midpoint of bin Bj. For z Bj, by Taylor s Theorem around yj: f(z) = f(yj) + (z yj)f (yj) + (z yj)2 2 f (ξz) for some ξz between z and yj. Integrating over Bj: Bj f(z)dz = Z f(yj) + (z yj)f (yj) + (z yj)2 2 f (ξz) dz Bj dz + f (yj) Z Bj (z yj)dz + Z = εf(yj) + 0 + Rj, where the remainder term Rj = R 2 f (ξz)dz. Since |z yj| ε/2 and |f (ξz)| M2, we have 2 M2dz = M2ε2 Bj dz = M2ε3 8 . Thus, Rj = O(ε3). The bias for y Bj is Bias(y) = E[f k N (y)] f(y) = pj ε f(y) = εf(yj)+Rj ε f(y) = f(yj) + Rj Expanding f(y) around yj: f(y) = f(yj) + (y yj)f (yj) + (y yj)2 2 f (ηy) for ηy between y and yj. Bias(y) = f(yj) + O(ε2) f(yj) + (y yj)f (yj) + O(ε2) = (y yj)f (yj) + O(ε2). Published in Transactions on Machine Learning Research (08/2025) Now, integrate the squared bias over bin Bj: Bj Bias(y)2dy = Z (y yj)f (yj) + O(ε2) 2 dy (y yj)2(f (yj))2 2(y yj)f (yj)O(ε2) + O(ε4) dy = (f (yj))2 Z Bj (y yj)2dy O(ε2)f (yj) Z Bj (y yj)dy + Z = (f (yj))2 Z ε/2 ε/2 u2du O(ε2)f (yj) 0 + O(ε4) ε (let u = y yj) = (f (yj))2 u3 ε/2 + O(ε5) = (f (yj))2 ε3 12 + O(ε5). Summing over all bins: 0 Bias(y)2dy = Bj Bias(y)2dy = (f (yj))2 ε3 j=0 (f (yj))2ε + 0 (f (y))2dy + O(ε2) + 2k O(ε5) 0 (f (y))2dy + O(ε4), where the third line uses a known approximation error for the Riemann sum with midpoint rule applied to (f )2 C1. Variance Analysis: The variance for y Bj is Variance(y) = V(f k N (y)) = V(Nj/(Nε)) = 1 (Nε)2 V(Nj). Since Nj Binomial(N, pj), V(Nj) = Npj(1 pj). Variance(y) = Npj(1 pj) N 2ε2 = pj(1 pj) = (εf(yj) + O(ε3))(1 εf(yj) O(ε3)) = εf(yj) ε2f(yj)2 + O(ε3) N + O(ε/N). Published in Transactions on Machine Learning Research (08/2025) Integrating the variance: 0 Variance(y)dy = Nε + O(1/N) dy Nε + O(ε/N) j=0 f(yj)ε + 2k O(ε/N) 0 f(y)dy + O(ε2) + O(1/N) (Riemann sum error for f C2) = 1 Nε(1 + O(ε2)) + O(1/N) = 1 Nε + O(ε/N) + O(1/N). Since we typically consider asymptotics where N and ε 0 such that Nε , the dominant variance term is 1/(Nε). Total Risk: Combining the integrated squared bias and integrated variance: R(f, f k N ) = Z 1 0 Bias(y)2dy + Z 1 0 Variance(y)dy 0 (f (y))2dy + O(ε4) + 1 Nε + O(ε/N) + O(1/N) 0 (f (y))2dy + 1 Nε + O(ε4) + O(1/N). (assuming ε/N is smaller than 1/N) Substituting ε = 2 k: R(f, f k N ) = 2 2k 0 (f (y))2dy + 2k N + O(2 4k + 1/N). This gives the asymptotic risk. The O(2 4k + 1/N) term is negligible, and can be disregarded. Published in Transactions on Machine Learning Research (08/2025) C Exact Experimental Details For all models, we sweeped the encoder (basic MLP with Re LU activation) by varying the number of layers within [2,3,4,5] and hidden units within [256, 512, 2048]. For x-normalization, we apply a mean and std scaling, i.e. x (x xmean)/xstd where xmean, xstd are coordinate-wise mean and standard deviations over all x s in the training set. The preprocessed tensor is then fed directly into the encoder. For y-normalization, we apply min/max linear scaling, i.e. y (y ymin)/(ymax ymin) where ymin, ymax are computed from the training set. This is applicable to models representing [0, 1] output range (i.e. Riemann and Normalized Decoder). For Pointwise and Mixture Density heads, we further apply a shift y y 0.5 to center the values within [ 0.5, 0.5]. All models were trained with a maximum of 300 epochs. To prevent overfitting, we apply early stopping (patience=5) where the validation split is 0.1 on the training set. Adam learning rates were sweeped over [1e-4, 5e-4]. We further describe hyperparameters and sweeps for individual heads below: Pointwise: Uses Re LU activations on every hidden layer. Weight decay: [0.0, 0.1, 1.0] Unnormalized Decoder: Uses vanilla temperature sampling. Base B: [4, 8, 10] Exponent Digit Count E: [1, 2, 4] Mantissa Digit Count M: [2, 4, 8] Transformer size: (3 layers, 128 units, 4 heads) or (1 layer, 32 units, 1 head). Normalized Decoder: Sampling same as unnormalized decoder. Base B: [2, 4, 8] Length K: [4, 8, 6] Transformer size: Same as unnormalized decoder. Riemann/Histogram Distribution: We specify a bin count, which uniformly partitions the range [0, 1] into equally spaced bins. Output is parameterized using softmax. Bin Count: [16, 64, 256, 1024, 4096, 16384] Mixture Density Network: Given a mixture count M, the distribution head consists of mixture πM M, mean µM RM, and standard deviation σM RM. Mixtures were parameterized using softmax, while standard deviations were via ELU(x) + 1 activation to enforce positivity. Mixtures M: [1, 2, 5, 10, 20, 50, 1000]