# safari_statespace_models_for_frameagnostic_representation__62be1682.pdf Published in Transactions on Machine Learning Research (10/2025) Sa FARi: State-Space Models for Frame-Agnostic Representation Hossein Babaei hb26@rice.edu Department of Electrical and Computer Engineering Rice University Mel White mel.white@rice.edu Department of Electrical and Computer Engineering Rice University Sina Alemohammad sa86@rice.edu Department of Electrical and Computer Engineering Rice University Richard G. Baraniuk richb@rice.edu Department of Electrical and Computer Engineering Rice University Reviewed on Open Review: https: // openreview. net/ forum? id= UAgx U8g Btv& note Id= 2849tqh Vi A State-Space Models (SSMs) have re-emerged as a powerful tool for online function approximation, and as the backbone of machine learning models for long-range dependent data. However, to date, only a few polynomial bases have been explored for this purpose, and the state-of-the-art implementations were built upon a few limited options. In this paper, we present a generalized method for building an SSM with any frame or basis. This framework encompasses the approach known as Hi PPO, but also permits an infinite diversity of other possible species of SSM, paving the way for improved performance of SSM-based machine learning models. We dub this approach Sa FARi: SSMs for Frame-Agnostic Representation. 1 Introduction Modeling sequential data is a cornerstone of modern machine learning, with applications spanning natural language processing, speech recognition, video analysis, and beyond (Zubić et al., 2024; Alemohammad et al., 2021; Nguyen et al., 2022). A fundamental challenge in these domains is the efficient representation of long-range dependence in time-series data, where the goal is to capture and preserve the essential features of the input signal necessary for downstream tasks over extended time horizons while maintaining computational tractability (Hochreiter & Schmidhuber, 1997). Machine learning approaches, such as recurrent neural networks (RNNs), struggle to learn long-range dependencies due to limited memory horizons (Elman, 1990; Hochreiter & Schmidhuber, 1997; Schuster & Paliwal, 1997; Pascanu et al., 2013). During backpropagation, gradients are repeatedly multiplied by the same weight matrix, causing them to either shrink exponentially (vanish) or grow exponentially (explode). Vanishing gradients prevent the network from updating weights effectively, while exploding gradients lead to unstable training. Although variants like LSTMs (Graves & Schmidhuber, 2005) and GRUs (Cho et al., 2014) address some of these limitations, they often require task-specific parameterization, and cannot generalize across different sequence lengths or timescales. State-space models (SSMs) present a powerful alternative for online representation of sequential data. By design, SSMs enable the online computation of compressive representations, maintaining a constant-size Published in Transactions on Machine Learning Research (10/2025) memory footprint regardless of sequence length. The seminal work of Gu et al. introduced High-Order Polynomial Projection Operators (Hi PPO), which leverages orthogonal function bases to enable theoretically grounded, real-time updates of sequence representations. This framework, and its subsequent extensions into learned architectures, have demonstrated remarkable performance in tasks involving long-range dependencies, such as language modeling and signal processing (Gu & Dao, 2023; Gu et al., 2022b; 2023; Gupta et al., 2024; Gu et al., 2022a; Smith et al., 2023; Hasani et al., 2023). By formulating sequence representation as an online function approximation problem, Hi PPO provides a unified perspective on memory mechanisms, offering both theoretical guarantees and practical efficiency. However, despite its successes, the Hi PPO framework has been limited to specific families of orthogonal polynomials. While these bases are well-suited for certain applications, they are not universally optimal for all signal classes. Fourier bases, for instance, are optimal for smooth, periodic signals due to their global frequency representation. Polynomial bases, such as orthogonal polynomials (e.g., Legendre or Chebyshev), are particularly effective for approximating smooth functions over compact intervals. The absence of a more flexible basis selection restricts the adaptability of the Hi PPO framework. In this work, we address this restriction by presenting a generalized method for constructing SSMs using any frame or basis of choice, which we term Sa FARi (SSMs for Frame-Agnostic Representation). Our approach extends and generalizes the Hi PPO framework using a numerical (as opposed to closed-form) method, which enables us to relax the requirements on the basis, such as orthogonality of the components. Our key contributions are as follows: Generalized SSM construction: We present Sa FARi, a frame-agnostic method for deriving SSMs associated with any basis or frame, generalizing the Hi PPO framework to a broader class of function representations. Error Analysis: We provide a comprehensive discussion of Sa FARi s error sources and derive error bounds, offering theoretical insights into its performance and limitations. This paper is organized as follows. In Section 2, we review the Hi PPO framework and its limitations, motivating the need for a generalized approach. Section 3 provides the required mathematical preliminaries. Section 4 introduces our frame-agnostic method for SSM construction, and then Sections 5 and 6 address implementation considerations and strategies for Sa FARi, including error analysis and computational efficiency. Section 7 demonstrates empirical validation of the method and the theoretical claims presented in this paper. Finally, Section 8 discusses the broader implications of our work and outlines directions for future research. 2 Background Recent advances in machine learning, computer vision, and LLMs have exploited the ability to collect and contextualize more and more data over longer time frames. Handling such sequential data presents three main challenges: 1) generating a compact and low-dimensional representation of the sequence, 2) effectively preserving information within that representation, and 3) enabling real-time updates for streaming data. The classic linear method of obtaining the coefficients of a compressed representation of a signal is through a transform (e.g. Fourier) (Oppenheim, 1999; Abbate et al., 2012; Box et al., 2015; Proakis, 2001; Prandoni & Vetterli, 2008). However, a significant limitation of this method is its inefficiency in handling real-time updates. When new data arrives, the representation must be recalculated in its entirety, necessitating the storage of all prior data, which is inefficient in terms of both computation and storage requirements. This limits the horizon of the captured dependencies within the sequence. Nonlinear models, such as recurrent neural networks (RNNs) and their variants, have been introduced more recently (Elman, 1990; Hochreiter & Schmidhuber, 1997; Cho et al., 2014; Schuster & Paliwal, 1997). Since these learned representations are task-specific, they are not easily utilized for other circumstances or applications. Furthermore, RNNs struggle to capture long-range dependencies due to issues such as vanishing and exploding gradients. Published in Transactions on Machine Learning Research (10/2025) u(t) + ODE Update c(t) Figure 1: An SSM block-diagram, with the necessary ODE update step included. 2.1 State-space models The state-space representation itself is not new; it was introduced by Kálmán (1960) via the eponymous Kalman Filter. For an input u(t), output y(t), and a state representation called x(t), many systems and their properties can be described and controlled with the following system of ODEs, illustrated in Fig. 1: x(t) = Ax(t) + Bu(t) y(t) = Cx(t) + Du(t). (1) In many classic applications, we iteratively update the matrices A, B, C, and D to control or predict the output y based on previous values of u. For online function approximation, however, we instead define the matrices A and B such that they iteratively update a vector of coefficients c over a particular basis. For the moment, we can ignore C and D (or, equivalently, consider C to be an identity matrix and D = 0). For stability, A must have only negative eigenvalues, so we explicitly include a negative sign here. A and B may or may not be constant over time, so for completeness, we call these A(t), B(t). Eq. 1 is now c = A(t)c(t) + B(t)u(t). (2) The challenge in the problem of the approximation of online functions is to derive appropriate matrices A and B. Legendre Memory Units (LMUs) were proposed for this task by (Voelker et al., 2019), and subsequently expanded upon by (Gu et al., 2020). We note that the term SSM in machine learning literature in recent years has become a synecdoche, referring both to the classical A, B matrix structure described here, as well as the larger models that utilize these SSMs as a layer, such as S4 Gu et al. (2022b). For the purposes of this paper, we disambiguate these structures, and we will use the term SSM to describe only the former. 2.2 Hi PPO: High-order Polynomial Projection Operators The LMU and Hi PPO frameworks (Voelker et al., 2019; Gu et al., 2020) enable online function approximation using pre-determined SSM parameters derived from a basis of orthogonal polynomials G. It optimizes u T g(T )(t) µ for g(T )(t) G, to find a set of coefficients for the orthogonal polynomials at every time T, which yields the approximation of the input function over [0, T]. In addition to choosing the set of polynomials, one must also select a measure, µ: the combination of a weighting function and a windowing function. The window indicates which samples are included, and the weighting assigns a weight to each sample. Hi PPO (Gu et al., 2020) considered several possible measures, two of which (illustrated in Fig. 2) are: θ1t [T θ,T ], µsc(t) = 1 T 1t [0,T ]. (3) The uniform translated measure, µtr(t), gives equal weight to all samples in the most recent window with a constant length θ, and zero weight to previous samples. The uniform scaled measure, µsc(t), gives equal weight to all the times observed since t = 0. This can be interpreted as squeezing or stretching the basis or frame to match the current length of the signal at any given time. Thus, the representation produced by this measure becomes less informative about the finer details of the signal as more history is observed, since Published in Transactions on Machine Learning Research (10/2025) the stretching of the basis will gradually increase the lowest observable frequency. This work considers only uniformly weighted measures, for the sake of clarity in our derivations and examples. However, one could implement any desired weighting scheme by applying any function of T in place of 1 T in Eq. 3. For the space of orthogonal polynomials, the Hi PPO ODE update (Gu et al., 2020) (see Fig. 1) follows a first-order linear differential equation: d d T c(T) = A(T ) c(T) + B(T )u(T) (4) where u(T) is the value of the input signal at the current time T, and c(T) is the vector containing the representation of the t [0, T] part of the input signal. A(T ) and B(T ) are also a time-varying matrix and vector that can be derived for the particular choice of the measure and orthogonal polynomial. Solving the differential equation to update the SSM: The differential equation (Eq. 4) can be solved incrementally, and there are several methods to do so Butcher (1987). We use the generalized bilinear transform (GBT) (Zhang et al., 2007), relying on findings from Gu et al. (2020) that the GBT produces the best numerical error in solving first-order SSMs. The GBT update rule is c(t + t) = (I + δtαA(t+δt)) 1 (I δt(1 α)A(t))c(t) + (I + δtαA(t+δt)) 1 δt B(t)u(t), (5) where 0 α 1 is a chosen constant. (For α = 0, the update rule becomes the forward Euler method.) Diagonalizing the transition matrix A: The incremental update given by the GBT requires a matrix inversion and matrix products at each increment. Having a diagonal A significantly reduces the computational cost of these operations. If the measure used for the SSM is such that the eigenvectors of A(t) remain constant (for example, if A(t) is a constant matrix multiplied by an arbitrary function of time), then it is possible to find a change of basis that makes the SSM matrix diagonal. To do this, one finds the eigenvalue decomposition of the matrix A(t) = V Λ(t)V 1 and re-write the SSM as tec = Λ(t)ec + e Bu(t), (6) where ec = V 1c and e B = V 1B(t). This means that one can solve Eq. 6 instead of Eq. 4, then find the representation c from ec with a single matrix multiplication. 2.3 Limitations of Hi PPO The original Hi PPO formulation and a subsequent follow-up (Gu et al., 2023) included a handful of orthogonal polynomial bases with specific measures. There is no method in the literature for arbitrary choice of basis and measure (see Table 1). 1S4 technically applies an exponential measure in the ODE; however, the A matrix of the SSM is generated based on a uniformly weighted scaled Legendre polynomial basis. 2Some implementations in the translated case also apply an exponential decay measure in order to ensure orthogonality. t0 t1 t2 t T Scaled (µsc) t0 t1 t2 t T t0 t1 t2 t T t0 t1 t2 t T Translated (µtr) t0 t1 t2 t T t0 t1 t2 t T Figure 2: Two different uniform measures (red) applied to a signal (blue). The red shaded area demonstrates how the measure changes as time evolves and more samples of the input are observed. Published in Transactions on Machine Learning Research (10/2025) Measure Scaled Translated Basis or Frame Legendre Gu et al. (2020), 4 Gu et al. (2020), 4 Gu et al. (2023)1 Fourier Gu et al. (2020), 4 Gu et al. (2020), 4 Laguerre 4 Gu et al. (2020)2, 4 Chebychev 4 Gu et al. (2020)2 4 Arbitrary 4 4 Table 1: An overview of the combinations of frames (or bases) and measures covered in the literature to date. Sa FARi fills in missing combinations in this table for the scaled and translated measures in Section 4. Performance of various bases were shown to be strongly task-dependent. Only one basis-measure combination, the scaled-Legendre (Leg S), performed empirically well across most tasks Gu et al. (2020), but it introduced additional challenges as its A matrix is not stably diagonalizable (Gu et al., 2022a). The majority of the follow-up work since Hi PPO has abandoned the task of function approximation by SSMs alone. Instead, most research has employed a diagonal approximation of the best-performing extant version (Leg S) as an initialization for machine learning architectures such as S4 (Gu et al., 2022b; Smith et al., 2023). Still, the Hi PPO framework still holds untapped potential for online function approximation and better initializations for learning tasks. 3 Mathematical Preliminaries Prior to introducing Sa FARi, we first cover some required theoretical background on the use of frames for function representation and approximation. This section will cover distinctions between approximations performed on the full signal all at once, and approximation performed in a sequential (online) fashion. 3.1 Function representation using frames Given a function u(t) of a time domain t DT , we aim to represent u(t) using a collection of functions over DT . This representation is performed with respect to a measure that determines the relative importance of the function s value at different times. We formulate the task of function representation using frames via the definitions below. Frame: The sequence Φ = {ϕn}n Γ is a frame in the Hilbert space H if there exist two constants A > 0 and B > 0 such that for any f H: Aframe f 2 X n Γ | f, ϕn |2 Bframe f 2 (7) where , is the inner product of the Hilbert space H, and Γ denotes the indices of the frame elements in Φ. If Aframe = Bframe, then the frame is said to be tight (Mallat, 2008; Christensen, 2003; Gröchenig, 2001). Frame operator: If the sequence Φ = {ϕn}n Γ is a frame, then its associated operator UΦ is defined as: UΦf = c, cn = f, ϕn , n Γ. (8) It can be shown that the frame condition (Eq. 7) is necessary and sufficient for the frame operator (Eq. 8) to be invertible on its image with a bounded inverse. This makes the frame a complete and stable (though potentially redundant) framework for signal representation. Dual frame: The set eΦ = {eϕn}n Γ is the dual of the frame Φ, eΦ = Dual{Φ} if: eϕn = (U ΦUΦ) 1ϕn (9) where U Φ is the adjoint of the frame operator: UΦf, c = f, U Φc . Published in Transactions on Machine Learning Research (10/2025) It can be shown that the composition of the dual frame and the frame is the identity operator in the Hilbert space H: U e ΦUΦf = f (Mallat, 2008; Christensen, 2003; Gröchenig, 2001). Thus, we can think of the dual frame as the operator that transforms the signal from frame representation back into the signal space. Function approximation using frames: The compressive power of frame-based function approximation lies in its ability to efficiently represent functions using a relatively small number of frame elements. Different classes of functions exhibit varying effectiveness in capturing the essential features of different signal classes. This efficiency is closely tied to the decay rate of frame coefficients, which can differ significantly between frames for a given input function class. As a result, selecting an appropriate frame is critical for optimal approximations while minimizing computational resources and storage space. In the task of online function representation, we aim to represent a function u T := {u(t), t [0, T]} for any time T, using a frame Φ that has the dual eΦ and the domain t D. To do so, we need an invertible warping function so that the composition of the frame and the warping function Φ(T ) includes the domain [0, T]. Without loss of generality, we assume that the frame has the domain DΦ = [0, 1], and use the warping ϕ(T )(t) = ϕ t T . (If this is not the case, then one can easily find another warping function to warp DΦ [0, 1] and apply it beforehand.) To calculate the representation, the frame operator of ΦT acts on u T : Projection: c = Φ(T )u T , cn = u T , ϕ(T ) n = Z T t=0 u(t)ϕ(T ) n (t)µ(t)dt, (10) where ϕ(t) is the complex conjugate of ϕ(t). Then, the dual frame operator transforms the discrete representation back to the function u T : Reconstruction: u T = eΦ(T )Φ(T )u T , u T (t) = X n Γ u T , ϕ(T ) n eϕ(T ) n (t) . (11) Measures: Throughout this paper, we work with the Hilbert space consisting of all functions over the domain D with µ(t) representing a measure as in Eq. 3, and the inner product defined as: t D u(t)v(t)µ(t)dt. (12) We will consider both scaling and translating windows with uniformly weighted measures as described in Eq. 3 and Fig. 2, and derive the SSMs for each. We only implement uniform weighting schemes in this work for the sake of clarity and generalizability, as the weighting does not impact the SSM s derivation. The windowing function does impact the derivation, however, because it changes the limits of integration, and so we provide frameworks for both scaled (with a changing number of prior inputs) and translated (with a constant number of prior inputs). Given the combination of these generic models, one could create any arbitrary measure by combining the appropriate window with any desired weighting scheme as a function of t in Eq. 12. We now introduce a generalized method for online function representation using arbitrary frames. We formulate the update rule for online approximation as a first-order linear differential equation following a state-space model, and provide templates for the uniform translated and scaled measures. 4.1 Formulation For the given frame Φ (without loss of generality, we assume Φ has the domain D = [0, 1]), and an input function u(T), the objective is to find the representation of the function using the frame Φ similar to Eq. 10: Projection: cn(T) = (UΦu)n = Z T t=t0 u(t)ϕ(T ) n (t)µ(t)dt . (13) The representation vector c(T) is a vector with its nth component defined as above. We next find the derivative of the representation with respect to the current time T, and show that it follows a particular SSM for a scaled or translated measure. Published in Transactions on Machine Learning Research (10/2025) 4.2 Uniform scaled measure: µsc When we use the scaled measure (µsc(t) in Eq. 3), the representation generated by applying the frame operator to the observed history of the input u T (t) is cn(T) = Z T T dt . (14) Definition 1. For a given frame Φ consisting of functions on D, we define the auxiliary derivative of the frame ΥΦ = {υn}n Γ as a collection having υn = t The auxiliary derivative of the frame is the result of the operator t t acting on each individual frame element. Note that the auxiliary derivative of the frame is not necessarily a frame itself and the frame condition in Eq. 7 does not necessarily hold for ΥΦ. Theorem 1. For the representation defined in Eq. 14, the partial derivative of c with respect to T is T A c(T) + 1 T Bu(T), A = I + UΥU e Φ (15) where B is the complex conjugate of a vector containing members of the main frame evaluated at T = 1 so that B = {ϕn(T = 1)}n Γ. The A operator can also be described as a matrix Ai,j = δi,j + Z 1 t=t eϕj(t )dt . (16) Proof is provided in the Appendix A.1. We will refer to the SSM with a scaling measure as scaled-Sa FARi. 4.3 Uniform translated measure: µtr When the translated measure is used (µtr(t) in Eq. 3), the representation resulting from applying the frame operator to the window of recently observed input u T (t) is cn(T) = Z T t=T θ u(t)ϕn Definition 2. For a given frame Φ = {ϕn}n Γ consisting of functions of time, we define the time derivative of the frame Φ = { ϕn = tϕn(t)}n Γ as a collection of time derivatives of ϕn components. Definition 3. For a given frame Φ = {ϕn}n Γ consisting of functions of time, the zero-time frame operator is similar to the frame operator but only acts on t = 0 instead of the integral over the entire domain: QΦf = x, xn = f(t = 0)ϕn(t = 0) . (18) Theorem 2. For the representation defined in Eq. 14, the partial derivative of c with respect to T is θA c(T) + 1 θBu(T), A = U ΦU e Φ + QΦQ e Φ (19) where B is the complex conjugate of a vector containing members of the main frame evaluated at T = 1 so we have B = {ϕn(T = 1)}n Γ. The A operator can also be described using the matrix representation Ai,j = ϕi(0)eϕj(0) + Z 1 t=t eϕj(t )dt . (20) Proof is provided in Appendix A.2. We will refer to the SSM with a scaling measure as translated-Sa FARi. Published in Transactions on Machine Learning Research (10/2025) L = 1000 Hi PPO-Leg S A Matrix, N = 10 SAFARi-Leg S A Matrix Sub-Block, N = 10 Figure 3: Hi PPO provides a closed-form solution for the scaled Legendre (Leg S) SSM. Sa FARi provides a computed solution, where the accuracy depends on the discretization of the N L frame. Larger L gives a finer discretization of the basis vectors and thus a better numerical result. 4.4 Sa FARi as generalization of Hi PPO Hi PPO provides exact, closed-form solutions for A and B for a few specific basis and measure combinations. Sa FARi replicates these A and B matrices to within some numerical error caused by discretization of the frame vectors into length L. Increasing L will provide matrices that converge toward the closed-form solution (see Fig. 3). When a closed-form solution exists for the desired basis and measure (e.g., Hi PPO-Leg S for the scaled Legendre basis), then it is preferable to use it Gu et al. (2020; 2023).1. Sa FARi provides a method for any basis/measure where the closed-form solution might not exist. We also show that Sa FARi preserves all of Hi PPO s robustness to timescale when applied to a general frame in Appendix A.3. While the numerical method described above could be applied to any differentiable set of vectors, we require that the vectors form a frame. If not, then projecting the input signal onto the given vectors is lossy and not invertible. More precisely, the frame condition (Eq. 7) is necessary and sufficient for the frame operator to be invertible on its image with a bounded inverse. This makes the frame a complete and stable (though potentially redundant) framework for signal representation. Second, and most importantly for this work, if ϕ does not meet the frame condition, then Eq. 9 does not hold, and the derivative of representation with respect to time cannot be calculated using only the current hidden state. 5 Error analysis This section describes the computational efficiency and accuracy concerns of Sa FARi, including strategies for producing the finite-dimensional approximation of the complete infinite-dimensional Sa FARi in Section 5.1. We analyze the errors introduced by these approximations in Section 5.2. 5.1 Truncation of frames Section 4 demonstrates how a particular SSM can be built from an arbitrary frame Φ. Since the input space for the SSM is the class of functions of time, no Φ with a finite number of elements can meet the frame condition (Eq. 7), since the true representation of the input signal is infinite-dimensional. In practice, the representation reduces to the truncated representation. In this section, we analyze the theoretical implications of truncated representation using Sa FARi. 5.1.1 Finite-dimensional approximation of Sa FARi In the finite-dimensional case, we will use only N elements of a frame. Partial representation of size N requires that the resulting representation approximates the infinite-dimensional representation. We call the SSM with its c having N coefficients Sa FARi(N). 1Note that Hi PPO used the convention of absorbing a negative sign from the ODE in Eq. 4 into the A matrix, whereas we do not. Published in Transactions on Machine Learning Research (10/2025) Definition 4. A Sa FARi(N) sequence is a sequence of the pairs [A(N), B(N)] where A(N) CN N and B(N) CN such that sequence converges to [A, B] of Sa FARi as lim N A(N) = A, lim N B(N) = B. (21) where convergence for A is the Strong Operator Topology (SOT) convergence, and convergence for B is the vector norm-2 convergence. This convergence means any arbitrary precision can be achieved by selecting the appropriate truncation level. See Appendix A.5 for details. Definition 4 is not a constructive definition; that is, it does not uniquely determine [A(N), B(N)]. In fact, there may be many such sequences that converge to [A, B]. Of course, this does not mean that all such sequences would produce equal representation error. For the following section, we assume the convergence and present and analyze two alternate constructions. See Appendix A.5 for details. 5.1.2 Truncation of dual (To D) The To D construct of a Sa FARi(N) sequence begins with the infinite-dimensional A and B, then truncates to size N as A(N) = A[0:N,0:N], B(N) = B[0:N]. (22) This construction results in a sequence that approximates the infinite-dimensional Sa FARi according to Definition 4. In practice, we find the truncated A, B via Ai,j as introduced in Eq. 16 and Eq. 20 for i, j < N. Calculating Ai,j requires finding eΦ, the dual of the (infinite-dimensional) frame Φ. For certain families of functions, the dual frame can be found analytically. However, if an analytical dual frame eΦ is not known, then one must use a numerical approximation of the dual frame. In this case, the construction for [A(N), B(N)] involves forming the truncated frame for a much bigger size N2 N, then finding eΦ(N2) = Dual{Φ[0:N2]} numerically as an approximation for the dual frame (eΦ(N2) eΦ). Next, we truncate the approximate dual and use its first N elements as an approximation for size N dual frame in Eq. 16 and Eq. 20. 5.1.3 Dual of truncation (Do T) The To D construction becomes numerically intractable for cases where the dual frame is not analytically known; this motivates the need for an alternate constructor for Sa FARi(N). To construct this sequence 1. Truncate the frame at level N and form Φ(N) = {ϕi}i N. This is not limited to SSMs, but is true for any basis representation of a signal. In Sa FARi, truncation errors correspond to discarding the green shaded region in the A(N) illustrated in Fig. 4. Figure 4: Error types due to frame truncation. Truncation errors arise from energy in coefficients of index n > N, while mixing errors result from energy blending during the operation A c. 5.2.2 Mixing errors Mixing errors arise from error propagation in the SSM update rule. Specifically, this update involves computing A c (as in Eq. 15 and Eq. 19), where the matrix A introduces unwanted mixing between the omitted components (n > N) and the retained components (n N) of the representation. Consequently, errors from the truncated portion of the representation propagate into the non-truncated portion. This is illustrated by the blue shaded regions in Fig. 4. For the operator A and truncation level N, the contaminating part of the operator is Ai,j (i N, j > N). In the case of a translated measure, this mixing error is exacerbated since each update step requires estimating the initial value in that window (recall Eq. 2). A key insight from our analysis below is that mixing errors have two sources: 1. nonzero components in the upper right quadrant of A, and 2. nonzero coefficients in c at indices greater than N. 5.2.3 Mitigating errors Truncation errors can never be eliminated, but may be alleviated by using a frame that exhibits a rapid decay in the energy carried by higher-order levels of representation. To counter mixing errors, we should ensure that values in the upper right quadrant of A are as close to zero as possible, and/or that coefficents of c in the blue region of Fig. 4 are as close to zero as possible. If the matrix A is lower-triangular, then any arbitrary truncation results in the contaminating part of A being zero, which guarantees the second type of error is always zero, regardless of any coefficients in c. This is the case for the Hi PPO-Leg S (scaled Legendre) A matrix, as shown in Fig. 5(a). Indeed, the zero coefficents in the upper right quadrant of A were considered strictly necessary in prior work (Gu et al., 2020; 2023). This restriction explains the continued use of Hi PPO-Leg S in follow-up works, regardless of whether or not scaled Legendre polynomials are an optimal choice for a given application. To summarize, there are two primary concerns when finding an appropriate frame for use in an SSM: (a) Scaled Legendre (b) Translated Legendre (c) Scaled Fourier (d) Translated Fourier Figure 5: Examples of the A matrices for several basis/measure combinations: (a) Scaled Legendre, (b) Translated Legendre, (c) Scaled Fourier, and (d) Translated Fourier. The dense non-zero elements in the upper right of (b) explain its poor performance compared to (a). The numerous small nonzero elements above the diagonal in (c) and (d) contribute to mixing errors over long sequences. Published in Transactions on Machine Learning Research (10/2025) 1. compatibility between the frame and the given class of input signals, and 2. the operator A that results from a given frame has a small contaminating part. Truncation and mixing errors have different sources but are linked. The optimal strategy to reduce both at the same time is to choose a basis that results in a representation where the energy is concentrated in the first N coefficients. This can only be achieved, however, if we have some prior knowledge of the input signal in order to choose the right basis and truncation level. In cases where little is known about the input signal or correct truncation level, it is advisable to instead choose an SSM that is zero in the upper right quadrant, such as Hi PPO-Leg S, as it will inherently negate mixing errors. 5.2.4 Error bound In order to quantify the mixing error, we show that the truncated representation follows the same differential equation with a perturbation defined by the theorem below. Theorem 3. (Poof in Appendix A.6) The truncated representation generated by scaled-Sa FARi follows a differential equation similar to the full representation, with the addition of a perturbation factor: T ϵ(T), (23) where ϵ(T) := u T , ξ , ξ = Υ(eΦΦ I). (24) The mixing error term ϵ(T) cannot be directly calculated; however, we can derive an upper bound for the mixing error and demonstrate that this bound can be made arbitrarily small with appropriate truncation. This generic error bound is not necessarily tight; tighter bounds could be identified for specific instantiations of SSMs for given parameters (frame or basis, measure, dimension, etc.) Theorem 4. If one finds an upper bound such that for all the times before T we have ϵ(T) 2 < ϵm, then the representation error is bound by c(T) 2 < ϵm λ2 i = ϵm A 1 F (25) where λi are the eigenvalues of A, and F indicates the Frobenius norm. See proof in Appendix A.6. 5.2.5 Error analysis for different Sa FARi(N) constructs In Section 5.1.1 we define Sa FARi(N) as a finite dimensional approximation for Sa FARi, and provide two particular constructions, Do T and To D. Armed with the quantification of representation error, we compare these different constructs. We provide Theorem 5 to demonstrate Do T has the optimal representation error between different choices for Sa FARi(N) using the same frame. Theorem 5. Given a frame Φ, the Dual of Truncation (Do T) construct introduced in Section 5.1.3 has optimal representation error when compared to any other Sa FARi(N) construct for the same frame Φ. See proof in Appendix A.6. As established in Theorem 5, Sa FARi(N) SSMs should be constructed with the Do T method. 6 Computational and runtime complexity This section develops the computational methods for obtaining sequence representations with Sa FARi, emphasizing its efficiency in both training and inference phases. We analyze the computational complexity of different update methods, and highlight the benefit of parallel computation with diagonalizable SSMs. These discussions provide a foundation for understanding the scalability of Sa FARi in practical applications. Published in Transactions on Machine Learning Research (10/2025) 6.1 Computational complexity for sequential updates To compute representations using scaled-Sa FARi, the GBT update requires solving an N N linear system at each step, leading to O(N 3L) complexity for a sequence of length L. In contrast, translated-Sa FARi reuses the same inverse across all steps, reducing the complexity to O(N 2L). If the state matrix A is diagonalizable, both scaled and translated variants can be accelerated. The sequence representation e C is computed for the diagonal SSM and transformed back via C = V e C, reducing the complexity to O(NL). On parallel hardware, such diagonalized systems decompose into N independent scalar SSMs, yielding O(L) runtime with sufficient parallel resources. Diagonalizability of an SSM depends on the frame or basis used in its construction. One limitation of Legendrebased SSMs such as Hi PPO is that its A matrix cannot be stably diagonalized, even for representation sizes as small as N = 20 (Gu et al., 2022a), leading to significantly higher cost. To address this limitation, (Gu et al., 2023) proposed a fast sequential update method for Hi PPO-Leg S, claiming O(N) computational complexity and O(1) runtime complexity per update on parallel processors. However, we observe that this method becomes numerically unstable at larger N, as discussed in Appendix A.7.1. To resolve this, we suggest a simple modification: computing lower-degree representations before higher-degree ones. While the modified approach preserves the O(N) overall computational cost, it increases runtime to O(N) per sequential update, as it is no longer parallelizable. Similar to Hi PPO-Leg S for the scaled measure, Hi PPO-Leg T for the translated measure also cannot be stably diagonalized. To our knowledge, Leg T has no analogue for the fast Leg S update method in Gu et al. (2020). Therefore, the computational complexity remains considerably higher than for diagonalizable SSMs. 6.2 Convolution kernels and diagonalization When using an SSM for a recognition or learning task, a training phase is required in which the downstream model is trained on the generated representation. Using sequential updates for training is prohibitively taxing on computational and time resources as the whole sequence is available in the training time. Ideally, we would perform computations of the sequential SSM in parallel. However, this is a challenge since each new update depends on the result of the previous. The authors of Gu et al. (2022b) discussed how to implement a parallel computation algorithm for SSMs produced by Hi PPO. To do so, one unrolls the SSM as: ck = Ak . . . A1B0u0 + + Ak Bk 1uk 1 + Bkuk , c = K u (26) Ai = (I + δtαAi) 1(I δt(1 α)Ai), Bi = (I + δtαAi+1) 1Bi. (27) The convolution kernel K in Eq. 26 removes the sequential dependency, enabling parallel computation on hardware such as GPUs, and significantly reducing training time. 6.3 Runtime complexity 6.3.1 Runtime complexity of scaled-Sa FARi If the SSM is not diagonalizable, then the kernel can still be computed in parallel by framing the problem as a scan over the associative prefix product. Since matrix multiplication is associative, all such prefix products can be computed efficiently using parallel scan algorithms (Blelloch, 1990; Hillis & Jr., 1986). When implemented on parallel hardware such as GPUs, this strategy achieves a time complexity of O(N 3 log L) if enough parallel processors are available. However, if the SSM is diagonalizable, all the matrix products Ak matrices in the matrix products appearing in the kernel expression become diagonal. As a result, the convolution kernel can be calculated with the time complexity of O(N log L). Furthermore, the below theorem suggests how to find the kernel in closed form. Theorem 6. For scaled-Sa FARi, if A is diagonalizable, then the convolution kernel K that computes the representations can be found in closed form. See Appendix A.7 for the closed-form solution. Published in Transactions on Machine Learning Research (10/2025) Name Spans L2 Orthogonal Redundant Category Legendre Orthonormal Basis Laguerre Orthonormal Basis Chebyshev Orthonormal Basis Fourier Orthonormal Basis Bernstein Non-orthogonal Basis Gabor Frame Random Harmonics Neither Fourier+Random Harmonics Frame Table 2: The first four elements in this table were studied in previous works (Gu et al. (2020; 2023)). We reproduce them here for comparison. We also introduce several new variants with characteristics such as non-orthogonality and redundancy, which can now be handled with Sa FARi. As noted above, Hi PPO-Leg S is not diagonalizable, complicating kernel computation. A heuristic method proposed by Gu et al. (2023) enables approximate kernel evaluation with time complexity O(N log L), but it introduces additional computation and lacks a closed-form solution. 6.3.2 Runtime complexity of translated-Sa FARi Similar to the scaled case, all the matrix products Ak . . . Ak m can be computed efficiently using parallel scan algorithms (Blelloch, 1990; Hillis & Jr., 1986). As a result, the convolution kernel K can be computed with an overall time complexity of O(N 2 log(L)) since Ak remains the same for all values of k. Similarly to the scaled case, if the A matrix is diagonalizable, then the A matrices become diagonal. With access to parallel processors, the runtime complexity can be reduced to O(N 2 log(L)). Furthermore, for diagonalizable SSMs, the convolution kernel for the translated measure has a closed form. Theorem 7. For translated-Sa FARi, if A is diagonalizable, then the convolution kernel K that computes the representations can be found in closed form. See Appendix A.7. 7 Empirical validation We demonstrate that Sa FARi can generate SSMs for function approximation over any frame or basis by choosing examples that are non-orthogonal, incomplete, or redundant. We then evaluate Sa FARi-generated state-space models on some sample datasets online function approximation, benchmarking against established baselines. Code to replicate the results of this section, as well as generate SSMs with arbitrary frames is provided at: https://github.com/echbaba/safari-ssm. 7.1 Instantiations with different frames No single frame is universally optimal all input classes. Different signal families exhibit different decay rates in representation error depending on the chosen frame or basis. We instantiate Sa FARi over several sets of functions as in Fig. 6, which may constitute bases, frames, or neither, as summarized in Table 2. Description and further discussion can be found in Appendix A.8. For each function family, we scale the functions to an interval of [0, 1]. 7.2 Diagonalization of A As discussed in Sec. 6.3, not all bases produce an A matrix that is stably diagonalizable. This section does not attempt to provide rules to guarantee a stably diagonalizable A. However, we consider two key metrics that illustrate how the choice of basis and measure can result in an A matrix with more (or less) desirable properties: how sensitive eigendecomposition is to perturbation, and the effective rank of A. Published in Transactions on Machine Learning Research (10/2025) From the Bauer-Fike theorem (Bauer & Fike (1960)), given a matrix A with eigenvalues λi and eigenvectors V , and a perturbed matrix e A = A + E with corresponding eigenvalues eλ, we have that2 |eλ λ| κ(V ) E 2 , (28) where κ(V ) = V 2 V 1 2. This theorem describes a relationship between a matrix perturbation and the impact on its eigenvalues, which has a bound determined by the condition number of the eigenvectors of A. The larger this value, the less numerically stable the operations on A will be. Varying N for different Sa FARi instantiations can result in unstable diagonalization of A matrices, as shown in Fig. 7. κ(V ) gives some information about the stability of the eigenvalue decomposition, but does not address structural issues such as degenerate eigenvalues. To address this, we consider the effective rank, defined by Roy & Vetterli (2007) as the Shannon entropy of the normalized singular values pk. The closer this value is to N, the less redundancy of eigenvalues. erank = exp k=1 pk log pk While these metrics may not account for all possible edge cases, in general, we would like the A matrix to have a very low value for κ(V ) and a high value for erank. Notably, in Fig. 7, several of the orthogonal polynomials of prior work are suboptimal, whereas some of our newly-introduced frames (which can only be constructed via Sa FARi, having no closed-form solution) show improvements in both metrics. 7.3 Datasets for experiments S&P 500: We use the daily S&P 500 index as a broad, large-cap U.S. equities benchmark over the last decade: from August 2015 to August 2025 (Yahoo Finance (2025)). The series consists of end-of-day levels for the price index. Overlapping sequences of 500 samples are collected as different instances of time series to form a dataset, and resampled into 4,000 samples to emulate longer time-series signals. 2The norm in Eq. 28 can be any p-norm, e.g. 1, 2, or . The 2-norm was used here, which denotes the largest singular value of E. Legendre Polynomials A Scaled A Translated Fourier Series A Scaled A Translated Chebyshev Polynomials A Scaled A Translated Random Harmonics A Scaled A Translated Laguerre Polynomials A Scaled A Translated Fourier+Random Frame A Scaled A Translated Bernstein Polynomials A Scaled A Translated Gabor Filters A Scaled A Translated Figure 6: For each of the SSMs instantiated in this work, we show a few elements of the frame, and the resulting A matrices for scaled and translated measures. The scale for the A matrices is the same as in Fig. 5 for consistency. Published in Transactions on Machine Learning Research (10/2025) 10 20 30 40 50 60 70 10 2 N (dimension of A) Condition Number of Eigenvectors of A 10 20 30 40 50 60 70 N (dimension of A) Effective Rank of A Leg S Leg T Fou S Fou T Cheby S Cheby T Lag S Lag T Bern S Bern T Gab S Gab T Figure 7: Comparison of A matrices produced by different frames via Sa FARi. The condition number of eigenvectors of A indicates how sensitive the diagonalization is to perturbations, and the effective rank relates to the distribution of eigenvalues. Our results support the findings of prior work in Gupta et al. (2024); Gu et al. (2022a), noting that at relatively low N, Legendre-based A matrices rapidly become difficult to stably diagonalize, but other frames (such as Fourier and Gabor) do not. M4: The M4 dataset consists of a collection of time-series data across different realms, including economic, financial, industrial, and demographic, at various intervals. 7.3.1 Function memorization and approximation 0 2,000 4,000 6,000 8,000 10,000 2 10 2 Reconstruction by Fourier Basis and Harmonics Signal Fourier Harmonics Fourier+ Harmonics 0 2,000 4,000 6,000 8,000 10,000 2 10 2 Reconstruction by Polynomial Families Signal Laguerre Chebyshev 0 2,000 4,000 6,000 8,000 10,000 2 10 2Reconstruction by Non-orthogonal Frames Signal Gabor Bernstein Figure 8: One sample from the M4 dataset reconstructed by different Sa FARi implementations. These are separated into multiple subfigures for visibility. Members of the M4 and S&P 500 datasets were passed to SSMs constructed from different bases and frames. At each iteration, the SSM generates a coefficient vector c (see Fig. 1, Eq. 13 and subsequent equations) which is a compressed representation of the time-series signal up to that point. The error between the reconstructed signal from c and the original input signal is a measure of how effectively the SSM can remember and represent the signal. An illustration based on a single instance of M4 is shown in Fig. 8. Scaled and translated SSMs: For the translated case, the window size is set at 10% of the input signal length. Since we are only estimating a small portion of the signal each time we evaluate the translated case, the MSE reported is the mean of all intermediate estimates. For consistency, the error in the scaled case is measured the same way. The MSEs between the reconstructed signal and the estimated signal are reported in Table 3. Size and Rank: Both scaled and translated versions were evaluated with N = 32, 64, 128, where N is the size of the signal representation. For orthonomal bases, basis elements are linearly independent, so both basis and A matrix have rank N. For frames, elements have redundancy, so the SSM has effective rank Neff < N. In Table 3, the two frames (Fourier+Random Harmonics and Gabor) are treated separately to highlight the distinction. We first instantiate these frames with N member vectors (Fou+, Gabor). We then add N > N member vectors and diagonalize the resulting A matrix, so that the resulting (Neff) matches the desired N (Fou+*, Gabor*). Published in Transactions on Machine Learning Research (10/2025) Scaled Translated N = 32 N = 64 N = 128 N = 32 N = 64 N = 128 SP M4 SP M4 SP M4 SP M4 SP M4 SP M4 Leg 0.039 0.328 0.039 0.322 0.039 0.322 0.166 0.288 0.166 0.287 0.166 0.289 Fou 0.018 0.218 0.013 0.159 0.009 0.109 0.314 0.271 0.314 0.270 0.314 0.271 Lag 0.678 0.712 0.678 0.709 0.678 0.712 0.313 0.271 0.314 0.207 Cheby 0.015 0.162 0.013 0.117 0.012 0.077 0.021 0.179 0.024 0.176 0.030 0.181 Bern 0.026 0.194 0.020 0.173 0.017 0.158 0.021 0.179 0.021 0.174 0.021 0.178 Rand 0.026 1.175 0.020 0.563 0.014 0.249 0.022 0.179 0.023 0.176 0.026 0.179 Fou+ 0.017 0.191 0.011 0.152 0.009 0.109 0.021 0.179 0.023 0.175 0.026 0.180 Fou+* 0.013 0.156 0.009 0.101 0.007 0.058 0.022 0.179 0.025 0.176 0.027 0.181 Gabor 0.024 0.266 0.018 0.191 0.714 0.206 0.021 0.179 0.022 0.175 0.023 0.179 Gabor* 0.018 0.223 0.016 0.181 0.051 0.144 0.022 0.179 0.022 0.175 0.024 0.179 Table 3: MSE of reconstructed signals with different instantiations of Sa FARi. The table is divided into polynomial (top) and non-polynomial (bottom) representations. Asterisks indicate that the frame was augmented with additional vector elements to achieve an effective rank. For each test, the minimum MSE for polynomial and non-polynomial SSMs is shaded. Missing entries for Laguerre could not be computed due to numerical errors arising from exponents in higher-order terms. The standard deviation is reported in Appendix A.9. Discussion: In Table 3, the lowest MSE depends on a combination of N, the windowing function, and the data being processed. However, some broad patterns emerge. Most of the best-performing SSMs have characteristics of redundancy (Fou+, Fou+*, Gabor, Gabor*) or non-orthogonality (Bernstein), which were never explored in prior work due to their lack of a closed-form solution for A and B. This suggests that alternative SSMs constructed via Sa FARi could improve the performance of SSM-based models. We also observe that the representative power of a given SSM depends in part on the length of the signal under consideration. In the translated case, the window size is constant, so any SSM that can adequately model features in a signal of length W will have negligible performance advantages. In the scaled case, however, the length of the signal changes at each iteration, and the choice of frame or basis for SSM can have a significant impact on MSE. Ultimately, the choice of basis or frame is task-dependent, and different representation structures will perform better or worse, depending on the signal s features. No single SSM is universally optimal for all signal types. However, given our observations regarding redundancy and non-orthogonality, wavelets are a natural choice. It was found in Babaei et al. (2025) that Daubechies wavelets can generally outperform polynomial-based Hi PPO SSMs for time-series signals, especially when those signals contain discontinuities or other non-idealities. 7.3.2 Comparing SSMs to learned structures in the long horizon delay task To compare the performance of Sa FARi SSMs as online memory units, Sa FARi features are computed on a signal by sequential update, then the observed scalar input at t = T d (where T is the current time sample and d is the delay) earlier is reconstructed from the representation via the dual frame. The MSE between the reconstruction and the ground truth delayed input is an indicator of the memory performance of the structure, as it demonstrates that information in the long history horizons is recoverable. For each test, we use the same delay amount and the same representation size of 32. The SSM-based models in our study do not include any learned parameters; their dynamics are fully determined by the state-space formulation, so their number of learnable parameters is effectively zero. The LSTM and GRU models are trained using an Adam optimizer (Kingma (2014)) until they converge, and the final validation performance is shown in Fig. 9. Discussion: The trends in Fig. 9 reflect how SSM-based models demonstrate better memory performance than learnable models for the same representation size. Models with learnable parameters (LSTMs, GRUs) Published in Transactions on Machine Learning Research (10/2025) 0 500 1,000 1,500 2,000 2,500 3,000 3,500 S&P 500 Dataset 500 1,000 1,500 2,000 10 2.4 10 2.2 10 2 GRU Sa FARi-Chebyshev Sa FARi-Gabor Hi PPO-Legendre Hi PPO-Fourier Figure 9: Memory recall performance of learned models, Hi PPO SSMs, and Sa FARi SSMs in the long horizon delay task. Legendre and Chebyshev SSMs are nearly indistinguishable. Gabor and Fourier also track closely. need large amounts of data to generalize well. When the delay horizon is long and the available data is limited, these models tend to perform worse. This is consistent with our observation that the gap between RNNs and SSMs is smaller for the M4 dataset, which has many more samples, compared to the S&P 500 dataset. Even in data-rich settings, RNNs prioritize short-term dependencies because of their internal gating mechanisms and training dynamics, which helps avoid vanishing or exploding gradients. This bias limits their ability to retain information from long time horizons. In contrast, SSM models are designed to maintain information flow over longer sequences. Another notable trend in the plots of Fig 9 is the varying MSE as the delay increases. Intuitively, at longer delays, the memorization task becomes more difficult, and we expect the MSE to increase. The MSE does increase overall, but not strictly monotonically, and this is especially salient for the SSM models. If the underlying time series exhibits non-monotonic autocorrelation structure, then periodic or quasi-periodic components cause the signal at specific lags to be more correlated (and thus more predictable) than at nearby lags, which is often the case for real-world signals Kantz & Schreiber (2003). SSM models also contain structured functions that exhibit periodicity (see Fig. 6), whereas learned models do not necessarily converge to periodic representations. Thus, the same structure that gives SSMs an advantage in MSE performance can also make them more susceptible to periodic correlation in delay and copying tasks. 8 Conclusions In this work, we demonstrate how our method, Sa FARi, generalizes Hi PPO (Gu et al. (2020)) to accommodate bases or frames where closed-form solutions are unavailable, paving the way for broader applicability in sequential modeling tasks. Key findings of this work regarding the choice of frame or basis of an SSM motivate the need for more flexibility than prior methods could provide: The underlying frame of the SSM should be compatible with the signal of interest; there is no one size fits all . The correct choice of frame reduces both the error and representation size (Sec. 7). Notably, SSMs using non-orthogonal and redundant frames which do not have closed-form solutions can outperform the standard polynomial-based instances (Sec. 7). Even for an optimal frame, an SSM s performance will also depend on structural features of the A matrix (Sec. 5.2). Different frames result in A matrices with better or worse numerical properties (Sec. 7.2), and stable diagonalization of the A matrix is critical for computational efficiency (Sec. 6.3). The groundwork laid in this paper lends itself naturally to several future research directions. One considers Sa FARi as a standalone representation module. While we have presented several new frames and bases in this work, there are many other structured frames, such as different wavelet types, that could be leveraged for localized and sparse representations. Published in Transactions on Machine Learning Research (10/2025) Another direction focuses on the integration of Sa FARi into learned models, specifically advanced SSM architectures such as S4 and Mamba. By embedding known temporal structures of dynamical systems directly into the SSM architecture with Sa FARi, we can improve memory and reconstruction performance of the core SSM, which in turn reduces the need to learn all parameters from scratch. We anticipate that this will reduce the computational burden of training, while increasing stability and parameter efficiency. Sa FARi is also a natural candidate for implementing gating-like behavior: the memory encoded in the state vector determines how historical information is selectively integrated into downstream computations. In this sense, Sa FARi modules could function analogously to gating layers in RNNs, but with an efficiently designed, context-aware structure. Sa FARi also enables us to investigate the synergy between ML models and SSM variants. The interplay between the internal structure of the SSM (from the frame and measure) and the architecture of the learned model that contains it may play an important role in overall performance. For example, a wavelet-based SSM that detects discontinuities in signals may have better synergy with models designed for sparse data, whereas polynomial-based SSMs techniques may be better suited for use with techniques involving logistic regression. By extending SSM construction beyond specific bases, Sa FARi provides a flexible foundation for efficient state-space representations by linking theoretical advancements to practical applications in sequential data processing. Acknowledgments The authors thank T. Mitchell Roddenberry for fruitful conversations over the course of this project, and Matt Ziemann for assistance with code development. This work was supported by NSF grants CCF-1911094, IIS-1838177, and IIS-1730574; ONR grants N00014-18-12571, N00014-20-1-2534, N00014-18-1-2047; MURI N00014-20-1-2787; AFOSR grant FA9550-22-1-0060; and DOE grant DE-SC0020345. Additional support came from a Vannevar Bush Faculty Fellowship, the Rice Academy of Fellows, and the Rice University and Houston Methodist 2024 Seed Grant Program. Agostino Abbate, Casimer De Cusatis, and Pankaj K Das. Wavelets and Subbands: Fundamentals and Applications. Springer, 2012. Sina Alemohammad, Hossein Babaei, Randall Balestriero, Matt Y. Cheung, Ahmed Imtiaz Humayun, Daniel Le Jeune, Naiming Liu, Lorenzo Luzi, Jasper Tan, Zichao Wang, and Richard G. Baraniuk. Wearing a mask: Compressed representations of variable-length sequences using recurrent neural tangent kernels. In 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 2950 2954, 2021. Hossein Babaei, Mel White, Sina Alemohammed, and Richard Baraniuk. Wa LRUS: Wavelets for long range representation using state space methods. In The Thirty-ninth Annual Conference on Neural Information Processing Systems, 2025. Friedrich L. Bauer and C. T. Fike. Norms and exclusion theorems. Numerische Mathematik, 2:137 141, 1960. Guy E. Blelloch. Prefix sums and their applications. Technical Report CMU-CS-90-190, School of Computer Science, Carnegie Mellon University, 1990. George EP Box, Gwilym M Jenkins, Gregory C Reinsel, and Greta M Ljung. Time Series Analysis: Forecasting and Control. John Wiley & Sons, 2015. John Charles Butcher. The Numerical Analysis of Ordinary Differential Equations: Runge-Kutta and General Linear Methods. Wiley-Interscience, 1987. Kyunghyun Cho, Bart Van Merriënboer, Caglar Gulcehre, Dzmitry Bahdanau, Fethi Bougares, Holger Schwenk, and Yoshua Bengio. Learning phrase representations using RNN encoder-decoder for statistical machine translation. ar Xiv preprint ar Xiv:1406.1078, 2014. Published in Transactions on Machine Learning Research (10/2025) O Christensen. An Introduction to Frames and Riesz Bases. Birkhauser, 2003. Jeffrey L. Elman. Finding structure in time. Cognitive Science, 14(2):179 211, 1990. Alex Graves and Jürgen Schmidhuber. Framewise phoneme classification with bidirectional LSTM and other neural network architectures. Neural Networks, 18(5):602 610, 2005. Karlheinz Gröchenig. Foundations of Time-Frequency Analysis. Springer, 2001. Albert Gu and Tri Dao. Mamba: Linear-time sequence modeling with selective state spaces. ar Xiv preprint ar Xiv:2312.00752, 2023. Albert Gu, Tri Dao, Stefano Ermon, Atri Rudra, and Christopher Ré. Hi PPO: Recurrent memory with optimal polynomial projections. In Advances in Neural Information Processing Systems, 2020. Albert Gu, Karan Goel, Ankit Gupta, and Christopher Ré. On the parameterization and initialization of diagonal state space models. In Advances in Neural Information Processing Systems, 2022a. Albert Gu, Karan Goel, and Christopher Re. Efficiently modeling long sequences with structured state spaces. In International Conference on Learning Representations, 2022b. Albert Gu, Isys Johnson, Aman Timalsina, Atri Rudra, and Christopher Re. How to train your Hi PPO: State space models with generalized orthogonal basis projections. In International Conference on Learning Representations, 2023. Ankit Gupta, Albert Gu, and Jonathan Berant. Diagonal state spaces are as effective as structured state spaces. In Advances in Neural Information Processing Systems, 2024. Ramin Hasani, Mathias Lechner, Tsun-Hsuan Wang, Makram Chahine, Alexander Amini, and Daniela Rus. Liquid structural state-space models. In International Conference on Learning Representations, 2023. W. Daniel Hillis and Guy L. Steele Jr. Data parallel algorithms. Communications of the ACM, 29(12): 1170 1183, 1986. Sepp Hochreiter and Jürgen Schmidhuber. Long short-term memory. Neural Computation, 9(8):1735 1780, 1997. Rudolf E. Kálmán. A new approach to linear filtering and prediction problems. J. Basic Eng., pp. 35 46, March 1960. Holger Kantz and Thomas Schreiber. Nonlinear Time Series Analysis. Cambridge University Press, 2 edition, 2003. Diederik P Kingma. Adam: A method for stochastic optimization. ar Xiv preprint ar Xiv:1412.6980, 2014. Stéphane Mallat. A Wavelet Tour of Signal Processing: The Sparse Way. Academic Press, 3rd edition, 2008. Eric Nguyen, Karan Goel, Albert Gu, Gordon Downs, Preey Shah, Tri Dao, Stephen Baccus, and Christopher Ré. S4ND: Modeling images and videos as multidimensional signals with state spaces. In Advances in Neural Information Processing Systems, 2022. Alan V Oppenheim. Discrete-Time Signal Processing. Pearson, 1999. Razvan Pascanu, Tomas Mikolov, and Yoshua Bengio. On the difficulty of training recurrent neural networks. In Proceedings of the 30th International Conference on Machine Learning, volume 28, pp. 1310 1318, 17 19 Jun 2013. Paolo Prandoni and Martin Vetterli. Signal Processing for Communications. EPFL Press, 2008. John G Proakis. Digital Signal Processing: Principles, Algorithms, and Applications. Pearson, 2001. Published in Transactions on Machine Learning Research (10/2025) Olivier Roy and Martin Vetterli. The effective rank: A measure of effective dimensionality. In 15th European Signal Processing Conference, pp. 606 610, 2007. Mike Schuster and Kuldip K Paliwal. Bidirectional recurrent neural networks. IEEE Transactions on Signal Processing, 45(11):2673 2681, 1997. Jimmy T.H. Smith, Andrew Warrington, and Scott Linderman. Simplified state space layers for sequence modeling. In International Conference on Learning Representations, 2023. Aaron Voelker, Ivana Kajić, and Chris Eliasmith. Legendre memory units: Continuous-time representation in recurrent neural networks. In Advances in Neural Information Processing Systems, volume 32, 2019. Yahoo Finance. S&P 500 index historical data. https://finance.yahoo.com/quote/%5EGSPC/history, 2025. Accessed: 2025-09-04. Guofeng Zhang, Tongwen Chen, and Xiang Chen. Performance recovery in digital implementation of analogue systems. SIAM Journal on Control and Optimization, 45(6):2207 2223, 2007. Nikola Zubić, Mathias Gehrig, and Davide Scaramuzza. State space models for event cameras. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2024. A.1 Derivation of Sa FARi with the scaled measure Theorem (1). Assuming elements in the frame Φ and the input signal u are right-continuous, for the representation defined in Eq. 14, the partial derivative is: T A c(T) + 1 T Bu(T) (A.1) where A is a linear operator defined as: A = I + UΥU e Φ (A.2) and B is the complex conjugate of a vector containing members of the main frame evaluated at T = 1 so we have B = {ϕn(T = 1)}n Γ . One can show that the A operator can also be described using the matrix multiplication: Ai,j = δi,j + Z 1 υi(t)eϕj(t)dt (A.3) Proof: Taking partial derivative with respect to T, we have: Now applying the Leibniz integral rule we have: T cn(T) = Z T T )dt + ϕn(1) T u, υn + ϕn(1) T u(T) (A.5) This is still not an SSM since the second term is not explicitly a linear form of c(T). To convert this to a linear form of c(T), we use the equality given in Eq. 11 to represent υn using the frame eΦ j Γ υn, ϕj eϕj = X j Γ υn, eϕj ϕj (A.6) Published in Transactions on Machine Learning Research (10/2025) u, υn = u, X j Γ υn, eϕj ϕj = X υn, eϕj u, ϕj = X υn, eϕj cj(T) (A.7) Putting Eq. A.7 in Eq. A.5 results in: T (I + UΥU e Φ) c(T) + ϕn(1) T u(T) (A.8) This proves the theorem. A.2 Derivation of Sa FARi with the translated measure Theorem (2). Assuming elements in the frame Φ and the input signal u are right-continuous, for the representation defined in Eq. 14, the partial derivative is: θA c(T) + 1 θBu(T) (A.9) where A is a linear operator defined as: A = U ΦU e Φ + QΦQ e Φ (A.10) and B is the complex conjugate of a vector containing members of the main frame evaluated at T = 1 so we have B = {ϕn(T = 1)}n Γ. One can show that the A operator can also be described using the matrix multiplication: Ai,j = ϕi(0)eϕj(0) + Z 1 t=t eϕj(t )dt (A.11) Proof: We can write the coefficients as: cn(T) = Z T Taking the partial derivative with respect to T, we have T θ u(t)ϕ n θϕn(1)u(T) 1 θϕn(0)u(T θ) (A.13) Similar to our approach for the previous theorem, we write ϕ n(z) as i Γ ϕ n, eϕi ϕi(z) = X i Γ Qn,i ϕi(z) (A.14) z=0 ϕ n(z)eϕi(z)dz (A.15) Now, if we use this expansion, and put it in Eq. A.13 we have θϕn(1)u(T) 1 θϕn(0)u(T θ) (A.16) θϕn(1)u(T) 1 θϕn(0)u(T θ) (A.17) We also write u(T θ) as a reconstruction using the current representation u(T θ) = P i ci eϕi(0) Qn,i ci + 1 θϕn(1)u(T) 1 i ci eϕi(0) (A.18) i (Qn,i + eϕi(0)ϕn(0)) ci + 1 θϕn(1)u(T) (A.19) If we put Ai,j = Qi,j + ϕi(0)eϕj(0), it proves the theorem. Published in Transactions on Machine Learning Research (10/2025) A.3 Mathematical properties of Sa FARi Proposition. For any scalar β > 0, if h(t) = u(βt) then for the scaled measure we have Sa FARi(h)(T) = Sa FARi(u)(βT) Proof: We start by writing the representation generated by the scaled Sa FARi for h(t). Sa FARi(h)(T) = Z T T dt (A.20) t=0 u(βt)ϕn T dt (A.21) t=0 u(t )ϕn β = Sa FARi(u)(βT) (A.22) Proposition. For any scalar β > 0, if h(t) = u(βt) then for the translated measure with parameter θ we have Sa FARiθ(h)(T) = Sa FARiβθ(u)(βT) Similar to the previous proposition, we start by writing the representation of the scaled Sa FARi: Sa FARiθ(h)(T) = Z T t=T θ h(t)ϕn t=T θ u(βt)ϕn t =βT βθ u(t )ϕn β = Sa FARiβθ(u)(βT) (A.25) A.4 The closed-form solution for Sa FARi differential equations Lemma 1. The closed form solution for the differential equation introduced in Eq. 15 is: t exp A ln t Bu(t)dt. (A.26) Proof: We begin by re-writing the differential equation for any time t t A c(t) = 1 t Bu(t) (A.27) Now we multiply both sides by exp(A ln(t)) exp(A ln(t)) t A exp (A ln(t)) c(t) = 1 t exp(A ln(t))Bu(t) (A.28) The left side of the equality is now a complete differential t (exp(A ln(t)) c(t)) = 1 t exp (A ln(t)) Bu(t) (A.29) exp (A ln(T)) c(T) = Z T t exp (A ln(t)) Bu(t) (A.30) c(T) = exp ( A ln(T)) Z T t exp (A ln(t)) Bu(t) (A.31) This proves Lemma 1. Published in Transactions on Machine Learning Research (10/2025) Lemma 2. The closed form solution for the differential equation introduced in Eq. 19 is: 1 θ exp At T Bu(t). (A.32) Proof: We begin by re-writing the differential equation for any time t as θA c(t) = 1 θBu(t). (A.33) Now we multiply both sides by exp(A t Bu(t) (A.34) The left side of the equality is now a complete differential Bu(t) (A.35) c(T) = exp AT 1 θ exp A t Bu(t) (A.36) This proves Lemma 2. A.5 Truncation of frames For a sequence of matrix operators to converge, there are different types of convergence, including operator norm convergence, Strong Operator Theory (SOT) convergence, and entry-wise convergence. A sequence of operators An is said to converge to A in the Strong Operator convergence sense if An A strongly if x l2: An Pnx Ax 0 (A.37) Where Pn projects the vector x into a vector containing only its first n coordinates. This type of convergence guarantees that for the SSM updates, An c(T) A c(T) 0. This means that the difference between the true infinite dimensional update and the update using truncated A vanishes to have zero norm-2, guaranteeing that the representation error using Sa FARi updates can be diminished to an arbitrarily small value. Our implementations and empirical evidence support that the Do T and To D constructs in Section 5.1 are Sa FARi(N) sequences. The fact that our Do T constructs reproduce the closed-form Hi PPO derivations for A and B also provide strong evidence for this framing of the construction. However, a rigorous proof that that the two introduced structures meet the SOT convergence criteria is still needed, and should be addressed in follow-up work. A.6 Error analysis Theorem (3). The truncated representation generated by the scaled-Sa FARi follows a differential equation similar to the full representation, with the addition of a perturbation factor. T ϵ(T). (A.38) where ϵ is defined as ϵ(T) = u T , ξ (A.39) ξ = Υ(eΦΦ I) (A.40) Published in Transactions on Machine Learning Research (10/2025) Proof: Repeat the steps taken in the proof of Theorem 1 until Eq. A.6. Truncating the frame results in an error in this step which can be written as j Γ υn, eϕj ϕj + ξn(t) (A.41) In fact, this is how ξ is defined. Adding ξ as a correction term here changes the SSM derivation: u, υn = u, ξn + X j Γ υn, eϕj ϕj = X υn, eϕj u, ϕj + u, ξn = X υn, eϕj cj(T) + u, ξn (A.42) Putting Eq. A.42 in Eq. A.5 results in: T ϵ(T). (A.43) Theorem (4). If one finds an upper bound such that t < T we have ϵ(t) 2 < ϵm, then the representation error can be bound by c(T) 2 < ϵm λ2 i = ϵm A 1 F (A.44) Proof: Using the result of Theorem 3: T (Bu(T) ϵ(T)) (A.45) We can use Lemma 1 to find the closed form solution of the perturbed SSM above t exp A ln t (Bu(t) ϵ(t)) dt (A.46) t exp A ln t Bu(t)dt Z T t exp A ln t ϵ(t)dt (A.47) Size N representation = True representation Error (A.48) The last term is indeed the second type error that we have discussed in the error analysis section of the paper. Using eigenvalue decomposition of A = V ΛV 1 we re-write the above error term as Error = V Z T t exp Λ ln t V 1 ϵ(t)dt (A.49) with a change of variable s = ln t V 1Error = Z T t=0 exp(Λs)V 1 ϵ(s)ds (A.50) According to the assumption of this theorem V 1ϵ(t) 2 = ϵ(t) 2 ϵm [V 1Error]j Z T t=0 exp(sλj)ϵmds = ϵm Error 2 = V 1Error 2 ϵm λ2 i = ϵm A 1 F (A.52) Theorem (5). Suppose a frame Φ is given. The Dual of Truncation (Do T) construct introduced in Section 5.1.3 has optimal representation error when compared to any other Sa FARi(N) construct for the same frame Φ. Published in Transactions on Machine Learning Research (10/2025) Proof: The proof for this theorem involves two steps. 1. First, we show that the optimal representation error in the theorem can be reduced to optimal error of the second type (mixing). 2. Then, we demonstrate that for a given frame Φ, and given truncation level N, the construct with the optimal second type error control is Do T. As discussed in Sec. 5.2, the first type of error is due to truncating the frame, and is independent of the SSM. In the scope of this theorem, all the Sa FARi(N) constructs use the same frame and the same truncation. Therefore, comparing the representation error between Sa FARi(N) in the theorem reduces to comparing the mixing error. The mixing error is shown in Theorem 3 to be proportional to ϵ(T) = u T , ξ (A.53) To minimize ϵ(T) irrespective of the input signal, one has to minimize ξ 2 F ξ = Υ[0:N](eΦ(N)Φ[0:N] I) (A.54) Where Υ[0:N] and Φ[0:N] are the first N elements of Υ and Φ. eΦ(N) is the approximation of the dual frame that determines the Sa FARi(N) construction. For the ease of notation, we rewrite Eq. A.54 as: ξ = Υ(eΦΦ I) (A.55) For a fixed Υ, and Φ, we aim to minimize ξ 2 F with respect to eΦ: Argmine Φ Υ(eΦΦ I) 2 F (A.56) For the optimal eΦ, the partial derivative is zero. eΦ Υ(eΦΦ I) 2 F = 2ΥΥT (eΦΦ I)ΦT = 0 (A.57) eΦ = (ΦΦT ) 1ΦT (A.58) One should note that the described eΦ is indeed the pseudo-inverse dual for the truncated frame Φ[0:N]. Therefore, among the possible Sa FARi(N) constructs for the same frame, the Dual of Truncation (Do T) has optimal representation error. A.7 Parallelization using the convolution kernel Theorem (6). For Sa FARi using the scaling measure, if A is diagonalizable, computing the scaled representation on a sequence with L samples can be done using a kernel multiplication. a) For the discretization using General Biliniear Transform (GBT) with parameter α, the kernel can be computed using: QL k=j+1 1 1 α QL k=j 1 + α k+1λi RN L (A.59) b) For long sequences, the kernel K can be approximated using KL(i, j) = 1 λi RN L (A.60) For either case a or b, the representation is computed as: c = MKL u, M = V diag(V 1B) (A.61) where V and λi are the eigenvector matrix and eigenvalues of A. Published in Transactions on Machine Learning Research (10/2025) Proof: a) rewriting the GBT update rule for the diagnoalzied SSM we have: e C[n + 1] = I + α n + 1Λ 1 I 1 α n + 1Λ e C[n] + I + α n + 1Λ 1 e Bu[n] (A.62) = An e C[n] + Bnu[n] e C[L + 1] = BLu[L] + AL BL 1u[L 1] + ... + AL . . . A1 B0u[0] (A.63) for the ease of computation and notation, we define KL such that: KL[i, j] = AL . . . Aj+1 1 + α j + 1λi QL k=j+1 1 1 α QL k=j 1 + α k+1λi (A.64) e Ci = e Bi X j KL[i, j]u[j] = e Bi[KL u]i (A.65) ec = e B (KL u) = diag(e B)KL u (A.66) c = V ec = V diag(e B)KL u = MKL u (A.67) Proof: b) Using Lemma 1 for the diagonalized version of Scaled-Sa FARi the closed form solution is t exp Λ ln t e Bu(t)dt. (A.68) t exp λi ln t e Biu(t)dt = e Bi λi u(t)dt = e Bi[KL u]i (A.69) ec = e B (KL u) = diag(e B)KL u (A.70) c = V ec = V diag(e B)KL u = MKL u (A.71) Theorem (7). For Sa FARi using translated measure with θL samples long sliding window,if A is diagonalizable, computing the translated representation on a sequence with L samples can be done using a kernel multiplication. a) for the discretization using General Bilinear Transform(GBT) with parameter α, the kernel can be computed using: KL[i, j] = 1 1 + α b) For long sequences, the kernel K can be approximated using KL(i, j) = 1 θL exp λi L j RN L (A.73) For either of case a or b, the representation is computed by c = MKL u, M = V diag(V 1B) . (A.74) where V and λi are the eigenvectors matrix, and eigenvalues of A. Proof: a) rewriting the GBT update rule for the diagonalized SSM we have: e C[n + 1] = I + α θ Λ 1 I 1 α θ Λ e C[n] + I + α θ Λ 1 e Bu[n] = A e C[n] + Bu[n] (A.75) Published in Transactions on Machine Learning Research (10/2025) we take a similar approach to the previous theorem. The only difference is that A and B remain the same for all the time indices. e C[L + 1] = Bu[L] + A Bu[L 1] + ... + AL Bu[0] (A.76) for the ease of computation and notation, we define KL such that: KL[i, j] = 1 1 + α e Ci = e Bi X j KL[i, j]u[j] = e Bi[KL u]i (A.78) ec = e B (KL u) = diag(e B)KL u (A.79) c = V ec = V diag(e B)KL u = MKL u (A.80) Proof: b) Using Lemma 2 for the diagonalized version of Translated-Sa FARi, the closed-form solution is: 1 θ exp Λt T e Bu(t). (A.81) 1 θ exp λi t T e Biu(t) = e Bi[KL u]i (A.82) ec = e B (KL u) = diag(e B)KL u (A.83) c = V ec = V diag(e B)KL u = MKL u (A.84) A.7.1 Numerical instability of the fast sequential Leg S solver As part of our experimental findings, we realized that the proposed method for sequential updates for Leg S SSM( in Gu et al. (2020),Appendix E) suffers from numerical instability when working with larger SSMs. x = cumsum(βcumprod 1 α) cumprod 1 where the introduced αk = dk 1+dk is a decreasing function. One can confirm that in the tth iteration, and for the kth degree Legendre polynomial αk = dk 1 + dk = 2(t + 1) k 2(t + 1) + k + 1 (A.86) Then the proposed solution requires finding cumulative product of 1 αk for k [1, N] in each step. log cumprodk 1 k =1 log 1 αk k =1 log 1 + 4(t + 1) + 1 k =1 log 1 + 4(t + 1) + 1 k =2(t+1)+1 log 1 + 4(t + 1) + 1 k =1 log 1 + 4(t + 1) + 1 k =1 log 1 + 4(t + 1) + 1 Published in Transactions on Machine Learning Research (10/2025) Figure 10: Fast Legs numerically diverges. left: For a system with N = 500, log | 1 αk | for different values of k [1, N] is plotted at different iterations t = 20, 40, 60, 80. right: log of the cumulative product which is equal to the cumulative sum of the left plot is plotted for different iterations. In the right plot, it is notable that for t = 80, the cumulative product reaches to 10300 for a k < 500 which is the largest value that a float-64 variable can handle. The studied Fast-Leg S method for an SSM having more then 500 coefficients diverges after only 80 sequential updates. Figure 11: As the given Leg S size N grows, the longest sequence length before observing numerical diversion is given in the above plot. For N < 400 we did not observe the numerical diversion. for N > 400 , the fast leg S sequential update method cannot handle sequences longer than a limited length before it becomes numerically unstable. For any specific iteration (fixed t), as K grows(higher representation index), the second summation above grows to infinity. Thus, for large enough N, cumprod( 1 αk ) diverges beyond machine precision. As a result, the proposed fast sequential leg S solver proposed in (Gu et al. (2020), Appendix E) fails. Figure. 10 Shows an example where for N = 500, fast Leg S numerically diverges for any sequence longer than 80 samples. It is crucial to note that this numerical instability is fundamental to leg S, and does not depend on the input signal at all. We also investigate the longest sequence length that the given fast leg S sequential solution can handle without numerical diversion. Figure 11 shows that as N grows, then length of sequence that fast leg S sequential solver can handle before becoming numerically unstable decrease to a limited length. A stable version of solving Leg S would be a similar approach as fast-Leg S, but in the last step, instead of introducing the proposed α and β, we find x1, then recursively find xi after finding all the pervious xis. This way, the overall computation complexity remains the same, while the run-time complexity increases, as one cannot compute xi until x0, . . . , xi 1 computed. Published in Transactions on Machine Learning Research (10/2025) A.8 Bases and frames Legendre Polynomials: Legendre polynomials form an orthonormal basis for square-integrable functions on the compact interval [ 1, 1]. Their recurrence relations and bounded values make them well-suited for tasks such as Gaussian quadrature, spectral methods for solving differential equations, and polynomial interpolation. However, SSMs built from Legendre polynomials can suffer from numerical instability for high degrees, require global support (making them less effective for localized features), and may produce oscillations near interval boundaries, limiting their practicality in approximating functions with sharp variations or discontinuities. Fourier Series: The familiar Fourier series is an orthonormal basis, and with integer n elements have the form: 2 cos(2πnt), 2 sin(2πnt). The Fourier series represents functions as infinite sums of sines and cosines, which form an orthonormal basis for square-integrable functions on a compact interval, making them highly effective for approximating periodic functions.However, they have important limitations for function approximation: they assume periodicity, and can exhibit the Gibbs phenomenon when approximating non-periodic or sharply varying functions; they provide global support, making them inefficient for capturing localized features; and high-frequency components are often needed to approximate functions with sharp transitions, which can be computationally expensive. Random Harmonics: Similar to the Fourier series, these functions have the form 2 cos(2πat), 2 sin(2πat), where a is a real number sampled from a uniform distribution, not integers. A random collection of these functions is not guaranteed to span L2, and therefore is not a frame or basis. In general, using random vectors in place of a frame or basis is a poor choice. We include it separately only as a counter-example to orthogonal bases and redundant frames. We can also use this set to augment a Fourier basis: since a Fourier basis spans L2, concatenating additional vectors introduces redundancy, and we have a frame. Laguerre Polynomials: Laguerre polynomials form an orthonormal basis for square-integrable functions on the semi-infinite interval [0, ] with respect to the weight function e x. These properties make them well-suited for approximating functions defined on unbounded domains, particularly when the function decays exponentially, and they are widely used in quantum mechanics, numerical analysis, and differential equations. Laguerre polynomials are less effective for approximating functions on compact intervals, as they exhibit increasingly oscillatory behavior for large degrees, and their reliance on a specific weight function limits their flexibility for approximating functions that do not exhibit exponential decay. Chebyshev Polynomials: Chebyshev polynomials form an orthonormal (or orthogonal, depending on normalization) basis for square-integrable functions on the compact interval [ 1, 1] with respect to the weight 1 1 x2 . They are known for their near-minimax property, meaning polynomial approximations using Chebyshev nodes minimize the maximum error. This makes them highly effective in polynomial interpolation, spectral methods for solving differential equations, and numerical approximation schemes that need high accuracy with fewer terms. However, Chebyshev polynomials are global basis functions, so they are less effective for functions with localized features or sharp discontinuities. They also require transformations or rescaling for domains other than [ 1, 1], and like other polynomial bases, they can be inefficient for very high-dimensional or highly irregular function approximations. Bernstein Polynomials: Bernstein polynomials form a basis for continuous functions on the compact interval [0, 1]. They are non-negative, form a partition of unity, and have excellent shape-preserving properties, making them particularly useful in computer graphics, geometric modeling (e.g., Bézier curves), and approximation theory. Bernstein polynomials are not orthogonal, which limits their efficiency in some numerical computations, and achieving high accuracy often requires large polynomial degrees, leading to higher computational cost. They are also less suitable for capturing oscillatory behavior or sharp transitions, as compared to orthogonal polynomial bases like Chebyshev or Legendre. Gabor: Gabor filters (or Morlets) capture localization in both time and frequency, making them popular for use in biomedical signal processing and image texture analysis. They can be constructed by modulating a complex sinusoid with a Gaussian as: fk(x) = e (x bk)2/w2e iak(x bk), where bk are shifts in time, ak are frequencies, and w is the scaling of a Gaussian envelope. To ensure that the resulting set of functions creates an oversampled lattice, we ensure that the time-frequency sampling overlaps; that is, a b < 1. Published in Transactions on Machine Learning Research (10/2025) A.9 Additional Experimental Results Tables 4 and 5 provide the standard deviation for the MSE results in Table 3. Most values are on the same order of magnitude. A notable exception is Rand . Since the collection of random sines and cosines do not form a complete frame or basis, it is a matter of luck how well they represent a particular signal. The more random components are added, the more likely it is that the vectors will support the signal of interest, and shorter signals require fewer components. Since the translated measure has a consistently small window, the variance here is on par with other true bases/frames. The worst case is when the signal is long (scaled measure) and the rank is low (N=32), as expected. N = 32 N = 64 N = 128 SP M4 SP M4 SP M4 Leg 0.0007 0.0062 0.0007 0.0060 0.0007 0.0060 Fou 0.0003 0.0040 0.0002 0.0029 0.0002 0.0020 Lag 0.0012 0.0100 0.0012 0.0100 0.0012 0.0100 Cheby 0.0003 0.0030 0.0003 0.0021 0.0003 0.0014 Bern 0.0005 0.0037 0.0004 0.0032 0.0003 0.0029 Rand 0.0005 0.0258 0.0004 0.0105 0.0003 0.0048 Fou+ 0.0003 0.0036 0.0002 0.0028 0.0002 0.0019 Fou+* 0.0002 0.0029 0.0002 0.0018 0.0001 0.0010 Gabor 0.0004 0.0050 0.0003 0.0033 0.0125 0.0040 Gabor* 0.0003 0.0041 0.0003 0.0035 0.0009 0.0026 Table 4: Standard deviation of reconstructed signals with different instantiations of Sa FARi and a scaled measure. The table is divided into polynomial (top) and non-polynomial (bottom) representations. N = 32 N = 64 N = 128 SP M4 SP M4 SP M4 Leg 0.0029 0.0069 0.0029 0.0068 0.0029 0.0069 Fou 0.0006 0.0052 0.0006 0.0051 0.0006 0.0052 Lag 0.0006 0.0052 0.0006 0.0051 Cheby 0.0006 0.0052 0.0008 0.0052 0.0011 0.0054 Bern 0.0006 0.0052 0.0006 0.0051 0.0006 0.0052 Rand 0.0006 0.0053 0.0007 0.0052 0.0009 0.0053 Fou+ 0.0006 0.0052 0.0007 0.0051 0.0009 0.0053 Fou+* 0.0006 0.0053 0.0008 0.0052 0.0009 0.0054 Gabor 0.0006 0.0052 0.0006 0.0051 0.0007 0.0053 Gabor* 0.0006 0.0053 0.0007 0.0051 0.0007 0.0053 Table 5: Standard deviation of reconstructed signals with different instantiations of Sa FARi and a translated measure. The table is divided into polynomial (top) and non-polynomial (bottom) representations. Missing entries for Laguerre could not be computed due to numerical errors arising from exponents in higher-order terms.