# lurie_networks_with_robust_convergent_dynamics__85706aae.pdf Published in Transactions on Machine Learning Research (04/2025) Lurie Networks with Robust Convergent Dynamics Carl R. Richardson1, 2 cr2g16@soton.ac.uk Matthew C. Turner1 m.c.turner@soton.ac.uk Steve R. Gunn1 srg@ecs.soton.ac.uk 1 School of Electronics and Computer Science, University of Southampton, Southampton, UK 2 The Alan Turing Institute, London, UK Reviewed on Open Review: https: // openreview. net/ forum? id= 3Jm4dbr KGZ The Lurie network is a novel and unifying time-invariant neural ODE. Many existing continuous-time models, including recurrent neural networks and neural oscillators, are special cases of the Lurie network in this context. Mild constraints on the weights and biases of the Lurie network are derived to ensure a generalised concept of stability is guaranteed. This generalised stability measure is that of k-contraction which permits global convergence to a point, line or plane in the neural state-space. This includes global convergence to one of multiple equilibrium points or limit cycles as observed in many dynamical systems including associative and working memory. Weights and biases of the Lurie network, which satisfy the k-contraction constraints, are encoded through unconstrained parametrisations. The novel stability results and parametrisations provide a toolset for training over the space of k-contracting Lurie network s using standard optimisation algorithms. These results are also leveraged to construct and train a graph Lurie network satisfying the same convergence properties. Empirical results show the improvement in prediction accuracy, generalisation and robustness on a range of simulated dynamical systems, when the graph structure and k-contraction conditions are introduced. These results also compare favourably against other well known stability-constrained models and an unconstrained neural ODE. 1 Introduction A Lurie1 system is a class of nonlinear ordinary differential equations (ODE) comprising a linear time-invariant (LTI) component interconnected with a, potentially time-varying, nonlinearity. Such systems are ubiquitous throughout the sciences and engineering, including machine learning (ML) and neuroscience Pauli et al. (2021); Lessard et al. (2016); Lanthaler et al. (2024); Wilson & Cowan (1972). When modelling time-invariant dynamical systems, many ML models including linear state space models (LSSM), recurrent neural networks (RNN) and some graph neural networks (GNN) are special cases of a Lurie system ( 3.1, B.2). Such models have proven to be highly expressive as demonstrated by their successful application on a wide range of tasks such as sequential processing Kozachkov et al. (2022a); Erichson et al. (2020); Chang et al. (2019); Gu et al. (2021), computer vision Erichson et al. (2020); Chang et al. (2019); Gu et al. (2021), language modelling Gu et al. (2021) and computational chemistry Rusch et al. (2022). Consider dynamical systems in neuroscience. The brain organises its representations of the world and carries out complex functions through collective interactions of simpler modules Kandel et al. (2000). More abstractly, this can be viewed as a graph structured dynamical system. Convergent dynamics in widespread regions of the central nervous system are thought to play a crucial role in: forming some of these representations Khona 1Named after Anatolii Isakovich Lurie and sometimes spelt Lur e or Lurye. Published in Transactions on Machine Learning Research (04/2025) & Fiete (2022), processing information over extended periods Vogt et al. (2022), learning Kozachkov et al. (2020); Centorrino et al. (2022), memory storage Hopfield (1982; 1984); Kozachkov et al. (2022b); Pals et al. (2024) and enhancing the robustness of each of these functions Khona & Fiete (2022). As summarised in Khona & Fiete (2022), convergent dynamics in the brain take several forms. For example, neural circuits with multiple equilibrium points (bistable and multi-stable) have been observed in the anterlor lateral motor cortex of a rat Piet et al. (2017) and are conjectured to appear in the mammalian hippocampus and auditory cortex. Some theories of associative memory also believe reconstruction of a learned pattern is obtained by flow to equilibrium points Krotov & Hopfield (2020); Sharma et al. (2022); Kozachkov et al. (2023). Limit cycles are another form of convergent dynamics for which there are numerous examples within the central nervous system. These include working memory Kozachkov et al. (2022b), which is thought to arise from the sustained spiking of neurons Ashwin et al. (2016), and sleep cycle generation Adamantidis et al. (2019). The graph structure and convergent dynamics are also shared with many other dynamical systems such as chemical processes Ofir et al. (2023), opinion dynamics and power systems Ofir et al. (2024). As a result, encoding such structural and dynamical properties as an inductive bias is motivated by neuroscience, for a general ML framework, and for learning models of dynamical systems. The convergence and stability analysis of dynamical systems has been well-studied in the control theory literature. A pertinent example is the absolute stability problem where the nonlinearity of the Lurie system is unknown, but assumed to be sector-bounded or slope-restricted. The goal is to find conditions on the model parameters which ensure the trajectories of all Lurie systems, with nonlinearities in the assumed class, uphold a chosen definition of convergence. Approaches to this problem can be classified as Lyapunov analysis Khalil (2002); Park (2002; 1997), Zames-Falb multipliers Zames & Falb (1968); Turner & Drummond (2019); Carrasco et al. (2016); Drummond et al. (2024) or k-contraction analysis Zhang & Cui (2013); Ofir et al. (2023; 2024). Lyapunov and Zames-Falb multiplier methods are primarily designed to analyse the convergence to an equilibrium point, whereas k-contraction methods analyse a variety of global convergence behaviours including convergence to points, lines and planes. As many ML models are examples of Lurie systems and many activation functions are sector-bounded or slope-restricted (Drummond et al., 2022, Table 1), much of the literature on the absolute stability problem is applicable to systems involving neural networks Pauli et al. (2021); Richardson et al. (2023; 2024); Fazlyab et al. (2020). Although designing networks with convergent dynamics is well motivated, ensuring such a property requires constraints on the network parameters, which can be detrimental if too restrictive. With this in mind, this work focuses on: (i) using k-contraction analysis to derive mild constraints on the weights of the Lurie network which ensure global convergence to a point, line or plane in the neural state space, for all Lurie networks with slope-restricted activation functions; (ii) establishing unconstrained parametrisations of these conditions which allows the Lurie network to be trained using gradient based optimisation algorithms whilst limiting the search space to weights which satisfy the k-contraction conditions; (iii) constructing a graph Lurie network (GLN) from individual Lurie networks and deriving constraints on the graph coupling matrix which ensure the k-contraction property is maintained. Similar unconstrained parametrisations are also derived. We empirically test the difference in prediction accuracy, generalisation and robustness of the unconstrained Lurie network, k-contracting Lurie network and GLN on a range of simulated dynamical systems datasets. We also compare against other well known stability-constrained architectures, which are special cases of the Lurie network, and an unconstrained neural ODE. 2 Preliminaries 2.1 Notation For two integers i < j, we define [i, j] := {i, i + 1, . . . , j}. The set of non-negative real numbers is denoted by ℜ+. Symmetric matrices of dimension n are denoted by Sn with the positive definite subset denoted by Sn +. All other positive definite subsets are denoted by a + subscript. Square diagonal matrices are denoted by Dn and n m diagonal matrices are symbolised by Dnm. A positive definite (semi-definite) matrix P is sometimes indicated by P 0 (P 0). Negative definite (semi-definite) matrices are indicated analogously. The set of n n orthogonal and skew-symmetric matrices are respectively denoted by O(n) and Skew(n). For W ℜn m, the ordered singular values are represented by σ1(W) σmin(n,m)(W) 0 Published in Transactions on Machine Learning Research (04/2025) Figure 1: Trajectories from three dynamical systems satisfying the k-contraction property. Crosses denote the initial condition and stars denote equilibirum points. and for W ℜn n, the ordered eigenvalues are denoted by λ1(W) λn(W). The k-multiplicative and k-additive compound matrices of W (see A.1 for definitions) are respectively denoted by W (k) and W [k]. The Jacobian of a function f(t, x) is denoted by Jf(t, x). The scaled 2-norm of a vector x ℜn with respect to (w.r.t) an invertible scaling matrix Θ ℜn n is defined by |x|2,Θ := |Θx|2, and the matrix measure induced by the scaled 2-norm is µ2,Θ(W) := µ2(ΘWΘ 1) = λ1 ΘWΘ 1 + (ΘWΘ 1) 2.2 k-contraction Analysis In this work, we leverage k-contraction analysis Wu et al. (2022); Muldowney (1990), the geometrical generalisation of contraction analysis Lohmiller & Slotine (1998), as a tool for controlling convergence in the neural state space. Intuitively, k-contraction implies the volume of k-dimensional bodies exponentially converges to zero when governed by the system dynamics. Alternatively, this could be thought of as exponential convergence to a (k 1)-dimensional subspace. When k = 1, this reduces to standard contraction Lohmiller & Slotine (1998), which implies that all trajectories exponentially converge to a single trajectory. For a general time-varying dynamical system, satisfying the k-contraction property does not guarantee stability. However, for time-invariant dynamical systems, it has been shown that for every bounded solution: 1-contraction implies global convergence to a unique equilibrium point Lohmiller & Slotine (1998), 2-contraction implies global convergence to an equilibrium point, which is not necessarily unique but must be connected along a line Muldowney (1990), and 3-contraction, under certain assumptions, implies convergence to a non-unique attractor in a 2d subspace Cecilia et al. (2023). Three examples of k-contracting dynamics are presented in Figure 1. Time-invariant dynamical systems which satisfy the k-contraction property for k {1, 2, 3} have several desirable properties for ML models. They can exhibit a wide range of complex convergent behaviours such as multi-stable and orbitally stable systems Zoboli et al. (2024). This suggests that a model can be expressive whilst satisfying the k-contraction conditions, particularly for higher values of k where the constraints are less restrictive, as highlighted in 3.2. The k-contraction property also implies an inherent robustness as the trajectories can only converge to a finite number of long term behaviours. Next, we present the fundamental k-contraction result from Wu et al. (2022), see A.3 for the formal definition of k-contraction. Theorem 1 Fix k [1, n] and consider the nonlinear system x = f(t, x) with f : ℜ+ ℜn ℜn continuously differentiable. If there exists η > 0 and an invertible matrix Θ ℜn n such that µ2,Θ(k) J[k] f (t, x) η x ℜn and t ℜ+ (1) then the nonlinear system is k-contracting in the 2-norm w.r.t the metric P := Θ Θ. Published in Transactions on Machine Learning Research (04/2025) This result has two features: (i) it requires the existence of an invertible matrix P. In the simplest case, one can expect a solution P = p In to exist. For other systems, such simple solutions will not exist and more general matrices such as P Sn + will be required, making the proofs more difficult; (ii) it requires the use of compound matrices ( A.1). For a matrix W ℜn m, the matrix W [k] with k [1, min(n, m)] will have the size n k m k which is typically much larger and more computationally difficult to work with. A technical introduction to compound matrices, k-contraction analysis and how they relate is presented in A. In 3.2 we derive results which verify (1) for the special case of nonlinear systems of the form (2). 3 Lurie Network A Lurie network is defined by (2) with weights A ℜn n, B ℜn m, C ℜm n and biases bx ℜn, by ℜm. x(t) = Ax(t) + BΦ y(t) + bx y(t) = Cx(t) + by x(0) = x0 (2) The model has a biased linear component interconnected with a nonlinearity of the form Φ(y) := ϕ1(y1) . . . ϕm(ym) where ϕi(yi) is assumed to be slope-restricted with an upper bound g > 0, such that 0 JΦ(y) g Im. This separation of the linear and nonlinear components is useful for analysis. Activation functions which satisfy this slope-restricted assumption include the hyperbolic tangent (tanh) and the rectified linear unit (Re LU). For simplicity, we assume the same scalar nonlinearity is applied element-wise and drop the subscript. The proposed Lurie network is a very general model as highlighted by the special cases in 3.1 and its relationship to deep feedforward neural networks as outlined in B.1. Finally, it is important to observe that the model is time-invariant, this implies that if Theorem 1 is satisfied, the model will inherit the appealing convergence and robustness properties stated in 2.2. This may be viewed as a limitation for modelling neural dynamics since the brain is subject to various external inputs; however, it is common to assume that, at least on the time scale of interest, the dynamics evolve in a time-invariant manner Khona & Fiete (2022). 3.1 Example Lurie Networks When applied to time-invariant dynamical systems, many models from the ML literature become special cases of (2). In this setting, the time-varying external inputs are replaced with trainable biases. A subset of examples are presented next with further examples included in B.2. As the results in 3.2 apply to any model of the form (2), they can also be applied to these special cases. Lipschitz RNN: A stability constrained RNN Erichson et al. (2020) where the parameters A, C are expressed as a weighted sum of symmetric and skew-symmetric terms in order to control the eigenvalues of the Jacobian. The remaining components are B = In, bx = 0 and ϕ( ) tanh( ). A, C {(1 β)(W + W ) + β(W W ) γIn | W ℜn n, 0.5 β 1, γ > 0} Antisymmetric RNN: A constrained RNN Chang et al. (2019) with the same motivations as above. Related to (2) by A = 0, B = In, C {W γIn | W Skew(n), γ > 0}, bx = 0 and ϕ( ) tanh( ). SVD Combo: A 1-contracting graph coupled RNN Kozachkov et al. (2022a) where q denotes the number of individual RNNs and n denotes the state dimension of each RNN. A special case of (2) with A = L a Iqn, C = Iqn and by = 0. The graph coupling matrix is denoted by L and B is block diagonal where each block contains the synaptic weights of the individual RNNs. Both of these matrices are expressed by special parametrised forms to ensure the individual RNNs and graph coupled RNN are 1-contracting. As in (2), the nonlinearity is required to be slope-restricted. 3.2 k-contraction Analysis of Lurie Networks Two sufficient results which satisfy Theorem 1 and guarantee (2) is k-contracting are presented next. Conditions were derived in (Ofir et al., 2024, Theorem 2) which verify Theorem 1 for a Lurie network with A Dn and by = 0. Theorem 2 extends them to account for A ℜn n and by = 0. Refer to C.1 for the proof. Published in Transactions on Machine Learning Research (04/2025) Theorem 2 Consider the Lurie network (2) with Φ(y) := ϕ1(y1) . . . ϕm(ym) being slope-restricted such that 0 JΦ(y) g Im. Fix k [1, n] and define αk := (2k) 1 Pk i=1 λi(A + A ). If αk < 0 and i=1 σ2 i (B)σ2 i (C) < α2 kk (3) then (2) is k-contracting in the 2-norm w.r.t the metric P = α 1 k In. The additional freedom permitted by k-contraction over standard contraction is highlighted by the summation of the eigenvalues and singular values. In 1-contraction, Theorem 2 requires the largest eigenvalue of the symmetric component of A to be negative whereas for k [2, n], this condition on A becomes incrementally more relaxed as k is increased. Equation (3) illustrates a similar relaxation of the constraints on B and C. Theorem 2 has several appealing features: (i) it does not require the computation of the troublesome compound matrices; (ii) it provides a way of embedding the k-contraction property into the structure of a Lurie network based on fairly simple unconstrained parametrisations of the weights, as shown in 4; (iii) the biases are not present in the condition, so are naturally unconstrained; (iv) the result does not rely on symmetries of the parameters. In many Hopfield-based models of associative memory, symmetry in the parameters is needed to make the existence of a global energy function mathematically tractable; however, this simplification is biologically unrealistic and limits the model s expressive power. The limitation of the result is that only Lurie networks which are k-contracting in a scalar metric can be verified. We now present a second result which addresses the scalar metric drawback; however, it comes at the cost of strong constraints on the weights B and C. See C.2 for the proof. Theorem 3 Consider the Lurie network (2) with Φ(y) := ϕ1(y1) . . . ϕn(yn) being slope-restricted such that 0 JΦ(y) g In. Fix k [1, n]. If B Dn, C = B 1 and P (k)A[k] + (A[k]) P (k) + 2kg P (k) 0 (4) then (2) is k-contracting in the 2-norm w.r.t the metric P Dn +. Theorem 3 improves upon Theorem 2 in the sense that Lurie networks which are k-contracting in a diagonal metric can now be verified; however, only when C = B 1. Due to this constraint, Theorem 3 has very limited practical use; however, it may prove to be theoretically insightful for addressing the scalar metric drawback. Finally, it is important to highlight that Theorem 2 and Theorem 3 apply to the class of slope-restricted nonlinearities, so these results address the absolute stability problem for the k-contraction property. 3.3 Graph Lurie Networks Many larger scale dynamical systems such as molecular, social, biological, and financial networks Hamilton et al. (2017) naturally have a graph structure. To make the Lurie network more applicable to these problems, a graph coupling term is introduced to model a set of q interacting Lurie networks. To illustrate this, q independent Lurie networks can be modelled by x(t) = AGx(t) + BGΦ y(t) + bx y(t) = CGx(t) + by x(0) = x0 (5) where the weights are AG := blockdiag(A1, . . . , Aq) ℜqn qn, BG := blockdiag(B1, . . . , Bq) ℜqn qm, CG := blockdiag(C1, . . . , Cq) ℜqm qn and the biases are bx ℜqn, by ℜqm. The graph Lurie network (GLN) is then defined by x(t) = (AG + L)x(t) + BGΦ y(t) + bx y(t) = CGx(t) + by x(0) = x0 (6) where L := Ljl is a block matrix with block Ljl ℜn n connecting Lurie network l to Lurie network j. The state and nonlinearity of the GLN are defined by x ℜqn and Φ : ℜqm ℜqm where the states of the independent Lurie networks have been stacked into a single state. Interestingly, the GLN (6) is actually a Published in Transactions on Machine Learning Research (04/2025) special case of a Lurie network; however, as the networks get larger, so does the search space. Thus, imposing any assumptions which respect the structure of the problem can reduce the search space and lead to more robust and generalisable models. Any further prior knowledge about the graph can be encoded through constraints on the graph coupling matrix. 3.4 k-contraction Analysis of Graph Lurie Networks In this section, we assume that q independent Lurie networks (2) are k-contracting in the 2-norm w.r.t metrics P1, . . . , Pq. This is equivalent to (5) k-contracting in the 2-norm w.r.t the metric P = blockdiag(P1, . . . , Pq). Theorem 2 and Theorem 3 provide two results which can, respectively, verify this w.r.t a scalar metric Pj = pj In and a diagonal metric Pj Dn + for j [1, q]. Other results may be used providing they apply to systems of the form (2). Under this assumption, Theorem 4 provides a constraint on the graph coupling term which ensures the GLN is k-contracting when constructed from q independently k-contracting Lurie networks. The proof is detailed in C.3. Theorem 4 Fix k [1, n]. Consider a GLN (6) where the q independent Lurie networks are collectively defined by the weights AG := blockdiag(A1, . . . , Aq), BG := blockdiag(B1, . . . , Bq), CG := blockdiag(C1, . . . , Cq) and biases bx ℜqn, by ℜqm and are k-contracting in the 2-norm w.r.t the metric P := blockdiag(P1, . . . , Pq). If the graph coupling matrix L ℜqn qn satisfies P (k)L[k] + (L[k]) P (k) 0 (7) then (6) is also k-contracting in the 2-norm w.r.t the metric P. Remark 1 It should be noted that Theorem 4 could be applied more generally for constructing k-contracting graph coupled systems. Providing the q subsystems are k-contracting in the 2-norm w.r.t some metric P = blockdiag(P1, . . . , Pq), then Theorem 4 is applicable when coupling them through the graph coupling term. It is not necessary for the q subsystems to be Lurie networks. 4 Parametrisation of k-contracting Lurie Networks To train a k-contracting Lurie network using gradient based optimisers, parametrisations which express the constrained weights in terms of unconstrained variables must be found. To formalise this idea, we define the sets Ω2(g, k) and Ω4(k, P). As the biases do not appear in these sets, they are naturally unconstrained. The set Ω2(g, k) contains all the weights of the Lurie network which satisfy Theorem 2. Ω2(g, k) := n ( A, B, C) αk := 1 i=1 λi(A + A ) < 0, zk := g2 k X i=1 σ2 i ( B)σ2 i ( C) < α2 kk o (8a) Ω4(k, P) := n L ℜqn qn P (k) L[k] + ( L[k]) P (k) 0 o (8b) The set Ω4(k, P) contains the graph coupling matrices which satisfy Theorem 4. A parametrisation associated with Theorem 3 was not established due to its limitations mentioned earlier. The next results present two different parametrisations of the set Ω2(g, k). See C.4 for the proofs which leverage the eigenvalue and singular value decompositions. Theorem 5 Given g > 0, k [1, n], UA, UB, VC O(n), VB, UC O(m), ΣB Dnm + , ΣC Dmn + , YA Skew(n), GA Dn + and define 2UAΣAU A + 1 k In GA (9a) B := UBΣBV B C := UCΣCV C (9b) then (A, B, C) Ω2(g, k). Published in Transactions on Machine Learning Research (04/2025) Theorem 6 Given g > 0, k [1, n], UA, UB, VC O(n), VB, UC O(m), ΣB Dnm + , ΣC Dmn + , YA Skew(n), ΣA1 Dk 1, GA2 > 0, GA3 Dn k + and define 2UAΣAU A + 1 2YA B := UBΣBV B C := UCΣCV C (10a) ΣA := blockdiag(ΣA1, ΣA2, ΣA3) ΣA1 Dk 1 (10b) i (ΣA1)ii GA2 ΣA3 := min(ΣA1, ΣA2)In k GA3 (10c) then (A, B, C) Ω2(g, k). Remark 2 In both results, a mapping between the sets Skew( ) and O( ) was exploited to express the orthogonal matrices in terms of unconstrained skew-symmetric matrices Lezcano-Casado & Martınez-Rubio (2019). The remaining variables are unconstrained or simply require positive elements, which can be obtained by taking the absolute value of unconstrained variables. In Theorem 5 and Theorem 6, the B and C matrices are unconstrained since they are simply expressed by their singular value decomposition. The variables representing the singular values of B and C directly upper bound αk. The only source of conservatism in the parametrisations is introduced through the definition of ΣA. In Theorem 5, the constraint on αk is enforced by a uniform negative constraint on the diagonal elements of ΣA. If this assumption is true, this significantly speeds up the learning process; however, it can be prohibitive if not. Theorem 6 allows the largest (k 1) eigenvalues to be unconstrained (ΣA1), meaning non-Hurwitz A matrices are encapsulated in the parametrisation. The constraint on αk is implemented by the ΣA2 block and the ΣA3 block ensures the remaining eigenvalues are less than the other k. As k is a hyperparameter of the model, it is useful to note that the set Ω2(g, k 1) intersects with Ω2(g, k) when the following holds g2σ2 k(B)σ2 k(C) < 1 4k λ2 k(A + A ) + k 1 k αk 1λk(A + A ) k 1 k α2 k 1 (11) We expect this to hold for many realistic systems since (11) just requires λk(A + A ) to be sufficiently negative to counteract the product of the growth of B and C in the kth direction. Hence, if the best value of k is unknown for a given application, then setting k = 3 allows the model to search over many of the stable l-contracting systems parametrised by Theorem 5 or Theorem 6, where l [1, k]. See C.5 for the proof. The next result provides an unconstrained parametrisation of the set Ω4(k, P) in (8b). Coupling this with the earlier parametrisations allows the k-contracting GLN to be trained with gradient based optimisers. See C.6 for the proof. Theorem 7 Given k [1, n], GL1, GL2 ℜqn qn, Θ = blockdiag(Θ1, . . . , Θq) where Θj Sn + for j [1, q], P = Θ Θ and define L := GL1 P 1G L1P + Θ 1GL2Θ (12) where (GL2 + G L2)[k] 0, then L Ω4(k, P). Theorem 7 provides flexibility in the choice of GL1, GL2 to define different graph structures, where the constraint on GL2 is equally a constraint on the sum of it s k-largest eigenvalues (Fact 7). This constraint could be parametrised in a simlar manner to the symmetric component of A in Theorem 5 and Theorem 6; in which case, Theorem 7 is satisfied and the graph has a directed all-to-all structure, including self-loops. Remark 3 If GL2 = 0 along with the main diagonal blocks of GL1, then (12) is a boundary condition of (7). Thus, Theorem 7 is satisfied and the self-loops are removed from the graph structure. One consideration when constructing k-contracting Lurie networks is the number of additional parameters required for the parametrisation. A Lurie network, as defined in (2), has a total parameter count NL = Published in Transactions on Machine Learning Research (04/2025) n2 + 2nm + n + m; whereas a k-contracting Lurie network, constructed according to Theorem 5 or Theorem 6, has a total parameter count NK 2n2 + m2. The number of parameters only increases significantly when m is large compared to n; however, throughout the literature, many special cases of a Lurie network set n = m (see 3.1); in which case, NK becomes marginally smaller than NL. Furthermore, the all-to-all graph coupling term has NC = 2(qn)2 parameters or NC = qn2(q 1) if the self-loops are removed according to Remark 3. This analysis highlights that ensuring the Lurie network and GLN are k-contracting comes at a minimal computational expense. 5 Empirical Evaluation In 1, it was highlighted that a range of convergent dynamics are prevalent in many dynamical systems and neural processes. In the following section we test the impact of the additional expressivity of the Lurie network and the importance of encoding this general convergence property for modelling a range of dynamical systems. The prediction accuracy, generalisation and robustness of the proposed models were used to compare performance. Refer to Richardson et al. (2025) for additional experiments regarding the Lurie network s ability to form representations. 5.1 Datasets and Training We first consider three time-invariant dynamical systems: (i) an opinion dynamics model of a social network where all opinions agree and thus converge to a unique equilibrium point; (ii) a Hopfield network of associative memory with two stable equilibrium points and one unstable; (iii) a generic simple attractor which could be used to model the stored patterns in working memory. Each system has n = 3 states allowing for visualisation of the ground truth and predictions. For each dynamical system, we have a dataset including 1, 000 trajectories, sampled every 0.01s over a 20s interval. The test sets were formed by holding out 100 trajectories. The input to each model was the initial condition sampled from a uniform distribution with the domain ( 1, +1)3 for the opinion/Hopfield datasets and ( 3, +3)3 for the simple attractor. The full trajectory was then used as the target to train the model. An illustration of these datasets can be seen in Figure 1 and further details about the data generation can be found in Appendix D. To test the out of distribution generalisation and robustness, two additional datasets were generated for each of the opinion, Hopfield and attractor tasks. These differ from the training datasets in the following ways: (i) they include 100 trajectories over a 30s interval; (ii) the initial conditions were sampled from a uniform distribution over the intervals 1 < |xi(0)| < 4 for the opinion/Hopfield datasets and 3 < |xi(0)| < 6 for the simple attractor, where i {1, 2, 3}. These trajectories are 10s longer than the training data with initial conditions also sampled outside the training distribution. To test the robustness, noise sampled from the standard normal distribution was added to the initial conditions of one of these datasets, before generating the trajectories. For the second set of experiments, we consider two 30 dimensional dynamical systems. The first is a graphcoupled (GC) Hopfield network, formed by connecting 10 previously described Hopfield networks through a graph coupling matrix. To ensure the convergence property was preserved, the matrix was expressed by (12) with GL1 sampled from a uniform distribution and GL2 = 0. The second system was a graph-coupled attractor, constructed in the same way. The datasets were generated as described in the previous two paragraphs; however, they each included 30, 000 trajectories. The training settings used are explicitly detailed in Appendix D. For all models and all datasets, the mean squared error (MSE) loss was used alongside the Adam optimiser. All code was implemented in Py Torch and can be found at https://github.com/CR-Richardson/Lurie Network. 5.2 k-contracting Lurie Network In this section, we compare the k-contracting Lurie network against five other continuous-time models: (i) the unconstrained Lurie network, for testing the importance of the k-contraction constraints; (ii) an unconstrained neural ODE with two hidden layers, each comprised of 20 neurons and Re LU activations; (iii) the three Published in Transactions on Machine Learning Research (04/2025) constrained continuous-time models detailed in 3.1 where the SVD combo network is comprised of a single node. The numerical integration was performed using the Euler method with step size δ = 1 10 2. Besides the neural ODE, each model had tanh activations, which is slope-restricted with g = 1. For the opinion and Hopfield datasets, the k-contracting Lurie network was constructed according to Theorem 5 whereas Theorem 6 was used for the simple attractor. We set k = 1 for the opinion dynamics, k = 2 for the mulit-stable Hopfield network and k = 3 for the simple attractor. Similar results were obtained when setting k = 3 for all examples. Table 1: MSE on a test set of 100 trajectories. The lowest MSE is presented alongside the mean and standard deviation calculated after training each model N = 3 times on a single T4 GPU (Google Colab). Model MSE (mean std, best) Opinion Hopfield Attractor k-Lurie Network (8.0 3.0, 5.10) 10 5 (1.5 1.0, 0.26) 10 2 (3.5 1.0, 1.70) 10 3 Lurie Network (3.7 4.0, 0.53) 10 3 (3.6 2.0, 0.39) 10 1 (5.1 5.0, 0.57) 10 1 Neural ODE (2.0 2.0, 0.43) 10 4 (2.5 1.0, 1.50) 10 2 (2.0 1.0, 1.00) 10 2 Lipschitz RNN (3.9 3.0, 0.88) 10 2 (2.9 2.0, 0.30) 10 1 1.48 2.0, 1.10 10 2 SVD Combo (8.6 3.0, 5.70) 10 4 (2.9 0.6, 2.10) 10 1 3.12 0.5, 2.74 Antisym. RNN (30.0 0.2, 28.0) 10 2 (4.3 0.2, 4.11) 10 1 6.93 0.2, 6.74 Table 1 compares the MSE on the test set of each task. The k-contracting Lurie network achieved the best MSE on two out of three examples. The importance of the k-contraction conditions is particularly clear when comparing the mean and standard deviation with that of the unconstrained Lurie network. These conditions clearly reduce the search space to a tractable region to optimise over as the MSE of the unconstrained Lurie network is at least an order of magnitude worse than its k-contracting counterpart. The other models perform as one would expect: (i) the neural ODE demonstrates strong accuracy across all tasks; (ii) SVD combo performs well on the opinion dataset, where the 1-contraction assumption is valid, but struggles on the others; (iii) the Lipschitz RNN struggles on the attractor dataset whilst the antisymmetric RNN struggles across the board due to the A matrix being fixed at zero and the eigenvalues of the C matrix being fixed to almost purely imaginary values. Figures 3, 6, 9 show a random sample of trajectories from each test set, along with the predictions of each model. Table 2: MSE on a new test set of 100 trajectories where: (i) the initial conditions were sampled outside the range used for training and the trajectories were 10s longer than the training set; (ii) additionally, noise sampled from the standard normal distribution was added to the initial conditions. Model MSE MSE (noisy inputs) Opinion Hopfield Attractor Opinion Hopfield Attractor k-Lurie Network 2.9 10 3 5.6 10 2 2.3 10 1 2.0 10 2 3.2 10 1 1.28 Lurie Network 6.4 10 2 1.8 10 1 5.96 2.7 10 1 4.4 10 1 6.79 Neural ODE 1.2 10 2 1.09 2.31 3.9 10 2 1.63 4.76 Lipschitz RNN 2.3 10 1 7.3 10 1 6.2 10 1 2.9 10 1 9.7 10 1 1.71 SVD Combo 1.0 10 2 2.38 20.9 3.3 10 2 7.97 30.30 Antisym. RNN 6.43 5.25 52.1 7.29 6.33 52.9 Table 2 compares the generalisation and robustness of the models on each task. No new models were trained, instead the best models from Table 1 were directly applied to these out of distribution and noisy datasets ( 5.1). The k-contracting Lurie network performs the best on all of these datasets and for some, it still demonstrates a MSE of an order of magnitude lower than the next best model. The MSE of the unconstrained Lurie network and neural ODE tended to drop off for these datasets whereas the MSE of the constrained models, excluding the antisymmetric RNN, tended to stay fairly consistent when their assumptions were valid. Figures 4, 7, 10 show a random sample of noise-free trajectories and predictions for each dataset, whilst Figures 5, 8, 11 repeat the same for the noisy inputs. The k-contracting Lurie network predicted the Published in Transactions on Machine Learning Research (04/2025) correct long-term behaviour most accurately under all conditions; even when noise was added, the error was predominantly present during the initial transient. 5.3 k-contracting Graph Lurie Network This section repeats the same experiments as the previous section, but for two 30-dimensional graph-coupled (GC) dynamical systems: the GC Hopfield network and the GC simple attractor ( 5.1). The state of these datasets is significantly larger than those used in other dynamical systems datasets such as: (i) the LASA dataset Lemme et al. (2015) where the 2-d trajectories are typically stacked to form 4-d or 8-d trajectories; (ii) simulated datasets of the 2, 4 or 8 link pendulums which, respectively, have 4, 8 or 16 dimension trajectories. For both datasets, the GLN was constructed according to Theorem 7 and Remark 3 where n = m = 3 and q = 10. The individual Lurie networks were constructed according to Theorem 5 for the GC Hopfield network, with k = 2 and Theorem 6 for the GC attractor, with k = 3. The neural ODE was formed using two layers with 100 neurons and Re LU activations. The SVD combo leveraged a similar graph structure to the one used in this paper. The other models were the same as in the previous section but with a (qn)-dimensional state. Table 3: MSE on a test set of 100 trajectories. The lowest MSE is presented alongside the mean and standard deviation calculated after training each model N = 3 times on a single A100 GPU (Google Colab). Model MSE (mean std, best) GC Hopfield GC Attractor GLN 0.016 0.0003, 0.016 0.293 0.1969, 0.015 k-Lurie Network 0.238 0.0011, 0.237 0.737 0.0108, 0.723 Lurie Network 2.537 1.3327, 1.157 291.8 191.84, 21.83 Neural ODE 0.138 0.0291, 0.114 3.445 0.3626, 2.942 Lipschitz RNN 0.124 0.0161, 0.105 0.658 0.1604, 0.433 SVD Combo 0.339 0.1363, 0.229 3.024 1.1240, 1.435 Antisym. RNN 0.448 0.0023, 0.444 6.386 0.0066, 6.380 Table 3 shows the GLN had a lower MSE than all other models by a factor of 10. Comparing the GLN, k-contracting Lurie network and the unconstrained Lurie network highlights the improvements due to the k-contraction conditions and the graph structure. Table 4 also indicates that the GLN generalised and remained robust to noise, even in these high-dimensional systems. The same can be said for the k-contracting Lurie network which achieved the second lowest MSE on the generalisation and robustness tests. The inherent graph structure of the SVD combo may be the reason behind its improved ranking in the GC Hopfield tasks whereas the neural ODE particularly struggled to generalise for the GC attractor. Possible explanations behind the poor performance of the constrained benchmark models are suggested in the next section. Table 4: MSE on a new test set of 100 trajectories where: (i) the initial conditions were sampled outside the range used for training and the trajectories were 10s longer than the training set; (ii) additionally, noise sampled from the standard normal distribution was added to the initial conditions. Model MSE MSE (noisy inputs) GC Hopfield GC Attractor GC Hopfield GC Attractor GLN 0.08 1.67 0.26 2.85 k-Lurie Network 0.77 6.10 1.05 6.90 Lurie Network 20350 5638 25481 6368 Neural ODE 2.12 24.93 2.76 25.11 Lipschitz RNN 3.58 6.17 4.25 7.84 SVD Combo 0.84 11.40 1.07 12.30 Antisym. RNN 5.32 50.68 6.21 52.10 Published in Transactions on Machine Learning Research (04/2025) 6 Related Work Several constrained continuous-time RNN models exist in the ML literature. The model structure of three notable examples were presented in 3.1 as they happen to be special cases of a Lurie network, when modelling time-invariant dynamical systems. The antisymmetric RNN Chang et al. (2019) and the Lipschitz RNN Erichson et al. (2020) were designed to address the exploding and vanishing gradient problem Pascanu et al. (2013). The antisymmetric RNN did so by parametrising the RNN such that the real eigenvalues of the Jacobian were zero. It achieved this by setting A = 0 and restricting C to being skew-symmetric. Whilst this does prevent the gradients from exploding and vanishing, it restricts the dynamics which the model can learn to purely oscillatory behaviour. The Lipschitz RNN has a more relaxed parametrisation. This model constructs the A and C matrices such that they are both convex combinations of symmetric and skew-symmetric matrices. However, the weight of the symmetric matrix can only vary between 0 and 0.5, whereas the weight of the skew-symmetric matrix can vary between 0.5 and 1. Again, this addresses the vanishing and exploding gradient problem, but the model cannot encode dynamics which are predominately decaying or growing. Finally, the RNN proposed in Kozachkov et al. (2022a) has biological motivations and encodes 1-contracting dynamics. This implies that all possible trajectories will exponentially converge, making the model robust to input disturbances. Whilst the models above address a variety of problems, each model is quite limited in the range of dynamics they can learn, which is a problem when trying to design a generalised model for learning time-invariant dynamical systems. With respect to (2), the Lurie network has more flexibility than all of these models. Firstly, it includes all three weight matrices (A, B, C) and biases (bx, by), whereas the models mentioned above fix at least one of these parameters. Secondly, the constraints imposed, and the corresponding parametrisations, allow the model to learn a variety of dynamics including, but not limited to, those mentioned above. The only limitation is that the dynamics must converge in some way. This includes certain types of chaotic systems, for example Thomas cyclically symmetric attractor Thomas (1999). At the edge of the chaotic regime, the trajectories converge to a strange attractor, which could be modelled by a 3-contracting Lurie network. The Lurie network is also related to a class of feed-forward models named implicit or equilibrium networks. These models use an implicit equation to express the relationship between the model output, layer outputs and model input in a compact vectorised form El Ghaoui et al. (2021). Like the Lurie network, these models can be represented by the interconnection of a linear time-invariant system and a nonlinearity. This makes analysis tools from control theory, such as Lipschitz bounds, applicable to these models Fazlyab et al. (2019). An additional connection is that the solution to the implicit equations correspond to equilibrium points of a Lurie system Revay et al. (2020). As mentioned in the introduction, the k-contraction constraints used in this paper have an interesting connection to some properties observed in biological learning systems. A 2-contracting model can replicate the behaviour of associative memory, where every stored pattern corresponds to an equilibrium Kozachkov et al. (2023). Furthermore, a 3-contracting model can replicate the dynamics of working memory, where patterns are retained as attractor states Kozachkov et al. (2022b). Due to the 2 contraction (3-contraction) constraints, the equilibrium points (attractor states) must lie on a line (plane). As a result, the conditions developed in this paper could be of interest to neuroscientists and ML researchers interested in memory storage and retrieval Ramsauer et al. (2020); Hopfield (1984); Krotov & Hopfield (2020). The formation of the GLN also has connections to biology. For example, constructing larger systems from smaller modules can be motivated by evolutionary biology Simon (1962) where the name facilitated variation Gerhart & Kirschner (2007) is used to describe the development of traits in response to adaptation of the regulatory elements that connect modules, rather than the core components themselves. This was investigated in Kozachkov et al. (2022a) where the weights of the individual RNNs satisfied a 1-contraction condition but were fixed. Only the graph coupling weights were learnt during training. The relationship between properties guaranteed by k-contraction analysis and those observed in associative and working memory suggest the k-contracting Lurie network possesses a number of appealing properties for an ML model; hence, it may be suitable for a wider class of ML problems. This proposition is supported by the successful application of the stability-constrained RNN models from 3.1 (special cases of a Lurie network) on a wide range of ML tasks. Since the Lurie network is a more structured, time-invariant example Published in Transactions on Machine Learning Research (04/2025) of a neural ODE Chen et al. (2018), it will also be applicable to a similar array of tasks. Beyond modelling time-invariant dynamical systems, this includes image classification and continuous normalising flows Kidger (2022). The only limiting requirement is that the input must be passed in through the initial condition which in some cases, such as image classification, may require a pre-processing layer. 7 Conclusion The Lurie network was presented as a novel and unifying time-invariant neural ODE. Stability results were presented which verified a generalised form of stability of the Lurie network, namely k-contraction. These results were incorporated in a principled approach for constructing k-contracting Lurie networks and graph Lurie networks which, unlike existing stability-constrained models, could be trained to model multi-stable and orbitally stable systems, such as those observed in associative and working memory. Empirical results showed the benefit of the graph structure and the k-contraction constraints through improved prediction accuracy, out of distribution generalisation and robustness on a range of simulated dynamical systems. These same models also compared favourably against well known stability-constrained models, which have been shown to be special cases of the Lurie network, and an unconstrained neural ODE. Future theoretical work will try to expand the class of systems the k-contracting Lurie network can be optimised over, by obtaining similar results for systems which are k-contracting in a diagonal metric. It would also be interesting to find conditions on the inputs of a time-varying Lurie network for which the convergence properties of the k-contracting Lurie network are also upheld, making the Lurie network applicable to sequential processing tasks. Finally, due to the connections with biological neural processes, we would like to empirically investigate the performance of the model on a wider range of machine learning tasks. The recently proposed working memory benchmark Sikarwar & Zhang (2024) is of particular interest due to its relationship with 3-contracting dynamics. Acknowledgments This work was supported in part by the Defence Science and Technology Laboratory (DSTL) and the UK Research and Innovation (UKRI) Centre of Machine Intelligence for Nano-electronic Devices and Systems [EP/S024298/1]. Antoine R Adamantidis, Carolina Gutierrez Herrera, and Thomas C Gent. Oscillating circuitries in the sleeping brain. Nature Reviews Neuroscience, 20(12):746 762, 2019. Peter Ashwin, Stephen Coombes, and Rachel Nicks. Mathematical frameworks for oscillatory network dynamics in neuroscience. The Journal of Mathematical Neuroscience, 6:1 92, 2016. Eyal Bar-Shalom, Omri Dalin, and Michael Margaliot. Compound matrices in systems and control theory: a tutorial. Mathematics of Control, Signals, and Systems, pp. 1 55, 2023. Joaquin Carrasco, Matthew C Turner, and William P Heath. Zames-Falb multipliers for absolute stability: From O Shea s contribution to convex searches. European Journal of Control, 28:1 19, 2016. Andreu Cecilia, Samuele Zoboli, Daniele Astolfi, Ulysse Serres, and Vincent Andrieu. Generalized Lyapunov conditions for k-contraction: analysis and feedback design. 2023. Veronica Centorrino, Francesco Bullo, and Giovanni Russo. Contraction analysis of Hopfield neural networks with Hebbian learning. In 2022 IEEE 61st Conference on Decision and Control (CDC), pp. 622 627. IEEE, 2022. Bo Chang, Minmin Chen, Eldad Haber, and Ed H Chi. Antisymmetricrnn: A dynamical system view on recurrent neural networks. ar Xiv preprint ar Xiv:1902.09689, 2019. Ricky TQ Chen, Yulia Rubanova, Jesse Bettencourt, and David K Duvenaud. Neural ordinary differential equations. Advances in neural information processing systems, 31, 2018. Published in Transactions on Machine Learning Research (04/2025) Ross Drummond, Matthew C Turner, and Stephen R Duncan. Reduced-order neural network synthesis with robustness guarantees. IEEE Transactions on Neural Networks and Learning Systems, 2022. Ross Drummond, Chris Guiver, and Matthew C Turner. Exponential input-to-state stability for Lur e systems via integral quadratic constraints and Zames Falb multipliers. IMA Journal of Mathematical Control and Information, pp. dnae003, 2024. Laurent El Ghaoui, Fangda Gu, Bertrand Travacca, Armin Askari, and Alicia Tsai. Implicit deep learning. SIAM Journal on Mathematics of Data Science, 3(3):930 958, 2021. N Benjamin Erichson, Omri Azencot, Alejandro Queiruga, Liam Hodgkinson, and Michael W Mahoney. Lipschitz recurrent neural networks. ar Xiv preprint ar Xiv:2006.12070, 2020. Mahyar Fazlyab, Alexander Robey, Hamed Hassani, Manfred Morari, and George Pappas. Efficient and accurate estimation of lipschitz constants for deep neural networks. Advances in neural information processing systems, 32, 2019. Mahyar Fazlyab, Manfred Morari, and George J Pappas. Safety verification and robustness analysis of neural networks via quadratic constraints and semidefinite programming. IEEE Transactions on Automatic Control, 67(1):1 15, 2020. John Gerhart and Marc Kirschner. The theory of facilitated variation. Proceedings of the National Academy of Sciences, 104(suppl_1):8582 8589, 2007. Albert Gu, Tri Dao, Stefano Ermon, Atri Rudra, and Christopher Ré. Hippo: Recurrent memory with optimal polynomial projections. Advances in neural information processing systems, 33:1474 1487, 2020. Albert Gu, Karan Goel, and Christopher Ré. Efficiently modelling long sequences with structured state spaces. ar Xiv preprint ar Xiv:2111.00396, 2021. William L Hamilton, Rex Ying, and Jure Leskovec. Representation learning on graphs: Methods and applications. ar Xiv preprint ar Xiv:1709.05584, 2017. John J Hopfield. Neural networks and physical systems with emergent collective computational abilities. Proceedings of the national academy of sciences, 79(8):2554 2558, 1982. John J Hopfield. Neurons with graded response have collective computational properties like those of two-state neurons. Proceedings of the national academy of sciences, 81(10):3088 3092, 1984. Roger A Horn and Charles R Johnson. Topics in matrix analysis. Cambridge university press, 1994. Eric R Kandel, James H Schwartz, Thomas M Jessell, Steven Siegelbaum, A James Hudspeth, Sarah Mack, et al. Principles of neural science, volume 4. Mc Graw-hill New York, 2000. Hassan K Khalil. Nonlinear systems. Patience Hall, 115, 2002. Mikail Khona and Ila R Fiete. Attractor and integrator networks in the brain. Nature Reviews Neuroscience, 23(12):744 766, 2022. Patrick Kidger. On neural differential equations. ar Xiv preprint ar Xiv:2202.02435, 2022. Leo Kozachkov, Mikael Lundqvist, Jean-Jacques Slotine, and Earl K Miller. Achieving stable dynamics in neural circuits. PLo S computational biology, 16(8):e1007659, 2020. Leo Kozachkov, Michaela Ennis, and Jean-Jacques Slotine. RNNs of RNNs: Recursive construction of stable assemblies of recurrent neural networks. Advances in Neural Information Processing Systems, 35: 30512 30527, 2022a. Leo Kozachkov, John Tauber, Mikael Lundqvist, Scott L Brincat, Jean-Jacques Slotine, and Earl K Miller. Robust and brain-like working memory through short-term synaptic plasticity. PLOS Computational Biology, 18(12):e1010776, 2022b. Published in Transactions on Machine Learning Research (04/2025) Leo Kozachkov, Jean-Jacques Slotine, and Dmitry Krotov. Neuron-astrocyte associative memory. ar Xiv preprint ar Xiv:2311.08135, 2023. Dmitry Krotov and John Hopfield. Large associative memory problem in neurobiology and machine learning. ar Xiv preprint ar Xiv:2008.06996, 2020. Samuel Lanthaler, T Konstantin Rusch, and Siddhartha Mishra. Neural oscillators are universal. Advances in Neural Information Processing Systems, 36, 2024. Andre Lemme, Yaron Meirovitch, M Khansari-Zadeh, Tamar Flash, Aude Billard, and Jochen J Steil. Opensource benchmarking for learned reaching motion generation in robotics. Paladyn, Journal of Behavioral Robotics, 6(1):000010151520150002, 2015. Laurent Lessard, Benjamin Recht, and Andrew Packard. Analysis and design of optimization algorithms via integral quadratic constraints. SIAM Journal on Optimization, 26(1):57 95, 2016. Mario Lezcano-Casado and David Martınez-Rubio. Cheap orthogonal constraints in neural networks: A simple parametrization of the orthogonal and unitary group. In International Conference on Machine Learning, pp. 3794 3803. PMLR, 2019. Winfried Lohmiller and Jean-Jacques E Slotine. On contraction analysis for non-linear systems. Automatica, 34(6):683 696, 1998. James S Muldowney. Compound matrices and ordinary differential equations. The Rocky Mountain Journal of Mathematics, pp. 857 872, 1990. Ron Ofir, Jean-Jacques Slotine, and Michael Margaliot. k-contraction in a generalized Lurie system. ar Xiv preprint ar Xiv:2309.07514, 2023. Ron Ofir, Alexander Ovseevich, and Michael Margaliot. Contraction and k-contraction in Lurie systems with applications to networked systems. Automatica, 159:111341, 2024. Matthijs Pals, Jakob H Macke, and Omri Barak. Trained recurrent neural networks develop phase-locked limit cycles in a working memory task. PLOS Computational Biology, 20(2):e1011852, 2024. Poogyeon Park. A revisited Popov criterion for nonlinear Lur e systems with sector-restrictions. International Journal of Control, 68(3):461 470, 1997. Poo Gyeon Park. Stability criteria of sector-and slope-restricted Lur e systems. IEEE Transactions on Automatic Control, 47(2):308 313, 2002. Razvan Pascanu, Tomas Mikolov, and Yoshua Bengio. On the difficulty of training recurrent neural networks. In International conference on machine learning, pp. 1310 1318. Pmlr, 2013. Patricia Pauli, Dennis Gramlich, Julian Berberich, and Frank Allgöwer. Linear systems with neural network nonlinearities: Improved stability analysis via acausal Zames-Falb multipliers. In 2021 60th IEEE Conference on Decision and Control (CDC), pp. 3611 3618. IEEE, 2021. Kaare Brandt Petersen, Michael Syskind Pedersen, et al. The matrix cookbook. Technical University of Denmark, 7(15):510, 2008. Alex T Piet, Jeffrey C Erlich, Charles D Kopec, and Carlos D Brody. Rat prefrontal cortex inactivations during decision making are explained by bistable attractor dynamics. Neural computation, 29(11):2861 2886, 2017. Hubert Ramsauer, Bernhard Schäfl, Johannes Lehner, Philipp Seidl, Michael Widrich, Thomas Adler, Lukas Gruber, Markus Holzleitner, Milena Pavlović, Geir Kjetil Sandve, et al. Hopfield networks is all you need. ar Xiv preprint ar Xiv:2008.02217, 2020. Max Revay, Ruigang Wang, and Ian R Manchester. Lipschitz bounded equilibrium networks. ar Xiv preprint ar Xiv:2010.01732, 2020. Published in Transactions on Machine Learning Research (04/2025) Carl Richardson, Matthew Turner, Steve Gunn, and Ross Drummond. Strengthened stability analysis of discrete-time Lurie systems involving Re LU neural networks. In 6th Annual Learning for Dynamics & Control Conference, pp. 209 221. PMLR, 2024. Carl R Richardson, Matthew C Turner, and Steve R Gunn. Strengthened Circle and Popov Criteria for the stability analysis of feedback systems with Re LU neural networks. IEEE Control Systems Letters, 2023. Carl R Richardson, Matthew C. Turner, and Steve R. Gunn. Lurie networks with k-contracting dynamics. In New Frontiers in Associative Memories Workshop at ICLR, 2025. URL https://openreview.net/forum? id=Ra AYe Cxj1u. T Konstantin Rusch, Ben Chamberlain, James Rowbottom, Siddhartha Mishra, and Michael Bronstein. Graph-coupled oscillator networks. In International Conference on Machine Learning, pp. 18888 18909. PMLR, 2022. Sugandha Sharma, Sarthak Chandra, and Ila Fiete. Content addressable memory without catastrophic forgetting by heteroassociation with a fixed scaffold. In International Conference on Machine Learning, pp. 19658 19682. PMLR, 2022. Ankur Sikarwar and Mengmi Zhang. Decoding the enigma: benchmarking humans and AIs on the many facets of working memory. Advances in Neural Information Processing Systems, 36, 2024. Herbert A Simon. The architecture of complexity. Proceedings of the American philosophical society, 106(6): 467 482, 1962. René Thomas. Deterministic chaos seen in terms of feedback circuits: Analysis, synthesis, labyrinth chaos. International Journal of Bifurcation and Chaos, 9(10):1889 1905, 1999. Matthew C Turner and Ross Drummond. Analysis of MIMO Lurie systems with slope restricted nonlinearities using concepts of external positivity. In 2019 IEEE 58th Conference on Decision and Control (CDC), pp. 163 168. IEEE, 2019. Mathukumalli Vidyasagar. Nonlinear systems analysis. SIAM, 2002. Ryan Vogt, Maximilian Puelma Touzel, Eli Shlizerman, and Guillaume Lajoie. On Lyapunov exponents for RNNs: Understanding information propagation using dynamical systems tools. Frontiers in Applied Mathematics and Statistics, 8:818799, 2022. Hugh R Wilson and Jack D Cowan. Excitatory and inhibitory interactions in localized populations of model neurons. Biophysical journal, 12(1):1 24, 1972. Chengshuai Wu, Ilya Kanevskiy, and Michael Margaliot. k-contraction: Theory and applications. Automatica, 136:110048, 2022. George Zames and PL Falb. Stability conditions for systems with monotone and slope-restricted nonlinearities. SIAM Journal on Control, 6(1):89 108, 1968. Xiaojiao Zhang and Baotong Cui. Synchronization of Lurie system based on contraction analysis. Applied Mathematics and Computation, 223:180 190, 2013. Samuele Zoboli, Andreu Cecilia, and Sophie Tarbouriech. Quadratic abstractions for k-contraction. 2024. Published in Transactions on Machine Learning Research (04/2025) A Extended Preliminaries As many of the tools used in this paper are not well-known in the machine learning community, a condensed background is presented here, based on results from Bar-Shalom et al. (2023); Wu et al. (2022). The first section is on compound matrices, an important algebraic tool needed for generalising contraction analysis to k-contraction analysis. Following that, a geometric interpretation of the k-compound matrix is stated through its relationship to the volume of a parametrised set. Finally, the results in these two sections are utilised to define and provide intuition for the convergence of k-contracting dynamics. A.1 Compound Matrices In this section, we document several known definitions and algebraic results related to compound matrices. The results are included without proof; the interested reader should refer to Bar-Shalom et al. (2023) for a more detailed tutorial on the topic. Let n be a positive integer and fix k [1, n]. The ordered set of increasing sequences of k integers from [1, n] is denoted by Q(k, n). For example: Q(3, 4) = {(1, 2, 3), (1, 2, 4), (1, 3, 4), (2, 3, 4)}. Now consider a matrix W ℜn m. For α Q(k, n) and β Q(k, m), the matrix W[α|β] denotes the k k sub-matrix obtained by taking the entries of W along the rows indexed by α and columns indexed by β. As an example, if k = 2 and n = m = 4, then Q(2, 4) = {(1, 2), (1, 3), (1, 4), (2, 3), (2, 4), (3, 4)}. The sub-matrix W[(1, 2)|(3, 4)] would then be given by W[(1, 2)|(3, 4)] = w13 w14 w23 w24 The k-minors of the matrix W are defined as W(α|β) := det(W[α|β]). Definition 1 (k-multiplicative compound) Let W ℜn m and fix k [1, min(n, m)]. The kmultiplicative compound of W, denoted W (k), is the n k m k matrix containing all the k-minors of W ordered lexicographically. For example, if we have n = m = 3 and k = 2 then α, β Q(2, 3) = {(1, 2), (1, 3), (2, 3)} and W (1, 2)|(1, 2) W (1, 2)|(1, 3) W (1, 2)|(2, 3) W (1, 3)|(1, 2) W (1, 3)|(1, 3) W (1, 3)|(2, 3) W (2, 3)|(1, 2) W (2, 3)|(1, 3) W (2, 3)|(2, 3) Some important special cases include W (1) = W W (n) = det(W) (p In)(k) = pk Is W Dn W (k) Ds (13) with s := n k . Next, we present a series of algebraic results concerned with the k-multiplicative compound. Fact 1 (Cauchy-Binet Formula) If U ℜn m, V ℜm p and k [1, min(n, m, p)], then (UV )(k) = U (k)V (k) Fact 2 Fix k [1, min(n, m)]. As a consequence of Definition 1, if W ℜn m then (W )(k) = (W (k)) Fact 3 Fix k [1, n]. If W ℜn n is non-singular, then by Theorem 1 (W 1)(k) = (W (k)) 1 Fact 4 Fix k [1, min(n, m, p)]. If W ℜn n, U ℜp n and V ℜn p, then by Theorem 1 (UWV )(k) = U (k)W (k)V (k) Published in Transactions on Machine Learning Research (04/2025) Fact 5 Fix k [1, n]. An implication of Theorem 1 is that if W ℜn n with eigenvalues λ1, . . . , λn, then the eigenvalues of W (k) are the n k products l=1 λil : 1 i1 < < ik n o We now introduce the definition of a second compound matrix, the k-additive compound, and a set of algebraic results related to it. Definition 2 (k-additive compound) Let W ℜn n and k [1, n]. The k-additive compound of W is the n k n k matrix defined by dϵ In + ϵW (k)|ϵ=0 Special cases include W [1] = W W [n] = tr(W) (p In)[k] = kp Is W Dn W [k] Ds (14) with s := n k . Like before, we now present some useful algebraic results related to the k-additive compound. Fact 6 If W ℜn n and k [1, n], then as a consequence of Definition 2 (W )[k] = (W [k]) Fact 7 Fix k [1, n]. For W ℜn n with eigenvalues λ1, . . . , λn, the eigenvalues of W [k] are the n k sums l=1 λil : 1 i1 < < ik n} An important consequence of Fact 7 is that if W is positive definite (semi-definite), then this property is upheld by W [k]. Opposite conclusions can be drawn if W is negative definite (semi-definite). Fact 8 Fix k [1, n]. If U, V ℜn n, then (U + V )[k] = U [k] + V [k] Fact 9 Fix k [1, min(n, p)]. If W ℜn n, U ℜp n, V ℜn p and UV = Ip, then (UWV )[k] = U (k)W [k](U (k)) 1 A.2 Volume of k-sets This section aims to provide a clear geometric interpretation of the k-multiplicative compound. The section begins by defining a k-set (the codomain of a function dependent on k variables) before presenting Theorem 8, the key result which exposes the relationship between the volume of a k-set and the k-multiplicative compound of the Jacobian. A k-parallelotope is then shown as an example. Like before, these are existing results, so are presented without proof. Refer to Wu et al. (2022) for more information. Definition 3 (k-sets) Consider a compact set D ℜk and a continuous differentiable map Ψ : D ℜn, with k [1, n]. The codomain of Ψ is given by the parametrised set Ψ(D) := {Ψ(r) : r D} ℜn (15) Since D is compact and Ψ( ) is continuous, Ψ(D) is a closed set. Published in Transactions on Machine Learning Research (04/2025) Ψ(𝐷; 𝑥!, 𝑥", 𝑥#) Figure 2: The 3-parallelotope with vertices x1, x2, x3 ℜ3 parametrised by the unit cube, D. Theorem 8 (Volume of k-sets) Fix k [1, n]. Consider a compact set D ℜk and a continuously differentiable map Ψ : D ℜn. The volume of the parametrised set (15) is given by vol Ψ(D) = Z J(k) Ψ (r) dr where JΨ(r) = h Ψ(r) r1 . . . Ψ(r) i is the Jacobian of Ψ. Note that as JΨ : D ℜn k, it implies J(k) Ψ (r) : D ℜs with s = n k . Theorem 8 states that the volume of a k-set is governed by the k-multiplicative compound of the Jacobian. An important geometrical feature, is that for k {1, 2, 3} the volume of the k-set is equivalent to the standard notions of length, area and volume. We now consider the k-parallelotope as an example of a k-set. Definition 4 (k-parallelotope) Fix k [1, n] and let vectors x1, . . . , xk ℜn. The parallelotope generated by these vectors (and the zero vertex) is the set given by P(x1, . . . , xk) := n k X i=1 rixi : ri [0, 1] i o Based on Definition 4, the k-parallelotope is a k-set with the following compact domain D and continuous differentiable function Ψ(r; x1, . . . , xk), as illustrated for k = n = 3 in Figure 2. D := r ℜk : ri [0, 1] i [1, k] Ψ(r; x1, . . . , xk) := The Jacobian of the k-parallelotope is JΨ(r) = X and from Theorem 8, the volume of the k-parallelotope is given by |X(k)|. A.3 k-contraction Analysis Consider the time-varying nonlinear system x = f(t, x) (16) where f : ℜ+ ℜn ℜn. It is assumed throughout that f is continuously differentiable w.r.t x. Fix k [1, n] and let Sk denote the unit simplex. Sk := {r ℜk : ri 0 and r1 + + rk 1} The convex combination of a set of initial conditions x1, . . . , xk+1 ℜn is defined by h : Sk ℜn. h(r; x1, . . . , xk+1) := i=1 rixi + 1 i=1 ri xk+1 Published in Transactions on Machine Learning Research (04/2025) The set h(Sk) is a k-set and can be thought of as a k-dimensional body of states representing initial conditions of (16). We now define wi(t, r) as a measure of the sensitivity of a solution to (16) at time t, to a change in the initial condition h(r), caused by a change in ri. wi(t, r) := x(t, h(r)) ri where wi(0, r) = h(r) ri = xi xk+1 for all i [1, k] We are now ready to present the definition of k-contraction, followed by its geometric interpretation. Definition 5 (k-contraction) Fix k [1, n]. The nonlinear system (16) is k-contracting if there exists an η > 0 and a vector norm | | such that for any x1, . . . , xk+1 ℜn and any r Sk, the mapping W : ℜ+ Sk ℜn k defined by W(t, r) := w1(t, r) . . . wk(t, r) satisfies |W (k)(t, r)| exp( ηt)|W (k)(0, r)| t ℜ+ To explain the geometric meaning of this definition, pick a domain D Sk and recall that h(D) is a k-set representing k-dimensional bodies of initial conditions for (16); thus, x t, h(D) := {x(t, h(r)) : r D} is a k-set describing how k-dimensional bodies evolve over time. We now leverage Theorem 8, to show how the volume of these bodies evolves over time when governed by (16). vol x t, h(D) = Z Jx t, h(r) (k) dr r1 . . . x t,h(r) W (k)(t, r) dr When (16) is k-contracting, the volume of these bodies is upper bounded by the initial volume scaled by an exponentially decaying term. vol x t, h(D) exp( ηt) Z W (k)(0, r) dr = exp( ηt) (x1 xk+1) . . . (xk xk+1) (k) Z Therefore, k-contraction of (16) implies the volume of k-dimensional bodies x(t, h(D)) converges to zero at an exponential rate. This can also be interpreted as the volume of k-dimensional bodies is contracting or converging to a (k 1)-dimensional subspace. Figure 1 in the main paper gives an illustration of (1, 2, 3)- contracting systems where the 1-contracting system converges to a point, the 2-contracting system converges to a line and the 3-contracting system converges to a plane. Many existing k-contraction results, including Theorem 1, are expressed in terms of matrix measures. An overview of their definitions and properties can be found in (Vidyasagar, 2002, Section 2.2). Theorem 1 provides a sufficient condition for verifying k-contraction in the 2-norm w.r.t a metric P. The 2-norm was chosen for this work due to its relationship with the eigenvalues of its argument, but other norms could be chosen. Furthermore, one may apply an invertible linear transformation Θ to wi(t, r) and the k-contraction analysis may be performed in this new domain whilst implying the same property holds in the original domain. This idea is made clear in Lohmiller & Slotine (1998) and translates analogously to the k-contraction case. When using such an invertible transformation, the system is said to be k-contracting w.r.t the metric P = Θ Θ. Published in Transactions on Machine Learning Research (04/2025) B Lurie Network B.1 Relationship Between Lurie Networks and Deep Feedforward Models Consider the vector field z = f(z) defined by the following L-layer feedforward network z = WLΦ(u L 1) + b L u L 1 = WL 1Φ(u L 2) + b L 1 u L 2 = WL 2Φ(u L 3) + b L 2 ... u2 = W2Φ(u1) + b2 u1 = W1z + b1 To illustrate the superior expressivity of the Lurie network, we wish to show that a special case of (2) can approximate the deep feedforward network (17). An alternative expression for (17) is z = 0z + WLΦ(u L 1) + b L ϵ u L 1 = u L 1 + WL 1Φ(u L 2) + b L 1 ϵ u L 2 = u L 2 + WL 2Φ(u L 3) + b L 2 ... ϵ u2 = u2 + W2Φ(u1) + b2 u1 = W1z + b1 where ϵ 0. Defining a new state x := z u L 1 u L 2 . . . u3 u2 and an output vector y := u L 1 u L 2 . . . u2 u1 it is clear that (18) is a special case of the Lurie network (2) with state-space matrices and biases defined by the sparse structures below. 0 0 0 . . . 0 0 0 I 0 . . . 0 0 0 0 I . . . 0 0 ... ... ... ... ... ... 0 0 0 . . . I 0 WL 0 . . . 0 0 0 0 WL 1 . . . 0 0 0 0 0 . . . 0 0 ... ... ... ... ... ... 0 0 0 . . . 0 W2 0 I 0 . . . 0 0 0 0 I . . . 0 0 0 0 0 . . . 0 0 ... ... ... ... ... ... W1 0 0 . . . 0 0 b L b L 1 b L 2 ... b2 0 0 0 ... b1 This is just one realisation of a Lurie network which approximates a feedforward network. Other permutations of the state would result in different realisations of the Lurie networks weights and biases. Finally, due to the division by ϵ, it would not be possible to train a Lurie network to have the exact form of a feedforward network; however, this analysis shows that it is possible to approximate the strucuture of a deep feedforward network with a Lurie network. Published in Transactions on Machine Learning Research (04/2025) B.2 Further Examples Neural Oscillators: This example is from the graph ML literature Lanthaler et al. (2024). The state of the general neural oscillator is governed by a second order ODE; however, it s equivalent first order representation takes the form (2) with one possible realisation given by C21 ℜn n, bx = 0 and A = 0 I 0 0 B = 0 0 0 I C = 0 0 C21 0 The solution is then passed through an affine readout layer. Graph Coupled Oscillators: Another example of a second order ODE from the graph ML literature Rusch et al. (2022). The state is defined as a matrix, but this can simply be recast in a vectorised form which relates to (2) if a linear coupling function is chosen. This requires the weights to have a block matrix form, where one possible realisation is defined by C21 ℜn n, bx = by = 0 and A = 0 I γI αI B = 0 0 0 I C = 0 0 C21 0 LSSM: When the external input is replaced by nonlinear output feedback (i.e., u(t) Φ(y)) and D = 0, the linear state-space layer used in S4 Gu et al. (2021) and Hippo Gu et al. (2020) is related to (2) with A being a lower triangular Hippo matrix , B ℜn m, C ℜm n and bx = by = 0. Published in Transactions on Machine Learning Research (04/2025) C.1 Proof of Theorem 2 We aim to verify Theorem 1 for the particular case where the nonlinear system is described by the Lurie network (2). Our proof begins with Theorem 9, which restates (Ofir et al., 2024, Theorem 1). This result is sufficient to satisfy Theorem 1 for systems of the form (19). Theorem 9 Fix k [1, n] and consider the system below. x = Ax(t) BΨ(t, y) y = Cx (19) If there exists η1, η2 > 0 and an invertible Θ ℜn n such that P (k) A[k] + ( A[k]) P (k) + Θ(k) (Θ B B Θ)[k] + (Θ 1 C CΘ 1)[k] Θ(k) η1P (k) (20) and Θ 1 C (J Ψ (t, y)JΨ(t, y) Im) CΘ 1 [k] η2Is t ℜ+ and y ℜm (21) where s = n k , then (19) is k-contracting in the 2-norm w.r.t the metric P := Θ Θ. We first need to express the Lurie network in the form (19). By (3) there exists γ < 0 satisfying 0 < γ2 < α2 k and g2 k X i=1 σ2 i (B)σ2 i (C) < γ2k (22) Using γ, we can express (2) in the form (19) through the definitions below, where the dependence on t has been dropped from Ψ. A := A B := γIn C := In Ψ(x) := γ 1BΦ(Cx + by) γ 1bx (23) The next step is to verify (20). Subbing (23) into the left hand side of (20) and assuming Θ = Θ results in the first equality. Setting P := p In with p > 0 results in the second. Now we must leverage some of the facts presented in A.1. Using the relevant special cases from (13) and (14) leads to equality three and consequently applying Fact 6 and Fact 8 results in equality four. Re-applying (14) and Fact 8 results in the final equality. = P (k)A[k] + (A[k]) P (k) + Θ(k) (γ2P)[k] + (P 1)[k] Θ(k) = (p In)(k)A[k] + (A[k]) (p In)(k) + (p 1 2 In)(k) (γ2p In)[k] + (p 1In)[k] (p 1 2 In)(k) = pk A[k] + (A[k]) + k(γ2p + p 1)pk Is = pk (A + A )[k] + k(γ2p + p 1)Is = pk A + A + (γ2p + p 1)In [k] If the matrix above is negative definite, then (20) is satisfied for some suitably chosen η1 > 0. This is true when A + A + (γ2p + p 1)In [k] 0 By Fact 7, the inequality above can be equivalently expressed as a condition on the sum of the k largest eigenvalues of the matrix inside the k-compound operator. Leveraging (Petersen et al., 2008, Eq. 285) allows us to separate p from the eigenvalues of the symmetric component of A, resulting in the equality below. i=1 λi A + A + (γ2p + p 1)In = k(γ2p + p 1) + i=1 λi(A + A ) < 0 Published in Transactions on Machine Learning Research (04/2025) By the definition of αk in Theorem 2, this simplifies to γ2p2 + 2αkp + 1 < 0 For γ satisfying (22), the quadratic inequality always emits at least one solution p = α 1 k . The final step is to verify (21). The Jacobian of Ψ, as defined in (23), is JΨ(x) = γ 1BJΦC For Θ = p 1 2 In and the definitions from (23), the left hand side of (21) reduces to = p 1J Ψ JΨ p 1In [k] If the matrix above is negative definite, then (21) is satisfied for some suitably chosen η2 > 0. Repeating the same steps as above, this negative definite requirement reduces to the inequality below. i=1 λi(p 1J Ψ JΨ) kp 1 = p 1 k X i=1 σ2 i (JΨ) kp 1 < 0 Subbing in the definition of JΨ and applying the well-known property of singular values (Horn & Johnson, 1994, Theorem 3.3.14), then (21) is true if i=1 σ2 i (B)σ2 i (JΦ)σ2 i (C) < k By the assumption made on the slope of Φ, this inequality will always be satisfied if (3) holds. C.2 Proof of Theorem 3 For this proof, we aim to directly verify Theorem 1 for Θ Dn where f represents the Lurie network (2). We start by substituting the Jacobian of the Lurie network into the left hand side of (1), followed by the application of Fact 8 to obtain the second equality. The subadditivity property of the matrix measure µ2 was then leveraged to split the terms (Vidyasagar, 2002, Section 2.2). As the second term is difficult to manipulate, we rely on the simplifying assumption C = B 1 in order to apply Fact 9. As Θ, B, JΦ Dn, so are both of their k-compound counterparts (13) (14), which means the Θ, B terms cancel out. We then apply the property µ2( ) || ||2 (Vidyasagar, 2002, Theorem 16) and Fact 7, which allows us to leverage the slope restricted assumption on Φ. By the relevant special case of the k-additive compound (14), we can directly calculate the 2-norm. Finally, kg can be incorporated in µ2 as shown in the final inequality. µ2,Θ(k)(J[k] f (t, x)) = µ2,Θ(k) (A + BJΦC)[k] = µ2,Θ(k) A[k] + (BJΦ(y)C)[k] µ2,Θ(k)(A[k]) + µ2,Θ(k) (BJΦC)[k] = µ2,Θ(k)(A[k]) + µ2(Θ(k)B(k)J[k] Φ B (k)Θ (k)) = µ2,Θ(k)(A[k]) + µ2(J[k] Φ ) µ2,Θ(k)(A[k]) + ||(g In)[k]||2 = µ2,Θ(k)(A[k]) + kg = µ2(Θ(k)A[k]Θ (k) + kg In) If the final inequality is negative, then Theorem 1 will be satisfied for some suitably chosen η. This is equivalent to the matrix inequality below. 1 2 Θ(k)A[k]Θ (k) + Θ (k)(A[k]) Θ(k) + 2kg In 0 Multiplying on the left by 2Θ(k) and on the right by Θ(k) results in (4). Published in Transactions on Machine Learning Research (04/2025) C.3 Proof of Theorem 4 The aim of this proof is to verify Theorem 1 when f is the GLN (6). We begin by expressing the Jacobian as a sum of the Jacobian of the q independent Lurie networks, Jindep, and Jacobian of the coupling term Jcouple Jf(x) = Jindep(x) + Jcouple(x) where Jindep(x) = AG + BGJΦCG and Jcouple(x) = L. Subbing Jf into the left hand side of (1) and applying the subadditivity property of µ2 results in µ2,Θ(k)(J[k] f ) µ2,Θ(k)(J[k] indep) + µ2,Θ(k)(J[k] couple) As we assume the q independent Lurie networks are k-contracting in the 2-norm w.r.t the metric P = blockdiag(P1, . . . , Pq), we know that µ2,Θ(k)(J[k] indep) < 0. Under this assumption, Theorem 1 is satisfied if µ2,Θ(k)(J[k] couple) 0 This is equivalent to the matrix inequality below, where Jcouple has been subbed in. 1 2 Θ(k)L[k]Θ (k) + Θ (k)(L[k]) Θ(k) 0 Multiplying on the left by 2Θ(k) and on the right by Θ(k) results in (7). C.4 Proof of Theorem 5 and Theorem 6 The aim of this proof is to express the weights of the Lurie network (2) such that Theorem 2 is always satisfied. More formally, this requires A, B, C Ω2(g, k) to always hold. As both theorems share the same proof until the final step, where ΣA is defined, they have been combined into one proof. To expose the singular values of B, C, we leverage the singular value decomposition, as in (9b) and (10a). This requires the matrices UB, UC, VB, VC to be orthogonal. We can immediately use the unconstrained parametrisation of the orthogonal class from Lezcano-Casado & Martınez-Rubio (2019) to express these matrices as unconstrained symmetric matrices. The matrices ΣB, ΣC contain the singular values of the respective matrix, hence ΣB Dnm + and ΣC Dmn + . We also treat these as unconstrained sets since any element can be obtained by taking the absolute value of an unconstrained diagonal matrix with the same shape. To verify Theorem 2, we can combine αk < 0 and (3) into one inequality representing the intersection of the two sets. i=1 λi(A + A ) < 2k rzk k where zk := g2 k X i=1 σ2 i (B)σ2 i (C) (24) Thanks to the definition of B and C, the right hand side is a function of the hyperparameters g, k and elements of the parameters ΣB, ΣC, so can be easily computed using sort and sum functions. To impose this constraint directly on the eigenvalues of the symmetric component of A, we express A as a sum of symmetric and skew-symmetric matrices. The skew-symmetric matrix is unconstrained, so this can be left as it is. Finally, we apply the eigenvalue decomposition of a symmetric matrix to obtain (9a) (10a). This expression allows us to directly place the constraint above on the diagonal matrix ΣA. This is where Theorem 5 and Theorem 6 differ. Defining ΣA as in (9a) ensures λi(A + A ) < 2 rzk k for all i [1, n] which guarantees both conditions of Theorem 2 will be satisfied; however, conservatism is introduced as all the eigenvalues must be negative. Theorem 6 was established to address this issue. Defining ΣA as in (10b) Published in Transactions on Machine Learning Research (04/2025) guarantees both conditions of Theorem 2 will be satisfied via (24). The definition of ΣA is split into one unconstrained block for the first (k 1)-eigenvalues, a block for λk(A + A ) which is defined to ensure (24) holds, and finally a block for the remaining eigenvalues which must be defined to ensure the k eigenvalues involved in (24) are the largest. C.5 Intersection Proof The set Ω2(g, k) is defined by (8a). For k [2, n], the set of (k 1)-contracting systems is denoted by Ω2(g, k 1) = n (A, B, C) αk 1 < 0 , zk 1 < α2 k 1(k 1) o Theorem 10 For k {2, 3}, the sets Ω2(g, k 1) and Ω2(g, k) intersect when g2σ2 k(B)σ2 k(C) < 1 4k λ2 k + k 1 k αk 1λk k 1 k α2 k 1 (25) We first note the following useful relationships, where λk := λk(A + A ). 2k λk + k 1 k αk 1 zk = g2σ2 k(B)σ2 k(C) + zk 1 (26) From (26) it is clear that αk < 0 is true when λk < 2(k 1)αk 1 A subset of these solutions occurs when the first condition of Ω2(g, k 1) holds. That is αk 1 < 0. From (26) it is also clear that zk < α2 kk is true when zk 1 < α2 kk g2σ2 k(B)σ2 k(C) If α2 k 1(k 1) < α2 kk g2σ2 k(B)σ2 k(C) then a subset of these solutions occurs when the second condition of Ω2(g, k 1) holds. After rearranging, this is equivalent to g2σ2 k(B)σ2 k(C) < α2 kk α2 k 1(k 1) The RHS can equivalently be expressed by a quadratic equation in λk α2 kk α2 k 1(k 1) = 1 2k λk + k 1 k αk 1 2 k α2 k 1(k 1) 4k2 λ2 k + k 1 k2 αk 1λk + (k 1)2 k2 α2 k 1 k α2 k 1(k 1) 4k λ2 k + k 1 k αk 1λk + (k 1)2 k α2 k 1 α2 k 1(k 1) 4k λ2 k + k 1 k αk 1λk k 1 := Qk(λk, αk 1) Therefore, Ω2(g, k 1) intersects with Ω2(g, k) if (25) holds. Published in Transactions on Machine Learning Research (04/2025) Finding the roots of Qk(λk, αk 1) will allow us to check what intervals of λk the Qk(λk, αk 1) > 0. This is required since the LHS of (25) must be positive. λr k = 2(1 k)αk 1 2k k2 α2 k 1 + k 1 = 2(1 k)αk 1 2 p k(k 1) αk 1 The roots for k = 2 and k = 3 are given by λr 2 = ( 2 2 2)α1 λr 3 = ( 4 2 Given αk 1 < 0, we will test the intervals λ2 [ , ( 2+2 2)α1] and λ3 [ , ( 4+2 6)α2] to see if the co-domain of Qk(λk, αk 1) is positive. These intervals ensure λk < 0 (a requirement for αk < 0 and αk 1 < 0). Using λ2 = α1 and λ3 = α2 as the test points, we see that Q2(α1; α1) = 1 8α2 1 > 0 (28) Q3(α2; α2) = 1 12α2 2 > 0 (30) As a result, for k {2, 3}, the sets Ω2(g, k 1) and Ω2(g, k) intersect if λk [ , 2(1 k + p k(k 1))αk 1] and g2σ2 k(B)σ2 k(C) < Qk(λk, αk 1). If (A, B, C) Ω2(g, 2) then λ2 must be in this interval since λ2 λ1. By the same logic, if (A, B, C) Ω2(g, 3) then λ3 must be in this interval since λ3 λ2 λ1 . Hence, the requirements for intersection can be simplified to just (25). C.6 Proof of Theorem 7 This proof aims to show that Theorem 4 is satisfied when L is defined as in (12). First, multiply (7) on the left and right by Θ (k). Sequentially applying Fact 6, Fact 9 and Fact 8 results in ΘLΘ 1 + Θ 1L Θ [k] 0 Subbing in the definition of L from (12) and recalling that P = Θ Θ leads to GL2 + G L2 [k] 0 which is assumed to hold. Published in Transactions on Machine Learning Research (04/2025) D Extended Empirical Evaluation The data was synthetically generated by numerically integrating over the analytical models of each dynamical system. The integration was performed using the Euler method with step size δ = 1 10 2. D.1.1 Opinion Dynamics This model was presented in Ofir et al. (2024). It has the following state space equations x = 1.5I3 + 0.5Φ(Cx) + b +1 1 0 1 +1 1 0 1 +1 and the tanh function is applied element-wise. The model is 1-contracting and has a unique equilibrium point at b. D.1.2 Hopfield Network This model is a variation of the Hopfield network presented in Ofir et al. (2024). It has the following state space equations x = 2.5I3 + BΦ(x) 1 1 1 1 1 1 1 1 1 and the tanh function is also applied element-wise. The model is 2-contracting and has two stable equilibrium points: e1 = 0.79 0.79 0.79 , e2 = e1; and an unstable equilibrium point e3 = 0. D.1.3 Simple Attractor This model was presented in Cecilia et al. (2023). It has the following state space equations x = Ax + BΦ(Cx) 0 1 2 1 0 1 0.5 0 0.5 0 0 0 0 0 0 0.5 0 0 0 0 0 0 0 0 1 0 0 and ϕ(z) = z3 is the nonlinearity applied element-wise. This function is not slope-restricted, so the simple attractor does not satisfy the assumptions of the Lurie network. The model is 3-contracting and has several attractor states. Published in Transactions on Machine Learning Research (04/2025) D.2 Training The default training settings common to the opinion, Hopfield and attractor datasets are presented in Table 5. The only parameters which varied between models were the learning rate and epoch which it was cut. Deviations from the default settings are detailed in Table 6. These values were chosen based on observations during training. No hyperparameter sweep was performed. The same details are also presented for the graph-coupled Hopfield network and graph-coupled attractor datasets in Table 7 and Table 8. When training the models for the opinion, Hopfield and attractor datasets, a single T4 GPU (accessed through Google Colab) was used. A single A100 GPU (also accessed through Google Colab) was used for training the models on the graph-coupled Hopfield and graph-coupled attractor datasets. Table 5: Default training settings for opinion, Hopfield and attractor datasets. Parameter Value Batches 10 Batch size 100 Test split 0.1 Epochs 100 Criterion Mean squared error Optimiser Adam (default settings) Learning rate (LR) 1 10 2 (no decay) Table 6: Deviations from the default settings for the opinion, Hopfield and attractor datasets. LR decay includes the size of the decay followed by the epoch the decay was made. Dataset Model LR LR Decay Opinion Dynamics Lurie Network 5 10 3 Opinion Dynamics Antisymmetric RNN 5 10 3 - Hopfield Network k-Lurie Network 5 10 3 - Hopfield Network Neural ODE 1 10 3 0.1, 75 Hopfield Network Antisymmetric RNN 5 10 3 - Simple Attractor Neural ODE 1 10 3 - Simple Attractor Antisymmetric RNN 5 10 3 - Table 7: Default training settings for the graph-coupled Hopfield and attractor datasets. Parameter Value Batches 15 Batch size 2000 Test split 1 15 Epochs 100 Criterion Mean squared error Optimiser Adam (default settings) Learning rate (LR) 5 10 3 (no cuts) Published in Transactions on Machine Learning Research (04/2025) Table 8: Deviations from the default settings for the graph-coupled Hopfield and attractor datasets. LR decay includes the size of the decay followed by the epoch the decay was made. Dataset Model LR LR decay GC Hopfield Network Lipschitz RNN 5 10 3 0.2, 60 GC Hopfield Network GLN 3 10 3 1 3, 60 GC Simple Attractor GLN 1 10 2 - GC Simple Attractor k-Lurie Network 1 10 2 - GC Simple Attractor Neural ODE 5 10 3 0.5, 40 D.3 Further Results Figure 3: Random sample of trajectories from the opinion dynamics test set. Predictions are made by the model associated with the best MSE in Table 1. Crosses denote initial conditions, stars denote equilibrium points. Published in Transactions on Machine Learning Research (04/2025) Figure 4: Random sample of 30s trajectories from the noise-free, out of distribution opinion dynamics dataset. Predictions are made by the model associated with the best MSE in Table 1. Crosses denote initial conditions, stars denote equilibrium points. Figure 5: Random sample of 30s trajectories from the noisy, out of distribution opinion dynamics dataset. Predictions are made by the model associated with the best MSE in Table 1. Crosses denote initial conditions, stars denote equilibrium points. Published in Transactions on Machine Learning Research (04/2025) Figure 6: Random sample of trajectories from the Hopfield network test set. Predictions are made by the model associated with the best MSE in Table 1. Crosses denote initial conditions, stars denote equilibrium points. Figure 7: Random sample of 30s trajectories from the noise-free, out of distribution Hopfield network dataset. Predictions are made by the model associated with the best MSE in Table 1. Crosses denote initial conditions, stars denote equilibrium points. Published in Transactions on Machine Learning Research (04/2025) Figure 8: Random sample of 30s trajectories from the noisy, out of distribution Hopfield network dataset. Predictions are made by the model associated with the best MSE in Table 1. Crosses denote initial conditions, stars denote equilibrium points. Figure 9: Random sample of trajectories from the simple attractor test set. Predictions are made by the model associated with the best MSE in Table 1. Crosses denote initial conditions. Published in Transactions on Machine Learning Research (04/2025) Figure 10: Random sample of 30s trajectories from the noise-free, out of distribution simple attractor dataset. Predictions are made by the model associated with the best MSE in Table 1. Crosses denote initial conditions. Figure 11: Random sample of 30s trajectories from the noisy, out of distribution simple attractor dataset. Predictions are made by the model associated with the best MSE in Table 1. Crosses denote initial conditions.