# on_sparse_modern_hopfield_model__cb8ae34e.pdf On Sparse Modern Hopfield Model Jerry Yao-Chieh Hu Donglin Yang Dennis Wu Chenwei Xu Bo-Yu Chen Han Liu Department of Computer Science, Northwestern University, Evanston, IL 60208 USA Department of Physics, National Taiwan University, Taipei 10617, Taiwan Department of Statistics and Data Science, Northwestern University, Evanston, IL 60208 USA {jhu, dlyang, hibb, cxu}@u.northwestern.edu b12202023@ntu.edu.tw, hanliu@northwestern.edu We introduce the sparse modern Hopfield model as a sparse extension of the modern Hopfield model. Like its dense counterpart, the sparse modern Hopfield model equips a memory-retrieval dynamics whose one-step approximation corresponds to the sparse attention mechanism. Theoretically, our key contribution is a principled derivation of a closed-form sparse Hopfield energy using the convex conjugate of the sparse entropic regularizer. Building upon this, we derive the sparse memory retrieval dynamics from the sparse energy function and show its one-step approximation is equivalent to the sparse-structured attention. Importantly, we provide a sparsity-dependent memory retrieval error bound which is provably tighter than its dense analog. The conditions for the benefits of sparsity to arise are therefore identified and discussed. In addition, we show that the sparse modern Hopfield model maintains the robust theoretical properties of its dense counterpart, including rapid fixed point convergence and exponential memory capacity. Empirically, we use both synthetic and real-world datasets to demonstrate that the sparse Hopfield model outperforms its dense counterpart in many situations. 1 Introduction We address the computational challenges of modern Hopfield models by introducing a sparse Hopfield model. Our sparse continuous Hopfield model equips a memory-retrieval dynamics that aligns with the sparse-structured attention mechanism. By establishing a connection to sparse attention, the proposed model not only offers a theoretically-grounded energy-based model for associative memory but also enables robust representation learning and seamless integration with deep learning architectures. This approach serves as an initial attempt of pushing the correspondence1 between Hopfield models and attention mechanism [Ramsauer et al., 2021] toward sparse region, both theoretically and empirically, resulting in data-dependent sparsity for meaningful and robust pattern representations, and a focus on the most relevant information for each specific instance. Hopfield models are classic associative memory models for both biological and artificial neural networks [Hopfield, 1982, 1984]. These models are designed to store and retrieve memory patterns. They achieve these by embedding the memories in the energy landscape of a physical system (e.g., the Ising model in [Hopfield, 1982, Peretto and Niez, 1986]; see Figure 3 for a visualization), where each memory corresponds to a local minimum. When a query is presented, the model initiates energy-minimizing retrieval dynamics at the query, which then navigate the energy landscape to find the nearest local minimum, effectively retrieving the memory most similar to the query. Code is available at Git Hub; future updates are on ar Xiv. [Version: Novermber, 28, 2023] 1While this equivalence only holds when the retrieval dynamics is applied exactly once, as originally shown in [Ramsauer et al., 2021] and later emphasized in [Krotov and Hopfield, 2021], it allows us to view modern Hopfield models as generalized attentions with additional functionalities and hence opens new avenues for Hopfield-based architecture designs. See Appendix C for more discussions. 37th Conference on Neural Information Processing Systems (Neur IPS 2023). In the same vein, Ramsauer et al. [2021] propose the modern Hopfield model and integrate it into deep learning architectures via a strong connection with transformer attention, offering enhanced performance, theoretically guaranteed exponential memory capacity, and the ability to handle continuous patterns. In addition, the modern Hopfield models have found success in various applications, such as immunology [Widrich et al., 2020] and large language model [Fürst et al., 2022]. Apart from the elegant connection to attention, theoretical advantages and empirically successes, the modern Hopfield models have been shown to be computationally heavy and vulnerable against noisy queries [Millidge et al., 2022]. In particular, the dense output alignments of the retrieval dynamics in modern Hopfield models [Ramsauer et al., 2021] can be computationally inefficient, making models less interpretable and noise-sensitive by assigning probability mass to many implausible outputs (patterns/keys). To combat above, incorporating sparsity is an essential and common strategy. While there is a vast body of work on sparsifying attention mechanisms [Tay et al., 2022, Beltagy et al., 2020, Qiu et al., 2019, Child et al., 2019, Peters et al., 2019, Martins and Astudillo, 2016], similar developments for the Hopfield models remain less explored. To bridge this gap, we present a sparse Hopfield model that corresponds to the sparsemax attention mechanism [Martins and Astudillo, 2016]. In this paper, we study the sparsification of the modern Hopfield model. The challenges are three-fold: (C1) Non-Trivial Sparsification Sparse Hopfield Sparse Attention: To enable the use of sparse Hopfield models as computational devices (DNN learning models) akin to [Ramsauer et al., 2021], it is essential to achieve non-trivial sparsifications that exhibit equivalence to specific sparse attention models. In other words, any meaningful sparsification should extend the established equivalence [Ramsauer et al., 2021] between modern Hopfield models and attention to encompass the sparse domain. While generalizing such equivalence is potentially impactful as it may lay the groundwork for future Hopfield-based methodologies, architecture designs and bio-computing systems (as in [Kozachkov et al., 2023]), the heuristic design of the modern Hopfield model poses great difficulty to developing desired sparse models. (C2) Introducing Sparsity into Hopfield Models: Unlike attention mechanisms where sparsification is typically achieved either on the attention matrix (e.g., structured-sparsity [Tay et al., 2020, Child et al., 2019]) or on the element-wise normalization map (e.g., sparsity-inducing maps [Correia et al., 2019, Peters et al., 2019, Martins and Astudillo, 2016]), the sparsification of Hopfield models is applied to both the energy function and the memory-retrieval dynamics, where the latter monotonically decreases the Hopfield energy over time. Since attention mechanisms (transformers) are typically not equipped with such a dynamical description, introducing sparsity into Hopfield models while retaining the connection to attention is a less straightforward process. (C3) Properties of the Sparse Hopfield Model: Further, it is unclear how the introduced sparsity may affect different aspects of the model, such as memory capacity, fixed point convergence, retrieval accuracy, and so on. Ideally, we are looking for sparsities that offer provable computational benefits, such as enhanced robustness and increased memory capacity, among others. Challenges (C1) and (C2) are inherent in Hopfield model, and certain requirements on the design of energy function and retrieval dynamics are inevitable to obtain non-trivial sparse models. Hence, we suppose the sparsified models should satisfy some conditions and verify them accordingly. Concretely, a formulation for deriving desired sparse Hopfield energy via convex conjugation of entropic regularizers is proposed. Furthermore, by applying Danskin s theorem and convex-concave procedure [Yuille and Rangarajan, 2003, 2001] on the sparse Hopfield energy function, we obtain sparse retrieval dynamics linked to sparse attention. For (C3), the convergence of energy stationary points and retrieval dynamics fixed points are connected via Zangwill s method [Zangwill, 1969]. The sparse retrieval error bound is derived and used to determined the well-separation condition for successful memory storage and retrieval. Lastly, the fundamental limit of memory capacity is derived using the expected separation of random points on spheres [Cai and Jiang, 2012, Brauchart et al., 2018, Ramsauer et al., 2021]. In summary, this work handles sparsification of modern Hopfield models while linking them to sparse attention by addressing the following question: Is it possible to develop a theoretically-grounded (non-trivial) sparse Hopfield model capable of storing information or learned prototypes throughout various layers of DNN models? Contributions. We propose the Sparse Modern Hopfield Model. Our contributions are as follows: We propose a novel sparse Hopfield model whose retrieval dynamics corresponds to sparsemax attention mechanism. It leads to sparse patterns by design, inheriting both noise robustness and potential computational efficiency2 from [Martins and Astudillo, 2016], compared to its dense counterparts. This work extends the theoretical understanding of the correspondence between artificial and biological neural networks to sparse region. In addition, the sparse Hopfield layer, a new deep learning component, is introduced with data-dependent sparsity. Theoretically, we establish provably advantages from sparsity and identify the conditions under which these benefits arise. We begin by deriving the closed-form sparse Hopfield energy from the convex conjugation of sparse entropic regularizer. Next, we demonstrate the correspondence between sparse Hopfield retrieval dynamics and sparsemax attention. In addition, we prove the fast convergence of the fixed points (also known as memory patterns, attractor states in literature) for the retrieval dynamics and establish the exponential (in pattern size) memory capacity lower bound with tighter retrieval error bound, compared with modern Hopfield models. Empirically, we conduct synthetic and realistic experiments to verify our theoretical results and proposed methodology. Specifically, the sparse Hopfield model outperforms the dense Hopfield model and machine learning baselines in sparse Multiple Instance Learning (MIL), time series prediction and neural machine translation problems. This is observed with both sparse synthetic and real-world datasets, where the baselines tend to fall short. Moreover, even in cases without data sparsity, our proposed model delivers performance on par with its dense counterpart. To the best of our knowledge, we are the first to propose a sparse Hopfield model whose retrieval dynamics is equivalent to sparse attention mechanism with provably computational advantages. Methodologically, the proposed model complements existing Hopfield-based DNN architectures [Hoover et al., 2023, Paischer et al., 2022, Seidl et al., 2022, Fürst et al., 2022, Ramsauer et al., 2021] by introducing a sparse Hopfield layer into deep learning models. Organization. In Section 2, the sparse Hopfield model is introduced. In Section 3, the memory capacity is discussed. In Section 4, experimental studies are conducted. In Section 5, concluding discussions are provided. Additionally, related works and limitations are discussed in Appendix C. Notations. We write a, b := a Tb as the inner product for vectors a, b Rd. The index set {1, , I} is denoted by [I], where I N+. The spectral norm is denoted by , which is equivalent to the l2-norm when applied to a vector. Throughout this paper, we denote the memory patterns (keys) by ξ Rd and the state/configuration/query pattern by x Rd, and Ξ := (ξ1, , ξM) Rd M as shorthand for stored memory (key) patterns {ξµ}µ [M]. Moreover, we set norm n := x be the norm of the query pattern, and m := Maxµ [M] ξµ be the largest norm of memory patterns. We also provide a nomenclature table (Table 3) in the appendix. 2 Sparse Hopfield Model In this section, we introduce the sparse Hopfield energy from convex conjugate of entropic regularizer, and then the sparse retrieval dynamics. In this paper we only consider the Gini entropic regularizer corresponding to the sparsemax distribution [Martins and Astudillo, 2016]. Let x Rd represent the query pattern, and let Ξ := (ξ1, , ξM) Rd M denote the memory patterns. The objective of the Hopfield models is to store the memory patterns Ξ and then retrieve a specific memory pattern ξµ based on a given query x. Consequently, any Hopfield model consist of two main components: an energy function H(x), encoding memories into its local minima, and a retrieval dynamics T (x), which retrieves a memory by iteratively minimizing H(x) when initialized with a query. We provide a visualization of this procedure in Figure 3. The construction of the energy function H(x) is straightforward. As emphasized in [Krotov and Hopfield, 2016], the memories can be easily encoded into H(x) through the overlap-construction: H(x) = F(ΞTx), where F : RM R is a smooth function. This ensures that the memories {ξµ}µ [M] are located at the stationary points of H(x), since x F(ΞTx)|ξµ = 0 for all µ [M]. Different choices of F lead to different Hopfield models, as demonstrated in [Krotov and Hopfield, 2016, Demircigil et al., 2017, Ramsauer et al., 2021, Krotov and Hopfield, 2021]. However, finding a corresponding 2Note that, the proposed model s sparsity falls under the category of sparsity-inducing normalization maps. Consequently, the forward pass still requires O(n2) space complexity. Here, potential computational efficiency refers that the computational efficiency can be enhanced if one employs efficient implementations that leverage sparsity, such as sort operations or median-finding algorithms, to circumvent unnecessary computations, see Appendix C and [Martins and Astudillo, 2016, Section 2] for more discussions. retrieval dynamics, T , for a given energy H(x), is generally more challenging. This is because T needs to satisfy two conditions to ensure successful memory retrieval: (i) To ensure consistent retrieval, an appropriate T should monotonically minimize H(x) when iteratively applied. (ii) To ensure accurate retrieval, an appropriate T should align its fixed points (the points where iterative application terminates) with the stationary points of H(x). To this end, we introduce the sparse Hopfield model, providing a principled construction for H and T . This model not only fulfills the aforementioned desirable properties, but also enables more robust and faster memory retrieval compared to the modern Hopfield model [Ramsauer et al., 2021]. 2.1 Sparse Hopfield Energy Let x Rd be the query pattern, and Ξ := (ξ1, , ξM) Rd M be the memory patterns. We introduce the sparse Hopfield energy as H(x) = Ψ βΞTx + 1 2 x, x , (2.1) with Ψ (z) := 1 2 Sparsemax(z) z 2 + 1 2, where Sparsemax( ) is defined as follows. Let z, p RM, and M := {p RM + | PM µ pµ = 1} be the (M 1)-dimensional unit simplex. Definition 2.1 (Sparsemax in Variational Form [Martins and Astudillo, 2016], also see Remark F.1). Sparsemax(z) := Arg Min p M p z 2 = Arg Max p M p Tz Ψ(p) , (2.2) where Ψ(p) := 1 2 PM ν pν(1 pν) is the negative Gini entropy or Gini entropic regularizer. Remark 2.1. Recall that, the variational form (2.2) is in fact general, that applies to various entropic regularizers, as discussed in [Peters et al., 2019, Wainwright et al., 2008]. The choice of Ψ determines the resulting sparse probability distribution. For instance, if we choose the Gibbs entropic regularizer ΨGibbs = PM ν pν ln pν, (2.2) reduces to the standard softmax distribution. Overview of Theoretical Results. At first glance, the energy function (2.1) may seem peculiar. However, it indeed represents a non-trivial sparse Hopfield model with appealing properties, including: (i) In response to challenge (C1) & (C2), as we shall see in Section 2.2, the energy (2.1) leads to a sparse retrieval dynamics that not only retrieves memory by monotonically decreasing (Lemma 2.1) to its stationary points (Lemma 2.2), but also associates with sparsemax attention through its single-step approximation (Remark 2.2); (ii) In response to challenge (C3), as we shall see in Section 3, it indulges fast convergence of retrieval (Corollary 3.1.2), exponential-in-d memory capacity akin to modern Hopfield models (Lemma 3.1). Notably, it accomplishes these with a tighter retrieval error bound (Theorem 2.1). We reveal each of these properties in the following sections. 2.2 Sparse Retrieval Dynamics and Connection to Sparse Attention The optimization problem Arg Maxp M p Tz Ψ(p) does not necessarily have a closed-form solution for arbitrary Ψ. However, a family of Ψ has been investigated in literature [Correia et al., 2019, Martins and Astudillo, 2016] with closed-form solutions derived, including the Sparsemax( ). Sparsemax in Closed-Form (Proposition 1 of [Martins and Astudillo, 2016]). Let z RM. Denote [a]+ := Max{0, a}, z(ν) the ν th element in a sorted descending z-sequence zsorted := z(1) z(2) . . . z(M), and κ(z) := Max k [M] 1 + kz(k) > P ν k z(ν) . The optimization problem(s) (2.2) has closed-form solution Sparsemax(z) = [z τ(z)1M]+ , (2.3) where τ : RM R is the threshold function τ(z) = h P ν κ(z) z(ν) 1 i /κ(z), satis- fying PM µ=1 [zµ τ(z)]+ = 1 for all z. Notably, κ(z) = |S(z)| where S(z) = {µ [M] | Sparsemaxµ(z) > 0} is the support set of Sparsemax(z). In this case, we present the following theorem to derive the convex conjugate of Ψ in closed-form: Theorem 2.1 (Convex Conjugate of Negative Gini Entropy). Let F(p) := p, z Ψ(p) with Ψ being the negative Gini entropy, Ψ(p) = 1 2. The convex conjugate of Ψ(p) is Ψ (z) := Max p M F(p, z) = 1 2 p z 2 + 1 where p = Sparsemax(z) is given by (2.3). Corollary 2.1.1. By Danskin s Theorem, Ψ (z) = Arg Maxp M F(p, z) = Sparsemax(z). Proof. A detailed proof is shown in Appendix E.1. Theorem 2.1 and Corollary 2.1.1 not only provide the intuition behind the sparse Hopfield energy (2.1) the memory patterns are stored in local minima aligned with the overlap-function constructions (i.e. ΞTx 2 = PM µ=1 ξµ, x 2) in [Ramsauer et al., 2021, Demircigil et al., 2017, Krotov and Hopfield, 2016] but also prepare us for the following corresponding sparse retrieval dynamics. Lemma 2.1 (Sparse Retrieval Dynamics). Let t be the iteration number. The energy (2.1) can be monotonically decreased by the following sparse retrieval dynamics over t: T (xt) := xΨ βΞTx xt = Ξ Sparsemax βΞTxt = xt+1. (2.5) Proof Sketch. To show monotonic decreasing property, we first derive the sparse retrieval dynamics by utilizing Theorem 2.1, Corollary 2.1.1, along with the convex-concave procedure [Yuille and Rangarajan, 2003, 2001]. Then, we show the monotonicity of H by constructing a iterative upper bound of H which is convex in xt+1 and thus, can be lowered iteratively by the convex-concave procedure. A detailed proof is shown in the Appendix E.2. Remark 2.2. Similar to [Ramsauer et al., 2021], (2.5) is equivalent to sparsemax attention [Martins and Astudillo, 2016] when the T is applied only once, see Appendix D for more details. Importantly, β acts as a scaling factor for the energy function, often referred to as the inverse temperature . It influences the sharpness of energy landscape Equation (2.1), thereby controlling the dynamics. High β values, corresponding to low temperatures, encourage that the basins of attraction for individual memory patterns remain distinct, leading to easier retrieval. Notably, since ΞTx 2 = PM µ=1 ξµ, x 2, (2.5) implies that the local optimum of H are located near the patterns ξµ. Different from previous studies on binary Hopfield models [Demircigil et al., 2017, Krotov and Hopfield, 2016], for continuous patterns, we adopt the relaxed definition from [Ramsauer et al., 2021]3 to rigorously analyze the memory retrieval, and the subsequent lemma. Definition 2.2 (Stored and Retrieved). Assuming that every pattern ξµ surrounded by a sphere Sµ with finite radius R := 1 2 Minµ =ν;µ,ν [M] ξµ ξν , we say ξµ is stored if there exists a generalized fixed point of T , x µ Sµ, to which all limit points x Sµ converge to, and Sµ Sν = for µ = ν. We say ξµ is ϵ-retrieved by T with x for an errora ϵ, if T (x) ξµ ϵ. a The retrieval error has a naive bound ϵ := Max x ξµ , ξµ x µ by interpolating from x to ξµ. Definition 2.2 sets the threshold for a memory pattern ξµ to be considered stored at a fixed point of T , x µ. However, this definition does not imply that the fixed points of T are also stationary points of the energy function H. In fact, monotonicity of (2.5) does not assure the existence of stationary points of energy H [Sriperumbudur and Lanckriet, 2009]. To establish a well-defined Hopfield model, we need two types of convergence guarantees. The first is the convergence between x µ and ξµ, which ensures that the retrieved memory is close to the stored memory. The second is the convergence of H to its stationary points through the dynamics of T , which ensures that the system reaches a state of minimal energy. The following lemma provides the convergence results for both. Lemma 2.2 (Convergence of Retrieval Dynamics T ). Suppose H is given by (2.1) and T (x) is given by (2.5). For any sequence {xt} t=0 defined by xt +1 = T (xt ), all limit points of this sequence are stationary points if they are obtained by iteratively applying T to H. 3Recall that a fixed point of T with respect to H is a point where x = T (x), and a generalized fixed point is a point where x T (x). For more details, refer to [Sriperumbudur and Lanckriet, 2009]. Proof Sketch. We verify and utilize Zangwill s global convergence theory [Zangwill, 1969] for iterative algorithms T , to first show that all the limit points of {xt} t=0 are generalized fixed points and limt H (xt) = H (x ), where x are some generalized fixed points of T . Subsequently, by [Sriperumbudur and Lanckriet, 2009, Lemma 5], we show that {x } are also stationary points of Minx [H], and hence H converges to local optimum. A detailed proof is shown in Appendix E.4. Intuitively, Lemma 2.2 indicates that the energy function converges to local optimum, i.e. limt H (xt) H (x ) , where x are stationary points of H. Consequently, it offers formal justifications for the retrieval dynamics (2.5) to retrieve stored memory patterns {ξµ}µ [M]: for any query (initial point) x, T monotonically and iteratively approaches stationary points of H, where the memory patterns {ξµ}µ [M] are stored. As for the retrieval error, we provide the following theorem stating that T achieves a lower retrieval error compared to its dense counterpart. Theorem 2.2 (Retrieval Error). Let TDense be the retrieval dynamics of the dense modern Hopfield model [Ramsauer et al., 2021]. It holds T (x) ξµ TDense(x) ξµ for all x Sµ. Moreover, T (x) ξµ m + d 1/2mβ κ Max ν [M] ξν, x ΞTx (κ) is the κth-largest element of ΞTx RM following the sparsemax definition (2.3). Proof. A detailed proof is shown in Appendix E.3. Interestingly, (2.6) is a sparsity dependent bound4. By denoting n := x , the second term on the RHS of (2.6) is dominated by the sparsity dimension κ as it can be expressed as κ 1 [ΞTx](κ)/(nm) ακ with a constant 0 α 2. When ΞTx is sparse (i.e. κ is small), the bound is tighter, vice versa. Remark 2.3 (Faster Convergence). Computationally, Theorem 2.2 implies that T requires fewer iterations to reach fixed points with the same amount of error tolerance compared to Tdense. Namely, T retrieves stored memory patterns faster and therefore more efficiently, as evidenced in Figure 2. Remark 2.4 (Noise-Robustness). Moreover, in cases of contaminated patterns with noise η, i.e. ex = x + η (noise in query) or eξµ = ξµ + η (noise in memory), the impact of noise η on the sparse retrieval error (2.6) is linear, while its effect on the dense retrieval error (2.7) is exponential. This suggests the robustness advantage of the sparse Hopfield model, as evidenced in Figure 1. 2.3 Sparse Hopfield Layers for Deep Learning The sparse Hopfield model can serve as a versatile component for deep learning frameworks, given its continuity and differentiability with respect to parameters. Corresponding to three types of Hopfield Layers proposed in [Ramsauer et al., 2021], we introduce their sparse analogs: (1) Sparse Hopfield, (2) Sparse Hopfield Pooling, (3) Sparse Hopfield Layer. Layer Sparse Hopfield has memory (stored or key) patterns Ξ and query (state) pattern x as inputs, and associates these two sets of patterns via the sparse retrieval dynamics (2.5). This layer regards the transformer attention layer as its one-step approximation, while utilizing the sparsemax [Martins and Astudillo, 2016] on attention matrix. Layer Sparse Hopfield Pooling and Layer Sparse Hopfield Layer are two variants of Sparse Hopfield, whose input patterns are memory patterns and query patterns from previous layers or external plugin, respectively. Sparse Hopfield Pooling, whose query patterns are learnable parameters, can be interpreted as performing a pooling operation over input memory patterns. Sparse Hopfield Layer, by contrast, has learnable memory patterns that maps query patterns to hidden states with sparsemax activation. Thus it can substitute a fully connected layer within deep learning architectures. See (D.12) and the implementation Algorithm 1 in Appendix D, and [Ramsauer et al., 2021, Section 3] for more details of these associations. In Section 4, we apply these layers and compare them with their dense counterparts in [Ramsauer et al., 2021] and other baseline machine learning methods. 3 Fundamental Limits of Memory Capacity of Sparse Hopfield Models How many patterns can be stored and reliably retrievable in the proposed model? We address this by decomposing it into to two sub-questions and answering them separately: (A) What is the condition for a pattern ξµ considered well stored in H, and correctly retrieved? 4Notably, T (x) ξµ is also upper-bounded by a sparsity-independent but M, β-dependent bound T (x) ξµ TDense(x) ξµ 2m(M 1) exp β ξµ, x Max ν [M],ν =µ ξµ, ξν . (2.7) (B) What is the number, in expectation, of the the patterns satisfying such condition? For (A), we first introduce the notion of separation of patterns following [Ramsauer et al., 2021], Definition 3.1 (Separation of Patterns). The separation of a memory pattern ξµ from all other memory patterns Ξ is defined as its minimal inner product difference to any other patterns: µ := Min ν,ν =µ [ ξµ, ξµ ξµ, ξν ] = ξµ, ξµ Max ν,ν =µ [ ξµ, ξν ] . (3.1) Similarly, the separation of ξµ at a given x from all memory patterns Ξ is given by e µ := Min ν,ν =µ [ x, ξµ x, ξν ] . (3.2) and then the well-separation condition for a pattern being well-stored and retrieved. Theorem 3.1 (Well-Separation Condition). Given the definition of stored and retrieved memories in Definition 2.2, suppose the memory patterns {ξµ}µ [M] are located within the sphere Sµ := x x ξµ R , where the radius R is finite and defined in Definition 2.2 for all µ. Then, the retrieval dynamics T maps the sphere Sµ onto itself under the following conditions: 1. The initial query x is located within the sphere Sµ, i.e., x Sµ. 2. The well-separation condition is satisfied, which is given by: µ mn + 2m R ΞTx Corollary 3.1.1. Let δ := TDense ξµ T ξµ . The well-separation condition can be expressed β ln 2(M 1)m R+δ + 2m R, which reduces to that of the dense Hopfield model when δ = 0. Proof Sketch. The proofs proceed by connecting µ with T (x) ξµ . To do so, we utilize Theorem 2.2 to incorporate the µ-dependent bound on the retrieval error of both sparse and dense Hopfield models [Ramsauer et al., 2021]. A detailed proof is shown in Appendix E.5. Together with Lemma 2.2, the well-separated condition serves as the necessary condition for pattern ξµ to be well-stored at the stationary points of H, and can be retrieved with at most ϵ = R by T , as per Definition 2.2. We make the following three observations about the blessings from sparsity. 1. In general, to appreciate the blessings of sparsity, we rearrange the well-separation condition as µ 2m R + mn ΞTx | {z } :=αnm with 0 α 2 and observe the two competing terms, αnm and (R m md 1/2)/(κmβd 1/2). Sparsity proves advantageous when the latter term surpasses the former, i.e. the sparse well-separation condition is consistently lower than its dense counterpart. The condition under which sparsity benefits are more likely to emerge (i.e., when the well-separation condition is more readily satisfied) is thereby: 1 2 Min µ,ν [M] ξµ ξν md 1/2 (1 + αβnmκ) + m, with 0 α 2. (3.4) Intuitively, the sparser ΞTx is, the easier it is for the above condition to be fulfilled. 2. Large M limit: For large M, the dense well-separation condition (Corollary 3.1.1) explodes while the sparse one (Theorem 3.1) saturates to the first three M-independent terms. This suggests that the hardness of distinguishing patterns can be tamed by the sparsity, preventing an increase of µ with M as observed in the dense Hopfield model. We numerically confirm this in Figure 1. 3. β Limit: In the region of low temperature, where β and hence all patterns can be error-free retrieved as per (2.7), we have µ 2m R + αnm with 0 α 2. Here, the second term on the RHS represents the sparsity level of ΞTx, i.e. a smaller α indicates a higher degree of sparsity in ΞTx. Hence, the higher the sparsity, the easier it is to separate patterns. For (B), equipped with Theorem 3.1 and Corollary 3.1.1, we provide a lower bound for the number of patterns being well-stored and can be at least R-retrieved in the next lemma5: 5Following the convention in memory capacity literature [Ramsauer et al., 2021, Demircigil et al., 2017, Krotov and Hopfield, 2016], we assume that all memory patterns {ξµ} are sampled from a d-sphere of radius m. Lemma 3.1 (Memory Capacity Lower Bound). Let 1 p be the probability of successfully storing and retrieving a pattern. The number of patterns randomly sampled from a sphere of radius m that the sparse Hopfield model can store and retrieve is lower-bounded by M p C d 1 4 , (3.5) where C is the solution to C = b/W0(exp{a+ln b}) with W0( ) being the principal branch of Lambert W function, a := (4/d 1) (ln [2m( p 1)/(R+δ)] + 1) and b := 4m2β/5(d 1). For sufficiently large β, the sparse Hopfield model exhibits a larger lower bound on the exponential memory capacity compared to its dense counterpart [Ramsauer et al., 2021]: M MDense. Proof Sketch. Our proof is built on [Ramsauer et al., 2021]. The high-level idea is to utilize the separation of random patterns sampled from spheres [Cai and Jiang, 2012, Brauchart et al., 2018] and the asymptotic expansion of the Lambert W function [Corless et al., 1996]. Firstly, we link the well-separation condition to cosine similarity distance, creating an inequality for the probability of a pattern being well-stored and retrieved. Next, we identify and prove conditions for the exponential memory capacity M = p C (d 1)/4 to hold. Finally, we analyze the scaling behaviors of C using its asymptotic expansion and show that M MDense. A detailed proof is shown in Appendix E.6. Intuitively, the benefits of sparsity arises from the increased energy landscape separation provided by the sparse Hopfield energy function, which enables the separation of closely correlated patterns, resulting in a tighter well-separation condition for distinguishing such patterns and hence a larger lower bound on the memory capacity. Moreover, the sparse Hopfield model also enjoys the properties of fast convergence and exponentially suppressed retrieval error provided by the following corollary. Corollary 3.1.2 (Fast Convergence and Exponentially Suppressed Retrieval Error). For any query x, T approximately retrieves a memory pattern ξµ with retrieval error ϵ exponentially suppressed by µ: T (x) ξµ 2m(M 1) exp β µ 2m Max x ξµ , x x µ . Proof. This results from Theorem 2.2, Lemma 2.2, and [Ramsauer et al., 2021, Theorem 4]. Corollary 3.1.2 suggests that, with a sufficient µ, T can approximately retrieve patterns after a single activation, allowing the integration of sparse Hopfield models into deep learning architectures similarly to [Hoover et al., 2023, Seidl et al., 2022, Fürst et al., 2022, Ramsauer et al., 2021]. 4 Proof of Concept Experimental Studies We demonstrate the validity of our theoretical results and method by testing them on various experimental settings with both synthetic and real-world datasets. 4.1 Experimental Validation of Theoretical Results We conduct experiments to verify our theoretical findings, and report the results in Figure 1. For the memory capacity (the top row of Figure 1), we test the proposed sparse model on retrieving half-masked patterns comparing with the Dense (Softmax) and 10th order polynomial Hopfield models [Millidge et al., 2022, Krotov and Hopfield, 2016] on MNIST (high sparsity), Cifar10 (low sparsity) and Image Net (low sparsity) datasets. For all Hopfield models, we set β = 1.6 A query is regarded as correctly retrieved if its cosine similarity error is below a set threshold. In addition, for the robustness against noisy queries (the bottom row of Figure 1), we inject Gaussian noises with varying variances (σ) into the images. Plotted are the means and standard deviations of 10 runs. The results show that the proposed sparse Hopfield model excels when memory patterns exhibit a high degree of sparsity and the signal-to-noise ratio in patterns is low, aligning with our theoretical results. 4.2 Multiple Instance Learning Tasks Ramsauer et al. [2021] point out that the memory-enhanced Hopfield layers present a promising approach for Multiple Instance Learning (MIL) tasks. Multiple Instance Learning (MIL) [Ilse et al., 2018, Carbonneau et al., 2018] is a variation of supervised learning where the training set consists of labeled bags, each containing multiple instances. The goal of MIL is to predict the bag labels based on the instances they contain, which makes it particularly useful in scenarios where labeling individual instances is difficult or impractical, but bag-level labels are available. Examples of such scenarios include medical imaging (where a bag could be an image, instances could be patches of the 6However, as pointed out in [Millidge et al., 2022], this is in fact not fair to compare modern Hopfield with β = 1 with higher order polynomial Hopfield models. Figure 1: Top: Memory Capacity measured by successful half-masked retrieval rates. Bottom: Memory Robustness measured by retrieving patterns with varying levels of Gaussian noise. For all Hopfield models, we set β = .01/0.1/0.1 (for MNIST/CIFAR10/Image Net) for better visualization. A query pattern is deemed correctly retrieved if its squared Euclidean distance is below a set threshold. For MNIST/CIFAR10/Image Net datasets, we set the error thresholds to be 10/20/20 to cope with different sparse levels in data. Plotted are the means and standard deviations of 10 runs. The results suggest that the sparse Hopfield model excels when memory patterns exhibit a high degree of sparsity and the signal-to-noise ratio in patterns is low. image, and the label could indicate the presence or absence of disease) and document classification (where a bag could be a document, instances could be the words or sentences in the document, and the label could indicate the topic or sentiment of the document). In this subsection, we implement our sparse Hopfield layers and applied them to MIL tasks on one synthetic and four real-world settings. 4.2.1 Synthetic Experiments We use a synthetic MIL dataset, the bit pattern dataset, to demonstrate the effectiveness of the sparse Hopfield model. Each bag in this synthetic dataset contains a set of binary bit strings. The positive bag includes at least one of the positive bit patterns. We compare the performance of the Sparse Hopfield and Sparse Hopfield Pooling to their dense counterparts and vanilla attention [Vaswani et al., 2017]. We report the mean test accuracy of 10 runs. To demonstrate the effectiveness of sparse Hopfield model, we vary two hyperparameters of the bit pattern dataset corresponding to two perspectives: bag sparsity (sparsity in data) and bag size (number of memory patterns, M). For bag sparsity, we fix the bag size as 200, and inject from 2 to 80 positive patterns in a positive bag, results in 1 to 40 percent of positive patterns in each positive bag. For bag size, we fix the number of positive pattern in a bag to be 1, and vary bag size from 20 to 300. We report results of Sparse Hopfield Pooling in Table 1, and implementation details in Appendix H.1.1. A more complete version of Table 1, including the results of Hopfield and attention, is in Appendix G. The sparse Hopfield model demonstrates a better performance across all sparsity and all bag sizes. Table 1: Top (Bag Size): Accuracy comparison on bit pattern dataset for sparse and dense Hopfield model. We report the average accuracy over 10 runs. The results suggest that the sparse Hopfield model demonstrates a better performance when facing a bag size increase. Bottom (Bag Sparsity): Performance comparison on bit pattern dataset for sparse and dense Hopfield model with varying bag sparsity. We report the average accuracy over 10 runs. The results suggest that the sparse Hopfield model demonstrates a better performance across all sparsity. Bag Size 20 50 100 150 200 300 Dense Hopfield Pooling 100.0 0.00 100.0 0.00 100.0 0.00 76.44 0.23 49.13 0.01 52.88 0.01 Sparse Hopfield Pooling 100.0 0.00 100.0 0.00 100.0 0.00 99.76 0.00 99.76 0.00 99.76 0.00 Bag Sparsity 1% 5% 10% 20% 40% Dense Hopfield Pooling 49.20 0.00 85.58 0.10 100.0 0.00 100.0 0.00 99.68 0.00 Sparse Hopfield Pooling 73.40 0.06 99.68 0.00 100.0 0.00 100.0 0.00 100.0 0.00 Convergence Analysis. In Figure 2, we numerically examine the convergence of the sparse and dense Hopfield models, plotting their loss and accuracy for the bag size tasks in above on the bit pattern dataset. We include multiple bag sizes to assess the effect of increasing memory patterns (i.e. M) on the loss curve. The plotted are the loss and accuracy curves of Sparse Hopfield Pooling. We refer results of Hopfield and more details to Appendix G.3. The results (Figure 2) show that, sparse Hopfield model surpasses its dense counterpart in all bag sizes. Moreover, for the same bag size, the sparse Hopfield model always reaches the minimum validation loss faster than dense Hopfield model, validating our Theorem 2.2. Sparsity Generalization. We also evaluate the models generalization performance with shifting information sparsity, by training dense and sparse Hopfield models with a specific bag sparsity and testing them on the other. We report the results in Table 5 and refer more details to Appendix G.3. 0 20 40 60 80 100 120 140 Epoch Training Loss dense M = 100 sparse M = 100 dense M = 150 sparse M = 150 dense M = 200 sparse M = 200 dense M = 300 sparse M = 300 0 20 40 60 80 100 120 140 Epoch Training Accuracy dense M = 100 sparse M = 100 dense M = 150 sparse M = 150 dense M = 200 sparse M = 200 dense M = 300 sparse M = 300 0 20 40 60 80 100 120 140 Epoch Validation Loss dense M = 100 sparse M = 100 dense M = 150 sparse M = 150 dense M = 200 sparse M = 200 dense M = 300 sparse M = 300 0 20 40 60 80 100 120 140 Epoch Validation Accuracy dense M = 100 sparse M = 100 dense M = 150 sparse M = 150 dense M = 200 sparse M = 200 dense M = 300 sparse M = 300 Figure 2: Top: The training loss and accuracy curve of dense and sparse Hopfield models with different bag sizes. Bottom: The validation loss and accuracy curve of dense and sparse Hopfield models with different bag sizes. The plotted are the mean of 10 runs. The results indicate that the sparse Hopfield model converges faster than the dense model and also yields superior accuracy. 4.2.2 Real-World MIL Tasks Next, we demonstrate that the proposed method achieves near-optimal performance on four realistic (non-sparse) MIL benchmark datasets: Elephant, Fox and Tiger for image annotation [Ilse et al., 2018], UCSB breast cancer classification [Kandemir et al., 2014]. We use Hopfield and Sparse Hopfield to construct a similar model architecture proposed in [Ramsauer et al., 2021] and a detailed description of this experiment as well as its training and evaluating process can be found in Appendix H.1.2. As shown in Table 2, both Sparse and Dense Hopfield achieve nearbest results on Tiger, Elephant and UCSB datasets, despite the low sparsity in data. The sparse Hopfield model outperforms the dense Hopfield model by a small margin on three out of four datasets. Table 2: Results for MIL benchmark datasets in terms of AUC score. The baselines are Path encoding [Küçüka scı and Baydo gan, 2018], MIn D [Cheplygina et al., 2015], MILES [Chen et al., 2006], APR [Dietterich et al., 1997], Citation KNN [Wang and Zucker, 2000] and DD [Maron and Lozano Pérez, 1997]. Results for baselines are taken from [Ramsauer et al., 2021]. The results suggest the proposed model achieves near-optimal performance even when the data is not sparse. Method Tiger Fox Elephant UCSB Dense Hopfield 0.878 0.028 0.600 0.011 0.907 0.022 0.880 0.013 Sparse Hopfield 0.892 0.021 0.611 0.010 0.912 0.016 0.877 0.009 Path encoding 0.910 0.010 0.712 0.014 0.944 0.007 0.880 0.022 MIn D 0.853 0.011 0.704 0.016 0.936 0.009 0.831 0.027 MILES 0.872 0.017 0.738 0.016 0.927 0.007 0.833 0.026 APR 0.778 0.007 0.541 0.009 0.550 0.010 Citation-k NN 0.855 0.009 0.635 0.015 0.896 0.009 0.706 0.032 DD 0.841 0.631 0.907 5 Conclusion We present a sparse Hopfield model with a memory-retrieval dynamics that corresponds to the sparsestructured attention mechanism. This model is capable of merging into deep learning architectures with datadependent sparsity. Theoretically, we introduce a principled construction for modern Hopfield models, based on the convex conjugate of different entropy regularizers. It allows us to easily recover the dense modern Hopfield model [Ramsauer et al., 2021] using Gibbs entropy. Moreover, we introduce the sparse Hopfield model using the Gini entropic regularizer, and explore its theoretical advantages, delineating conditions that favor its use. Empirically, we demonstrate our theoretical results and methodology to be effective on various synthetic and realistic settings. This work extends the correspondence between artificial and biological neural networks to sparse domain, potentially paving the way for future Hopfield-based methodologies and bio-inspired computing systems. Post-Acceptance Note [November 28, 2023]. After the completion of this work, the authors learn of two upcoming works [Anonymous, 2023] at ICLR 24 and [Martins et al., 2023] in the Associative Memory & Hopfield Networks Workshop at Neur IPS 23 both addressing similar topics. Both of these works explore theoretical generalizations of this work. In addition, [Anonymous, 2023] further presents a (sparse) Hopfield-based deep learning model for multivariate time series prediction. We thank the authors of [Martins et al., 2023] for enlightening discussions and for sharing their preliminary manuscript. Acknowledgments JH would like to thank Tim Tsz-Kit Lau, Robin Luo and Andrew Chen for enlightening discussions on related topics, and Jiayi Wang for invaluable support in facilitating experimental deployments. The authors would like to thank the anonymous reviewers and program chairs for constructive comments. JH is partially supported by the Walter P. Murphy Fellowship. HL is partially supported by NIH R01LM1372201, NSF CAREER1841569, DOE DE-AC02-07CH11359, DOE LAB 20-2261 and a NSF TRIPODS1740735. This research was supported in part through the computational resources and staff contributions provided for the Quest high performance computing facility at Northwestern University which is jointly supported by the Office of the Provost, the Office for Research, and Northwestern University Information Technology. The content is solely the responsibility of the authors and does not necessarily represent the official views of the funding agencies. Anonymous. STanhop: Sparse tandem hopfield model for memory-enhanced time series prediction. In Submitted to The Twelfth International Conference on Learning Representations, 2023. URL https://openreview.net/forum?id=6iwg437CZs. under review. Shaojie Bai, J Zico Kolter, and Vladlen Koltun. Deep equilibrium models. Advances in Neural Information Processing Systems, 32, 2019. URL https://arxiv.org/abs/1909.01377. Iz Beltagy, Matthew E Peters, and Arman Cohan. Longformer: The long-document transformer. ar Xiv preprint ar Xiv:2004.05150, 2020. URL https://arxiv.org/abs/2004.05150. Ondˇrej Bojar, Rajen Chatterjee, Christian Federmann, Yvette Graham, Barry Haddow, Shujian Huang, Matthias Huck, Philipp Koehn, Qun Liu, Varvara Logacheva, et al. Findings of the 2017 conference on machine translation (wmt17). Association for Computational Linguistics, 2017. Johannes Brandstetter. Blog post: Hopfield networks is all you need, 2021. URL https:// ml-jku.github.io/hopfield-layers/. Accessed: April 4, 2023. Johann S Brauchart, Alexander B Reznikov, Edward B Saff, Ian H Sloan, Yu Guang Wang, and Robert S Womersley. Random point sets on the sphere hole radii, covering, and separation. Experimental Mathematics, 27(1):62 81, 2018. URL https://arxiv.org/abs/1512. 07470. Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et al. Language models are few-shot learners. Advances in neural information processing systems, 33:1877 1901, 2020. URL https://arxiv.org/abs/1512.07470. T Tony Cai and Tiefeng Jiang. Phase transition in limiting distributions of coherence of highdimensional random matrices. Journal of Multivariate Analysis, 107:24 39, 2012. URL https: //arxiv.org/abs/1102.2926. Marc-André Carbonneau, Veronika Cheplygina, Eric Granger, and Ghyslain Gagnon. Multiple instance learning: A survey of problem characteristics and applications. Pattern Recognition, 77: 329 353, 2018. URL https://arxiv.org/abs/1612.03365. Minghao Chen, Houwen Peng, Jianlong Fu, and Haibin Ling. Autoformer: Searching transformers for visual recognition. In Proceedings of the IEEE/CVF international conference on computer vision, pages 12270 12280, 2021. URL https://arxiv.org/abs/2107.00651. Yixin Chen, Jinbo Bi, and James Ze Wang. Miles: Multiple-instance learning via embedded instance selection. IEEE transactions on pattern analysis and machine intelligence, 28(12):1931 1947, 2006. Veronika Cheplygina, David MJ Tax, and Marco Loog. Dissimilarity-based ensembles for multiple instance learning. IEEE transactions on neural networks and learning systems, 27(6):1379 1391, 2015. URL https://arxiv.org/abs/1402.1349. Rewon Child, Scott Gray, Alec Radford, and Ilya Sutskever. Generating long sequences with sparse transformers. ar Xiv preprint ar Xiv:1904.10509, 2019. URL https://arxiv.org/abs/ 1904.10509. Aakanksha Chowdhery, Sharan Narang, Jacob Devlin, Maarten Bosma, Gaurav Mishra, Adam Roberts, Paul Barham, Hyung Won Chung, Charles Sutton, Sebastian Gehrmann, et al. Palm: Scaling language modeling with pathways. ar Xiv preprint ar Xiv:2204.02311, 2022. URL https: //arxiv.org/abs/2204.02311. Robert M Corless, Gaston H Gonnet, David EG Hare, David J Jeffrey, and Donald E Knuth. On the lambert w function. Advances in Computational mathematics, 5:329 359, 1996. Gonçalo M Correia, Vlad Niculae, and André FT Martins. Adaptively sparse transformers. ar Xiv preprint ar Xiv:1909.00015, 2019. URL https://arxiv.org/abs/1909.00015. Herbert A David and Haikady N Nagaraja. Order statistics. John Wiley & Sons, 2004. Mete Demircigil, Judith Heusel, Matthias Löwe, Sven Upgang, and Franck Vermet. On a model of associative memory with huge storage capacity. Journal of Statistical Physics, 168:288 299, 2017. URL https://arxiv.org/abs/1702.01929. Thomas G Dietterich, Richard H Lathrop, and Tomás Lozano-Pérez. Solving the multiple instance problem with axis-parallel rectangles. Artificial intelligence, 89(1-2):31 71, 1997. Michael Elad. Sparse and redundant representations: from theory to applications in signal and image processing, volume 2. Springer, 2010. Peter Földiak. Forming sparse representations by local anti-hebbian learning. Biological cybernetics, 64(2):165 170, 1990. Andreas Fürst, Elisabeth Rumetshofer, Johannes Lehner, Viet T Tran, Fei Tang, Hubert Ramsauer, David Kreil, Michael Kopp, Günter Klambauer, Angela Bitto, et al. Cloob: Modern hopfield networks with infoloob outperform clip. Advances in neural information processing systems, 35: 20450 20468, 2022. URL https://arxiv.org/abs/2110.11316. Bolin Gao and Lacra Pavel. On the properties of the softmax function with application in game theory and reinforcement learning. ar Xiv preprint ar Xiv:1704.00805, 2017. URL https: //arxiv.org/abs/1704.00805. Asela Gunawardana, William Byrne, and Michael I Jordan. Convergence theorems for generalized alternating minimization procedures. Journal of machine learning research, 6 (12), 2005. URL https://www.jmlr.org/papers/volume6/gunawardana05a/ gunawardana05a.pdf. Benjamin Hoover, Yuchen Liang, Bao Pham, Rameswar Panda, Hendrik Strobelt, Duen Horng Chau, Mohammed J Zaki, and Dmitry Krotov. Energy transformer. ar Xiv preprint ar Xiv:2302.07253, 2023. URL https://arxiv.org/abs/2302.07253. John J Hopfield. Neural networks and physical systems with emergent collective computational abilities. Proceedings of the national academy of sciences, 79(8):2554 2558, 1982. John J Hopfield. Neurons with graded response have collective computational properties like those of two-state neurons. Proceedings of the national academy of sciences, 81(10):3088 3092, 1984. Maximilian Ilse, Jakub Tomczak, and Max Welling. Attention-based deep multiple instance learning. In International conference on machine learning, pages 2127 2136. PMLR, 2018. URL https: //arxiv.org/abs/1802.04712. Yanrong Ji, Zhihan Zhou, Han Liu, and Ramana V Davuluri. Dnabert: pre-trained bidirectional encoder representations from transformers model for dna-language in genome. Bioinformatics, 37 (15):2112 2120, 2021. URL https://www.biorxiv.org/content/10.1101/2020. 09.17.301879v1. Melih Kandemir, Chong Zhang, and Fred A Hamprecht. Empowering multiple instance histopathology cancer diagnosis by cell graphs. In Medical Image Computing and Computer-Assisted Intervention MICCAI 2014: 17th International Conference, Boston, MA, USA, September 14-18, 2014, Proceedings, Part II 17, pages 228 235. Springer, 2014. Leo Kozachkov, Ksenia V Kastanenka, and Dmitry Krotov. Building transformers from neurons and astrocytes. Proceedings of the National Academy of Sciences, 120(34):e2219150120, 2023. URL https://www.biorxiv.org/content/10.1101/2022.10.12.511910v1. Dmitry Krotov and John J Hopfield. Dense associative memory for pattern recognition. Advances in neural information processing systems, 29, 2016. URL https://arxiv.org/abs/1606. 01164. Dmitry Krotov and John J. Hopfield. Large associative memory problem in neurobiology and machine learning. In International Conference on Learning Representations, 2021. URL https: //arxiv.org/abs/2008.06996. Emel Seyma Küçüka scı and Mustafa Gökçe Baydo gan. Bag encoding strategies in multiple instance learning problems. Information Sciences, 467:559 578, 2018. YC Lee, Gary Doolen, HH Chen, GZ Sun, Tom Maxwell, and HY Lee. Machine learning using a higher order correlation network. Technical report, Los Alamos National Lab.(LANL), Los Alamos, NM (United States); Univ. of Maryland, College Park, MD (United States), 1986. Ze Liu, Yutong Lin, Yue Cao, Han Hu, Yixuan Wei, Zheng Zhang, Stephen Lin, and Baining Guo. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF international conference on computer vision, pages 10012 10022, 2021. URL https://arxiv.org/abs/2103.14030. Ilya Loshchilov and Frank Hutter. Decoupled weight decay regularization. ar Xiv preprint ar Xiv:1711.05101, 2017. URL https://arxiv.org/abs/1711.05101. Julien Mairal, Francis Bach, Jean Ponce, and Guillermo Sapiro. Online learning for matrix factorization and sparse coding. Journal of Machine Learning Research, 11(1), 2010. URL https://arxiv.org/abs/0908.0050. Alireza Makhzani and Brendan J Frey. Winner-take-all autoencoders. Advances in neural information processing systems, 28, 2015. URL https://arxiv.org/abs/1409.2752. Oded Maron and Tomás Lozano-Pérez. A framework for multiple-instance learning. Advances in neural information processing systems, 10, 1997. Andre Martins and Ramon Astudillo. From softmax to sparsemax: A sparse model of attention and multi-label classification. In International conference on machine learning, pages 1614 1623. PMLR, 2016. URL https://arxiv.org/abs/1602.02068. Andre F. T. Martins, Vlad Niculae, and Daniel Mc Namee. Sparse modern hopfield networks. Associative Memory & Hopfield Networks in 2023. Neur IPS 2023 workshop., 2023. URL https: //openreview.net/pdf?id=zwql V7Hoa T. Beren Millidge, Tommaso Salvatori, Yuhang Song, Thomas Lukasiewicz, and Rafal Bogacz. Universal hopfield networks: A general framework for single-shot associative memory models. In International Conference on Machine Learning, pages 15561 15583. PMLR, 2022. URL https://arxiv.org/abs/2202.04557. Charles M Newman. Memory capacity in neural network models: Rigorous lower bounds. Neural Networks, 1(3):223 238, 1988. Bruno A Olshausen and David J Field. Sparse coding with an overcomplete basis set: A strategy employed by v1? Vision research, 37(23):3311 3325, 1997. Frank WJ Olver, Daniel W Lozier, Ronald F Boisvert, and Charles W Clark. NIST handbook of mathematical functions hardback and CD-ROM. Cambridge university press, 2010. Fabian Paischer, Thomas Adler, Vihang Patil, Angela Bitto-Nemling, Markus Holzleitner, Sebastian Lehner, Hamid Eghbal-Zadeh, and Sepp Hochreiter. History compression via language models in reinforcement learning. In International Conference on Machine Learning, pages 17156 17185. PMLR, 2022. URL https://arxiv.org/abs/2205.12258. Günther Palm. Neural associative memories and sparse coding. Neural Networks, 37:165 171, 2013. Pierre Peretto and Jean-Jacques Niez. Long term memory storage capacity of multiconnected neural networks. Biological Cybernetics, 54(1):53 63, 1986. Ben Peters, Vlad Niculae, and André FT Martins. Sparse sequence-to-sequence models. ar Xiv preprint ar Xiv:1905.05702, 2019. URL https://arxiv.org/abs/1905.05702. Jiezhong Qiu, Hao Ma, Omer Levy, Scott Wen-tau Yih, Sinong Wang, and Jie Tang. Blockwise self-attention for long document understanding. ar Xiv preprint ar Xiv:1911.02972, 2019. URL https://arxiv.org/abs/1911.02972. Hubert Ramsauer, Bernhard Schäfl, Johannes Lehner, Philipp Seidl, Michael Widrich, Lukas Gruber, Markus Holzleitner, Thomas Adler, David Kreil, Michael K Kopp, Günter Klambauer, Johannes Brandstetter, and Sepp Hochreiter. Hopfield networks is all you need. In International Conference on Learning Representations, 2021. URL https://arxiv.org/abs/2008.02217. Ron Rubinstein, Alfred M Bruckstein, and Michael Elad. Dictionaries for sparse representation modeling. Proceedings of the IEEE, 98(6):1045 1057, 2010. Philipp Seidl, Philipp Renz, Natalia Dyubankova, Paulo Neves, Jonas Verhoeven, Jorg K Wegner, Marwin Segler, Sepp Hochreiter, and Gunter Klambauer. Improving few-and zero-shot reaction template prediction using modern hopfield networks. Journal of chemical information and modeling, 62(9):2111 2120, 2022. Bharath K Sriperumbudur and Gert RG Lanckriet. On the convergence of the concave-convex procedure. In Advances in neural information processing systems, volume 9, pages 1759 1767, 2009. URL https://papers.nips.cc/paper_files/paper/2009/file/ 8b5040a8a5baf3e0e67386c2e3a9b903-Paper.pdf. Yi Tay, Dara Bahri, Liu Yang, Donald Metzler, and Da-Cheng Juan. Sparse sinkhorn attention. In International Conference on Machine Learning, pages 9438 9447. PMLR, 2020. Yi Tay, Mostafa Dehghani, Dara Bahri, and Donald Metzler. Efficient transformers: A survey. ACM Computing Surveys, 55(6):1 28, 2022. URL https://arxiv.org/abs/2009.06732. Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. Advances in neural information processing systems, 30, 2017. URL https://arxiv.org/abs/1706.03762. Martin J Wainwright, Michael I Jordan, et al. Graphical models, exponential families, and variational inference. Foundations and Trends in Machine Learning, 1(1 2):1 305, 2008. Jun Wang and Jean-Daniel Zucker. Solving multiple-instance problem: A lazy learning approach. 2000. Michael Widrich, Bernhard Schäfl, Milena Pavlovi c, Hubert Ramsauer, Lukas Gruber, Markus Holzleitner, Johannes Brandstetter, Geir Kjetil Sandve, Victor Greiff, Sepp Hochreiter, et al. Modern hopfield networks and attention for immune repertoire classification. Advances in Neural Information Processing Systems, 33:18832 18845, 2020. URL https://arxiv.org/abs/ 2007.13505. Yongyi Yang, Zengfeng Huang, and David Wipf. Transformers from an optimization perspective. In Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho, editors, Advances in Neural Information Processing Systems, 2022. URL https://openreview.net/forum? id=VT0Y4Pl V2m0. Alan L Yuille and Anand Rangarajan. The concave-convex procedure (cccp). Advances in neural information processing systems, 14, 2001. URL https: //proceedings.neurips.cc/paper_files/paper/2001/file/ a012869311d64a44b5a0d567cd20de04-Paper.pdf. Alan L Yuille and Anand Rangarajan. The concave-convex procedure. Neural computation, 15(4): 915 936, 2003. Willard I Zangwill. Nonlinear programming: a unified approach, volume 52. Prentice-Hall Englewood Cliffs, NJ, 1969. Yunhao Zhang and Junchi Yan. Crossformer: Transformer utilizing cross-dimension dependency for multivariate time series forecasting. In The Eleventh International Conference on Learning Representations, 2022. URL https://openreview.net/forum?id=v SVLM2j9eie. Haoyi Zhou, Shanghang Zhang, Jieqi Peng, Shuai Zhang, Jianxin Li, Hui Xiong, and Wancai Zhang. Informer: Beyond efficient transformer for long sequence time-series forecasting. In Proceedings of the AAAI conference on artificial intelligence, volume 35, pages 11106 11115, 2021. URL https://arxiv.org/abs/2012.07436. Tian Zhou, Ziqing Ma, Qingsong Wen, Liang Sun, Tao Yao, Wotao Yin, Rong Jin, et al. Film: Frequency improved legendre memory model for long-term time series forecasting. Advances in Neural Information Processing Systems, 35:12677 12690, 2022. URL https://arxiv.org/ abs/2205.08897.