# hygene_a_diffusionbased_hypergraph_generation_method__f5d8681c.pdf HYGENE: A Diffusion-Based Hypergraph Generation Method Dorian Gailhard, Enzo Tartaglione, Lirida Naviner, Jhony H. Giraldo LTCI, T el ecom Paris, Institut Polytechnique de Paris, France {name.surname}@telecom-paris.fr Hypergraphs are powerful mathematical structures that can model complex, high-order relationships in various domains, including social networks, bioinformatics, and recommender systems. However, generating realistic and diverse hypergraphs remains challenging due to their inherent complexity and lack of effective generative models. In this paper, we introduce a diffusion-based Hypergraph Generation (HYGENE) method that addresses these challenges through a progressive local expansion approach. HYGENE works on the bipartite representation of hypergraphs, starting with a single pair of connected nodes and iteratively expanding it to form the target hypergraph. At each step, nodes and hyperedges are added in a localized manner using a denoising diffusion process, which allows for the construction of the global structure before refining local details. Our experiments demonstrated the effectiveness of HYGENE, proving its ability to closely mimic a variety of properties in hypergraphs. To the best of our knowledge, this is the first attempt to employ diffusion models for hypergraph generation. Code https://github.com/Dorian Gailhard/HYGENE Extended version https://arxiv.org/abs/2408.16457 1 Introduction Hypergraphs are higher-order extensions of graphs. They comprise a set of nodes, also called vertices, and a set of hyperedges. Unlike regular graphs, where edges connect only two nodes, hyperedges can connect any number of nodes. These structures have demonstrated their ability to capture more complex relationships than graphs and have been applied in various domains (Gong, Higham, and Zygalakis 2023; Estrada and Rodr ıguez-Vel azquez 2006). For instance, hypergraphs have been applied in drug discovery (Kajino 2019), modeling contagion spread (Higham and De Kergorlay 2021), and electronics (Starostin and Balashov 2008; Luo et al. 2024; Grzesiak-Kope c, Oramus, and Ogorzałek 2017). Besides, they have also proven useful in recommender systems (Lin et al. 2023), molecular biology (Murgas, Saucan, and Sandhu 2022; Rahman et al. 2012), and urban planning (Dupagne and Teller 1998). The Copyright 2025, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. Training Examples Generated Hypergraphs Variational Autoencoder HYGENE DCGAN Figure 1: Examples of generated ego hypergraphs by a variational autoencoder, a Deep Convolutional Generative Adversarial Network (DCGAN), and our model (HYGENE). versatility of hypergraphs in representing multi-way relationships makes them a powerful tool across these diverse fields; consequently, hypergraph generation (the ability to sample from specific hypergraph distributions) holds significant promise. Despite its wide applicability, research in hypergraph generation has primarily focused on algorithmic approaches, aiming to develop methodologies that produce hypergraphs with specific, predefined structural properties (Do et al. 2020; Arafat et al. 2020). In contrast, the exploration of hypergraph generation using deep learning models remains largely understudied. This gap represents a significant opportunity to advance this field since deep learning approaches may capture complex patterns and generate more realistic and diverse hypergraphs. Such methods could enhance the modeling of intricate relationships beyond the scope of traditional graph structures. In contrast to hypergraph generation, deep learning-based graph generation has been extensively studied (Zhu et al. 2022). Learning-based graph generation can be broadly categorized into two approaches: one-shot approaches that generate the entire graph simultaneously (Simonovsky and Komodakis 2018; You et al. 2018), and iterative models that The Thirty-Ninth AAAI Conference on Artificial Intelligence (AAAI-25) generate the graphs incrementally, predicting edges for each new node (Vignac et al. 2023; Chen et al. 2023). While graph generation techniques have shown promise, their adaptation to hypergraphs remains challenging. The variable size of hyperedges and their higher-order relationships increase the difficulty of the task, making direct application of graph methods non-trivial. Figure 1 shows that na ıvely generating the incidence matrix using classical image generation architectures is not sufficient either, as it lacks the correct understanding of the underlying data structure. By leveraging the spectral equivalence between a hypergraph and two carefully chosen representations (the clique and star expansions), we generalize the work by Bergmeister et al. (2024) for hypergraph generation. Their method is based on an iterative local expansion scheme, where graph generation is performed hierarchically, first building the global structure before refining the details. This process is seen as the inverse of a coarsening operation, which involves the reduction of a graph while preserving its relevant properties. For hypergraphs, the variable size of hyperedges increases the complexity of the problem, as they are exponentially more numerous than classical edges because every non-empty set of nodes is a possible hyperedge. In order to mitigate this, we introduce an iterative expansion and refinement process for Hyperedge Generation (HYGENE), rather than predicting all possible hyperedges at once. We train a denoising diffusion model (Karras et al. 2022) using this framework, and validate our method on four synthetic and three real-world datasets, demonstrating its effectiveness in replicating important structural properties. At a glance, our main contributions are the following: To the best of our knowledge, we introduce the first diffusion-based method for generating hypergraphs sampled from specific distributions (Sec. 3.2). We generalize important concepts in the graph domain to hypergraph generation, like hypergraph coarsening and diffusion (Sec. 3.3, Sec. 3.4, Sec. 3.5). We provide rigorous theoretical justifications for our technical choices. We validate HYGENE on four synthetic and three realworld datasets, showcasing its ability to capture and reproduce subtle structural properties of hypergraphs (Sec. 4). 2 Related Work Graph Generation Using Deep Learning. The field of graph generation using deep learning models was pioneered by Graph VAE (Simonovsky and Komodakis 2018). This approach employs an autoencoder to embed graphs into a latent space, from which new graphs could be generated by sampling and decoding. Subsequently, You et al. (2018) significantly enhanced generation quality by utilizing a recurrent neural network to produce the adjacency matrix column-wise. Recent advancements include the work by Kong et al. (2023), which formalized the generation process as an inverse discrete absorption. In this method, nodes from a training graph are sequentially absorbed, and a mixture of multinomials is trained to predict the necessary edge additions when reintroducing a node. Diffusion models, which have shown remarkable success in image generation, were adapted for graph generation in (Niu et al. 2020). This approach was further refined by Vignac et al. (2023) and expanded in (Chen et al. 2023) by incorporating target degree distributions. A departure from typical methods in graph generation was proposed by Bergmeister et al. (2024). Instead of sequentially adding nodes and predicting their connections, this approach reverses a coarsening process. During training, graphs are reduced by merging nodes into clusters. The model then learns to identify these clusters, decompose them, and reconstruct the original connections between their nodes. Graph generation begins with a single-node graph and progressively expands it using the trained model, mirroring the diffusion approaches seen in modern image generation techniques. Similar hierarchical concepts have been explored in molecule generation, with Zhu et al. (2023) applying normalizing flows to this domain. Related ideas can also be found in the work of Guo, Zou, and Lerman (2023). In contrast to previous work, we focus on the problem of hypergraph generation, which extends the concept of graph generation to higher-order structures. Furthermore, we employ a hierarchical view of the problem instead of the sequential edge-by-edge generation commonly employed. We also depart from the classical view of edge prediction, where a model outputs the probability of existence for all possible edges, and, instead, predict both the number of hyperedges and their composition. 3 Method 3.1 Preliminaries Notation. In this work, calligraphic letters such as V denote sets, and |V| represents the cardinality of the set. Uppercase boldface letters such as A represent matrices, while lowercase boldface letters such as x denote vectors. The superscripts ( )T correspond to transposition. diag(x) denotes a diagonal matrix with entries given by the vector x = [x1, x2, . . . , xn] RN. Finally, Sp(A) denotes the set of eigenvalues of a matrix A. Basic Definitions. A graph G is defined as a pair (V, E), where V is a set of vertices and E V V is a set of edges. Each edge e E is a pair of vertices (u, v), representing a connection between nodes u and v. A bipartite graph B is a special case of a graph, defined as (VL, VR, E), where VL and VR are disjoint sets of vertices, and E VL VR. Every edge in a bipartite graph connects a node in VL to a node in VR. Note that a bipartite graph can be identified as a graph where V = VL VR. We define the Laplacian of a graph LG R|V| |V| as LG = D W, where D is the diagonal degree matrix and W is the weighted adjacency matrix with W[i,j] = 0 if (i, j) E. The normalized Laplacian is defined as LG = I D 1/2WD 1/2, where I is the identity. A hypergraph H is defined as a pair (V, E), where V is a set of vertices and E is a set of hyperedges, with each e E being a subset of V. Unlike in graphs, hyperedges can connect any number of vertices. We define two graph Expansion and Refinement Corresponding Bipartite View Figure 2: Starting from a hypergraph, our method computes different views of it at increasingly coarser resolutions. An equivalent bipartite representation is maintained in parallel, and a graph neural network model (in our case a PPGN (Maron et al. 2019)) is trained to recover a bipartite representation from its coarser version. representations of hypergraphs: the clique and star expansions. The clique expansion of a hypergraph H is a graph C = (Vc, Ec), where Ec = {(u, v) | e E : u, v e}. The star expansion of a hypergraph H is a bipartite graph B = (VL, VR, Eb), where VL = V, VR = E, and Eb = {(v, e) | v VL, e VR, v e in H}. Furthermore, we define the Bolla s Laplacian (Bolla 1993) of a hypergraph as LH R|V| |V| as LH = DV HD 1 E HT , where DV N|V| |V| is the diagonal degree matrix for the nodes, DE N|E| |E| is the diagonal matrix of edge orders, and H {0, 1}|V| |E| is the incidence matrix. The normalized version of LH, known as Zhou s Laplacian (Zhou, Huang, and Sch olkopf 2005), is defined as LH = I D 1/2 V HD 1 E HT D 1/2 V . The goal of this work is to train a model capable of sampling from the underlying distribution of a given dataset of hypergraphs (H1, . . . , HN), i.e., learning to generate hypergraphs from data. All the proofs of propositions and lemmas of this paper are provided in Appendix A. Please refer to the extended version of this paper for the appendices. 3.2 Overview The workflow of our approach is illustrated in Figure 2. Our method utilizes two distinct representations of hypergraphs: the weighted clique expansion and the star expansion. The weighted clique expansion (not depicted in the figure) facilitates the downward traversal of the resolution scales. This representation enables the application of the algorithm proposed by Loukas (2019) to generate reduced versions of the hypergraph while maintaining its spectral properties. Concurrently, we maintain the hypergraph s star expansion, representing it as a bipartite graph. In this representation, one partition corresponds to the hypergraph s nodes (left side), while the other represents its hyperedges (right side). Each node is connected to the hyperedges containing it. The bipartite representation proves particularly convenient for training a deep learning model capable of ascending the resolution scales. Starting from a coarser representation, the model learns to identify merged nodes and hyperedges, subsequently reconstructing the original bipartite representation. In our implementation, we employ the denoising diffusion model framework (Karras et al. 2022) to model the learning problem: the truth values for the nodes and edges requiring expansion and deletion are noised, and a model is trained to recover the original values. We chose Provably Powerful Graph Network (PPGN) (Maron et al. 2019) as the architecture of this model. Hypergraph generation can then be achieved through an iterative process of increasing resolution, starting from the coarsest bipartite representation, which consists of a pair of connected nodes. 3.3 Descending through the Resolution Scales: Coarsening Sequences Definition 1 (Graph coarsening). Let G = (V, E) be an arbitrary graph and P = {V(1), . . . , V( n)} be a partitioning1 of the node set V such that each set V(p) P induces a connected subgraph in G. We construct a coarsening G(G, P) = ( V, E) of G by representing each part V(p) P as a single node v(p) V. We add an edge e{p,q} E, between distinct nodes v(p) = v(q) V in the coarsened graph if and only if there exists an edge e{i,j} E between the corresponding disjoint clusters in the original graph, i.e., v(i) V(p) and v(j) V(q). 1This implies that V(p) V, S n i=1 V(i) = V, and V(i) V(j) = 1 i, j n. (a) Weighted clique expansion (b) Coarsening (c) Bipartite representation update Figure 3: Coarsening: (a) Compute the weighted clique expansion by collapsing each hyperedge into an appropriately weighted clique. (b) Coarsen the clique expansion while preserving the spectral properties of the hypergraph (dark blue nodes). (c) Update the bipartite view: corresponding left side nodes (in dark blue) are merged, then right side nodes representing the same hyperedge (circled in black) are merged. Remark 2. This definition implies partitioning the nodes into different connected sets and merging each part into a cluster. Two clusters are connected if and only if there exists an edge between some node in the first cluster and some node in the second cluster. We extend Definition 1 to the bipartite representations of hypergraphs: Definition 3 (Bipartite representation coarsening). Let H be an arbitrary hypergraph, C = (Vc, Ec) its weighted clique expansion, B = (VL, VR, E) its bipartite representation, and PL = {V(1), . . . , V(n)} a partitioning of the node set Vc of C (and equivalently of the node set VL of B), such that each part V(p) PL induces a connected subgraph in C. We construct an intermediate coarsening B(B, PL) = ( VL, VR, E) of B by representing each part V(p) PL as a single node v(p) VL. We add an edge e{p,q} E, between distinct nodes v(p) VL and v(q) VR in the coarsened graph if and only if there exists an edge e{i,q} E between a node v(i) V(p) and right side node v(q) in the original graph. Now let v1 v2 N(v1) = N(v2) define an equivalence relation for the right side nodes in VR, where N(v) denotes the set of neighbors of v. This equivalence relation induces a partitioning PR = {V(1) R , . . . , V(m) R } of VR. Finally, we construct the fully coarsened bipartite representation by representing each part V(p) PR as a single node v(p) VR, in a similar way to VL. Remark 4. Here, nodes are merged into clusters, and then hyperedges appearing multiple times are merged. This process is illustrated in Figure 3. We now describe the construction of coarsening sequences for a hypergraph H = (V, E). This process maintains the weighted clique expansion and the bipartite representation of H in parallel. We leverage the following result: Proposition 5 (Adapted from Section 6.4 in (Agarwal, Branson, and Belongie 2006) and proved in Appendix A.4). For an unweighted hypergraph, Bolla s unnormalized Laplacian LH = DV HD 1 E HT is equal to the unnormalized Laplacian of the associated clique expansion C, where each edge euv is weighted by P e u,v; e E 1 |e|. This proposition establishes that the spectral properties of the hypergraph are equivalent to those of an appropriately weighted clique expansion. Loukas (2019) introduced an algorithm to construct a coarsened version of a graph while preserving a subset of its eigenvalues and eigenvectors. As the weighted clique expansion is a graph, this algorithm can be applied to select groups of nodes for merging. This allows the construction of a coarser view of the hypergraph while preserving relevant spectral properties, which are known to capture important characteristics of the underlying topology. Figure 3a illustrates an example of such a weighted clique expansion. The coarsening process consists of three steps (illustrated in Figures 3b and 3c): 1. The algorithm by Loukas (2019) operates on the weighted clique expansion to identify a suitable partitioning of nodes, also referred to as contraction sets . 2. These contraction sets are merged in the weighted clique expansion, and the corresponding left-side nodes in the bipartite representation of the hypergraph are also merged. 3. Finally, the right-side nodes in the bipartite representation of the hypergraph representing the same hyperedge are merged. This procedure is applied iteratively until we obtain a single-node graph for the weighted clique expansion and a corresponding bipartite graph with a single node on each side for the bipartite representation. It is important to note that in this setting, controlling the merging of hyperedges (right-side nodes for the bipartite representation) is challenging. Empirical experiments have shown that even with few left-node mergings, the right side can easily merge tens of hyperedges into one cluster at once in dense hypergraphs. To avoid this issue, we select the contraction family to be the set of all pairs of adjacent nodes in the clique representation. Therefore, we obtain the following result: Proposition 6. For a single merging of two adjacent nodes in the clique representation, at most three hyperedges can be involved in each hyperedge merging in the bipartite representation. Remark 7. This proposition holds only for a single merging of a node pair at each coarsening step. In the case of multiple simultaneous mergings, the proposition does not necessarily hold. However, it can be enforced by ensuring that the different contraction sets have disjoint neighborhoods. In our experiments, we instead consider each relevant node merging individually and proceed with it only if every rightside cluster does not exceed three hyperedges. The complete Figure 4: Our model starts from a single pair of linked nodes (in the bipartite representation) and iteratively expands the left-side nodes (in dark blue) and right-side nodes (in red), where each duplicate keeps the connections of its parent node. Then, our method refines the resulting bipartite graph filtering edges to recover an appropriate local structure. coarsening sampling procedure incorporating this approach is detailed in Algorithm 1 of Appendix E. 3.4 Ascending through the Resolution Scales: Expansion and Refinement We now describe the expansion and refinement of the bipartite representation of a hypergraph, which is the inverse of the coarsening process (the proof can be found in Appendix A.3). This process operates exclusively on the bipartite representation. At each step, starting from the bipartite representation for a specific resolution level B = (VL, VR, E), we first select which nodes to duplicate. This selection is encoded via two vectors: v L N|VL| for left-side nodes, and v R N|VR| for right-side nodes. These vectors specify the number of times each node needs to be duplicated. Therefore, we expand the bipartite graph by duplicating each node by the specified number. Each duplicate retains the same connectivity as its parent cluster node. The following definition formalizes this process (illustrated in the two center figures of Figure 4): Definition 8 (Bipartite graph expansion). Given a bipartite graph B = (VL, VR, E) and two cluster size vectors v L N|VL|, v R N|VR|, denoting the expansion size of each node, let B(B, v L, v R) = ( VL, VR, E) denote the expansion of B, whose node sets are given by: VL = V(1) L V(|VL|) L , where V(p) L = {v(p,i) L | 1 i v L[p]} for 1 p |VL|. VR = V(1) R V(|VR|) R , where V(p) R = {v(p,i) R | 1 i v R[p]} for 1 p |VR|. The edge set E includes all the cluster interconnecting edges: {e{p,i;q,j} | e{p,q} E, v(p,i) L V(p) L , v(q,j) R V(q) R }. After the expansion, we selectively keep or remove edges of the resulting bipartite graph using an edge selection vector e {0, 1}|E| (this corresponds to the rightmost step in Figure 4): Definition 9 (Bipartite representation refinement). Given a bipartite graph B = ( VL, VR, E) and an edge selection vector e {0, 1}|E|, let B( B, e) = (VL, VR, E) denote the refinement of B, where: VL = VL, VR = VR, E E such that the i-th edge e(i) E if and only if e[i] = 1. The process of generating a hypergraph consists of the following steps: 1. Start from a bipartite graph containing only two nodes linked by an edge: B(L) = ({1}, {2}, {(1, 2)}) (leftmost figure of Figure 4). 2. Iteratively expand and refine the current bipartite representation to add details until the desired size is attained: B(l) expand B(l 1) refine B(l 1) 3. Once the final bipartite representation is generated, recover the associated hypergraph by collapsing each node on the right side into a hyperedge. Remark 10. Our generation part differs from that by Bergmeister et al. (2024) in the expansion step. While they initially retain all possible edges between connected clusters before selection, we treat hyperedges analogously to nodes due to the exponential growth of potential hyperedges in hypergraphs. This constraint is necessary to maintain computational feasibility. 3.5 Probabilistic Modeling We now formalize our learning problem, generalizing the approach by Bergmeister et al. (2024). Given a dataset {H(1), . . . , H(N)} of i.i.d. hypergraph samples, our goal is to fit a distribution p(H) that closely approximates the unknown true generative process. We model the marginal likelihood of a hypergraph H as the sum of likelihoods over expansion sequences of its bipartite representation B: p(H) = p(B) = X ϖ Π(B) p(ϖ), (1) where Π(B) denotes the set of all possible expansion sequences (B(L) = ({1}, {2}, {(1, 2)}), B(1), . . . , B(0) = B) from a minimal bipartite graph to the target hypergraph s bipartite representation B. Each B(l 1) is a refined expansion of its predecessor, as per Definitions 8 and 9. Assuming a Markovian structure, we factorize the likelihood as: p(ϖ) = p(B(L)) | {z } 1 l=L p(B(l 1)|B(l)) (2) l=L p(e(l 1)| B(l 1))p(v(l) L , v(l) R |B(l)). (3) To avoid modeling two separate distributions p(e(l)| B(l)) and p(v(l) L , v(l) R |B(l)), we rearrange terms as: p(ϖ) = p(e0| B0) p(v(L) L , v(L) R |B(L)) | {z } p(v(L) L ,v(L) R ) l=L 1 p(v(l) L , v(l) R |B(l))p(e(l)| B(l)) We model v(l) L and v(l) R to be conditionally independent of B(l) given B(l), i.e., p(v(l) L , v(l) R |B(l), B(l)) = p(v(l) L , v(l) R |B(l)). This allows us to finally write: p(v(l) L , v(l) R |B(l))p(e(l)| B(l))=p(v(l) L , v(l) R , e(l)|B(l)). (5) SBM Hypergraphs Ego Hypergraphs Tree Hypergraphs (navg = 31.73, std = 0.55) (navg = 109.71, std = 10.23) (navg = 32, std = 0) Valid SBM Node Num Node Deg Edge Size Spectral Valid Ego Node Num Node Deg Edge Size Spectral Valid Tree Node Num Node Deg Edge Size Spectral Hyper PA 2.5% 0.075 4.062 0.407 0.273 0% 35.83 2.590 0.423 0.237 0% 2.350 0.315 0.284 0.159 VAE 0% 0.375 1.280 1.059 0.024 0% 47.58 0.803 1.458 0.133 0% 9.700 0.072 0.480 0.124 GAN 0% 1.200 2.106 1.203 0.059 0% 60.35 0.917 1.665 0.230 0% 6.000 0.151 0.469 0.089 Diffusion 0% 0.150 1.717 1.390 0.031 0% 4.475 3.984 2.985 0.190 0% 2.225 1.718 1.922 0.127 HYGENE 65% 0.525 0.321 0.002 0.010 90% 12.55 0.063 0.220 0.004 77.5% 0.000 0.059 0.108 0.012 Table 1: Comparison between HYGENE and other baselines for the SBM, Ego, and Tree hypergraphs. 3.6 Implementation We employ EDM denoising diffusion framework (Karras et al. 2022) to model the probability p(v L, v R, e|B): v L, v R, and e are treated as node and edge features of the bipartite representation; these features are noised and a model is trained to recover the original features. Our model consists of an embedding layer for each vector, followed by a PPGN (Maron et al. 2019) feeding into one final output layer. Appendix C.1 details the architecture. Additional details are provided as follows (see Appendix C.2 for more details). 1. Similarly to (Bergmeister et al. 2024), we use a deterministic expansion size procedure where only a predefined number of nodes are expanded at each iteration, those being the most probable according to the model, while the others are left unchanged. 2. We also reuse the perturbed expansion, where the bipartite graph is noised through the introduction of random edges connecting a node to others located within a predefined radius around it. 3. We extend spectral conditioning to hypergraphs, wherein spectral properties of the target hypergraph are used as conditioning during the prediction of B(l). Indeed, the spectrum of B(l+1) approximates the target spectrum due to the spectral preservation during coarsening: using Lemma 1 by Agarwal, Branson, and Belongie (2006), we can easily prove (see Appendix A.4): Sp(LB) = n 1 1 λ | λ Sp(LH) o [0, 2], (6) where Sp(LB) and Sp(LH) are the spectra of the normalized Laplacians of the bipartite representation and the hypergraph, respectively. There is also equivalence for the eigenspaces (see the proof in Appendix A.4). Consequently, preserving the k smallest non-zero eigenvalues of the unnormalized Laplacian of the weighted clique expansion also preserves the k smallest non-zero (and k largest non-equal to 2) eigenvalues of the normalized Laplacian of the bipartite representation. During hypergraph reconstruction from its bipartite representation, isolated nodes and empty hyperedges (corresponding to empty rows and columns in the incidence matrix) are discarded. 4 Experiments and Results In this section, we detail our experimental setup, covering datasets and evaluation metrics. Then, we compare our approach against the following baselines: Hyper PA (Do et al. 2020), a Variational Autoencoder (VAE) (Kingma and Welling 2013), a Generative Adversarial Network (GAN) (Goodfellow et al. 2020), and a standard 2D diffusion model (Ho, Jain, and Abbeel 2020) trained on incidence matrix images, where hyperedge membership is represented by white pixels and absence by black pixels. Finally, we ablate on the spectrum-preserving coarsening and the upper bound for the number of hyperedges defined in Proposition 6. Our goal is threefold: (i) proving that HYGENE can generate the desired hyperedge distribution, (ii) proving that HYGENE can closely mimic a range of strict structural properties, and (iii) proving the importance of our components, i.e., spectrum-preserving coarsening and upper bounding the size of hyperedge clusters during coarsening. Detailed numerical results are available in Appendix G, and Appendix H provides several visualizations of our generated hypergraphs. Datasets. We evaluate our method on four synthetic hypergraph datasets: Erd os R enyi (ER) (Erd os and R enyi 1960), Stochastic Block Model (SBM) (Kim, Bandeira, and Goemans 2018), Ego (Comrie and Kleinberg 2021), and Tree (Nieminen and Peltola 1999). Furthermore, we also test HYGENE on topologies of low-poly feature-less versions of three classes of Model Net40 (Wu et al. 2015) converted to hypergraphs: plant, piano, and bookshelf. Each dataset is split into 128 training, 32 validation, and 40 test hypergraphs. More details can be found in Appendix D.1. Metrics. Our metrics measure: (i) overall structural similarities like Node Num (difference in the number of nodes), Node Deg (difference in node degrees), and Edge Size (difference in the size of the hyperedges); (ii) topological properties by computing the average difference of the Spectral properties. For datasets with specific structural requirements, Valid metrics assess the fraction of generated samples satisfying these properties. Lower values indicate better performance for all metrics except Valid, where higher is better. More details can be found in Appendix D.2. Comparison with the Baselines. Tables 1 and 2 show the comparison of HYGENE against our baselines. We observe that the proposed method effectively captures the edge Erdos-Renyi Hypergraphs Model Net40 Piano Model Net40 Plant Model Net40 Bookshelf (navg = 32, std = 0.07) (navg = 177.29, std = 57.11) (navg = 124.86, std = 87.88) (navg = 119.38, std = 68.20) Node Num Node Deg Edge Size Spectral Node Num Node Deg Edge Size Spectral Node Num Node Deg Edge Size Spectral Node Num Node Deg Edge Size Spectral Hyper PA 0.000 5.530 0.183 0.177 0.825 9.254 0.023 0.067 10.83 6.566 0.046 0.061 8.025 7.562 0.044 0.048 VAE 0.100 2.140 0.540 0.035 75.35 8.060 1.686 0.396 76.15 3.895 1.573 0.205 47.45 6.190 1.520 0.190 GAN 0.675 2.560 0.657 0.048 0.000 409.0 86.38 0.697 0.000 378.1 56.35 0.364 0.000 397.2 46.30 0.476 Diffusion 0.050 2.225 0.781 0.014 0.050 20.90 4.192 0.113 0.025 21.03 3.439 0.069 0.000 20.36 2.346 0.079 HYGENE 0.775 0.445 0.012 0.006 42.52 6.290 0.027 0.117 68.38 2.428 0.027 0.034 69.73 1.050 0.034 0.068 Table 2: Comparison between HYGENE and other baselines for the ER and Model Net40 hypergraphs. Upper Bound Spec. Pr. Coarsen. SBM Hypergraphs Ego Hypergraphs Tree Hypergraphs Erd os R enyi Hypergraphs Valid SBM Node Deg Edge Size Spectral Valid Ego Node Deg Edge Size Spectral Valid Tree Node Deg Edge Size Spectral Node Deg Edge Size Spectral 57.5% 1.234 0.006 0.009 77.5% 0.861 0.654 0.149 20% 0.134 0.088 0.014 1.018 0.045 0.009 47.5% 0.148 0.005 0.011 77.5% 0.115 0.146 0.004 77.5% 0.072 0.015 0.026 1.015 0.124 0.014 65% 0.321 0.002 0.010 90% 0.063 0.220 0.004 77.5% 0.059 0.108 0.012 0.445 0.012 0.006 Table 3: Ablation studies on the upper bound in the number of hyperedges and the spectrum-preserving coarsening. size distribution and successfully imitates structural properties across datasets. This is particularly evident in the Ego dataset, where our approach uniquely generates correct ego hypergraphs with a high success rate of 90%. In some cases, Hyper PA has a similar or better performance compared to HYGENE, but this baseline requires to know a-priori the true node degree and hyperedge distributions on a per-hypergraph basis, whereas our model only requires the desired number of nodes. The primary advantage of HYGENE over other baseline approaches lies in its comprehension of hypergraph structure. This is particularly evident in the Valid metrics, where only HYGENE achieves satisfactory results. Indeed, node degrees and hyperedge sizes can be replicated by merely outputting the correct density of white pixels on the incidence matrix images. This explains the satisfactory results sometimes achieved by these baselines for Node Deg and Edge Size. However, the underlying structure cannot be captured this way, and our baselines fail the Valid metrics. While one might consider using graph generators to produce hypergraph representations, this approach is infeasible because: (i) generating the clique expansion and recovering the associated hypergraph is an NP-hard problem as it requires the enumeration of all cliques; and (ii) for the bipartite representation, determining which side corresponds to nodes and which to hyperedges is non-trivial. Additionally, we empirically observe that graph-based models often struggle to generate valid bipartite graphs. For example, we only obtained 30% of correct bipartite graphs for both models by Bergmeister et al. (2024) and Vignac et al. (2023). Ablation Studies. Table 3 shows the results of our ablation studies. First, we observe that not enforcing an upper limit on the hyperedge cluster sizes makes the hyperedge generation task more difficult: for the four datasets, Node Deg greatly increases. This is especially harmful to the Valid Tree and the Ego s Spectral metrics as their structure requires very specific sparsity properties. For the four datasets, the model overestimates the correct number of hyperedges and produces denser hypergraphs. Then, we also observe that the effects of spectrumpreserving coarsening are more subtle. SBM and Ego hypergraphs suffer the most from its absence. This is expected since failing to preserve their structure during coarsening reduces the available information the model possesses during generation. Tree hypergraphs are not heavily impacted due to their relative absence of global structure, whereas ER hypergraphs suffer in the Node Deg and Edge Size metrics as the model struggles to correctly generate the hyperedges. This is also expected since the spectrum of the hypergraph contains information on the hyperedge distribution. Limitations. The mesh datasets reveal that HYGENE faces difficulties in accurately producing the specified number of nodes and hyperedges. The model appears to sample from the underlying distribution of hypergraph sizes rather than adhering to the node count directive provided during the generation process. Notably, the Node Num metric approximates the standard deviation (std) of node counts for each dataset. We hypothesize it stems from an inability to correctly estimate the number of hyperedges. The model then discards the excess nodes by disconnecting them to keep the targeted properties in the remainder of the hypergraph. 5 Conclusions In this work, we introduced HYGENE which is, to the best of our knowledge, the first attempt at diffusion-based hypergraph generation. We generalized the iterative local expansion scheme by Bergmeister et al. (2024) and the coarsening process by Loukas (2019) for hypergraphs. Therefore, we introduced an iterative local expansion procedure for the generation of hyperedges. We trained a denoising diffusion model and successfully tested the ability of our method to generate hypergraphs sampled from specific distributions. This work provides a key contribution to graph generation, being one of the first able of directly generating hypergraphs. Acknowledgments The authors acknowledge the ANR France (French National Research Agency) for its financial support of the System On Chip Design leveraging Artificial Intelligence (SODA) project under grant ANR-23-IAS3-0004 and the JCJC project De SNAP ANR-24-CE23-1895-01. This project has also been partially funded by the Hi!PARIS Center on Data Analytics and Artificial Intelligence. References Agarwal, S.; Branson, K.; and Belongie, S. 2006. Higher order learning with graphs. In International Conference on Machine Learning. Arafat, N. A.; Basu, D.; Decreusefond, L.; and Bressan, S. 2020. Construction and random generation of hypergraphs with prescribed degree and dimension sequences. In International Conference on Database and Expert Systems Applications. Bergmeister, A.; Martinkus, K.; Perraudin, N.; and Wattenhofer, R. 2024. Efficient and scalable graph generation through iterative local expansion. In International Conference on Learning Representations. Bolla, M. 1993. Spectra, Euclidean representations and clusterings of hypergraphs. Discrete Mathematics. Chen, X.; He, J.; Han, X.; and Liu, L.-P. 2023. Efficient and degree-guided graph generation via discrete diffusion modeling. In International Conference on Machine Learning. Comrie, C.; and Kleinberg, J. 2021. Hypergraph egonetworks and their temporal evolution. In IEEE International Conference on Data Mining. Do, M. T.; Yoon, S.-e.; Hooi, B.; and Shin, K. 2020. Structural patterns and generative models of real-world hypergraphs. In ACM SIGKDD International Conference on Knowledge Discovery & Data Mining. Dupagne, A.; and Teller, A. 1998. Hypergraph formalism for urban form specification. In COST C4 Final Conference, Kiruna. Erd os, P.; and R enyi, A. 1960. On the evolution of random graphs. Publ. Math. Inst. Hungar. Acad. Sci. Estrada, E.; and Rodr ıguez-Vel azquez, J. A. 2006. Subgraph centrality and clustering in complex hyper-networks. Physica A: Statistical Mechanics and its Applications. Gong, X.; Higham, D. J.; and Zygalakis, K. 2023. Generative hypergraph models and spectral embedding. Scientific Reports. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; and Bengio, Y. 2020. Generative adversarial networks. Communications of the ACM. Grzesiak-Kope c, K.; Oramus, P.; and Ogorzałek, M. 2017. Hypergraphs and extremal optimization in 3D integrated circuit design automation. Advanced Engineering Informatics. Guo, Y.; Zou, D.; and Lerman, G. 2023. An Unpooling Layer for Graph Generation. In International Conference on Artificial Intelligence and Statistics. Higham, D. J.; and De Kergorlay, H.-L. 2021. Epidemics on hypergraphs: Spectral thresholds for extinction. Proceedings of the Royal Society A. Ho, J.; Jain, A.; and Abbeel, P. 2020. Denoising diffusion probabilistic models. In Advances in Neural Information Processing Systems. Kajino, H. 2019. Molecular hypergraph grammar with its application to molecular optimization. In International Conference on Machine Learning. Karras, T.; Aittala, M.; Aila, T.; and Laine, S. 2022. Elucidating the design space of diffusion-based generative models. In Advances in Neural Information Processing Systems. Kim, C.; Bandeira, A. S.; and Goemans, M. X. 2018. Stochastic block model for hypergraphs: Statistical limits and a semidefinite programming approach. ar Xiv preprint ar Xiv:1807.02884. Kingma, D. P.; and Welling, M. 2013. Auto-encoding variational Bayes. In International Conference on Learning Representations. Kong, L.; Cui, J.; Sun, H.; Zhuang, Y.; Prakash, B. A.; and Zhang, C. 2023. Autoregressive diffusion model for graph generation. In International Conference on Machine Learning. Lin, Z.; Yan, Q.; Liu, W.; Wang, S.; Wang, M.; Tan, Y.; and Yang, C. 2023. Automatic hypergraph generation for enhancing recommendation with sparse optimization. IEEE Transactions on Multimedia. Loukas, A. 2019. Graph reduction with spectral and cut guarantees. Journal of Machine Learning Research. Luo, Z.; Hy, T. S.; Tabaghi, P.; Defferrard, M.; Rezaei, E.; Carey, R. M.; Davis, R.; Jain, R.; and Wang, Y. 2024. DEHNN: An effective neural model for circuit netlist representation. In International Conference on Artificial Intelligence and Statistics. Maron, H.; Ben-Hamu, H.; Serviansky, H.; and Lipman, Y. 2019. Provably powerful graph networks. In Advances in Neural Information Processing Systems. Murgas, K. A.; Saucan, E.; and Sandhu, R. 2022. Hypergraph geometry reflects higher-order dynamics in protein interaction networks. Scientific Reports. Nieminen, J.; and Peltola, M. 1999. Hypertrees. Applied mathematics letters. Niu, C.; Song, Y.; Song, J.; Zhao, S.; Grover, A.; and Ermon, S. 2020. Permutation invariant graph generation via scorebased generative modeling. In International Conference on Artificial Intelligence and Statistics. Rahman, A.; Poirel, C. L.; Badger, D. J.; and Murali, T. 2012. Reverse engineering molecular hypergraphs. In ACM Conference on Bioinformatics, Computational Biology and Biomedicine. Simonovsky, M.; and Komodakis, N. 2018. Graph VAE: Towards generation of small graphs using variational autoencoders. In International Conference on Artificial Neural Networks. Starostin, N.; and Balashov, V. 2008. The use of hypergraphs for solving the problem of orthogonal routing of large-scale integrated circuits with an irregular structure. Journal of Communications Technology and Electronics. Vignac, C.; Krawczuk, I.; Siraudin, A.; Wang, B.; Cevher, V.; and Frossard, P. 2023. Digress: Discrete denoising diffusion for graph generation. In International Conference on Learning Representations. Wu, Z.; Song, S.; Khosla, A.; Yu, F.; Zhang, L.; Tang, X.; and Xiao, J. 2015. 3D shapenets: A deep representation for volumetric shapes. In IEEE/CVF Conference on Computer Vision and Pattern Recognition. You, J.; Ying, R.; Ren, X.; Hamilton, W.; and Leskovec, J. 2018. Graph RNN: Generating realistic graphs with deep auto-regressive models. In International Conference on Machine Learning. Zhou, D.; Huang, J.; and Sch olkopf, B. 2005. Beyond pairwise classification and clustering using hypergraphs. Technical Report 143, Max Planck Institute for Biological Cybernetics. Zhu, Y.; Du, Y.; Wang, Y.; Xu, Y.; Zhang, J.; Liu, Q.; and Wu, S. 2022. A survey on deep graph generation: Methods and applications. In Learning on Graphs Conference. Zhu, Y.; Ouyang, Z.; Liao, B.; Wu, J.; Wu, Y.; Hsieh, C.-Y.; Hou, T.; and Wu, J. 2023. MOLHF: A hierarchical normalizing flow for molecular graph generation. In International Joint Conference on Artificial Intelligence.