# open_materials_generation_with_stochastic_interpolants__d085e024.pdf Open Materials Generation with Stochastic Interpolants Philipp H ollmer * 1 Thomas Egg * 1 Maya M. Martirossyan * 1 Eric Fuemmeler * 2 Zeren Shui 2 Amit Gupta 2 Pawan Prakash 3 Adrian Roitberg 3 Mingjie Liu 3 George Karypis 2 Mark Transtrum 4 Richard G. Hennig 3 Ellad B. Tadmor 2 Stefano Martiniani 1 The discovery of new materials is essential for enabling technological advancements. Computational approaches for predicting novel materials must effectively learn the manifold of stable crystal structures within an infinite design space. We introduce Open Materials Generation (OMat G), a unifying framework for the generative design and discovery of inorganic crystalline materials. OMat G employs stochastic interpolants (SI) to bridge an arbitrary base distribution to the target distribution of inorganic crystals via a broad class of tunable stochastic processes, encompassing both diffusion models and flow matching as special cases. In this work, we adapt the SI framework by integrating an equivariant graph representation of crystal structures and extending it to account for periodic boundary conditions in unit cell representations. Additionally, we couple the SI flow over spatial coordinates and lattice vectors with discrete flow matching for atomic species. We benchmark OMat G s performance on two tasks: Crystal Structure Prediction (CSP) for specified compositions, and de novo generation (DNG) aimed at discovering stable, novel, and unique structures. In our ground-up implementation of OMat G, we refine and extend both CSP and DNG metrics compared to previous works. OMat G establishes a new state of the art in generative modeling for materials discovery, outperforming purely flow-based and diffusion-based implementations. These results underscore the importance of designing flexible deep learning frameworks to accelerate progress in materials science. The OMat G code is available at https: //github.com/FERMat-ML/OMat G. *Equal contribution 1New York University 2University of Minnesota 3University of Florida 4Brigham Young University. Correspondence to: Stefano Martiniani . Proceedings of the 42 nd International Conference on Machine Learning, Vancouver, Canada. PMLR 267, 2025. Copyright 2025 by the author(s). 1. Introduction A core objective of materials science is the discovery of new synthesizable structures and compounds with the potential to meet critical societal demands. The development of new materials such as room-temperature superconductors (Boeri et al., 2022), high-performance alloys with exceptional mechanical properties (Gludovatz et al., 2014; 2016; George et al., 2019), advanced catalysts (Strmcnik et al., 2016; Nakaya & Furukawa, 2023), and materials for energy storage and generation (Liu et al., 2010; Snyder & Toberer, 2008) holds the potential to drive technological revolutions. Exploring the vast compositional and structural landscape of multicomponent materials with novel properties is essential, yet exhaustive experimental screening is infeasible (Cantor, 2021). Quantum and classical molecular simulation offer a powerful alternative, enabling a more targeted and efficient exploration. In recent decades, both experimental (Potyrailo et al., 2011; Maier, 2019) and computational (Jain et al., 2011; Curtarolo et al., 2013) high-throughput pipelines have led to a proliferation of materials databases for crystal structures (Bergerhoff et al., 1983; Mehl et al., 2017) and simulations (Blaiszik et al., 2016; Vita et al., 2023; Fuemmeler et al., 2024). These advances have already facilitated the development of more accurate machine-learned interatomic potentials (Batzner et al., 2022; Batatia et al., 2022; Chen & Ong, 2022). Still, efficiently sampling the manifold of stable materials structures under diverse constraints such as composition and target properties remains a major challenge. Traditional approaches to materials discovery have relied on first-principles electronic structure methods such as density functional theory (DFT)1 or more sophisticated theory, depending on the property (Booth et al., 2013; Zaki et al., 2014; Isaacs & Marianetti, 2020) which, while powerful and fairly accurate, are very computationally expensive. These methods include ab initio random structure searching (AIRSS) (Pickard & Needs, 2011) or genetic algorithms for structure and phase prediction (Tipton & Hennig, 2013), both of which have successfully predicted new crystal struc- 1See Appendix A for a list of acronyms used throughout this paper. Open Materials Generation with Stochastic Interpolants tures and some of which have even been experimentally realized (Oganov et al., 2019). However, the high computational cost of these approaches has limited the scope and speed of material exploration, highlighting the need for cutting-edge ML techniques to significantly accelerate the discovery of stable inorganic crystalline materials. 1.1. Related Works Recent advances in machine learning techniques have generated significant interest in applying data-driven approaches for inorganic materials discovery. Among these, Graph Networks for Materials Exploration (GNo ME) has demonstrated remarkable success by coupling coarse sampling strategies for structure and composition with AIRSS that leverages a highly accurate machine-learned interatomic potential (MLIP) to predict material stability, leading to the identification of millions of new candidate crystal structures (Merchant et al., 2023). Other frameworks have approached the generation of composition and structure jointly through fully ML-based methods. Crystal Diffusion Variational Autoencoder (CDVAE) leverages variational autoencoders and a graph neural network representation to sample new crystal structures from a learned latent space (Xie et al., 2022). To date, state-of-the-art performance in both crystal structure prediction for given compositions and de novo generation of novel stable materials has been achieved by diffusion models such as Diff CSP (Jiao et al., 2023) and Matter Gen (Zeni et al., 2025), as well as conditional flowmatching frameworks such as Flow MM (Miller et al., 2024). While these approaches have demonstrated that ML can push the boundaries of computational materials discovery, it remains uncertain whether score-based diffusion or flowmatching represents the definitive methodological frameworks for this problem. Furthermore, the extent to which the optimal approach depends on the training data remains an open question. Thus far, each new method has typically outperformed its predecessors across datasets. 1.2. Our Contribution The work we present in this paper is the first implementation and extension of the stochastic interpolants (SIs) framework (Albergo et al., 2023) for the modeling and generation of inorganic crystalline materials. SIs are a unifying framework that encompasses both flow-matching and diffusion-based methods as specific instances, while offering a more general and flexible framework for generative modeling. In this context, SIs define a stochastic process that interpolates between pairs of samples from a known base distribution and a target distribution of inorganic crystals. By learning the velocity term of an ordinary differential equation (ODE) or the drift term of a stochastic differential equations (SDE), new samples can be generated by numerically integrating these equations. The flexibility of the SI framework stems from the ability to tailor the choice of interpolants, and the incorporation of an additional random latent variable, further enhancing its expressivity. With their rich parameterization, SIs thus provide an ideal framework for optimizing generative models for materials design. We implement the SI framework in the open-source Open Materials Generation (OMat G) package, released alongside this paper. OMat G allows to train and benchmark models for two materials generation tasks: Crystal structure prediction (CSP) which only learns to generate atomic positions and lattice vectors for a given composition, and de novo generation (DNG) which learns to generate both crystal structure and composition to predict novel materials. We discover that optimizing interpolation schemes for different degrees of freedom of the crystal unit cell substantially improves performance across diverse datasets. As a result, our approach achieves a new state of the art outperforming both Diff CSP (Jiao et al., 2023) and Flow MM (Miller et al., 2024) in CSP and DNG, as well as Matter Gen (Zeni et al., 2025) in DNG across all evaluated datasets under existing, revised, and new performance measures. 2. Background 2.1. Diffusion Models A widely used approach in generative modeling uses diffusion models (Sohl-Dickstein et al., 2015), which define a stochastic process that progressively transforms structured data into noise via a predefined diffusion dynamic. A model is then trained to approximate the reverse process, enabling the generation of new samples, typically by integrating a corresponding SDE. Score-based diffusion models (SBDMs) are an instantiation of diffusion models that learn a score function the gradient of the log probability density to guide the reversal of the diffusion process via numerical integration (Song et al., 2021). SBDMs have demonstrated remarkable success in generating high-quality and novel samples across a wide range of applications where the target distribution is complex and intractable, such as photorealistic image generation (Saharia et al., 2022) and molecular conformation prediction (Corso et al., 2023). 2.2. Conditional Flow Matching Conditional flow matching (CFM) (Liu, 2022; Lipman et al., 2023; Albergo & Vanden-Eijnden, 2023) is a generative modeling technique that learns a flow which transports samples from a base distribution at time t = 0 to a target distribution at time t = 1. This process defines a probability path that describes how samples are distributed at any intermediate time t [0, 1]. The velocity field associated with Open Materials Generation with Stochastic Interpolants this flow governs how individual samples evolve over time. CFM learns the velocity indirectly by constructing conditional vector fields that are known a priori. Once trained, samples drawn from the base distribution can be evolved numerically to generate new samples from the target distribution. Originally, CFM was formulated using Gaussian conditional probability paths, but Tong et al. (2024) later extended this framework to allow for arbitrary probability paths and couplings between base and target distributions. A further extension, particularly relevant to physics and chemistry, is Riemannian flow matching (RFM), which generalizes CFM to Riemannian manifolds (Chen & Lipman, 2024). This allows in particular to use the flow-matching framework for systems with periodic boundary conditions as they appear in unit cell representations of inorganic crystals (Miller et al., 2024). 3. Open Materials Generation 3.1. Stochastic Interpolants SIs provide a unifying mathematical framework for generative modeling, generalizing both SBDMs and CFM (Albergo et al., 2023). The SI x(t, x0, x1, z) bridges the base distribution ρ0 with a target distribution ρ1 by learning a time-dependent map. In this work, we focus on stochastic interpolants of the form: xt x(t, x0, x1, z) = α(t)x0 + β(t)x1 + γ(t)z. (1) Here, t [0, 1] represents time and (x0, x1) are paired samples drawn from ρ0 and ρ1, respectively. The random variable z is drawn from a standard Gaussian N(0, I) independently of x0 and x1. The functional forms of α, β, and γ are flexible, subject to few constraints (see Appendix B.2). The inclusion of the latent variable γ(t)z allows sampling of an ensemble of paths around the mean interpolant I(t, x) = α(t)x0 + β(t)x1, and is theorized to improve generative modeling by promoting smoother and more regular learned flows (Albergo et al., 2023). The time-dependent density ρt of the stochastic process xt in Eq. (1) can be realized either via deterministic sampling through an ODE (derived from a transport equation) or stochastic sampling through an SDE (derived from a Fokker Planck equation) only requiring x0 ρ0 (see Appendix B.1). This enables generative modeling by evolving samples from a known base distribution ρ0 to the target distribution ρ1. For both ODEand SDE-based sampling, the required velocity term bθ(t, x) is learned by minimizing the loss function Lb(θ) = Et,z,x0,x1 |bθ(t, xt)|2 2 tx(t, x0, x1, z) bθ(t, xt) , (2) where the expectation is taken independently over t U(0, 1), the uniform distribution between 0 and 1, z N(0, I), x0 ρ0, and x1 ρ1. For SDE-based sampling, an additional denoiser zθ(t, x) must be learned by minimizing an additional loss Lθ z(θ) = Et,z,x0,x1 |zθ(t, xt)|2 2 zθ(t, xt) z . (3) The velocity term, along with the denoiser in the case of SDE-based sampling, enables the generation of samples from the target distribution (Albergo et al., 2023). Note that minimizing with respect to these loss functions amounts to minimizing with respect to a mean-squared error loss function (see Appendix C.2). For ODE-based sampling, γ(t) = 0 in the interpolant x(t, x0, x1, z) is a possible choice. However, for SDE-based sampling, γ(t) > 0 is required for all t (0, 1) (see Appendix B.1). By appropriately selecting interpolation functions α, β, γ and choosing between deterministic (ODE) and stochastic (SDE) sampling schemes (see Appendix B.2), the SI framework not only recovers CFM and SBDM as special cases (see Appendix B.9) but also enables the design of a broad class of novel generative models. The strength of OMat G s SI implementation for materials discovery lies in its ability to tune both the interpolation and sampling schemes, as illustrated in Fig. 1 for a pair of structures sampled from ρ0 and ρ1. By systematically optimizing over this large design space, we achieve superior performance for CSP and DNG tasks across datasets, as discussed in Section 5. 3.2. Crystal Representation and Generation A crystalline material is defined by its idealized repeat unit, or unit cell, which encodes its periodicity. In the OMat G representation, a unit cell is described by separating the material s chemical composition given by its atomic species A ZN >0, where N is the number of atoms in the unit cell from its structural representation its fractional coordinates X [0, 1)3 N with periodic boundaries and lattice vectors L R3 3. During training, all three components {A, X, L} are considered simultaneously. We apply the SI framework only to the continuous structural representations {X, L} with loss functions defined in Eqs (2) and (3), and use discrete flow matching (DFM) on the chemical species A (see Section 3.2.3) (Gat et al., 2024). The number of atoms N in the structure x0 sampled from the base distribution ρ0 is determined by the number of atoms in the corresponding structure x1 sampled from the target distribution ρ1. 3.2.1. ATOMIC COORDINATES For treating fractional coordinates, we implement a variety of periodic interpolants that connect the base and target data distributions (see Section 4.1). We specify the base distribution for the fractional coordinates x [0, 1) for all x X via a uniform distribution (except for the score-based dif- Open Materials Generation with Stochastic Interpolants Linear Enc. Dec. a = 0 a = 0.2 = 0 γ γ(t) = sin2(πt) Enc. Dec. Interpolant Interpolant choice Tuning γ(t) = at(1 t) ODE SDE Sampling scheme Figure 1. Visualization of the tunable components of the SI framework for bridging samples x0 (gray particles) and x1 (purple particles). Interpolation paths are shown only for one pair of highlighted particles. (a) The choice of the interpolant changes the path of the time-dependent interpolation trajectory. (b) During inference, the learned velocity term bθ(t, x) and denoiser zθ(t, x) generate new samples via ODE or SDE integration, here for a linear interpolant with γ = p 0.07t(1 t). (c) The inclusion of a latent variable γ(t)z changes the interpolation path. For SDE-based sampling, γ(t) > 0 is required. (d) The function γ(t, a) = p at(1 t) depends on a that also influences the interpolation path. fusion interpolant that, following the approach of Jiao et al. (2023), uses a wrapped normal distribution ρ0(x) which becomes a uniform distribution in the limit of large variance). To satisfy periodic boundary conditions on the paths defined by the interpolants, we extend the SI framework to the surface of a four-dimensional torus in this paper. Reminiscent of RFM (Chen & Lipman, 2024), the linear interpolant on the torus traverses a path equivalent to the shortest-path geodesic which is always well-defined.2 Other interpolants, however, are more complex. In order to uniquely define them, we always define the interpolation with respect to the shortest-path geodesic. That is, for interpolation between x0 and x1 with a periodic boundary at 0 and 1, we first unwrap x1 to the periodic image x 1 which has the shortest possible distance from x0. Following this, the interpolation between x0 and x 1 is computed given a choice of interpolant, and the traversed path is wrapped back into the boundary from 0 to 1. This approach is illustrated in Appendix B.5. 2The only exception being when two points are precisely half the box length apart. However, this case is not relevant for the given base distribution. 3.2.2. LATTICE VECTORS Lattice vectors L are treated with a wide range of (nonperiodic) stochastic interpolants (see Section 4.1 again). To construct the base distribution, we follow Miller et al. (2024) and construct an informative base distribution ρ0(L) by combining a uniform distribution over the lattice angles with a log-normal distribution fitted to the empirical distribution of the lattice lengths in each target dataset. This choice brings the base distribution closer to the target distribution. Unlike SBDM, which requires a Gaussian base distribution, the SI framework allows such flexibility. Importantly, the model still has to learn to generate a joint, correlated distribution of lattice vectors, fractional coordinates, and atomic species. 3.2.3. ATOMIC SPECIES The discrete nature of chemical compositions A in atomic crystals requires a specialized approach for generative modeling. To address this, we implement discrete flow matching (DFM) (Campbell et al., 2024). In our implementation of the DFM framework, each atomic species a A can take values in {1, 2, . . . , 100} {M}; where {1 100} are atomic element numbers and M is a masking token used during training. The base distribution is defined as ρ0(a) = [M]N, meaning that initially all N atoms are masked. As sampling progresses, the identities of the atoms evolve via a continuous-time Markov Chain (CTMC), and are progressively unmasked to reveal valid atomic species. At t = 1, all masked tokens are replaced. To learn this process, we define a conditional flow pt|1(at|a1) that linearly interpolates in time from the fully masked state a0 toward a1 and thus yields the composition at of the interpolated structure xt. Based on these conditional flows, a neural network is trained to approximate the denoising distribution pθ 1|t(a1|xt), which yields the probability for the composition a1 given the entire structure xt, by minimizing a cross-entropy loss LDFM(θ) = Et,x1,xt h log pθ 1|t(a1|xt) i . (4) In doing this, we are able to directly construct the marginal rate matrix Rθ t (at, i) for the CTMC that dictates the rate of at at time t jumping to a different state i during generation (see Appendix B.6). It is important to note that the learned probability path is a function of the entire atomic configuration {A, X, L} which is necessary for the prediction of chemical composition from structure. 3.3. Joint Generation with Stochastic Interpolants For both CSP and DNG tasks, we seek to generate samples from a joint distribution over multiple coordinates. For DNG, this joint distribution ρ1 encompasses all elements of a crystal unit cell. For CSP we similarly model the joint Open Materials Generation with Stochastic Interpolants distribution, ρ1, but with atom types fixed to compositions sampled from the target dataset. For both tasks, the total loss function is formulated as a weighted sum of the individual loss functions for each variable (see Appendix C.2), and their relative weights are optimized (see Appendix C.3). We illustrate both types of models and their structure generation process in Fig. 2a. Additionally, for DNG, we consider a two-step process in which composition is learned separately from structure, as seen in Fig. 2b. In this approach, we first train a chemical formula prediction (CFP) model (see Appendix C.1) to generate compositions optimized for SMACT stability (Davies et al., 2019), similarity in the distribution of Narity of known structures, as well as uniqueness and novelty. The predicted compositions are then used as input for a pretrained CSP model, which generates the corresponding atomic configurations. 4. Methodology 4.1. Choice of Interpolant In training OMat G, we optimize the choice of the interpolating function that is used during training for the lattice vectors (without periodic boundary conditions) and the fractional coordinates (with periodic boundary conditions). We consider four interpolants of the form defined in Eq. (1), each shaping the interpolation trajectory differently (see also Appendices B.2 and B.9 for further details). The linear interpolant defines a constant velocity trajectory from x0 to x1. When combined with an ODE sampling scheme and γ = 0, this reproduces the specific instantiation of CFM implemented in Flow MM (see Appendix B.9). However, combining the linear interpolant with an SDE sampling scheme or nonzero γ already introduces key differences. The inclusion of the latent variable can promote smoother learned flows (Albergo et al., 2023), while stochastic sampling alters the generative dynamics compared to the deterministic formulation in Flow MM. The trigonometric interpolant prescribes trajectories with more curvature than the linear interpolant. The encoder-decoder interpolant first evolves samples from ρ0 at t = 0 to follow an intermediate Gaussian distribution at a switch time Tswitch, before mapping them to samples from the target distribution ρ1 at t = 1. This approach has been found to interpolate more smoothly between distributions, potentially mitigating the formation of spurious features in the probability path at intermediate times (Albergo et al., 2023). Lastly, we consider variance-preserving score-based diffusion (VP SBD) and variance-exploding score-based diffusion (VE SBD) interpolants. When paired with an SDE sampling scheme, these interpolants are mathematically equivalent to the corresponding SBDM, but on the continuous time interval [0, 1]. Different noise schedules of the variancepreserving and variance-exploding SBDMs can likewise be encoded in different variants of the VP and VE SBD interpolants (see Appendix B.9). For the results presented in this paper, we only consider a constant noise schedule for the VP SBD interpolant, and a geometric noise schedule for the VE SBD interpolant. The SBD interpolants assume that ρ0 is a Gaussian distribution, and unlike the previous interpolants it involves no explicit latent variable; instead the α(t)x0 term takes on this role. The incorporation of VP and VE SBD interpolants enables OMat G to reproduce similar conditions to those in Diff CSP and Matter Gen. The trajectory of the encoder-decoder interpolant between times t = Tswitch and t = 1 resembles that of the SBD interpolants between times t = 0 and t = 1. For the example of using the encoder-decoder interpolant only for the coordinates, however, we emphasize that the Gaussiandistributed coordinates at t = Tswitch are conditioned on other coordinates that are partially interpolated at this point. Conversely, for SBD interpolation, the Gaussian distributed coordinates at t = 0 are only conditioned on other random variables since, at this point, all elements of x0 are randomly distributed. To investigate how different interpolants affect generative performance, we consider all interpolants outlined above for both the atomic positions X and the lattice vectors L. We noted that learning accurate velocities and denoisers for the atomic positions, X is more challenging than for the other degrees of freedom. Accordingly, we optimize all hyperparameters including the choice of interpolant for L separately for each interpolant applied to X. This results in a set of experiments specific to the positional interpolants, where the best performing lattice interpolant may vary. 4.2. Equivariant Representation of Crystal Structures Imposing inductive biases on the latent representation of the crystal structure can promote data efficiency and improve learning. The CSPNet architecture (Jiao et al., 2023), originally adopted in Diff CSP, is an equivariant graph neural network (EGNN) (Satorras et al., 2021) that produces a permutationand rotation-equivariant, as well as translationinvariant representation of the crystal structures. In the current OMat G implementation, we employ CSPNet as an encoder that is trained from scratch. The CSPNet architecture encodes atomic types using learnable atomic embeddings and represents fractional coordinates through sinusoidal positional encodings (see Appendix C.1). These features are processed through six layers of message-passing, after which the encoder produces the velocity bθ(t, x) of both the lattice and the fractional coordinates, as well as potentially predicting the denoiser zθ(t, x). For DNG, the Open Materials Generation with Stochastic Interpolants (A0, X0, L0) ρ0 (A0.33, X0.33, L0.33) (A0.66, X0.66, L0.66) (A1, X1, L1) Figure 2. Illustration of crystal structure prediction (CSP) and de novo generation (DNG) tasks. (a) For CSP, the species A are fixed with known compositions from t = 0. From this, we predict X and L from randomly sampled initial values. For DNG, we predict (A, X, L) jointly. Our implementation of discrete flow matching (DFM) initializes A as a sequence of masked particles that are unmasked through a series of discrete jumps to reveal a physically reasonable composition. (b) Two avenues for performing DNG of materials. The first uses two steps: a chemical formula prediction (CFP) model predicts compositions and then uses a CSP model to find accompanying stable structures. The second trains a DNG model over cell, species, and fractional coordinates jointly as shown in (a). network must also predict log pθ 1|t(a1|xt). The resulting outputs inherently preserve the permutation, rotational, and translational symmetries embedded in CSPNet. The output of CSPNet is invariant with respect to translations of the fractional coordinates in the input. Thus, one should, in principle, use a representation of the fractional coordinates that does not contain any information about translations. While this is straightforward in Euclidean space by removing the mean of the coordinates of the given structure, this cannot be done with periodic boundary conditions where the mean is not uniquely defined. We follow Miller et al. (2024) and instead remove the well-defined center-of-mass motion when computing the ground-truth velocity tx(t, x0, x1, z) in Eq. (2). Alternative EGNNs such as Nequ IP (Batzner et al., 2022), M3GNet (Chen & Ong, 2022), or MACE (Batatia et al., 2022) which have been widely used for the development of MLIPs can also serve as plug-and-play encoders within OMat G s SI framework. Integrating different architectures is a direction that we plan to explore in future iterations of the framework. 4.3. Comparison to other Frameworks We compare our results to Diff CSP and Flow MM models for both CSP and DNG. For DNG, we additionally consider the Matter Gen-MP model that was trained on the same MP20 dataset as OMat G s DNG models. We detail in Section 5 how we improve the extant benchmarks used in the field and therefore recompute all CSP and DNG benchmarks for these models. In nearly all cases, we were able to generate structures using the Diff CSP, Flow MM, and Matter Gen source codes whose metrics closely matched the previously reported metrics in their respective manuscripts. For Diff CSP and Matter Gen, we relied on published checkpoints while we retrained Flow MM from scratch. The observed differences can be attributed to the use of a newer version of SMACT composition rules3 (Davies et al., 2019) and to natural fluctuations during generation and model retraining. Since the focus of this work is to assess our model s ability to learn unconstrained and unconditioned flows, we do not compare against symmetry-constrained generation methods (AI4Science et al., 2023; Cao et al., 2024; Zhu et al., 2024; Kazeev et al., 2024; Jiao et al., 2024). Symmetry constraints can be incorporated in future extensions of the flexible OMat G framework. 5. Experiments 5.1. Performance Metrics Crystal structure prediction. We assess the performance of OMat G s and competing models using a variety of standard (introduced in (Xie et al., 2022; Zeni et al., 2025)), refined, and contributed benchmarks. For the CSP task, we generate a structure for every composition in the test dataset. We then attempt to match every generated structure with the corresponding test structure using Pymatgen s 3The SMACT Python library updated its default oxidation states with the release of version 3.0. Open Materials Generation with Stochastic Interpolants Table 1. Results from crystal structure prediction. Match rate and RMSE of matched structures without (left) and with (right) filtering for structural and compositional validity are reported for all models. For OMat G s models, the choice of positional interpolant, latent variable component γ, and sampling scheme are noted. For all SDE sampling schemes the inclusion of γ is assumed and not noted; for SBD interpolants γ is not relevant. Further details and complete results for perov-5 and MP-20 can be found in Appendix C.3. Method perov-5 MP-20 MPTS-52 Alex-MP-20 Match (%) RMSE Match (%) RMSE Match (%) RMSE Match (%) RMSE Diff CSP 53.08 / 51.94 0.0774 / 0.0775 57.82 / 52.51 0.0627 / 0.0600 15.79 / 14.29 0.1533 / 0.1489 - Flow MM 53.63 / 51.86 0.1025 / 0.0994 66.22 / 59.98 0.0661 / 0.0629 22.29 / 20.28 0.1541 / 0.1486 - Linear (ODE) w/o γ 51.86 / 50.62 0.0757 / 0.0760 69.83 / 63.75 0.0741 / 0.0720 27.38 / 25.15 0.1970 / 0.1931 72.02 / 64.23 0.0683 / 0.0671 Linear (SDE) w/ γ 74.16 / 72.87 0.3307 / 0.3315 68.20 / 61.88 0.1632 / 0.1611 23.95 / 21.70 0.2402 / 0.2353 61.07 / 54.45 0.1870 / 0.1860 Trig (SDE) w/ γ 73.37 / 71.60 0.3610 / 0.3614 68.90 / 62.65 0.1249 / 0.1235 24.51 / 22.26 0.1867 / 0.1804 72.50 / 64.71 0.1261 / 0.1251 Enc-Dec (ODE) w/ γ 68.08 / 64.60 0.4005 / 0.4003 55.15 / 49.45 0.1306 / 0.1260 14.65 / 13.53 0.2543 / 0.2500 68.11 / 60.58 0.0957 / 0.0938 VP SBD (ODE) 83.06 / 81.27 0.3753 / 0.3755 45.57 / 39.48 0.1880 / 0.1775 9.66 / 8.36 0.3088 / 0.3041 46.23 / 39.96 0.1718 / 0.1618 VE SBD (ODE) 60.18 / 52.97 0.2510 / 0.2337 63.79 / 57.82 0.0809 / 0.0780 21.42 / 19.57 0.1740 / 0.1702 67.79 / 60.25 0.0674 / 0.0649 Table 2. Results from de novo generation of 10 000 structures with models trained on the MP-20 dataset. The integration steps for OMat G is chosen based on best overall performance. For OMat G s model, the choice of positional interpolant, latent variable component γ, and sampling scheme are noted. Best scores in each category are bolded. Method Integration steps Validity (% ) Coverage (% ) Property ( ) Structural Composition Combined Recall Precision wdist (ρ) wdist (Nary) wdist ( CN ) Diff CSP 1000 99.91 82.68 82.65 99.67 99.63 0.3133 0.3193 0.3053 Flow MM 1000 92.26 83.11 76.94 99.34 99.02 1.0712 0.1130 0.4405 Matter Gen-MP 1000 99.93 83.89 83.89 96.62 99.90 0.2741 0.1632 0.4155 Linear (SDE) w/ γ 710 99.04 83.40 83.40 99.47 98.81 0.2583 0.0418 0.4066 Trig (ODE) w/ γ 680 95.05 82.84 82.84 99.33 94.75 0.0607 0.0172 0.1650 Enc-Dec (ODE) w/ γ 840 97.25 86.35 84.19 99.62 99.61 0.1155 0.0553 0.0465 VP SBD (SDE) 870 93.38 80.66 80.66 98.95 92.76 0.1865 0.0768 0.1637 CFP + CSP [Linear (ODE) w/o γ] 130+210 97.95 79.68 78.21 99.67 99.50 0.5614 0.2008 0.6256 We do not bold any values in the structural validity category as the CDVAE model reports the state of the art with 100% structural validity. For the Wasserstein distances of the density and Nary distributions, we only bold values lower than 0.075 and 0.079 respectively, as these were the values reported by Flow MM for their model with 500 integration steps (not included in this table). Structure Matcher module (Ong et al., 2013) with tolerances (stol = 0.5, ltol = 0.3, angletol = 10). We finally report the match rate and the average root-mean square displacement (RMSE) between the test structures and matched generated structures. Here, the RMSEs computed by Pymatgen are normalized by (V/N)1/3, where V is the (matched) volume and N is the number of atoms. During hyperparameter optimization, we only attempt to maximize the match rate (see Appendix C.3). Previously reported match rates filtered the matched generated structures by their structural and compositional validity (see Appendix D.2). We note, however, that the datasets themselves contain invalid structures for example, the MP20 test dataset has 10% compositionally invalid structures. Thus, we argue that the removal of these invalid structures for computation of match rate and RMSE is not reasonable for assessing learning performance; we do, however, provide both match rates (with and without validation filtering). De novo generation. For the DNG task, metrics include validity (structural and compositional), coverage (recall and precision), and Wasserstein distances between distributions of properties including density ρ, number of unique elements N (i.e., an Nary material), and average coordination number by structure CN . We newly introduce the average coordination number benchmark due to the difficulty of generating symmetric structures; a structure s average coordination number is a useful fingerprint, and higher-coordinated structures tend to be more symmetric. The previous DNG metrics are used in conjunction during optimization of hyperparameters (see Appendix C.3). For the best models, we then structurally relax the generated structures in order to calculate the stability and the S.U.N. (stable, unique, and novel) rates using Matter Gen s code base (Matter Gen, 2025). The S.U.N. rate is defined as the percentage of generated structures that are stable with respect to a reference convex hull (within 0.1 e V/atom), are not found within the reference set (novel), and are not dupli- Open Materials Generation with Stochastic Interpolants cated within the generated set itself (unique). The machinelearned interatomic potential Matter Sim (Yang et al., 2024) is utilized for initial structural relaxation, which is subsequently followed by a more computationally expensive DFT relaxation. Full workflow details are given in Appendix D.4. In addition to the stability-based metrics, we also report the root mean squared displacement (RMSD) between the generated and relaxed structures (unnormalized, in units of A). 5.2. Benchmarks and Datasets We use the following datasets to benchmark the models: perov-5 (Castelli et al., 2012), a dataset of perovskites with 18 928 samples with five atoms per unit cell in which only lattice lengths and atomic types change; MP-20 (Jain et al., 2013; Xie et al., 2022) from the Materials Project that contains 45 231 structures with a maximum of N = 20 atoms per unit cell, and MPTS-52 (Baird et al., 2024) which is a chronological data split of the Materials Project with 40 476 structures with up to N = 52 atoms per unit cell and is typically the most difficult to learn. We use the same 6020-20 splits as Xie et al. (2022); Jiao et al. (2023); Miller et al. (2024). Additionally, we consider the Alex-MP-20 dataset (Zeni et al., 2025), where we used an 80-10-10 split constructed from Matter Gen s 90-10 split, in which we removed 10% of the training data to create a test dataset. This dataset contains 675 204 structures with 20 or fewer atoms per unit cell from the Alexandria (Schmidt et al., 2022a;b) and MP-20 datasets. We do not include the carbon-24 dataset (Pickard, 2020) in our results, as the current match rate metric is ill-defined for this dataset; because all elements are carbon, it is not clear how many generated structures are unique and producing a structure that matches one in the reference dataset is trivial.4 5.3. Results Crystal structure prediction. We report the CSP performance of Diff CSP, Flow MM, and six OMat G models on the four benchmark datasets in Tab. 1. Further ablation results across OMat G variants can be found in Tables 6 and 8 in the Appendix. OMat G significantly outperforms previous approaches with respect to match rate on all datasets. We highlight the strong match rates on the perov-5 dataset achieved using the VP SBD and trigonometric positional interpolants with ODE sampling schemes, as shown in Table 1 and Table 6, that greatly surpass (by a factor of 1.6) the match rates of previous models. We note that the relative performance for ODE vs. SDE sampling schemes depends on the positional interpolant. OMat G also outperforms pre- 4Previous papers (Xie et al., 2022; Jiao et al., 2023; Miller et al., 2024) report match rate for carbon-24, but they do not compare each generated structure to the entirety of the reference dataset; their results suggest the match tolerance is larger than the differences between the carbon-24 structures themselves. Table 3. Stability (defined as 0.1 e V/atom above hull), uniqueness, and novelty results from de novo generation on the MP-20 dataset computed for the same models as in Tab. 2. All evaluations are calculated with respect to 1000 relaxed structures (see Appendix D.4), utilizing the Matter Gen code base (Matter Gen, 2025) and the included reference Alex-MP-20 dataset. The average RMSD is between the generated and the relaxed structures, and the average energy above hull is reported in units of e V/atom. Method E /N ( ) RMSD Novelty Stability S.U.N. above hull ( A, ) Rate (%, ) Rate (%, ) Rate (%, ) Diff CSP 0.1751 0.3861 70.04 43.43 15.95 Flow MM 0.1917 0.6316 69.13 41.20 11.73 Matter Gen-MP 0.1772 0.1038 72.40 44.79 20.30 Linear (SDE) w/ γ 0.1808 0.6357 73.31 46.18 22.48 Trig (ODE) w/ γ 0.1670 0.6877 66.45 52.81 21.10 Enc-Dec (ODE) w/ γ 0.1482 0.4187 55.21 60.04 18.77 VP SBD (SDE) 0.2120 0.7851 76.25 40.83 20.73 CFP + CSP 0.2302 0.5375 79.08 39.38 20.17 vious models match rate for CSP on the MP-20 and MPTS52 datasets with the linear (both ODE and SDE sampling scheme) and the trigonometric positional interpolants. Finally, OMat G establishes the first performance baseline for the CSP task on the Alex-MP-20 dataset. De novo generation. For DNG models, we show the validity, coverage, and property metrics of the models in Tab. 2 and the stability, uniqueness, and novelty results in Tab. 3. Further ablation results across OMat G variants can be found in Tab. 10 in the Appendix, and qualitative plots of various distributions are compared in Appendix D.5. OMat G achieves state-of-the-art performance over Diff CSP, Flow MM, and Matter Gen-MP for multiple positional interpolants thanks to the broader design space brought by the SIs. Figure 3 compares the distributions of the average energies above the hull for generated structures, exhibiting OMat G s superior performance for the generation of stable structures. OMat G consistently produces lower energy structures compared to previous models, and they are also generated close to their relaxed configuration. This, together with high novelty rates, begets improved S.U.N. rates. OMat G also outperforms Flow MM in settings where large language models are used as base distributions (Sriram et al., 2024), with results shown in Appendix E. 5.4. Discussion Crystal structure prediction. We generally observe during hyperparameter optimization that the main learning challenge lies in the accurate prediction of the atomic coordinates, which tended to have a higher relative weight in calculating the full loss function in Eq. (32) (see Tabs 7 and 9 in the Appendix). However, the two best-performing models for the perov-5 dataset instead exhibited the opposite lending the most weight to learning of the cell vectors. Open Materials Generation with Stochastic Interpolants 0.2 0.0 0.2 0.4 0.6 0.8 Energy Above Hull (e V/Atom) Diff CSP Flow MM Matter Gen-MP OMat G (Enc-Dec) Figure 3. Histogram of the computed energies above the convex hull for structures generated by Flow MM, Diff CSP, Matter Gen-MP, and OMat G (Enc-Dec interpolant). The OMat G model consistently produces lower energy structures compared to competing models. See Appendix D.4 for calculation details. Our CSP results indicate a tradeoff between match rate and RMSE which is only computed if matched. We find that as the number of generated structures matching known compositions in the test dataset increases, the structural fidelity quantified by the accompanying RMSE also tends to increase. This tradeoff most strongly influences our results on the perov-5 dataset where all positional interpolants but the linear one can beat the previous state-of-the-art match rate. Here, particles generally find the correct local chemical configurations to flow towards during generation, but are not able to end up in the precise symmetric sites. The ODE-based linear interpolant without a latent variable, in contrast, has the lowest RMSE because the particles flow to more symmetric positions, but the local environments are not correct due to species mismatch. We quantify this effect in Fig. 7 in the Appendix. For the CSP task on the perov-5 dataset, we highlighted the particularly strong performance of the VP SBD and trigonometric interpolants in achieving a high match rate. Unlike other datasets, perov-5 has a fixed number of N = 5 atoms per unit cell and a fixed (cubic) cell with varying side lengths and similar fractional positions a combination which should not expose the model to a large variety of unit cell choices during interpolation or generation. By contrast, in other datasets, no singular representation of the periodic repeat unit is imposed on flows, meaning the model cannot learn the invariance (or even equivariance) to the choice of periodic repeat unit.5 This likely contributes to the difficulty of unconstrained flow-based models in generating highly symmetric structures. Thus, the perov-5 dataset presents a unique case where the invariance to unit cell choice does 5Using Niggli reduction during learning to enforce a unique choice of unit cell on structures from our datasets is not sufficient for enforcing this invariance during generation of structures. not need to be learned, making this dataset a useful benchmark for evaluating positional interpolant performance. It is possible that the superior performance of the VP SBD and trigonometric interpolants arise from their ability to generate more circuitous flow trajectories compared to the strictly geodesic paths imposed by the linear interpolant akin to the reasoning behind using latent variables to enhance learning in SIs (Albergo et al., 2023). De novo generation. Thorough hyperparameter optimization enables us to note trends among the best-performing DNG models. In Tabs 10 and 11 in the Appendix, we show the performance metrics and hyperparameters for each model by choice of positional interpolant, sampling scheme, and γ(t) in the latent variable. We observe that several of our best performing models (with respect to S.U.N. and RMSD) possess lower levels of species noise η which sets the probability that an atom will change its identity if already in an unmasked state (see Appendix B.6). Additionally, we find that linear and trigonometric interpolants favor an element-order permutation as a data-dependent coupling during training (see Appendix B.8), while the encoderdecoder and SBD interpolants prefer to not use this coupling. Finally, we note that VP SBD models require a similar magnitude of velocity annealing during generation for positions and lattices (see Appendix B.7). This is in stark contrast to all other models, where a significantly larger velocity annealing parameter is required for generating the positions. 6. Conclusion We adapt stochastic interpolants (SIs) for material generation tasks and propose Open Materials Generation (OMat G), a material-generation framework that unifies score-based diffusion and conditional flow-matching approaches under the umbrella of SIs. By incorporating an equivariant graph representation of crystal structures and explicitly handling periodic boundary conditions, OMat G jointly models spatial coordinates, lattice vectors, and discrete atomic species in a cohesive flow-based pipeline. Our extensive experiments on crystal structure prediction and de novo generation tasks demonstrate that OMat G sets a new state of the art in generative modeling for inorganic materials discovery, yielding more stable, novel, and unique structures than either pure diffusion or pure conditional flow-matching counterparts. We underscore the importance of flexible ML frameworks like OMat G, which can adapt to different types of materials datasets by optimizing the generative model accordingly. Our work represents a key step forward in applications of machine-learning methods to materials discovery. Looking forward, we plan to extend the flexibility of OMat G to additional interpolating functions, improve on the evaluation metrics and datasets, and investigate how different SIs influence the discovery of suitable materials. Open Materials Generation with Stochastic Interpolants Acknowledgements The authors thank Shenglong Wang at NYU IT High Performance Computing and Gregory Wolfe for their resourcefulness and valuable support. The authors acknowledge funding from NSF Grant OAC-2311632. P. H. and S. M. also acknowledge support from the Simons Center for Computational Physical Chemistry (Simons Foundation grant 839534, MT). The authors gratefully acknowledge use of the research computing resources of the Empire AI Consortium, Inc, with support from the State of New York, the Simons Foundation, and the Secunda Family Foundation. Moreover, the authors gratefully acknowledge the additional computational resources and consultation support that have contributed to the research results reported in this publication, provided by: IT High Performance Computing at New York University; the Minnesota Supercomputing Institute (http://www.msi.umn.edu) at the University of Minnesota; UFIT Research Computing (http://www.rc.ufl.edu) and the NVIDIA AI Technology Center at the University of Florida in part through the AI and Complex Computational Research Award; Drexel University through NSF Grant OAC-2320600. Impact Statement This paper presents work whose goal is to advance the field of Machine Learning. There are many potential societal consequences of our work, none which we feel must be specifically highlighted here. AI4Science, M., Hernandez-Garcia, A., Duval, A., Volokhova, A., Bengio, Y., Sharma, D., Carrier, P. L., Benabed, Y., Koziarski, M., and Schmidt, V. Crystal GFN: Sampling crystals with desirable properties and constraints, December 2023. URL http://arxiv. org/abs/2310.04925. Albergo, M. S. and Vanden-Eijnden, E. Building Normalizing Flows with Stochastic Interpolants, March 2023. URL http://arxiv.org/abs/2209.15571. Albergo, M. S., Boffi, N. M., and Vanden-Eijnden, E. Stochastic Interpolants: A Unifying Framework for Flows and Diffusions, November 2023. URL http://arxiv. org/abs/2303.08797. Albergo, M. S., Goldstein, M., Boffi, N. M., Ranganath, R., and Vanden-Eijnden, E. Stochastic interpolants with data-dependent couplings, September 2024. URL http: //arxiv.org/abs/2310.03725. Aranguri, S., Biroli, G., Mezard, M., and Vanden-Eijnden, E. Optimizing Noise Schedules of Generative Models in High Dimensions, January 2025. URL http://arxiv. org/abs/2501.00988. Baird, S. G., Sayeed, H. M., Montoya, J., and Sparks, T. D. Matbench-genmetrics: A Python library for benchmarking crystal structure generative models using time-based splits of Materials Project structures. Journal of Open Source Software, 9(97):5618, May 2024. ISSN 24759066. doi: 10.21105/joss.05618. URL https://joss. theoj.org/papers/10.21105/joss.05618. Batatia, I., Kovacs, D. P., Simm, G., Ortner, C., and Csanyi, G. MACE: Higher Order Equivariant Message Passing Neural Networks for Fast and Accurate Force Fields. Advances in Neural Information Processing Systems, 35:11423 11436, December 2022. URL https://arxiv.org/abs/2206.07697. Batzner, S., Musaelian, A., Sun, L., Geiger, M., Mailoa, J. P., Kornbluth, M., Molinari, N., Smidt, T. E., and Kozinsky, B. E(3)-equivariant graph neural networks for data-efficient and accurate interatomic potentials. Nature Communications, 13(1): 2453, May 2022. ISSN 2041-1723. doi: 10.1038/ s41467-022-29939-5. URL https://www.nature. com/articles/s41467-022-29939-5. Bergerhoff, G., Hundt, R., Sievers, R., and Brown, I. D. The inorganic crystal structure data base. J. Chem. Inf. Comput. Sci., 23(2):66 69, May 1983. ISSN 0095-2338. doi: 10.1021/ci00038a003. URL https://doi.org/ 10.1021/ci00038a003. Bergstra, J., Yamins, D., and Cox, D. D. Making a Science of Model Search: Hyperparameter Optimization in Hundreds of Dimensions for Vision Architectures. TProc. of the 30th International Conference on Machine Learning (ICML 2013), pp. pp. I 115 to I 123, 2013. Blaiszik, B., Chard, K., Pruyne, J., Ananthakrishnan, R., Tuecke, S., and Foster, I. The Materials Data Facility: Data Services to Advance Materials Science Research. JOM, 68(8):2045 2052, August 2016. ISSN 15431851. doi: 10.1007/s11837-016-2001-3. URL https: //doi.org/10.1007/s11837-016-2001-3. Boeri, L., Hennig, R., Hirschfeld, P., Profeta, G., Sanna, A., Zurek, E., Pickett, W. E., Amsler, M., Dias, R., Eremets, M. I., Heil, C., Hemley, R. J., Liu, H., Ma, Y., Pierleoni, C., Kolmogorov, A. N., Rybin, N., Novoselov, D., Anisimov, V., Oganov, A. R., Pickard, C. J., Bi, T., Arita, R., Errea, I., Pellegrini, C., Requist, R., Gross, E. K. U., Margine, E. R., Xie, S. R., Quan, Y., Hire, A., Fanfarillo, L., Stewart, G. R., Hamlin, J. J., Stanev, V., Gonnelli, R. S., Piatti, E., Romanin, D., Daghero, D., and Valenti, R. The 2021 room-temperature superconductivity roadmap. J. Phys.: Condens. Matter, Open Materials Generation with Stochastic Interpolants 34(18):183002, March 2022. ISSN 0953-8984. doi: 10.1088/1361-648X/ac2864. URL https://dx.doi. org/10.1088/1361-648X/ac2864. Booth, G. H., Gr uneis, A., Kresse, G., and Alavi, A. Towards an exact description of electronic wavefunctions in real solids. Nature, 493(7432):365 370, 2013. Bose, A. J., Akhound-Sadegh, T., Huguet, G., Fatras, K., Rector-Brooks, J., Liu, C.-H., Nica, A. C., Korablyov, M., Bronstein, M., and Tong, A. SE(3)-Stochastic Flow Matching for Protein Backbone Generation, April 2024. URL http://arxiv.org/abs/2310.02391. Campbell, A., Yim, J., Barzilay, R., Rainforth, T., and Jaakkola, T. Generative Flows on Discrete State-Spaces: Enabling Multimodal Flows with Applications to Protein Co-Design, June 2024. URL http://arxiv.org/ abs/2402.04997. Cantor, B. Multicomponent high-entropy Cantor alloys. Progress in Materials Science, 120:100754, July 2021. ISSN 0079-6425. doi: 10.1016/j.pmatsci.2020.100754. URL https://www.sciencedirect.com/ science/article/pii/S0079642520301183. Cao, Z., Luo, X., Lv, J., and Wang, L. Space Group Informed Transformer for Crystalline Materials Generation, March 2024. URL https://arxiv.org/abs/ 2403.15734v2. Castelli, I. E., Landis, D. D., Thygesen, K. S., Dahl, S., Chorkendorff, I., Jaramillo, T. F., and Jacobsen, K. W. New cubic perovskites for oneand two-photon water splitting using the computational materials repository. Energy & Environmental Science, 5(10):9034 9043, September 2012. ISSN 1754-5706. doi: 10.1039/C2EE22341D. URL https://pubs.rsc.org/en/content/ articlelanding/2012/ee/c2ee22341d. Chen, C. and Ong, S. P. A universal graph deep learning interatomic potential for the periodic table. Nature Computational Science, 2(11):718 728, November 2022. ISSN 2662-8457. doi: 10.1038/ s43588-022-00349-3. URL https://www.nature. com/articles/s43588-022-00349-3. Chen, R. T. Q. and Lipman, Y. Flow Matching on General Geometries, February 2024. URL http://arxiv. org/abs/2302.03660. Corso, G., St ark, H., Jing, B., Barzilay, R., and Jaakkola, T. Diff Dock: Diffusion Steps, Twists, and Turns for Molecular Docking, February 2023. Curtarolo, S., Hart, G. L. W., Nardelli, M. B., Mingo, N., Sanvito, S., and Levy, O. The high-throughput highway to computational materials design. Nat. Mater., 12(3):191 201, March 2013. ISSN 1476-4660. doi: 10.1038/nmat3568. URL https://www.nature. com/articles/nmat3568. Davies, D. W., Butler, K. T., Jackson, A. J., Skelton, J. M., Morita, K., and Walsh, A. SMACT: Semiconducting Materials by Analogy and Chemical Theory. Journal of Open Source Software, 4(38):1361, June 2019. ISSN 24759066. doi: 10.21105/joss.01361. URL https://joss. theoj.org/papers/10.21105/joss.01361. Fuemmeler, E., Wolfe, G., Gupta, A., Vita, J. A., Tadmor, E. B., and Martiniani, S. Advancing the colabfit exchange towards a web-scale data source for machine learning interatomic potentials. In AI for Accelerated Materials Design-Neur IPS 2024, 2024. Ganose, A. M., Sahasrabuddhe, H., Asta, M., Beck, K., Biswas, T., Bonkowski, A., Bustamante, J., Chen, X., Chiang, Y., Chrzan, D., Clary, J., Cohen, O., Ertural, C., Gallant, M., George, J., Gerits, S., Goodall, R., Guha, R., Hautier, G., Horton, M., Kaplan, A., Kingsbury, R., Kuner, M., Li, B., Linn, X., Mc Dermott, M., Mohanakrishnan, R. S., Naik, A., Neaton, J., Persson, K., Petretto, G., Purcell, T., Ricci, F., Rich, B., Riebesell, J., Rignanese, G.-M., Rosen, A., Scheffler, M., Schmidt, J., Shen, J.-X., Sobolev, A., Sundararaman, R., Tezak, C., Trinquet, V., Varley, J., Vigil-Fowler, D., Wang, D., Waroquiers, D., Wen, M., Yang, H., Zheng, H., Zheng, J., Zhu, Z., and Jain, A. Atomate2: Modular workflows for materials science. Chem Rxiv, 2025. URL https://chemrxiv. org/engage/chemrxiv/article-details/ 678e76a16dde43c9085c75e9. Gat, I., Remez, T., Shaul, N., Kreuk, F., Chen, R. T. Q., Synnaeve, G., Adi, Y., and Lipman, Y. Discrete Flow Matching, July 2024. URL http://arxiv.org/ abs/2407.15595. George, E. P., Raabe, D., and Ritchie, R. O. Highentropy alloys. Nat. Rev. Mater., 4(8):515 534, August 2019. ISSN 2058-8437. doi: 10.1038/ s41578-019-0121-4. URL https://www.nature. com/articles/s41578-019-0121-4. Gludovatz, B., Hohenwarter, A., Catoor, D., Chang, E. H., George, E. P., and Ritchie, R. O. A fracture-resistant high-entropy alloy for cryogenic applications. Science, 345(6201):1153 1158, September 2014. doi: 10.1126/ science.1254581. URL https://www.science. org/doi/10.1126/science.1254581. Open Materials Generation with Stochastic Interpolants Gludovatz, B., Hohenwarter, A., Thurston, K. V. S., Bei, H., Wu, Z., George, E. P., and Ritchie, R. O. Exceptional damage-tolerance of a medium-entropy alloy Cr Co Ni at cryogenic temperatures. Nat. Commun., 7(1): 10602, February 2016. ISSN 2041-1723. doi: 10.1038/ ncomms10602. URL https://www.nature.com/ articles/ncomms10602. Gruver, N., Sriram, A., Madotto, A., Wilson, A. G., Zitnick, C. L., and Ulissi, Z. Fine-Tuned Language Models Generate Stable Inorganic Materials as Text, February 2024. URL http://arxiv.org/abs/2402.04379. Ho, J., Jain, A., and Abbeel, P. Denoising Diffusion Probabilistic Models, December 2020. Isaacs, E. B. and Marianetti, C. A. Compositional phase stability of correlated electron materials within DFT+DMFT. Phys. Rev. B, 102: 045146, Jul 2020. doi: 10.1103/Phys Rev B.102. 045146. URL https://link.aps.org/doi/10. 1103/Phys Rev B.102.045146. Jain, A., Hautier, G., Moore, C., Ong, S., Fischer, C., Mueller, T., Persson, K., and Ceder, G. A high-throughput infrastructure for density functional theory calculations. Comput. Mater. Sci., 50:2295 2310, June 2011. doi: 10.1016/j.commatsci.2011.02.023. Jain, A., Ong, S. P., Hautier, G., Chen, W., Richards, W. D., Dacek, S., Cholia, S., Gunter, D., Skinner, D., Ceder, G., and Persson, K. A. Commentary: The Materials Project: A materials genome approach to accelerating materials innovation. APL Materials, 1(1):011002, July 2013. doi: 10.1063/1.4812323. URL https://aip. scitation.org/doi/10.1063/1.4812323. Jiao, R., Huang, W., Lin, P., Han, J., Chen, P., Lu, Y., and Liu, Y. Crystal Structure Prediction by Joint Equivariant Diffusion, July 2023. URL https://arxiv.org/ abs/2309.04475v2. Jiao, R., Huang, W., Liu, Y., Zhao, D., and Liu, Y. Space Group Constrained Crystal Generation, April 2024. URL http://arxiv.org/abs/2402.03992. Kazeev, N., Zhu, R., Romanov, I., Ustyuzhanin, A. E., Yamazaki, S., Nong, W., and Hippalgaonkar, K. Wyckoff Transformer: Generation of Symmetric Crystals. In AI for Accelerated Materials Design - Neur IPS 2024, November 2024. URL https://openreview.net/forum? id=Jcy1b POqr Y. Liaw, R., Liang, E., Nishihara, R., Moritz, P., Gonzalez, J. E., and Stoica, I. Tune: A research platform for distributed model selection and training. ar Xiv preprint ar Xiv:1807.05118, 2018. Lipman, Y., Chen, R. T. Q., Ben-Hamu, H., Nickel, M., and Le, M. Flow Matching for Generative Modeling, February 2023. URL http://arxiv.org/abs/2210. 02747. Liu, C., Li, F., Ma, L.-P., and Cheng, H.-M. Advanced Materials for Energy Storage. Advanced Materials, 22(8):E28 E62, 2010. ISSN 15214095. doi: 10.1002/adma.200903328. URL https://onlinelibrary.wiley.com/doi/ abs/10.1002/adma.200903328. Liu, Q. Rectified Flow: A Marginal Preserving Approach to Optimal Transport, September 2022. Maier, W. F. Early Years of High-Throughput Experimentation and Combinatorial Approaches in Catalysis and Materials Science. ACS Comb. Sci., 21(6):437 444, June 2019. ISSN 2156-8952. doi: 10.1021/ acscombsci.8b00189. URL https://doi.org/10. 1021/acscombsci.8b00189. Matter Gen. Mattergen. Microsoft, January 2025. URL https://github.com/microsoft/ mattergen. Mehl, M. J., Hicks, D., Toher, C., Levy, O., Hanson, R. M., Hart, G., and Curtarolo, S. The AFLOW Library of Crystallographic Prototypes: Part 1. Comput. Mater. Sci., 136:S1 S828, August 2017. ISSN 0927-0256. doi: 10.1016/j.commatsci.2017.01. 017. URL https://www.sciencedirect.com/ science/article/pii/S0927025617300241. Merchant, A., Batzner, S., Schoenholz, S. S., Aykol, M., Cheon, G., and Cubuk, E. D. Scaling deep learning for materials discovery. Nature, 624:80 85, November 2023. ISSN 1476-4687. doi: 10.1038/ s41586-023-06735-9. URL https://www.nature. com/articles/s41586-023-06735-9. Miller, B. K., Chen, R. T. Q., Sriram, A., and Wood, B. M. Flow MM: Generating Materials with Riemannian Flow Matching, June 2024. URL http://arxiv.org/ abs/2406.04713. Nakaya, Y. and Furukawa, S. Catalysis of Alloys: Classification, Principles, and Design for a Variety of Materials and Reactions. Chem. Rev., 123(9):5859 5947, May 2023. ISSN 0009-2665. doi: 10.1021/acs.chemrev. 2c00356. URL https://doi.org/10.1021/acs. chemrev.2c00356. Nichol, A. Q. and Dhariwal, P. Improved denoising diffusion probabilistic models. In Meila, M. and Zhang, T. (eds.), Proceedings of the 38th International Conference on Machine Learning, volume 139 of Proceedings Open Materials Generation with Stochastic Interpolants of Machine Learning Research, pp. 8162 8171. PMLR, 18 24 Jul 2021. URL https://proceedings.mlr. press/v139/nichol21a.html. Oganov, A. R., Pickard, C. J., Zhu, Q., and Needs, R. J. Structure prediction drives materials discovery. Nat Rev Mater, 4(5):331 348, May 2019. ISSN 2058-8437. doi: 10.1038/ s41578-019-0101-8. URL https://www.nature. com/articles/s41578-019-0101-8. Ong, S. P., Richards, W. D., Jain, A., Hautier, G., Kocher, M., Cholia, S., Gunter, D., Chevrier, V. L., Persson, K. A., and Ceder, G. Python Materials Genomics (pymatgen): A robust, open-source python library for materials analysis. Computational Materials Science, 68:314 319, February 2013. ISSN 0927-0256. doi: 10.1016/j.commatsci.2012.10. 028. URL https://www.sciencedirect.com/ science/article/pii/S0927025612006295. Pickard, C. J. Airss data for carbon at 10gpa and the c+n+h+o system at 1gpa, 2020. URL https://archive.materialscloud.org/ record/2020.0026/v1. Pickard, C. J. and Needs, R. J. Ab Initio random structure searching. J. Phys.: Condens. Matter, 23 (5):053201, February 2011. ISSN 0953-8984, 1361648X. doi: 10.1088/0953-8984/23/5/053201. URL https://iopscience.iop.org/article/10. 1088/0953-8984/23/5/053201. Potyrailo, R., Rajan, K., Stoewe, K., Takeuchi, I., Chisholm, B., and Lam, H. Combinatorial and High-Throughput Screening of Materials Libraries: Review of State of the Art. ACS Comb. Sci., 13(6):579 633, November 2011. ISSN 2156-8952. doi: 10.1021/co200007w. URL https://doi.org/10.1021/co200007w. Saharia, C., Chan, W., Saxena, S., Li, L., Whang, J., Denton, E., Ghasemipour, S. K. S., Ayan, B. K., Mahdavi, S. S., Lopes, R. G., Salimans, T., Ho, J., Fleet, D. J., and Norouzi, M. Photorealistic Text-to-Image Diffusion Models with Deep Language Understanding, May 2022. Satorras, V. G., Hoogeboom, E., and Welling, M. E(n) equivariant graph neural networks. In Meila, M. and Zhang, T. (eds.), Proceedings of the 38th International Conference on Machine Learning, volume 139 of Proceedings of Machine Learning Research, pp. 9323 9332. PMLR, 18 24 Jul 2021. URL https://proceedings.mlr. press/v139/satorras21a.html. Schmidt, J., Hoffmann, N., Wang, H.-C., Borlido, P., Carric o, P. J. M. A., Cerqueira, T. F. T., Botti, S., and Marques, M. A. L. Large-scale machine-learning-assisted exploration of the whole materials space, October 2022a. URL http://arxiv.org/abs/2210.00579. Schmidt, J., Wang, H.-C., Cerqueira, T. F. T., Botti, S., and Marques, M. A. L. A dataset of 175k stable and metastable materials calculated with the PBEsol and SCAN functionals. Scientific Data, 9(1):64, March 2022b. ISSN 2052-4463. doi: 10.1038/s41597-022-01177-w. Snyder, G. J. and Toberer, E. S. Complex thermoelectric materials. Nature Mater, 7(2):105 114, February 2008. ISSN 1476-4660. doi: 10.1038/nmat2090. URL https: //www.nature.com/articles/nmat2090. Sohl-Dickstein, J., Weiss, E. A., Maheswaranathan, N., and Ganguli, S. Deep Unsupervised Learning using Nonequilibrium Thermodynamics, November 2015. URL http://arxiv.org/abs/1503.03585. Song, Y. and Ermon, S. Generative modeling by estimating gradients of the data distribution. In Wallach, H., Larochelle, H., Beygelzimer, A., d'Alch e-Buc, F., Fox, E., and Garnett, R. (eds.), Advances in Neural Information Processing Systems, volume 32. Curran Associates, Inc., 2019. URL https://proceedings.neurips. cc/paper_files/paper/2019/file/ 3001ef257407d5a371a96dcd947c7d93-Paper. pdf. Song, Y., Sohl-Dickstein, J., Kingma, D. P., Kumar, A., Ermon, S., and Poole, B. Score-Based Generative Modeling through Stochastic Differential Equations, February 2021. URL http://arxiv.org/abs/2011.13456. Sriram, A., Miller, B., Chen, R. T., and Wood, B. Flowllm: Flow matching for material generation with large language models as base distributions. Advances in Neural Information Processing Systems, 37:46025 46046, 2024. Strmcnik, D., Lopes, P. P., Genorio, B., Stamenkovic, V. R., and Markovic, N. M. Design principles for hydrogen evolution reaction catalyst materials. Nano Energy, 29:29 36, November 2016. ISSN 2211-2855. doi: 10.1016/j.nanoen.2016.04. 017. URL https://www.sciencedirect.com/ science/article/pii/S2211285516300738. Tipton, W. W. and Hennig, R. G. A grand canonical genetic algorithm for the prediction of multi-component phase diagrams and testing of empirical potentials. Journal of Physics: Condensed Matter, 25(49):495401, November 2013. ISSN 0953-8984. doi: 10.1088/0953-8984/25/49/ 495401. URL https://dx.doi.org/10.1088/ 0953-8984/25/49/495401. Togo, A., Shinohara, K., and Tanaka, I. Spglib: A software library for crystal symmetry search, March 2024. URL http://arxiv.org/abs/1808.01590. Open Materials Generation with Stochastic Interpolants Tong, A., Fatras, K., Malkin, N., Huguet, G., Zhang, Y., Rector-Brooks, J., Wolf, G., and Bengio, Y. Improving and generalizing flow-based generative models with minibatch optimal transport, March 2024. URL http://arxiv.org/abs/2302.00482. Vita, J. A., Fuemmeler, E. G., Gupta, A., Wolfe, G. P., Tao, A. Q., Elliott, R. S., Martiniani, S., and Tadmor, E. B. Colab Fit exchange: Open-access datasets for data-driven interatomic potentials. J. Chem. Phys., 159(15):154802, October 2023. ISSN 0021-9606. doi: 10.1063/5.0163882. URL https://doi.org/10.1063/5.0163882. Xie, T., Fu, X., Ganea, O.-E., Barzilay, R., and Jaakkola, T. Crystal Diffusion Variational Autoencoder for Periodic Material Generation, March 2022. Yang, H., Hu, C., Zhou, Y., Liu, X., Shi, Y., Li, J., Li, G., Chen, Z., Chen, S., Zeni, C., Horton, M., Pinsler, R., Fowler, A., Z ugner, D., Xie, T., Smith, J., Sun, L., Wang, Q., Kong, L., Liu, C., Hao, H., and Lu, Z. Matter Sim: A Deep Learning Atomistic Model Across Elements, Temperatures and Pressures, May 2024. URL http://arxiv.org/abs/2405.04967. Yim, J., Campbell, A., Foong, A. Y. K., Gastegger, M., Jim enez-Luna, J., Lewis, S., Satorras, V. G., Veeling, B. S., Barzilay, R., Jaakkola, T., and No e, F. Fast protein backbone generation with SE(3) flow matching, October 2023. URL http://arxiv.org/abs/2310. 05297. Zaki, N., Park, H., Osgood, R. M., Millis, A. J., and Marianetti, C. A. Failure of dft-based computations for a stepped-substrate-supported correlated co wire. Phys. Rev. B, 89:205427, May 2014. doi: 10.1103/Phys Rev B. 89.205427. URL https://link.aps.org/doi/ 10.1103/Phys Rev B.89.205427. Zeni, C., Pinsler, R., Z ugner, D., Fowler, A., Horton, M., Fu, X., Wang, Z., Shysheya, A., Crabb e, J., Ueda, S., Sordillo, R., Sun, L., Smith, J., Nguyen, B., Schulz, H., Lewis, S., Huang, C.-W., Lu, Z., Zhou, Y., Yang, H., Hao, H., Li, J., Yang, C., Li, W., Tomioka, R., and Xie, T. A generative model for inorganic materials design. Nature, January 2025. ISSN 0028-0836, 1476-4687. doi: 10.1038/s41586-025-08628-5. Zhu, R., Nong, W., Yamazaki, S., and Hippalgaonkar, K. Wy Cryst: Wyckoff Inorganic Crystal Generator Framework, March 2024. URL http://arxiv.org/abs/ 2311.17916. Open Materials Generation with Stochastic Interpolants A. Acronyms Table 4 provides a list of acronyms used throughout this paper for reference. Table 4. Acronym definitions. Acronym Full Name AIRSS Ab initio Random Structure Searching CDVAE Crystal Diffusion Variational Autoencoder CFP Chemical Formula Prediction CFM Conditional Flow Matching CTMC Continuous-Time Markov Chain CSP Crystal Structure Prediction DFT Density Functional Theory DFM Discrete Flow Matching DNG De Novo Generation EGNN Equivariant Graph Neural Network GNo ME Graph Networks for Materials Exploration MLIP Machine-Learned Interatomic Potential Nequ IP Neural Equivariant Interatomic Potentials ODE Ordinary Differential Equation OMat G Open Materials Generation RFM Riemannian Flow Matching SBD Score-Based Diffusion SBDM Score-Based Diffusion Model SDE Stochastic Differential Equation SI Stochastic Interpolant SMACT Semiconducting Materials from Analogy and Chemical Theory SUN Stable, Unique, and Novel B. Implementation Details of Stochastic Interpolants B.1. OMat G Framework Figures 4 and 5 summarize the training and the integration pipeline of the OMat G framework, respectively. Depending on the specific task, there are several stochastic interpolants at once. For CSP, one stochastic interpolant considers lattice vectors L, and another one considers fractional coordinates X. The model output of CSPNet (see Appendix C.1) depends on the full structural representation {A, X, L} and time t, where A are the atomic species. For the DNG task, we additionally use discrete flow matching for the atomic species A (Campbell et al., 2024). During the numerical integration in the CSP task, X and L are integrated jointly while A is fixed. For DNG, A is evolved according to discrete flow matching (Campbell et al., 2024) (see Appendix B.6). For the SDE sampling scheme in Fig. 5, one chooses a time-dependent noise ε(t) that only appears during integration and not during training (see Appendix B.4). Also, γ(t) has to be unequal zero in order to prevent the divergence in 1/γ(t). However, since γ(t) necessarily vanishes at times t = 0 and t = 1 (see Appendix B.2), one should choose a time-varying ε(t) that vanishes near these endpoints (see Appendix B.4) (Albergo et al., 2023). B.2. Interpolant Choice In this work, we are concerned with spatially linear interpolants of the form specified in Eq. (1). The following conditions must be met (Albergo et al., 2023): α(0) = β(1) = 1, α(1) = β(0) = γ(0) = γ(1) = 0, γ(t) > 0 t (0, 1). (5) Under these constraints, the form of the SI is relatively flexible, and many different interpolants can be defined. Also, the base distribution can be arbitrary as in CFM (Liu, 2022; Albergo & Vanden-Eijnden, 2023; Tong et al., 2024). In this work, Open Materials Generation with Stochastic Interpolants Dataset (Perov-5, MP-20, MPTS-52, Alex-MP) For structure Stochastic Interpolants: MSE Loss between and Update batch loss Minimization step on C over CSPNet Time and initial structure Draw batch After computation of batch loss Figure 4. Training pipeline of the OMat G framework: A batch of structures is drawn from a dataset with target distribution ρ1. Every structure x1 ρ1 is connected with a structure x0 from the base distribution ρ0 with stochastic interpolants that yield the interpolated structure xt = x(t, x0, x1, z) and the drift bt = txt at time t U(0, 1), possibly using a random variable z N(0, I). The model CSPNet predicts bθ t = bθ(t, xt) and zθ t = z(t, xt) and its parameters are minimized based on the MSE losses in Eqs (2) and (3) [see also Eq. (32)]. Numerical integration ODE: Initial structure Final structure Figure 5. Numerical integration pipeline of the OMat G framework: An initial structure x0 from the base distribution ρ0 is numerically integrated following either an ODE or an SDE based on the model predictions bθ t and zθ t . For an SDE, one can choose a noise ε(t) during integration. Open Materials Generation with Stochastic Interpolants Table 5. SI parameters from Albergo et al. (2023). Stochastic Interpolant α(t) β(t) γ(t) linear 1 t t p Arbitrary ρ0 trig cos π enc-dec cos2(πt)1[0, 1 2 )(t) cos2(πt)1( 1 2 ,1](t) sin2(πt) Gaussian ρ0 VP SBD we mostly rely on interpolants originally defined in Albergo et al. (2023) and listed in Tab. 5. For the VP SBD interpolant, we allow for different widths σ0 of the Gaussian base distribution ρ0. In Appendix B.9, we introduce additional VP and VE SBD interpolants beyond the one in Tab. 5 that are implemented in OMat G. The encoder-decoder (enc-dec) interpolant as defined in Tab. 5 evolves samples from the base distribution ρ0 to follow an intermediate Gaussian distribution with variance 1 at the switch time Tswitch = 0.5, before mapping them to a sample from ρ1. This can be generalized to arbitrary variances a > 0 and switch times Tswitch (0, 1): α(t) = cos2 π(t Tswitcht)p (Tswitch Tswitcht)p + (t Tswitcht)p 1[0,Tswitch)(t), β(t) = cos2 π(t Tswitcht)p (Tswitch Tswitcht)p + (t Tswitcht)p 1(Tswitch,1](t), γ(t) = a sin2 π(t Tswitcht)p (Tswitch Tswitcht)p + (t Tswitcht)p where p 1/2. We consider the cases p {1/2, 1} and note that the general interpolant in Eq. (6) reduces to the interpolant in Tab. 5 for a = 1, p = 1, and Tswitch = 0.5. B.3. Antithetic Sampling As shown by Albergo et al. (2023), the loss function can become unstable around t = 0 and t = 1 for certain choices of γ(t). To account for this, we implement antithetic sampling. This requires simultaneously computing the loss at both x+ and x where x+(t, x0, x1, z) = α(t)x0 + β(t)x1 + γ(t)z, (7) x (t, x0, x1, z) = α(t)x0 + β(t)x1 γ(t)z. (8) Both losses are computed using the same value of z and subsequently averaged. B.4. Diffusion Coefficient An important inference-time parameter for models integrated with an SDE is the choice of ϵ(t) 0 which plays the role of a diffusion coefficient. Albergo et al. (2023) note that the presence of γ 1(t) in the drift term seen in Fig. 5 can pose a numerical instability at the endpoints t = 0 and t = 1 during integration. For the choice ϵconst(t) = c, they consider integrating from some nonzero time t 0 to t 1 in order to avoid the singularity. Alternatively, one can design a form for ϵ(t) such that it vanishes at these endpoints. In OMat G, we opt for the latter approach and consider a diffusion coefficient, ϵvanish(t), which vanishes at the endpoints ϵvanish(t) = c 1 + e t µ σ 1 + e 1 µ t Here, c dictates the magnitude of the diffusion, µ sets the times at which the midpoints between ϵ(t) = 0 and ϵ(t) = c are reached, and σ controls the rate of this increase from ϵ(t) = 0 to ϵ(t) = c. The only constraints on these parameters are that c 0, µ > 0, and σ > 0. Importantly, these parameters should be chosen such that they are near zero at the endpoints. Open Materials Generation with Stochastic Interpolants B.5. Interpolation with Periodic Boundary Conditions We adopt a task-specific formulation for handling periodic boundary conditions with SIs tailored to flat tori, which are the relevant manifolds for fractional coordinates in crystal generation. We do not attempt to generalize stochastic interpolants (SIs) to arbitrary manifolds as in Riemannian flow matching (Chen & Lipman, 2024). As in Flow MM (Miller et al., 2024), in order to uniquely define the interpolating paths, we rely on shortest geodesic interpolation paths between pairs of fractional coordinates from x0 and x1, ensuring that interpolants are well-defined and differentiable. As noted in Section 3.2.1, this shortest geodesic path is computed by first unwrapping one of the coordinates (say x1) into its periodic image x 1, such that it is the closest image to x0. We then compute, for example, the linear interpolant x(t, x0, x 1) = (1 t)x0 + tx 1, as if in Euclidean space and finally wrap the interpolated path back onto the torus. (The geodesic is the same as the linear interpolant wrapped back into the box.) We perform this procedure for all choices of interpolants. The reason for unwrapping according to the closest image first for all interpolants is because there are multiple ways to connect two points on a torus (e.g., in a periodic box one can connect two points with or without crossing the box boundaries). All periodic stochastic interpolants are then defined this way, by computing x(t, x0, x 1, z) = α(t)x0 + β(t)x 1 + γ(t)z in the unwrapped (Euclidean) space and wrapping back onto the torus. We emphasize that this procedure is important not only for the choice of interpolant, but also for the addition of the latent variable γ(t)z which also moves the interpolation trajectory away from the geodesic. Our process yields exactly the same shortest-path geodesic as in Flow MM if using the linear interpolant, and thus recovers its corresponding conditional flow-matching loss. We depict our implementation of periodic stochastic interpolants in Fig. 6. We also demonstrate in Fig. 6c that averaging over the latent variable γ(t)z recovers the deterministic base interpolant path, as required by the SI framework. Figure 6. Extending interpolants to incorporate periodic boundary conditions. (a b) The path for a score based diffusion interpolant is calculated by first computing the shortest-path geodesic (blue) between the initial (green dot) and final positions (red dot). Next, the path of the interpolant moving the final position outside the bounding box is computed (green), and finally the path is wrapped back into the bounding box to produce the interpolant trajectory (orange). (c) The effect of adding a latent variable to any interpolant must be handled similarly to calculating the path of a non-linear interpolant. For a linear interpolant with a nonzero γ, we show samples of possible paths (blue) and their averaged path (orange) which collapses onto the path of the linear interpolant. B.6. DFM Details DFM allows for generative modeling of discrete sequences of tokens while respecting the discrete nature of the design space. As discussed, a parameterized neural network pθ 1|t(x1|xt) is learned, which attempts to predict the final sequence from the sequence at time t. Borrowing from Campbell et al. (2024), we choose a conditional rate matrix Rt(xt, i|x1) giving the rate of xt jumping to a different state i given x1 which generates the conditional flow pt|1(xt|x1) of the form: Rt(xt, i|x1) = Re LU tpt|1(i|x1) tpt|1(xt|x1) S pt|1(xt|x1) , (10) where S is the number of possible tokens a sequence element can take on. This conditional rate matrix can be modified by including a term that introduces stochasticity in the form of a detailed balance rate matrix RDB t by writing Rη t = Rt+ηRDB t . Open Materials Generation with Stochastic Interpolants Here, (Campbell et al., 2024): RDB t (i, j|x1) = ηδ{i, x1}δ{j, M} + ηt 1 tδ{i, M}δ{j, x1}, (11) where M is the masking token. The parameter η R+ represents the level of stochasticity that only appears during generation. During generation, our objective is to compute Rθ t (xt, i) based on the learned distribution pθ 1|t(x1|xt). Formally, we have Rθ t (xt, i) = Epθ 1|t(x1|xt) Rη t (xt, i|x1) (12) In practice, Campbell et al. (2024) show that we need not compute a full expectation, but rather, simply draw x1 pθ 1|t(x1|xt), evaluate the conditional rate matrix Rη t (xt, i|x1), and perform an update of xt to xt+ t with discrete time step t directly from this by sampling xt+ t according to pt+ t|t(xt+ t|x1, xt) = δ{xt, xt+ t} + Rη t (xt, i|x1) t. (13) B.7. Velocity Annealing Velocity annealing rescaling the learned velocity field during generation to increase velocity over time as bθ(t, x) (1 + st) bθ(t, x) with s as an hyperparameter during integration has been empirically shown to improve performance in a number of studies that apply CFM to physical systems (Yim et al., 2023; Bose et al., 2024; Miller et al., 2024). For instance, Miller et al. (2024) demonstrated that applying velocity annealing significantly improves performance in CSP and DNG benchmarks for materials. Motivated by these findings, we include velocity annealing in OMat G as a tunable hyperparameter, while emphasizing that this technique lacks a formal theoretical justification within the mathematical frameworks underlying flow models and stochastic interpolants. B.8. Data-Dependent Coupling SIs have been used with data-dependent couplings (Albergo et al., 2024), where a coupling function ν(x0, x1) enables biasing of x0 based on the sampled x1. In OMat G, we incorporate an optional data-dependent coupling that enforces an ordering (i.e., a permutation on the order of atomic elements within a structure) that produces the minimum fractionalcoordinate distance between each particle pair (xi 0, xi 1) from structures x0 ρ0 and x1 ρ1. We find that the inclusion of this data-dependent coupling is optimal during hyperparameter tuning depending on the type of model: CSP models typically performed better without this coupling, but DNG models (see Tab. 11) can benefit in certain cases from minimizing traveled distance via permutation of elements. Formally, our coupling is conditional on the sampled (x0, x1) and is defined as i d(p(xi 0), xi 1). (14) Here, d( , ) is a distance metric which we define on a periodic manifold in fractional-coordinate space (i.e., a fourdimensional torus) and p is some permutation function that permutes the discrete indices i. Under this coupling, we still sample (x0, x1) independently but then bias the sampled x0 to travel the minimum permutational distance necessary to reach the target structure. B.9. SI unifies CFM and SBDM The SI framework implemented in OMat G unifies the frameworks of CFM, as implemented in Flow MM (Miller et al., 2024), and SBDM, as implemented in Diff CSP (Jiao et al., 2023) and Matter Gen (Zeni et al., 2025). Flow MM is naturally subsumed by OMat G. For the choice of ODE-based sampling, the velocity term bθ(t, x) is learned by minimizing the loss function in Eq. (2). By using γ(t) = 0 in the linear interpolant x(t, x0, x1) = (1 t)x0 + tx1 (see Appendix B.2), Eq. (2) becomes identical to the Flow MM loss (see Eq. (15) in Miller et al. (2024)). Furthermore, the treatment of periodic boundary conditions for the linear interpolant (see Appendix B.5) leads to the same geodesic paths as in Flow MM, and the center-of-mass motion of the ground-truth velocity is removed similarly in both frameworks. Open Materials Generation with Stochastic Interpolants The connection between SIs and SBDM requires the discussion of both variance-preserving (VP) and variance-exploding (VE) cases. In the VP case (Sohl-Dickstein et al., 2015; Ho et al., 2020), a sequence of N noise increments with variance βi perturbs data y0 as yi = p 1 βi yi 1 + p βi zi 1, i = 1, . . . , N, (15) where zi N(0, I). As N , this converges to the SDE 2β(s)ys dt + p β(s) dws, (16) where ws is the standard Wiener process (Song et al., 2021). Under this stochastic process, the data distribution at time s = 0 is transformed into a Gaussian base distribution as s . This differs from the time convention in SI where samples from the base distribution at time t = 0 are transported to samples from the data distribution at t = 1. With a corresponding change of variables s(t) = log(t), Aranguri et al. (2025) show within the SI framework that ys is equal in law to the one-sided interpolant x(t, x0, x1) = p 1 τ 2(t) x0 + τ(t)x1, (17) In accordance with SBDM, the base distribution ρ0 is Gaussian, that is, x0 N(0, σ2 0I). Here, we introduced the width σ0 of the base distribution as a tunable hyperparameter. The SDE in Eq. (16) and the variance-schedule β(s) are often considered on the time interval s [0, 1], which implies that τ(t) from Eq. (18) is only used in t [1/e, 1] (Aranguri et al., 2025). OMat G implements three such schedules. The linear schedule (Ho et al., 2020; Aranguri et al., 2025) is βlin(s) = βmin + s(βmax βmin) where βmin and βmax are chosen empirically. This implies τ lin(t) = exp 1 2βmin log(t) 1 4(βmax βmin) log2(t) . (19) This function is well-behaved for the entire time range t [0, 1] and implemented as such in OMat G. For the cosine schedule, which is used in Diff CSP for the lattice vectors, the noise variance in Eq. (15) is given by βi = 1 αi/ αi 1 with f(0), f(i) = cos2 π where d is a small constant offset (Nichol & Dhariwal, 2021). As N , one gets for s [0, 1] βcos(s) = d cos2 π 2 s+d 1+d cos2 π 2 d 1+d = π 1 + d tan π 2 s + d 1 + d For t [1/e, 1], this leads to τ cos(t) = csc π 2 + 2d sin π + π log(t) and for t [0, 1/e), we use τ cos(t) = 0. The schedule τ const(t) = t on t [0, 1] corresponds to a constant schedule βconst(s) = 2 and yields the SBD interpolant of Appendix B.2 derived in Albergo et al. (2023). This is the VP SBD interpolant considered throughout this paper. In the VE case (Song & Ermon, 2019), a sequence of N noise increments with variance σi perturbs data y0 as yi = yi 1 + q σ2 i σ2 i 1 zi 1. (23) As N , this converges to the SDE (Song et al., 2021) ds dws. (24) Open Materials Generation with Stochastic Interpolants The corresponding one-sided interpolant in the SI framework is given by σ2(1 t) σ2(0) x0 + x1, (25) where again x0 N(0, σ2 0I) (Aranguri et al., 2025). As in Diff CSP for the fractional coordinates, the schedule σ(s) on s [0, 1] is typically given by σ(s) = σmin(σmax/σmin)s which we implement in OMat G. The parameters σmin and σmax are optimizeable hyperparameters. The reported match rate and RMSE of the CSP task for the VE SBD positional interpolant for the MP-20 dataset in Tab. 1 are close to the ones of Diff CSP. This highlights that OMat G is able to reproduce similar conditions to those in Diff CSP. B.10. Comparison of OMat G-Linear to Flow MM A subset of OMat G models, specifically those which use linear interpolants for both the fractional coordinates and lattice vectors, map closely onto the conditional flow-matching model Flow MM (Miller et al., 2024). The notable differences between OMat G-Linear models and Flow MM are as follows: (1) Discrete flow matching on species for OMat G vs. analog bits for Flow MM. (2) Lattice matrix representation for OMat G vs. lattice parameter representation (lengths and angles) for Flow MM. (3) Original CSPNet encoder for OMat G vs. slightly modified CSPNet for Flow MM. OMat G s CSP results improve upon Flow MM s. Since the CSP task does not utilize any species learning, the different handling of species is not sufficient to fully explain the differences in model performance for CSP. For DNG models, the handling of species is also a relevant difference between OMat G-Linear and Flow MM. C. Model Architecture C.1. Graph Neural Network We implement a message-passing graph neural network (GNN) with CSPNet as introduced in Jiao et al. (2023): hi (0) = ϕh(0)(ai) (26) mij (s) = φm hi s 1, hj s 1, l, Sinusoidal Embedding(xj xi) (27) j=1 mij (s) (28) hi (s) = hi (s 1) + φh(h(s 1), mi (s)) (29) bx = φx hi (max s) (30) i=1 hi (max s) Here, node embeddings hj (s) of node j at layer s are initialized as a function of the atom types, a. Embeddings are then updated by a message passing scheme through a series of graph convolution layers. Messages are computed with a parameterized neural network, φm, from neighboring node embeddings as well as information about the lattice, l, and distance between the fractional coordinates x. All necessary drift and denoiser terms are computed from single layer MLPs applied to the final node embeddings. For the CFP model that should only predict compositions, we simply remove the input of the lattice l and the fractional coordinates x from the computation of the message in Eq. (27). This ensures that the output pθ 1|t(a1|xt) of CSPNet for the composition does not depend on lattice vectors or fractional coordinates, while preserving permutational equivariance. Open Materials Generation with Stochastic Interpolants C.2. Loss Function With Eqs (2), (3), and (4), we can construct a loss function for the modeling of our joint distribution of interest for the DNG task, L(θ) =Et,z,x0,x1 λx,b |bθ x(t, xt)|2 2 tx(t, x0, x1, z) bθ x(t, xt) + λx,z |zθ x(t, xt)|2 2zθ x(t, xt) z + λl,b |bθ l (t, xt)|2 2 tx(t, x0, x1, z) bθ l (t, xt) + λl,z |zθ l (t, xt)|2 2zθ l (t, xt) z + λa h log pθ 1|t(a1|xt) i . For the CSP task, the last line is left out. The λ terms correspond to the relative weights of each term in the loss function. These weighting factors are hyperparameters that are included in our hyperparameter sweep. The respective terms for the fractional coordinates and lattice vectors corresponding to Eqs (2) and (3) are equivalent to a mean-squared error (MSE) loss function as, for instance, LMSE b (θ) = Et,z,x0,x1 |bθ(t, xt) tx(t, x0, x1, z)|2 (33) for the velocity term. They only differ by a constant term that does not influence gradients. We do not include that constant term because the possible divergence of tγ(t) near t = 0 and t = 1 can artificially inflate the absolute value of the loss, even when antithetic sampling is applied (see Section B.3). C.3. Hyperparameter Optimization For every choice of the positional interpolant, sampling scheme, and latent variable γ, an independent hyperparameter optimization was performed using the Ray Tune package (Liaw et al., 2018) in conjunction with the Hyper Opt Python library (Bergstra et al., 2013) for Bayesian optimization. The tuned hyperparameters include both those relevant during training the relative loss weights λ, the choice of stochastic interpolant for the lattice vectors, the parameters for chosen γ(t) (if necessary), the sampling scheme, the usage of data-dependent coupling, the batch size, and the learning rate and during inference the number of integration steps, the choice of the noises ε(t) and η, and the magnitude of the velocity annealing parameter s for both lattice vectors and atomic coordinates. Hyperparameters are sampled according to the distributions below6: Number of integration timesteps Uniform(100, 1000). Batch size Choice(32, 64, 128, 256, 512, 1024). Min. permutational distance data coupling Choice(True, False). Relative weigths λx,b, λx,z, λa Log Uniform(0.1, 2000.0). Relative weight λl,z Log Uniform(0.1, 100.0). Niggli reduction of cell during training Choice(True, False). DFM Stochastictity Uniform(0, 50.0). Learning rate Log Uniform(10 5, 10 2). Weight decay Log Uniform(10 5, 10 3). Velocity annealing coefficient (both for x and l) Uniform(0.0, 15.0). Diffusion coefficient parameter c (both for x and l) Uniform(0.1, 10.0). Diffusion coefficient parameter µ (both for x and l) Uniform(0.05, 0.3). Diffusion coefficient parameter σ (both for x and l) Uniform(0.005, 0.05). 6Relative loss weights for a are only swept over for DNG. Otherwise only weight parameters for x and l are optimized. Relative loss weights for the denoiser are only included when SDE integration is used. Open Materials Generation with Stochastic Interpolants Parameter Tswitch of encoder-decoder interpolant Uniform(0.1, 0.9). Parameter p of encoder-decoder interpolant Choice(0.5, 1.0). Parameter a of γ(t) functions Log Uniform(0.01, 10.0). Standard deviation σ0 of Gaussian ρ0 for SBD interpolants Log Uniform(0.01, 10.0). Parameter σmin of the σ(s) schedule of the VE SBD interpolant Uniform(0.001, 0.01). Parameter σmax of the σ(s) schedule of the VE SBD interpolant Uniform(0.1, 1.0). The relative loss weight for the velocity of the lattice vectors is fixed: λl,b = 1. By ensuring that the sum of all relevant loss weights λ is one, one can transform the relative weights λ to the weights λ in Eq. (32). For the CSP models, the hyperparameter optimization attempts to maximize the match rate. For the DNG models, we combine the metrics in Tab. 2 to a single evaluation metric eval DNG that is supposed to be minimized by the hyperparameter optimization: eval DNG = avg combined validity, avg h wdist(ρ), wdist(Nary), wdist( CN ) i , avg h 1 coverage recall, 1 coverage precision i# Here, the function avg returns the average of its arguments. We perform hyperparameter optimization for the DNG task only for the MP-20 dataset. For the CSP task, we optimize hyperparameters for the perov-5 and MP-20 datasets. For the CSP task on the MPTS-52 and Alex-MP-20 datasets, we simply transfer the hyperparameters of the optimized MP-20 models. We provide hyperparameter-tuned models with the relevant performance metrics and hyperparameters for perov-5 CSP in Tabs 6 and 7, MP-20 CSP in Tabs 8 and 9, and MP-20 DNG in Tabs 10 and 11. Many models were partially trained and compared in the process of hyperparameter tuning: on average 27 models (perov-5) and 32 models (MP-20) for each choice of positional interpolant, sample scheme, and latent variable. Open Materials Generation with Stochastic Interpolants Table 6. Study for the perov-5 dataset comparing CSP performance metrics for choice of positional interpolant, sample scheme, and γ(t) in the latent variable (or width σ0 of the Gaussian base distribution ρ0 for the SBD interpolants, and parameters σmin and σmax of the σ(s) schedule of the VE SBD interpolant). Positional Positional Positional Match rate RMSE interpolant sampling scheme γ(t) (%, Full / Valid) (Full / Valid) Linear ODE None: γ = 0 51.86% / 50.62% 0.0757 / 0.0760 Linear ODE Latent Sqrt: γ = p 0.034 t (1 t) 72.21% / 62.54% 0.3510 / 0.3444 Linear SDE Latent Sqrt: γ = p 0.028 t (1 t) 74.16% / 72.87% 0.3307 / 0.3315 Trigonometric ODE None: γ = 0 81.51% / 52.36% 0.3674 / 0.3628 Trigonometric ODE Latent Sqrt: γ = p 0.011 t (1 t) 80.85% / 79.55% 0.3864 / 0.3873 Trigonometric SDE Latent Sqrt: γ = p 0.063 t (1 t) 73.37% / 71.60 % 0.3610 / 0.3614 Encoder-Decoder ODE Enc-Dec: γ = 0.66 sin2 π(t 0.80t) (0.80 0.80t)+(t 0.80t) 68.08% / 64.60% 0.4005 / 0.4003 Encoder-Decoder SDE Enc-Dec: γ = 8.45 sin2 π(t 0.61t) (0.61 0.61t)+(t 0.61t) 78.28% / 76.80% 0.3616 / 0.3620 VP Score-Based Diffusion ODE σ0 = 0.28 83.06% / 81.27% 0.3753 / 0.3755 VP Score-Based Diffusion SDE σ0 = 0.13 76.54% / 64.46% 0.3529 / 0.3402 VE Score-Based Diffusion ODE σ0 = 8.96; σmin = 0.0078, σmax = 0.5165 60.18% / 52.97% 0.2510 / 0.2337 Table 7. Study for the perov-5 dataset CSP comparing hyperparameters for each choice of positional interpolant, sample scheme, and γ(t) (as reported in Tab. 6). Pos. interpolant, Cell interpolant Annealing param. s Integration Min. dist. Niggli λx,b/λl,b/λx,z/λl,z Sampling scheme, γ Sampling scheme, γ(t) (Pos. / Cell) steps permutation Linear, ODE, None Linear, ODE, γ = 0 14.11 / 2.90 820 False False 0.9729 / 0.0271 / - / - Linear, ODE, Latent Sqrt Linear, ODE, γ = 0 0.008 / 12.19 820 True True 0.9724 / 0.0276 / - / - Linear, SDE, Latent Sqrt Linear, ODE, 8.20 / 1.46 910 True True 0.0024 / 0.0051 / 0.9925 / - 0.013 t (1 t) Trig, ODE, None Linear, ODE, 14.99 / 14.97 880 True False 0.9983 / 0.0017 / - / - 0.021 t (1 t) Trig, ODE, Latent Sqrt Linear, ODE, γ = 0 9.68 / 2.42 110 False False 0.1130 / 0.8870 / - / - Trig, SDE, Latent Sqrt Linear, ODE, 3.43 / 0.03 900 True True 0.6868 / 0.0643 / 0.2489 / - 0.051 t (1 t) Enc-Dec, ODE, Enc-Dec Linear, ODE, γ = 0 14.94 / 0.318 460 True True 0.8563 / 0.1437 / - / - Enc-Dec, SDE, Enc-Dec Linear, ODE, 14.55 / 0.075 930 True False 0.2828 / 0.0004 / 0.7168 / - 0.154 t (1 t) VP SBD, ODE SBD, SDE, σ = 0.61 12.79 / 2.69 130 True True 0.0035 / 0.0121 / - / 0.9844 VP SBD, SDE Trig, SDE, 11.54 / 11.53 350 True False 0.2898 / 0.1960 / 0.3259 / 0.1883 0.029 t (1 t) VE SBD, ODE Trig, SDE, 0.003 / 14.93 380 False False 0.9800 / 0.0187 / - / 0.0014 0.024 t (1 t) Open Materials Generation with Stochastic Interpolants Table 8. Study for the MP-20 dataset comparing CSP performance metrics for choice of positional interpolant, sample scheme, and γ(t) in the latent variable (or width σ0 of the Gaussian base distribution ρ0 for the SBD interpolants, and parameters σmin and σmax of the σ(s) schedule of the VE SBD interpolant). Positional Positional Positional Match rate RMSE interpolant sampling scheme γ(t) (%, Full / Valid) (Full / Valid) Linear ODE None: γ = 0 69.83% / 63.75% 0.0741 / 0.0720 Linear ODE Latent Sqrt: γ = p 0.258 t (1 t) 55.60% / 50.04% 0.1531 / 0.1494 Linear SDE Latent Sqrt: γ = p 0.063 t (1 t) 68.20% / 61.88% 0.1632 / 0.1611 Trigonometric ODE None: γ = 0 65.30% / 58.94% 0.1184 / 0.1149 Trigonometric ODE Latent Sqrt: γ = p 0.033 t (1 t) 66.19% / 59.81% 0.1002 / 0.0968 Trigonometric SDE Latent Sqrt: γ = p 0.049 t (1 t) 68.90% / 62.65% 0.1249 / 0.1235 Encoder-Decoder ODE Enc-Dec: γ = 1.99 sin2 π(t 0.65t) (0.65 0.65t)+(t 0.65t) 55.15% / 49.45% 0.1306 / 0.1260 Encoder-Decoder SDE Enc-Dec: γ = 0.04 sin2 π(t 0.42t)0.5 (0.42 0.42t)0.5+(t 0.42t)0.5 57.69% / 52.44% 0.1160 / 0.1125 VP Score-Based Diffusion ODE σ0 = 0.22 45.57% / 39.48% 0.1880 / 0.1775 VP Score-Based Diffusion SDE σ0 = 2.29 42.29% / 38.08% 0.2124 / 0.2088 VE Score-Based Diffusion ODE σ0 = 9.77; σmin = 0.0047, σmax = 0.9967 63.79% / 57.82% 0.0809 / 0.0780 Table 9. Study for the MP-20 dataset comparing CSP hyperparameters for choice of positional interpolant, sample scheme, and γ(t) (as reported in Tab. 8). Pos. interpolant, Cell interpolant Annealing param. s Integration Min. dist. Niggli λx,b/λl,b/λx,z/λl,z Sampling scheme, γ Sampling scheme, γ(t) (Pos. / Cell) steps permutation Linear, ODE, None Linear, ODE, γ = 0 10.18 / 1.82 210 False False 0.9994 / 0.0006 / - / - Linear, ODE, Latent Sqrt Trig, ODE, 7.76 / 4.12 690 False True 0.9976 / 0.0024/ - / - 2.976 t (1 t) Linear, SDE, Latent Sqrt Linear, SDE, 11.58 / 5.08 310 False False 0.0073 / 0.0642 / 0.9154 / 0.0131 0.132 t (1 t) Trig, ODE, None Enc-Dec, SDE, 12.34 / 3.61 170 False False 0.9967 / 0.0023 / - / 0.0010 5.27 sin2 π(t 0.41t)0.5 (0.41 0.41t)0.5+(t 0.41t)0.5 Trig, ODE, Latent Sqrt Linear, SDE, 13.54 / 2.38 780 False True 0.9830 / 0.0167 / - / 0.0003 0.017 t (1 t) Trig, SDE, Latent Sqrt Trig, ODE, γ = 0 11.48 / 0.43 740 True True 0.2468 / 0.0301 / 0.7231 / - Enc-Dec, ODE, Enc-Dec Trig, SDE, 12.29 / 4.30 820 False True 0.6892 / 0.1235 / - / 0.1873 0.219 t (1 t) Enc-Dec, SDE, Enc-Dec Linear, ODE, 3.78 / 1.14 710 False True 0.6143 / 0.0063 / 0.3794 / - 4.961 t (1 t) VP SBD, ODE Linear, ODE, γ = 0 6.61 / 2.45 890 True True 0.9598 / 0.0402 / - / - VP SBD, SDE Linear, ODE, 6.46 / 0.67 600 True True 0.6060 / 0.0112 / 0.3828 / - 3.684 t (1 t) VE SBD, ODE Linear, SDE, 8.28 / 0.43 660 False False 0.9813 / 0.0005 / - / 0.0182 0.017 t (1 t) Open Materials Generation with Stochastic Interpolants Table 10. Study for the MP-20 dataset comparing DNG performance metrics for choice of positional interpolant, sample scheme, and γ(t) in the latent variable (or width σ0 of the Gaussian base distribution ρ0 for the SBD interpolants, and parameters σmin and σmax of the σ(s) schedule of the VE SBD interpolant). S.U.N. rates are computed according to the Matter Sim potential. Positional Positional Positional S.U.N. RMSD interpolant sampling scheme γ(t) Rate Linear ODE None: γ = 0 18.59% 0.2939 Linear ODE Latent Sqrt: γ = p 1.450 t (1 t) 9.95% 1.6660 Linear SDE Latent Sqrt: γ = p 0.018 t (1 t) 22.07% 0.6148 Trigonometric ODE None: γ = 0 19.63% 0.8289 Trigonometric ODE Latent Sqrt: γ = p 0.027 t (1 t) 19.96% 0.6570 Trigonometric SDE Latent Sqrt: γ = p 0.023 t (1 t) 17.60% 0.7763 Encoder-Decoder ODE Enc-Dec: γ = sin2(πt) 17.59% 0.3899 Encoder-Decoder SDE Enc-Dec: γ = 0.10 sin2 π(t 0.73t)0.5 (0.73 0.73t)0.5+(t 0.73t)0.5 16.27% 1.1795 VP Score-Based Diffusion ODE σ0 = 0.23 17.30% 1.1376 VP Score-Based Diffusion SDE σ0 = 7.14 22.10% 0.7631 VE Score-Based Diffusion ODE σ0 = 0.45; σmin = 0.0021, σmax = 0.8319 20.38% 0.6644 Table 11. Study for the MP-20 dataset comparing DNG hyperparameters for choice of positional interpolant, sample scheme, and γ(t) (as reported in Tab. 10). Pos. interpolant, Cell interpolant Annealing param. s Integration Min. dist. Niggli Species λx,b/λl,b/λx,z/λl,z/λa Sampling scheme, γ Sampling scheme, γ(t) (Pos. / Cell) steps permutation noise η Linear, ODE, None Linear, ODE, γ = 0 13.62 / 1.07 150 True False 7.08 0.9775 / 0.0006 / - / - / 0.0218 Linear, ODE, Latent Sqrt Enc-Dec, SDE, 14.83 / 5.91 130 True False 23.87 0.7683 / 0.0089 / - / 0.0012 / 0.2216 7.88 sin2 π(t 0.14t) (0.14 0.14t)+(t 0.14t) Linear, SDE, Latent Sqrt Linear, ODE, γ = 0 6.33 / 1.07 710 True False 0.19 0.1309 / 0.0065 / 0.2708 / - / 0.5918 Trig, ODE, None Trig, ODE, 8.59 / 0.29 860 True False 32.69 0.3302 / 0.0023 / - / - / 0.6675 1.183 t (1 t) Trig, ODE, Latent Sqrt Linear, SDE, 7.79 / 0.30 680 True True 27.25 0.2322 / 0.0035 / - / 0.3338 / 0.4306 0.848 t (1 t) Trig, SDE, Latent Sqrt Trig, ODE, 12.80 / 4.36 760 True False 13.15 0.6304 / 0.1582 / 0.0753 / - / 0.1360 0.316 t (1 t) Enc-Dec, ODE, Enc-Dec Linear, ODE, γ = 0 10.27 / 0.08 840 False False 0.85 0.7268 / 0.0084 / - / - / 0.2648 Enc-Dec, SDE, Enc-Dec Linear, ODE, 7.87 / 3.92 610 False False 19.78 0.2143 / 0.1547 / 0.1968 / - / 0.4341 1.651 t (1 t) VP SBD, ODE Trig, ODE, 2.30 / 2.74 710 False False 20.27 0.4053 / 0.0447 / - / - / 0.5500 7.797 t (1 t) VP SBD, SDE Trig, SDE, 9.06 / 11.77 870 False False 8.52 0.5184 / 0.0044 / 0.0008 / 0.1180 / 0.3584 3.100 t (1 t) VE SBD, ODE Linear, SDE, 12.72 / 0.98 330 False True 5.87 0.2209 / 0.0430 / - / 0.6371 / 0.0990 0.913 t (1 t) Open Materials Generation with Stochastic Interpolants D. Evaluation Metrics In this section, we provide details and discussion of the various metrics we use to evaluate CSP and DNG models (see Section 5.1). D.1. Match Rate and RMSE The tradeoff between match rate and RMSE most strongly influences the perov-5 dataset. We show in Fig. 7 how different positional interpolants for the atomic coordinates (trigonometric vs. linear with ODE sampling schemes) learn to generate matched structures differently. For the linear case, the change in matching tolerance (via the ltol parameter of Pymatgen s Structure Matcher) makes little difference. For the trigonometric interpolant, it makes a far more significant difference and leads to a much higher match rate, suggesting that the trigonometric interpolant learns structures more reliably but less accurately. OMat G (Trig) OMat G (Linear) ltol=0.2 ltol=0.3 mr_all = 20.73 mr_valid = 20.13 mr_all = 80.85 mr_valid = 79.55 mr_all = 51.86 mr_valid = 50.75 mr_all = 46.58 mr_valid = 45.60 All Valid only RMSE (dimensionless) RMSE (dimensionless) All Valid only All Valid only RMSE (dimensionless) RMSE (dimensionless) All Valid only Figure 7. We show here the effect of making matching more difficult by decreasing the length tolerance used by Pymatgen s Structure Matcher. We plot the density of the normalized RMSE distributions from CSP models trained on the perov-5 dataset (Linear and Trigonometric positional interpolants with ODE sampling schemes). We note that the curves for all generated structures and only valid generated structures overlap significantly. D.2. Validity Metrics The structural validity of generated structures is defined according to the bond lengths present in the structure all lengths must be >0.5 A to be considered valid. The compositional validity is defined according to the SMACT software package (Davies et al., 2019). We note that the default oxidation states have been updated with the release of SMACT version 3.0 which changed the DNG compositional validity rates by several percent. This also impacts the CSP match rate when filtered by valid structures. As such, all values for all models were recomputed with the most up-to-date version (3.0) of the SMACT software. Open Materials Generation with Stochastic Interpolants D.3. Coverage and Property Statistics As in (Xie et al., 2022), we evaluate coverage recall and precision (reported as rates) by measuring the percentage of crystals in the test set and in the generated samples that match each other within a defined fingerprint distance threshold. For structural matching, we use Crystal NN, and for compositional matching, we use Magpie fingerprints. Additionally, we report Wasserstein distances between property distributions of the generated and reference datasets. The considered properties are the mass density (ρ), the number of unique elements (Nary), and the average coordination number for each element in the unit cell ( CN ). D.4. Calculation of S.U.N. Rates Evaluation of DNG structures was performed using scripts provided by the developers of Matter Gen (Matter Gen, 2025). A total of 10,000 structures were generated from each of OMat G, Diff CSP, Flow MM, and Matter Gen-MP. These structures were then filtered to remove any that contained elements not supported by the Matter Sim potential (version Matter Sim-v1.0.0-1M) (Yang et al., 2024) or the reference convex hull. These included heavy elements with atomic numbers >89, radioactive elements, and the noble gases (specifically: Ac , U , Th , Ne , Tc , Kr , Pu , Np , Xe , Pm , He , Pa ).7 Stability and novelty were computed with respect to the default dataset provided by Matter Gen which contains 845 997 structures from the MP-20 (Jain et al., 2013; Xie et al., 2022) and Alexandria (Schmidt et al., 2022b;a) datasets. This provides a more challenging reference for computing novelty as each model was trained only on the 27 000 structures from the MP-20 training set. The Matter Sim potential was utilized for the first structural relaxation, requiring far less compute resources compared to full DFT. Results derived from the Matter Sim-relaxed structures are shown in Tab. 12. Following the Matter Sim relaxtions, 1000 structures were relaxed with DFT. All DFT relaxations utilized MPGGADouble Relax Static flows from the Atomate2 (Ganose et al., 2025) package to produce MP20-compatible data. Comparing the results from Tab. 12 to Tab. 3, we find that overall there is reasonable agreement between the metrics computed at the machine learning potential and DFT level. Relative performance ordering between methods remains fairly consistent, allowing for qualitative trends to be made at the much cheaper ML potential level. Nevertheless, for a full quantitative understanding, DFT is essential. Table 12. Stability (defined as 0.1 e V/atom above hull), uniqueness, and novelty results from de novo generation on the MP-20 dataset computed for the same models as in Tab. 2. All evaluations are performed with the Matter Gen code base (Matter Gen, 2025) with respect to the included reference Alex-MP-20 dataset. The average RMSD is between the generated and the relaxed structures, and the average energy above hull is reported in units of e V/atom. All structures were relaxed with the Matter Sim potential. Note that the results in Tab. 3 relied on a subsequent DFT relaxation. Method E /N ( ) RMSD Novelty Stability S.U.N. above hull ( A, ) Rate (%, ) Rate (%, ) Rate (%, ) Diff CSP 0.1984 0.367 72.73 43.04 19.00 Flow MM 0.2509 0.651 72.76 37.47 13.86 Matter Gen-MP 0.1724 0.142 72.17 47.07 22.66 Linear 0.1823 0.615 72.00 45.00 22.07 Trig 0.1857 0.657 65.35 51.40 19.96 Enc-Dec 0.1699 0.390 54.97 58.56 17.59 SBD 0.2189 0.763 75.80 42.60 22.10 CFP + CSP 0.2340 0.488 75.85 42.21 20.50 D.5. Stability and Structural Analysis of Generated Structures In Fig. 8, we show the distribution of computed energies above the convex hull across various OMat G models, showing best stability of generated structures for linear, encoder-decoder, trigonometric, and VP SBD positional interpolants. 7These elements were not removed from any datasets during training. Open Materials Generation with Stochastic Interpolants 0.2 0.0 0.2 0.4 0.6 0.8 Energy Above Hull (e V/Atom) Linear Enc-Dec Trig SBD CFP + CSP Figure 8. Histogram of the computed energies above the convex hull for structures generated by the five OMat G DNG models highlighted in the main text (see Table 3). We show that all positional interpolants are effective at generating structures close to the convex hull, with the VP SBD interpolant and CFP + CSP method performing slightly worse than other interpolants. By evaluating the distribution of Nary structures (Fig. 9), the distribution of average coordination numbers (both by structure in Fig. 10 and by species in Fig. 11), as well as distribution of crystal systems (Fig. 12) which are related to a structure s Bravais lattice, we provide qualitative analysis for model performance. Space groups (and thus crystal systems) were determined using the spglib software (Togo et al., 2024) and choosing the most common space group identification with exponentially (geometrically) decreasing tolerance. We find that OMat G models have superior performance on the Nary metric across the board, with Diff CSP performing the most poorly. Specific OMat G models (with positional Encoder-Decoder interpolant and CFP+CSP with Linear interpolant) and Diff CSP showed the best performance in matching the distribution of average coordination number for each structure, particularly for high-coordinated structures. The average coordination number for species were best-matched by OMat G models across the board (with the exception of the OMat G-VPSBD model which tended to overpredict the coordination environments), and broadly underpredicted by Diff CSP, Flow MM, and Matter Gen-MP. Finally, the best matching results for distribution across crystal systems was for the OMat G CFP+CSP model, with all other interpolants (and Diff CSP and Flow MM) showing a propensity for generating low-symmetry (triclinic and monoclinic) crystal structures. Overall, we find that OMat G models closely reproduce the elemental and structural diversity present in the data. Open Materials Generation with Stochastic Interpolants Diff CSP Flow MM OMat G-Enc Dec OMat G-Linear OMat G-Trig OMat G-VPSBD OMat G-CFP+CSP Unique element count 1 2 3 4 5 6 7 8 Unique element count 1 2 3 4 5 6 7 Unique element count 1 2 3 4 5 6 Unique element count 1 2 3 4 5 6 Unique element count 1 2 3 4 5 6 Unique element count 1 2 3 4 5 6 Unique element count 1 2 3 4 5 6 Unique element count 1 2 3 4 5 6 Matter Gen-MP Unique element count 1 2 3 4 5 6 7 OMat G-VESBD Figure 9. Qualitative performance of the distribution of Nary crystals for (a) Non-OMat G models and (b) OMat G models across structural benchmarks computed on generated structures and test set structures from the MP-20 dataset. Atomic elements are listed in increasing atomic number from left to right. Open Materials Generation with Stochastic Interpolants OMat G-Enc Dec OMat G-Linear Average CN by structure 0 10 5 15 Average CN by structure 0 8 4 12 Average CN by structure 0 8 4 12 Average CN by structure 0 8 4 12 Average CN by structure 0 8 4 12 OMat G-Trig Average CN by structure 0 8 4 12 OMat G-VPSBD Average CN by structure 0 8 4 12 OMat G-CFP+CSP Average CN by structure 0 8 4 12 Matter Gen-MP OMat G-VESBD Average CN by structure 0 8 4 12 Figure 10. Qualitative performance of the distribution of average coordination number by structure for (a) Non-OMat G models and (b) OMat G models across structural benchmarks computed on generated structures and test set structures from the MP-20 dataset. Open Materials Generation with Stochastic Interpolants OMat G-Linear OMat G-Trig OMat G-VPSBD OMat G-CFP+CSP Atomic element Atomic element Atomic element Atomic element Atomic element OMat G-Enc Dec Atomic element Atomic element Atomic element Matter Gen-MP Atomic element OMat G-VESBD Figure 11. Qualitative performance of the distribution of average coordination number by species (listed left to right in order of atomic number) for (a) Non-OMat G models and (b) OMat G models across structural benchmarks computed on generated structures and test set structures from the MP-20 dataset. Open Materials Generation with Stochastic Interpolants OMat G-Enc Dec OMat G-Linear OMat G-Trig OMat G-VPSBD OMat G-CFP+CSP Crystal system Crystal system Crystal system Crystal system Crystal system Crystal system Crystal system Matter Gen-MP Crystal system OMat G-VESBD Crystal system Figure 12. Qualitative performance of the distribution of crystal system by structure for (a) Non-OMat G models and (b) OMat G models across structural benchmarks computed on generated structures and test set structures from the MP-20 dataset. Open Materials Generation with Stochastic Interpolants E. Large Language Models as Base Distributions Flow LLM (Sriram et al., 2024) combines large language models (LLMs) with the conditional flow-matching framework Flow MM (Miller et al., 2024) to design novel crystalline materials in the DNG task. A fine-tuned LLM serves as the base distribution and samples initial structures; Flow MM then refines the fractional coordinates and lattice parameters as in the CSP task. This idea can be similarly applied to OMat G, which allows the use of LLMs within the general SI framework for materials generation. We extend OMat G to OMat G-LLM by allowing for LLM-generated structures as the initial structures. We evaluate both Flow LLM and OMat G-LLM on the LLM dataset released by Flow LLM. Specifically, we use the training (containing 40 000 structures) and validation sets (6000 structures) from https://github.com/facebookresearch/flowmm and the LLM-generated initial structures (10 000 structures) from https://github.com/facebookresearch/ crystal-text-llm as the test set. These initial structures are generated by a fine-tuned Llama-70B model (Gruver et al., 2024). As shown in Tab. 13, OMat G-LLM s linear and trigonometric interpolants outperform Flow LLM in almost all DNG metrics. Since the Wasserstein distance with respect to the Narity distributions and the compositional validity only depend on the atomic species generated by the LLM, these two metrics are necessarily equal for Flow LLM and OMat G-LLM. Note that the original Flow LLM (Sriram et al., 2024) is trained on 3M LLM-generated structures while our experiments are conducted on the 40K structures open-sourced by the authors. The performance of Flow LLM in our experiments thus differs from the scores reported in (Sriram et al., 2024). Table 13. Flow LLM s and OMat G-LLM s performance (with linear and the trigonometric interpolants) when using the same fine-tuned LLM (Gruver et al., 2024) as the base distribution. The best performance for each metric is in bold. Method Validity (% ) Coverage (% ) Property ( ) S.U.N. Rate (% ) Structural Composition Combined Recall Precision wdist (ρ) wdist (Nary) wdist ( CN ) Flow LLM 96.27 86.40 83.55 97.98 96.55 0.9922 0.5427 0.5936 10.28 Linear 97.86 86.40 84.85 99.16 98.40 0.9100 0.5427 0.8600 12.61 Trigonometric (ODE) 95.70 86.40 83.25 98.57 98.24 0.7410 0.5427 0.6165 11.14 Trigonometric (SDE) 97.78 86.40 84.72 97.41 99.12 3.6214 0.5427 0.4448 11.86 F. Computational Costs In Tabs 14 and 15, we present the computational costs for both the CSP and DNG tasks. We compare the cost of training and integrating OMat G on the MP-20 dataset and show low computational costs for OMat G s ODE scheme for both training and inference. The SDE scheme is more expensive but competitive. For these experiments, we use an Nvidia RTX8000 GPU with a batch size of 512 and 1000 integration steps. Table 14. Computational costs for Diff CSP, Flow MM, OMat G (ODE) and OMat G (SDE) models trained on the CSP task. Task OMat G (ODE) Flow MM OMat G (SDE) Diff CSP Training (s / epoch) 56.8 0.75 70.35 1.38 89.0 1.41 21.89 0.31 Sampling (s / batch) 313.67 9.29 424.125 11.78 479.5 13.5 338.11 11.93 Table 15. Computational costs for Diff CSP, Flow MM, OMat G (ODE) and OMat G (SDE) models trained on the DNG task. Task OMat G (ODE) Flow MM OMat G (SDE) Diff CSP Training (s / epoch) 75.26 2.08 73.32 0.47 102.65 1.87 21.85 0.36 Sampling (s / batch) 473.14 13.20 469.93 6.12 617.2 18.2 322.63 10.28