# towards_symmetryaware_generation_of_periodic_materials__2913a9de.pdf Towards Symmetry-Aware Generation of Periodic Materials Youzhi Luo, Chengkai Liu, Shuiwang Ji Department of Computer Science & Engineering Texas A&M University College Station, TX 77843 {yzluo,liuchengkai,sji}@tamu.edu We consider the problem of generating periodic materials with deep models. While symmetry-aware molecule generation has been studied extensively, periodic materials possess different symmetries, which have not been completely captured by existing methods. In this work, we propose Sy Mat, a novel material generation approach that can capture physical symmetries of periodic material structures. Sy Mat generates atom types and lattices of materials through generating atom type sets, lattice lengths and lattice angles with a variational auto-encoder model. In addition, Sy Mat employs a score-based diffusion model to generate atom coordinates of materials, in which a novel symmetry-aware probabilistic model is used in the coordinate diffusion process. We show that Sy Mat is theoretically invariant to all symmetry transformations on materials and demonstrate that Sy Mat achieves promising performance on random generation and property optimization tasks. Our code is publicly available as part of the AIRS library (https://github.com/divelab/AIRS). 1 Introduction Designing or synthesizing novel periodic materials with target properties is a fundamental problem in many real-world applications, such as designing new periodic materials for solar cells and batteries [1]. For a long time, this challenging task heavily relies on either manually designing material structures based on the experience of chemical experts, or running very expensive and time-consuming density functional theory (DFT) based simulation. Recently, thanks to the progress of deep learning techniques, many studies have applied advanced deep generative models [22, 13] to generate or discover novel chemical compounds, such as molecules [10, 39, 50] and proteins [48, 16]. However, while deep learning has been widely used in periodic material representation learning [51, 5, 40, 4, 57, 25], generating novel periodic materials with deep generative models remains largely under-explored. The major challenge of employing deep generative models to the generation of periodic materials is capturing physical symmetries of periodic material structures. Ideally, deep generative models for periodic materials should maintain invariance to symmetry transformations of periodic materials, including permutation, rotation, translation, and periodic transformations [59, 9, 60]. To achieve this target, we propose Sy Mat, a novel symmetry-aware periodic material generation method. Sy Mat transforms the atom types and lattices of materials to symmetry-aware generation targets, which are generated by a variational auto-encoder model [22]. Besides, Sy Mat adopts novel symmetry-aware probabilistic diffusion process to generate atom coordinates with a score-based diffusion model [44]. Experiments show that Sy Mat can achieve promising performance on random generation and property optimization tasks. 37th Conference on Neural Information Processing Systems (Neur IPS 2023). Relations with Prior Methods. We note that several existing studies are related to our proposed Sy Mat, such as CDVAE [52]. Generally, the major difference of Sy Mat between them lies in the capture of symmetries. We will discuss the detailed differences in Section 3.5. 2 Background and Related Work 2.1 Molecule Generation Molecule Material Figure 1: An illustration of molecule (left) and periodic material (right) structures. Different from molecules, atoms in materials periodically repeat themselves infinitely in 3D space (we use a 2D visualization here for simplicity). Designing or discovering novel molecules with target properties has long been a challenging problem in many chemical applications. Recently, motivated from the significant progress in deep generative learning [22, 12, 44, 13], a lot of studies have applied advanced deep generative models to this problem. Some early studies [19, 58, 41, 32, 27] consider molecules as 2D topology graphs and design graph generation models to discover novel molecules. More recently, considering the importance of 3D structures in many molecular properties, many studies have also proposed advanced generative models that can generate 3D molecular conformations either from scratch [10, 39, 31, 15, 50], or from input conditional information [33, 56, 43, 53, 54, 42, 55, 20, 28]. However, all these existing studies are designed for non-periodic molecules while materials have periodic structures, in which atoms periodically repeat themselves in 3D space (see Figure 1 for an illustration of differences between molecules and periodic materials). Hence, they cannot be directly applied to the periodic material generation problem that we study in this work. 2.2 Periodic Material Generation While molecule generation has been extensively studied, currently the periodic material design or generation problem remains largely under-explored. Early periodic material design methods [34, 7] only generate compositions of chemical elements in periodic materials but do not generate 3D structures. More recent studies have proposed to use variational auto-encoders (VAEs) [22] or generative adversarial networks (GANs) [12] to generate 3D structures of periodic materials. However, these methods use coordinate matrices [36, 21, 61] or voxel images [14, 6] of materials as the direct generation targets, which results in the violation of physical symmetries of periodic material structures in their captured probabilistic distributions. Also, the VAE and GAN models used by them are not powerful enough to capture the distribution of complicated 3D periodic material structures. Different from them, our method employs a more powerful score-based diffusion model [44] to generate 3D periodic material structures, and the model is designed to capture physical symmetries in materials. 2.3 Score-Based Diffusion Models for 3D Structure Generation Diffusion models are a family of deep generative models that have achieved outstanding generation performance in a variety of data modalities, such as images [38, 37], audios [24], and molecular conformations [55, 15, 20]. Currently, score-based diffusion models [44] and denoising diffusion probabilistic models [13] are two most commonly used diffusion models, and our method employs score-based diffusion models for 3D periodic material generation. For a given data distribution p(x) in the high-dimensional data space, the score-based diffusion model trains a parameterized score model sθ( ) with parameters θ to accurately approximate the score function x log p(x) from x. To improve the score approximation accuracy in the regions where the training data is sparse, denoising score matching [46, 44] is proposed to more effectively train sθ( ). Specifically, let {σt}T t=1 be a set of positive scalars satisfying σ1 > σ2 > ... > σT , the denoising score matching method first perturbs the data point x sampled from p(x) by a sequence of Gaussian noises with noise magnitude levels {σt}T t=1, i.e., sampling x from pσt( x|x) = N( x|x, σ2 t I) for every noise level σt. The score model sθ( ) is optimized to predict the denoising score function x log pσt( x|x) from x at each noise level, where the ground truth of x log pσt( x|x) can be easily computed from the mathematical formula of N( x|x, σ2 t I). Formally, sθ( ) is optimized to minimize the following training objective L: t=1 σ2 t E x,x As shown in Vincent [46], when σt is small enough, after sθ( ) is well trained with the objective in Equation (1), the output sθ(x) can accurately approximate the exact score function for any input x. New data samples can be generated by running multiple steps of the Langevin dynamics sampling algorithm [49] with the score functions approximated by the score model. Several existing studies have applied score-based diffusion models to 3D molecular and material structure generation. Conf GF [42] is the first score-based diffusion model for 3D molecular conformation generation. It proposes a novel theoretic framework to achieve roto-translational equivariance in score functions. Based on the theoretic framework of Conf GF, DGSM [30] further improves the performance of Conf GF by dynamic graph score matching. Though they have achieved good performance in 3D molecules, they do not consider periodic invariance so they cannot be applied to 3D periodic materials. Recently, a novel approach CDVAE [52] is proposed for periodic material generation, and score-based diffusion models are used in CDVAE to generate atom coordinates in materials. However, as we discussed in Section 3.5, CDVAE fails to achieve translational invariance in calculating the denoising score function, so it does not capture all physical symmetries. 3 Sy Mat: Symmetry-Aware Generation of Periodic Materials While a variety of deep generative models have been proposed to capture all physical symmetries in 3D molecular conformations, it remains challenging to do it for periodic materials as they have more complicated 3D periodic structures. In this section, we present Sy Mat, a novel periodic material generation method. In the following subsections, we will first introduce the problem of symmetryaware material generation, then elaborate how our proposed Sy Mat approach achieves invariance to all symmetry transformations of periodic materials. 3.1 Symmetry-Aware Generation For any periodic material structure, we can consider it as periodic repetitions of one unit cell in 3D space, where any unit cell is the smallest repeatable structure of the material. Hence, for any periodic material M, we can describe its complete structure information with one of its unit cells and the three lattice vectors describing its periodic repeating directions. Specifically, assuming there are n atoms in any unit cell of M, we represent M as M = (A, P , L), where A Zn, P R3 n, and L R3 3 are the atom type vector, coordinate matrix, and lattice matrix, respectively. Here, the i-th element ai of A = [a1, ..., an] and the i-th column vector pi of P = [p1, ..., pn] denote the atom type, i.e., atomic number and the 3D Cartesian coordinate of the i-th atom in the unit cell, respectively. The lattice matrix L = [ℓ1, ℓ2, ℓ3] is formed by three lattice vectors ℓ1, ℓ2, ℓ3 indicating how the atoms in a unit cell periodically repeat themselves in 3D space. In this work, we consider the problem of generating periodic material structures with generative models. Formally, we are given a dataset M = {Mj}m j=1 where each periodic material data Mj is assumed to be sampled from the distribution p( ). Because of the physical symmetry properties in periodic materials [59, 9], p( ) is restricted to be invariant to the following symmetry transformations. Permutation. For any M = (A, P , L), if we permute A, i.e., exchange elements in A to obtain A , and use the same permutation order to permute the column vectors of P to obtain P , we actually change nothing but atom orders in unit cell representations. Hence, let M = (A , P , L), the probability densities of M and M are actually the same, i.e., p(M) = p(M ). Rotation and translation. For any M = (A, P , L), we can rotate and translate its atom positions in 3D space to obtain a new material M = (A, P , L ), where P = QP + b1T and L = QL. Here Q R3 3 is the rotation matrix satisfying QT Q = I, b R3 is the translation vector, and 1 is an n-dimensional vector whose elements are all 1. It can be intuitively understood that rotation and translation do not change any internal structures of periodic materials, so M and M are actually different representations of the same material, so p(M) = p(M ) should hold. VAE encoder VAE decoder Reconstruction loss KL-distance loss Score model Score matching loss Atom type set Lattice lengths Lattice angles + + (a) The overall process of training the Sy Mat model on a data sample M = (A, P , L) from the dataset. VAE decoder (b) The overall process of generating a new material with the Sy Mat model. Figure 2: Illustrations of training and generation process in our proposed Sy Mat method. Periodic transformation. Because every atom from a unit cell periodically repeats itself along lattice vectors in 3D space, any coordinate matrix P R3 n has infinite number of periodically equivalent coordinate matrices. Formally, a periodically equivalent coordinate matrix P K of P can be obtained by periodic transformation as P K = P + LK, where K Z3 n is an arbitrary integer matrix. Since P K and P are equivalent and the periodic transformations between them are invertible, we consider M = (A, P , L) and M K = (A, P K, L) as the same material so p(M) = p(M K) should hold. In the periodic material generation problem, our target is to learn a generative model pθ( ) with parameters θ from the given dataset M so that the model can capture the real data distribution p( ), and the learned generation model can generate a valid material structure M with a high probability pθ(M). In addition, we require the distribution pθ( ) captured by the generation model to have the same symmetry-related restrictions as the real data distribution p( ). The main motivation of modeling symmetries comes from property optimization. A significant target of developing generative models for periodic materials is to generate novel periodic materials with desirable chemical properties (e.g., low energies). Notably, chemical properties are invariant to physical symmetry transformations. For instance, rotating a material structure in 3D space does not change its energy. In other words, chemical properties only depend on the 3D geometric information that is invariant to symmetry transformations (e.g., interatomic distances). Using symmetry-aware modeling forces generative models to only capture the distribution of these 3D geometric information, which facilitates searching materials with desirable properties in the latent space. 3.2 An Overview of Sy Mat Generally, to generate a new material M = (A, P , L), our proposed Sy Mat method first generates the atom type vector A and lattice matrix L, then generates the coordinate matrix P conditioned on A and L. The probability of generating M = (A, P , L) can then be described as pθ(M) = pθ(A)pθ(L)pθ(P |A, L). (2) Since the structures of atom types and lattices are relatively simple, we find that a VAE [22] model already has adequate capability to capture their distributions. Hence, we use a VAE model to generate A and L. However, the structures of coordinates are much more complicated, so we use a powerful score-based diffusion model [44] for their generation. Next, we will elaborate the details of generating atom types and lattices in Section 3.3 and the details of generating coordinates in Section 3.4. An overview of the proposed Sy Mat method is illustrated in Figure 2. 3.3 Atom Type and Lattice Generation For the generation of atom types and lattices, a natural idea is using the VAE model to directly generate and capture the distribution of A and L. However, this fails to capture the invariance to symmetry transformations described in Section 3.1, including the permutation invariance of A and rotation invariance of L. To tackle this issue, we transform A and L to items that are invariant to symmetry transformations, and make these items be the direct generation targets of the VAE model. Specifically, for A = [a1, ..., an], we can count the number of atoms for every atom type existed in A, and represent A with an unordered k-element atom type set c = {(e1, n1), ..., (ek, nk)}, which indicates that for i = 1, ..., k, there exist ni atoms with the atom type ei in A. We can easily find that c is invariant to any permutations on A so we use c as the direct generation targets. Besides, for the lattice matrix L = [ℓ1, ℓ2, ℓ3], we do not generate the three lattice vectors ℓ1, ℓ2, ℓ3 directly. Instead, we use six rotation-invariant items as the generation targets, including three lattice lengths ℓ= [ℓ1, ℓ2, ℓ2], where ℓi = ||ℓi||2, and three lattice angles ϕ = [ϕ12, ϕ13, ϕ23], where ϕij is the angle between ℓi and ℓj. Overall, we transform A and L to generation targets c, ℓ, ϕ so that the generation process and captured distributions pθ(A), pθ(L) are ensured to be symmetry-aware. The VAE model used for generating c, ℓ, and ϕ consists of an encoder model and a decoder model. The encoder model is based on a 3D graph neural network (GNN) that takes a material M = (A, P , L) as inputs and outputs a d-dimensional latent variable z A Rd and an f-dimensional latent variable z L Rf. This 3D GNN employs the symmetry-aware Sphere Net [29] architecture so z A, z L will not change if applying any transformations described in Section 3.1 to the input material. We will introduce more details of this 3D GNN model in Appendix B. The decoder model is composed of four multi-layer perceptrons (MLPs) including MLPe, MLPk, MLPn, MLPL, which are used to predict c = {(e1, n1), ..., (ek, nk)}, ℓ= [ℓ1, ℓ2, ℓ3], and ϕ = [ϕ12, ϕ13, ϕ23] from the latent variables z A and z L. Specifically, let E be the number of all considered atom types and N be the largest possible number of atoms that can exist in a material, MLPe predicts a vector pe [0, 1]E which contains the existing probability for each of the E atom type from the input z A, and MLPk predicts the size k of the set c from the input z A. The k atom types e1, ..., ek with top-k probabilities in pe are chosen as the atom types in c, and MLPn predicts the number of atoms n1, ..., nk for them, where each ni is predicted from ei and z A. Here, the prediction of pe is considered as E binary classification tasks for each of E atom types, and the prediction of k and n1, ..., nk are both considered as N-category classification tasks. Besides, MLPL is used to predict lattice items [ℓ, ϕ] = [ℓ1, ℓ2, ℓ3, ϕ12, ϕ13, ϕ23] from z L, which is considered as a regression task. We now describe the training and generation process of the VAE model. When training the VAE model on a material dataset M, for any material M = (A, P , L) in M, the encoder model first maps it to latent variable z A and z L. Afterwards, the decoder model uses z A and z L to reconstruct the exact c, ℓ, and ϕ obtained from A and L. The VAE model is optimized to minimize the reconstruction error of c, ℓ, ϕ together with the KL-distance between z A, z L and the standard Gaussian distribution N(0, I). The reconstruction error losses are defined as cross entropy loss and minimum squared error loss for classification and regression tasks, respectively. After the model is trained, we can generate A and L by first sampling latent variables z A, z L from N(0, I), then using the decoder to map z A, z L to c, ℓ, ϕ. Finally, we can produce A from c = {(e1, n1), ..., (ek, nk)} as A = [e1, ..., e1 | {z } n1 , e2, ..., e2 | {z } n2 , ..., ek, ..., ek | {z } nk and produce L from ℓ, ϕ by constructing three lattice vectors with lengths in ℓand pairwise angles in ϕ. The detailed procedure of producing three lattice vectors is described in Appendix B. 3.4 Coordinate Generation After the atom type vector A and lattice matrix L are generated, our method will generate the coordinate matrix P conditioned on A and L. We use a score-based diffusion model for coordinate generation. However, because the original score-based diffusion model proposed by Song and Ermon [44] does not consider any invariance to symmetry transformations in data distributions, we cannot directly apply its theoretic framework to coordinate generation. To overcome this limitation, we propose a novel probabilistic modeling process to calculate the score function of P , i.e., the score matrix P log p(P |A, L), and design the architecture and training framework of the score model to achieve invariance to all symmetry transformations of crystal materials. We will elaborate the details of our proposed coordinate generation framework in the next paragraphs. We first elaborate the restrictions of the score matrix P log p(P |A, L) that is supposed to be captured by the score model. By the following Proposition 3.1 (see its proof in Appendix A.1), we can show that if the material distribution p(M) is invariant to all symmetry transformations described in Section 3.1, the score matrix P log p(P |A, L) will be invariant to translation and periodic transformations, and equivariant to permutation and rotation transformations. Proposition 3.1. If the material distribution p(M) is invariant to permutation, rotation, translation, and periodic transformations described in Section 3.1, P log p(P |A, L) will satisfy the following properties: (1) P log p(P |A, L) is equivariant to permutations on A and P ; (2) QP +b1T log p(QP + b1T |A, QL) = Q P log p(P |A, L) holds for any Q R3 3, QT Q = I and b R3; (3) P +LK log p(P + LK|A, L) = P log p(P |A, L) holds for any K Z3 n. To satisfy all properties in Proposition 3.1, we formulate P log p(P |A, L) as a function of score functions of edge distances in a graph created by multi-graph method [51]. For a periodic material M with coordinate matrix P = [p1, ..., pn] and lattice matrix L, multi-graph method produces an n-node undirected graph on M. In this graph, the node i corresponds to the i-th atom in the unit cell, and an edge is added between any two nodes i, j if one of the interatomic distances between them is smaller than a pre-defined cutoff r. Here, the interatomic distances between i, j are the distances between their corresponding atoms in the complete infinite structure of M, including both the distance within the unit cell and the distances crossing the unit cell boundary. Formally, the set of all edges in the graph constructed on M can be written as E(M) = {(i, j, k) : ||pi + Lk pj||2 r, k Z3, 1 i, j n}. Note that there may exist more than one edges connecting the nodes i, j if more than one interatomic distances between them is smaller than r. Let di,j,k = ||pi + Lk pj||2 be the distance of the edge (i, j, k), we consider log p(P |A, L) as a function of the distances of all edges in E(M). Denoting si as the score function of pi, then P log p(P |A, L) = [s1, ..., sn]. From the chain rule of derivatives, we can calculate si as (j,k) N(i) di,j,k log p(P |A, L) pidi,j,k = X (j,k) N(i) si,j,k pi + Lk pj di,j,k , (3) where the scalar si,j,k = di,j,k log p(P |A, L) is the score function of the distance di,j,k and N(i) = {(j, k) : (i, j, k) E(M)}. With this probabilistic modeling process, P log p(P |A, L) can be approximated by first approximating si,j,k for every edge in the multi-graph representation of M, then calculating each column vector of P log p(P |A, L) by Equation (3). By the following Proposition 3.2 (see its proof in Appendix A.2), we can show that if si,j,k is invariant to all symmetry transformations described in Section 3.1 for every edge (i, j, k), the score matrix P log p(P |A, L) will satisfy all properties in Proposition 3.1. Proposition 3.2. For P log p(P |A, L) = [s1, ..., sn] where each si is calculated by Equation (3) on a multi-graph representation of M = (A, P , L), if si,j,k is invariant to permutation, rotation, translation, and periodic transformations described in Section 3.1 for any edge (i, j, k), then the calculated P log p(P |A, L) will always satisfy the properties in Proposition 3.1. In our method, we use a 3D GNN model as the score model sθ( ) to approximate P log p(P |A, L). sθ( ) takes the multi-graph representation of M as input and outputs oi,j,k as the approximated si,j,k at every edge (i, j, k) in the input graph. We use Sphere Net [29] as the backbone architecture of sθ( ), which ensures that the output oi,j,k is always invariant to any symmetry transformation. More details about Sphere Net will be introduced in Appendix B. To train the score model, we follow the most common operations in denoising score matching framework [44]. Specifically, for a material M = (A, P , L) in the dataset, we will perturb it to multiple noisy materials by a sequence of Gaussian noises with different noise magnitude levels {σt}T t=1, in which σ1 > σ2 > ... > σT . Let the noisy material at the t-th noise level be f M = A, e P , L , the noisy coordinate matrix e P = [ p1, ..., pn] is obtained by adding Gaussian noise N(0, σ2 t I) to P . The score model sθ( ) is trained to predict the denoising score matrix e P log pσt e P |P , A, L = [ s1, ..., sn] from f M. Here, each si is calculated in a similar way as in Equation (3) with the approximated distance score si,j,k = di,j,k log pσt e P |P , A, L , where di,j,k = || pi + Lk pj||2. Motivated from Algorithm 1: Langevin Dynamics Sampling Algorithm for Coordinate Generation Input: A, L, atom number n, noise magnitudes {σt}T t=1, step size ϵ, step number q, score model sθ( ) Sample F0 R3 n from uniform distribution U(0, 1) and set P0 = LF0 for t = 1 to T do Pt = Pt 1 αt = ϵ σ2 t /σ2 T for j = 1 to q do Mt = (A, Pt, L) Sample Z R3 n from N(0, I) Obtain the approximated edge distance score from sθ( ) which takes Mt 1 as inputs Obtain sθ(Mt 1) = [s1, ..., sn] where each si is computed by Equation (3) with edge distance scores. Pt = Pt + αtsθ(Mt) + 2αt Z end for end for Output PT as the finally generated coordinate matrix the strategies of existing methods [30, 42, 55], we model the distribution of di,j,k as a Gaussian distribution N ˆdi,j,k, ˆσ2 t , where the computation process of the mean ˆdi,j,k and the variance ˆσ2 t is described in Appendix B. We can calculate si,j,k directly by taking the derivative of the mathematical formula of this Gaussian distribution. The score model sθ( ) is trained to minimize the difference between the output oi,j,k from sθ( ) and si,j,k for every edge (i, j, k) in f M with the loss L: t=1 ˆσ2 t E f M,M (i,j,k) E( f M) ˆσt + di,j,k ˆdi,j,k After the score model is trained, we can use it to generate coordinate matrix P from given A and L with annealed Langevin dynamics sampling algorithm [49]. Specifically, we first randomly initializes a coordinate matrix P0, then starting from M0 = (A, P0, L), we iteratively update Mt 1 = (A, Pt 1, L) to Mt = (A, Pt, L) (t = 1, ..., T) with the approximated score matrix sθ(Mt 1) by running multiple Langevin dynamics sampling steps. See Algorithm 1 for the detailed coordinate generation algorithm pseudocodes. 3.5 Discussions and Relations with Prior Methods Advantages and limitations. In our method, the atom type vector and lattice matrix are first transformed to items that are invariant to symmetry transformations, then a VAE model is used to capture the distributions of these items. In addition, our method adopts powerful score-based diffusion models to generate atom coordinates, and a novel probabilistic modeling framework for score approximation is employed to satisfy all symmetry-related restrictions. With all these strategies, our method successfully incorporates physical symmetries of periodic materials into generative models, thereby capturing the underlying distributions of material structures more effectively. We will show these advantages through experiments in Section 4. The major limitations of our method lie in that the speed of generating atom coordinates with score-based diffusion models is slow because running thousands of Langevin dynamics sampling steps is needed, and our method cannot be applied to non-periodic materials. We leave more discussions about limitations and broader impacts in Appendix C. Motivations in Model Design. The model design of our Sy Mat method has the following three motivations. (1) First, the structures of atom types and lattices are relatively simple and their symmetryinvariant representations (i.e., atom type sets, lattice lengths and angles) are low-dimensional vectors. Their distributions are simple enough to be captured by VAE models. Also, VAE models are faster than diffusion models in generation speed. Hence, we use VAE models for atom type and lattice generation. (2) Second, the structure of atom coordinates is much more complicated than atom types or lattices, so we use more powerful score-based diffusion models to generate them. More importantly, using score-based diffusion models for atom coordinates enables us to trickily convert the prediction of coordinate scores to that of interatomic distance scores by the chain rule of derivatives. This strategy maintains invariance to rotation, translation and periodic transformations. We think it is hard to develop such a symmetry-aware probabilistic modeling for other generative models. Hence, we use score-based diffusion models for coordinate generation. (3) Third, Sphere Net is chosen as the backbone network of our model. Sphere Net is a 3D graph neural network for 3D graph representation learning. It has powerful capacities in 3D structure feature extraction because it considers complete 3D geometric information contained in pairwise distances, line angles and plane angles (torsion angles). Also, the features extracted by Sphere Net are invariant to rotation, translation and periodic transformations when input 3D graphs are created by multi-graph method. Because of its powerful capacity and symmetry-aware feature extraction, we use Sphere Net as the backbone network. Relations with prior methods. Here we will discuss several studies that are related to our work. Some studies [42, 30] have also applied score-based diffusion models to generate atom coordinates from molecular graphs, but there are not periodic structures in 3D molecules. Differently, keeping invariant to periodic transformations is significant in periodic material generation, and our score-based diffusion model uses periodic invariant multi-graph representations of materials as inputs to satisfy this property. Besides, compared with a recently proposed periodic material generation method CDVAE [52], our Sy Mat has two major differences. (1) First, to generate the atom type vector, CDVAE generates total atom numbers and compositions (the proportion of atom numbers for every atom type to the total atom number) by a VAE model, in which compositions are vectors composed of continuous numbers. While compositions have complicated structures and are hard to predict, Sy Mat uses relatively simpler atom type sets as generation targets and the prediction of them can be easily converted to integer prediction or classification problems (see Section 3.3). (2) Second, in coordinate generation, CDVAE also uses the denoising score matching framework, but the denoising score matrix e P log pσt( e P |P , A, L) is calculated as the direct difference between e P and a coordinate matrix obtained by aligning P , which makes e P log pσt( e P |P , A, L) not invariant to translations on e P . However, Sy Mat calculates it from the edge distance score functions to ensure invariance to all symmetry transformations (see Section 3.4). In addition, another two studies, PGD-VAE [47] and Diff CSP [18], are also related to our Sy Mat approach. PGD-VAE proposes a VAE model for periodic graph generation. However, PGD-VAE is fundamentally different from Sy Mat in that it only generates 2D periodic topology graphs without 3D structures, while Sy Mat is developed for 3D periodic material structures. PGD-VAE cannot be applied to 3D periodic materials so we do not compare Sy Mat with PGD-VAE in our experiments. Diff CSP proposes a novel method for generating 3D periodic material structures from their atom types. Different from Sy Mat, Diff CSP uses denoising diffusion models to jointly generate lattices and atom coordinates. In addition, Diff CSP does not generate atom types of periodic materials, but lattices and atom coordinates from the input atom types, so Diff CSP cannot be applied to design novel periodic materials from scratch. Hence, we do not compare Sy Mat with Diff CSP in our experiments as all our experiments evaluate the performance of different methods in generating novel periodic materials. 4 Experiments In this section, we evaluate our proposed Sy Mat method in two periodic material generation tasks, including random generation and property optimization. We show that our proposed Sy Mat can achieve promising performance in both tasks. 4.1 Experimental Setup Data. We evaluate Sy Mat on three benchmark datasets curated by Xie et al. [52], including Perov5 [3, 2], Carbon-24 [35], and MP-20 [17]. Perov-5 is a dataset with a collection of 18,928 perovskite materials. The chemical element compositions of all materials in Perov-5 can be denoted by a general chemical formula ABX3, which means that there are 3 different atom types and 5 atoms in each unit cell. The Carbon-24 dataset we used has 10,153 materials. All these material structures consist of only carbon elements and have 6 - 24 atoms in each unit cell, and the 3D structures of them are obtained by DFT simulation. MP-20 is a dataset curated from Materials Project [17] library. It includes 45,231 materials with various structures and compositions, and in all materials, there exist at Table 1: Random generation performance on Perov-5, Carbon-24, and MP-20 datasets. Here means higher metric values lead to better performance, while means the opposite. Because all materials in the Carbon-24 dataset are composed of only carbon atoms, all methods easily achieve 100% in composition validity and 0.0 in # element EMD so we do not use these two metrics. Also, Cond-DFC-VAE can only be used to Perov-5 dataset in which materials have cubic structures. Bold and underline numbers highlight the best and second best performance, respectively. Dataset Method Composition validity Structure validity # Element EMD Density EMD Energy EMD COV-R COV-P FTCP 54.24% 0.24% 0.6297 10.27 156.0 0.00% 0.00% Cond-DFC-VAE 82.95% 73.60% 0.8373 2.268 4.111 77.80% 12.38% G-Sch Net 98.79% 99.92% 0.0368 1.625 4.746 0.25% 0.37% P-G-Sch Net 99.13% 79.63% 0.4552 0.2755 1.388 0.56% 0.41% CDVAE 98.59% 100.00% 0.0628 0.1258 0.0264 99.50% 98.93% Sy Mat (ours) 97.40 % 100.00% 0.0177 0.1893 0.2364 99.68% 98.64% FTCP 0.08% 5.206 19.05 0.00% 0.00% G-Sch Net 99.94% 0.9427 1.320 0.00% 0.00% P-G-Sch Net 48.39% 1.533 134.7 0.00% 0.00% CDVAE 100.00% 0.1407 0.2850 100.00% 99.98% Sy Mat (ours) 100.00% 0.1195 3.9576 100.00% 97.59% FTCP 48.37% 1.55% 0.7363 23.71 160.9 5.26% 0.23% G-Sch Net 75.96% 99.65% 0.6411 3.034 42.09 41.68% 99.65% P-G-Sch Net 76.40% 77.51% 0.6234 4.04 2.448 44.89% 99.76% CDVAE 86.70% 100.00% 1.432 0.6875 0.2778 99.17% 99.64% Sy Mat (ours) 88.26% 100.00% 0.5067 0.3805 0.3506 98.97% 99.97% most 20 atoms in each unit cell. For all three datasets, we split them with a ratio of 3:1:1 as training, validation, and test sets in our experiments. Tasks. We evaluate the performance of Sy Mat in random generation and property optimization tasks. In the random generation task, Sy Mat generates novel periodic materials from randomly sampled latent variables and we employ a variety of validity and statistic metrics to evaluate the quality of the generated periodic materials. In the property optimization task, Sy Mat is evaluated by the ability of discovering novel periodic materials with good properties. Besides, to show that latent representations are informative enough to reconstruct periodic materials, we also evaluate the performance by the material reconstruction task. See Appendix D.2 for experiment results of this task. Baselines. We compare Sy Mat with two early periodic material generation methods FTCP [36] and Cond-DFC-VAE [6], which generate material structures through generating Fourier-transformed crystal property matrices and 3D voxel images with VAE models, respectively. Note that Cond DFC-VAE can only be used to generate cubic systems, so it is only used for Perov-5 dataset in which materials have cubic structures. In addition, we compare with an autoregressive 3D molecule generation method G-Sch Net [10], and its variant P-G-Sch Net that incorporates periodicity in GSch Net. The latest periodic material generation method CDVAE [52] is also compared with our method. Note that all these baseline methods are compared with Sy Mat in the random generation task, but G-Sch Net and P-G-Sch Net are not compared in the property optimization task because they cannot produce latent representations with fixed dimensions for different periodic materials. 4.2 Random Generation Metrics. In the random generation task, we adopt seven metrics to evaluate physical and statistic properties of materials generated by our method and baseline methods. We evaluate the composition validity and structure validity, which are the percentages of generated materials with valid atom type vectors and 3D structures, respectively. Here, an atom type vector is considered as valid if its overall charge computed by SMACT package [8] is neutral, and following Court et al. [6], a 3D structure is considered as valid if the distances between any pairwise atoms is larger than 0.5Å. In addition, we evaluate the similarity between the generated materials and the materials in the test set by five statistic metrics, including the earth mover s distance (EMD) between the distributions in chemical element number, density (g/cm3), and energy predicted by a GNN model. Also, we evaluate the percentage of the test set materials that cover at least one of the generated materials (COV-R), and the percentage of generated materials that cover at least one of the test set materials (COV-P). All seven metrics are evaluated on 10,000 generated periodic materials. See Appendix D.1 for more information about evaluation metrics and experimental settings in the random generation task. Table 2: Property optimization performance on Perov-5, Carbon-24, and MP-20 datasets. We report the success rate (SR), i.e., the rate of periodic materials that can achieve top 5%, 10%, and 15% of the property distribution in the dataset after being optimized. Cond-DFC-VAE can only be used to Perov-5 dataset in which materials have cubic structures. Bold and underline numbers highlight the best and second best performance, respectively. Method Perov-5 Carbon-24 MP-20 SR5 SR10 SR15 SR5 SR10 SR15 SR5 SR10 SR15 FTCP 0.06 0.11 0.16 0.00 0.00 0.00 0.02 0.04 0.05 Cond-DFC-VAE 0.55 0.64 0.69 CDVAE 0.52 0.65 0.79 0.00 0.06 0.06 0.78 0.86 0.90 Sy Mat (ours) 0.73 0.80 0.87 0.06 0.13 0.13 0.92 0.97 0.97 Results. The random generation performance results of all methods on three datasets are presented in Table 1. Among the total 19 metrics over three datasets in the table, our Sy Mat method achieves top-1 performance in 11 metrics and top-2 performance in 17 metrics. Particularly, Sy Mat can achieve excellent performance in composition and structure validity, which demonstrates that Sy Mat can accurately learn the rules of forming chemically valid and stable periodic materials so it generates chemically valid periodic materials with high chances. Overall, Sy Mat achieves promising performance in the random generation task, showing that Sy Mat is an effective approach to capture the distribution of complicated periodic material structures. We visualize some generated periodic materials in Figure 3 of Appendix E, and use additional metrics to evaluate the diversity of the generated periodic materials in Appendix D.3. To demonstrate the effectiveness of coordinate generation module in Sy Mat, we conduct an ablation study using several evaluation metrics in the random generation task and present this part in Appendix D.4. 4.3 Property Optimization Metrics. In the property optimization task, we evaluate the ability of Sy Mat in discovering novel periodic materials with low energies. We jointly train the Sy Mat model with an energy prediction model that predicts the energies from the latent representations. Afterwards, we follow the evaluation procedure in [52] to optimize the latent representations of testing materials with 5000 gradient descent steps, decode 100 periodic materials from the optimized latent representations to periodic materials, and report the success rates (SR), i.e., the rate of decoded materials whose properties are in the top 5%, 10%, and 15% of the energy distribution in the dataset. We use the same evaluation procedure to baseline methods and compare them with Sy Mat in optimization success rates. Results. The property optimization performance results of all methods on three datasets are summarized in Table 2. On three datasets, Sy Mat outperforms all baseline methods in optimization success rates, showing that Sy Mat has the highest chances to generate periodic materials with low energies. This demonstrates that Sy Mat models have strong capacities in discovering novel periodic materials with good properties. The good performance of Sy Mat in the property optimization method shows that it can be an effective tool for designing novel periodic materials with desired properties. We visualize the changes of material structures and energies of some periodic materials after property optimization in Figure 4 of Appendix E. We also visualize the material energy distribution before and after property optimization by Sy Mat models in Figure 5 of Appendix E. 5 Conclusion We propose Sy Mat, a novel deep generative model for periodic material generation. Sy Mat achieves invariance to physical symmetry transformations of periodic materials, including permutation, rotation, translation, and periodic transformations. In Sy Mat, atom types and lattices of materials are generated in the form of atom type sets, lattice lengths and lattice angles with a VAE model. Additionally, Sy Mat uses a score-based diffusion model based on a symmetry-aware coordinate diffusion process to generate atom coordinates. Experiments on random generation and property optimization tasks show that Sy Mat is a promising approach for discovering novel materials. In the future, we will study the problem of accelerating atom coordinate generation in Sy Mat. Acknowledgments and Disclosure of Funding We thank Xiang Fu for his help on providing technical details about running baseline methods. This work was supported in part by National Science Foundation under grants IIS-1908220 and IIS-2006861. [1] Keith T Butler, Daniel W Davies, Hugh Cartwright, Olexandr Isayev, and Aron Walsh. Machine learning for molecular and materials science. Nature, 559(7715):547 555, 2018. [2] Ivano E Castelli, David D Landis, Kristian S Thygesen, Søren Dahl, Ib Chorkendorff, Thomas F Jaramillo, and Karsten W Jacobsen. New cubic perovskites for one-and two-photon water splitting using the computational materials repository. Energy & Environmental Science, 5(10): 9034 9043, 2012. [3] Ivano E Castelli, Thomas Olsen, Soumendu Datta, David D Landis, Søren Dahl, Kristian S Thygesen, and Karsten W Jacobsen. Computational screening of perovskite metal oxides for optimal solar light capture. Energy & Environmental Science, 5(2):5814 5819, 2012. [4] Chi Chen, Weike Ye, Yunxing Zuo, Chen Zheng, and Shyue Ping Ong. Graph networks as a universal machine learning framework for molecules and crystals. Chemistry of Materials, 31 (9):3564 3572, 2019. [5] Kamal Choudhary and Brian De Cost. Atomistic line graph neural network for improved materials property predictions. npj Computational Materials, 7(1):1 8, 2021. [6] Callum J Court, Batuhan Yildirim, Apoorv Jain, and Jacqueline M Cole. 3-D inorganic crystal structure generation and property prediction via representation learning. Journal of chemical information and modeling, 60(10):4518 4535, 2020. [7] Yabo Dan, Yong Zhao, Xiang Li, Shaobo Li, Ming Hu, and Jianjun Hu. Generative adversarial networks (GAN) based efficient sampling of chemical composition space for inverse design of inorganic materials. npj Computational Materials, 6(1):1 7, 2020. [8] Daniel W Davies, Keith T Butler, Adam J Jackson, Jonathan M Skelton, Kazuki Morita, and Aron Walsh. SMACT: Semiconducting materials by analogy and chemical theory. Journal of Open Source Software, 4(38):1361, 2019. [9] Mildred S Dresselhaus, Gene Dresselhaus, and Ado Jorio. Group theory: application to the physics of condensed matter. Springer Science & Business Media, 2007. [10] Niklas Gebauer, Michael Gastegger, and Kristof Schütt. Symmetry-adapted generation of 3d point sets for the targeted discovery of molecules. In H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 32. Curran Associates, Inc., 2019. [11] Justin Gilmer, Samuel S. Schoenholz, Patrick F. Riley, Oriol Vinyals, and George E. Dahl. Neural message passing for quantum chemistry. In Doina Precup and Yee Whye Teh, editors, Proceedings of the 34th International Conference on Machine Learning, volume 70 of Proceedings of Machine Learning Research, pages 1263 1272, International Convention Centre, Sydney, Australia, 2017. [12] Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial nets. In Z. Ghahramani, M. Welling, C. Cortes, N. Lawrence, and K. Q. Weinberger, editors, Advances in Neural Information Processing Systems, volume 27, pages 2672 2680. Curran Associates, Inc., 2014. [13] Jonathan Ho, Ajay Jain, and Pieter Abbeel. Denoising diffusion probabilistic models. In H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 6840 6851. Curran Associates, Inc., 2020. URL https://proceedings.neurips.cc/paper/2020/file/ 4c5bcfec8584af0d967f1ab10179ca4b-Paper.pdf. [14] Jordan Hoffmann, Louis Maestrati, Yoshihide Sawada, Jian Tang, Jean Michel Sellier, and Yoshua Bengio. Data-driven approach to encoding and decoding 3-D crystal structures. ar Xiv preprint ar Xiv:1909.00949, 2019. [15] Emiel Hoogeboom, Víctor Garcia Satorras, Clément Vignac, and Max Welling. Equivariant diffusion for molecule generation in 3D. In Kamalika Chaudhuri, Stefanie Jegelka, Le Song, Csaba Szepesvari, Gang Niu, and Sivan Sabato, editors, Proceedings of the 39th International Conference on Machine Learning, volume 162 of Proceedings of Machine Learning Research, pages 8867 8887. PMLR, 17 23 Jul 2022. URL https://proceedings.mlr.press/v162/hoogeboom22a. html. [16] John Ingraham, Max Baranov, Zak Costello, Vincent Frappier, Ahmed Ismail, Shan Tie, Wujie Wang, Vincent Xue, Fritz Obermeyer, Andrew Beam, and Gevorg Grigoryan. Illuminating protein space with a programmable generative model. bio Rxiv, 2022. doi: 10.1101/2022.12.01. 518682. URL https://www.biorxiv.org/content/early/2022/12/02/2022.12.01.518682. [17] Anubhav Jain, Shyue Ping Ong, Geoffroy Hautier, Wei Chen, William Davidson Richards, Stephen Dacek, Shreyas Cholia, Dan Gunter, David Skinner, Gerbrand Ceder, et al. Commentary: The materials project: A materials genome approach to accelerating materials innovation. APL materials, 1(1):011002, 2013. [18] Rui Jiao, Wenbing Huang, Peijia Lin, Jiaqi Han, Pin Chen, Yutong Lu, and Yang Liu. Crystal structure prediction by joint equivariant diffusion on lattices and fractional coordinates. In Workshop on Machine Learning for Materials ICLR 2023, 2023. URL https://openreview. net/forum?id=VPByphdu24j. [19] Wengong Jin, Regina Barzilay, and Tommi Jaakkola. Junction tree variational autoencoder for molecular graph generation. In Jennifer Dy and Andreas Krause, editors, Proceedings of the 35th International Conference on Machine Learning, volume 80 of Proceedings of Machine Learning Research, pages 2323 2332, 2018. [20] Bowen Jing, Gabriele Corso, Jeffrey Chang, Regina Barzilay, and Tommi S. Jaakkola. Torsional diffusion for molecular conformer generation. In Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho, editors, Advances in Neural Information Processing Systems, 2022. URL https://openreview.net/forum?id=w6fj2r62r_H. [21] Sungwon Kim, Juhwan Noh, Geun Ho Gu, Alan Aspuru-Guzik, and Yousung Jung. Generative adversarial networks for crystal structure prediction. ACS central science, 6(8):1412 1420, 2020. [22] Diederik P Kingma and Max Welling. Auto-Encoding variational bayes. In 2nd International Conference on Learning Representations, 2014. [23] Diederik P Kingma, Tim Salimans, Ben Poole, and Jonathan Ho. On density estimation with diffusion models. In A. Beygelzimer, Y. Dauphin, P. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, 2021. URL https://openreview.net/forum? id=2Ld Bqxc1Yv. [24] Zhifeng Kong, Wei Ping, Jiaji Huang, Kexin Zhao, and Bryan Catanzaro. Diff Wave: A versatile diffusion model for audio synthesis. In International Conference on Learning Representations, 2021. URL https://openreview.net/forum?id=a-x FK8Ymz5J. [25] Yuchao Lin, Keqiang Yan, Youzhi Luo, Yi Liu, Xiaoning Qian, and Shuiwang Ji. Efficient approximations of complete interatomic potentials for crystal property prediction. In Andreas Krause, Emma Brunskill, Kyunghyun Cho, Barbara Engelhardt, Sivan Sabato, and Jonathan Scarlett, editors, Proceedings of the 40th International Conference on Machine Learning, volume 202 of Proceedings of Machine Learning Research, pages 21260 21287. PMLR, 23 29 Jul 2023. URL https://proceedings.mlr.press/v202/lin23m.html. [26] Meng Liu, Youzhi Luo, Limei Wang, Yaochen Xie, Hao Yuan, Shurui Gui, Haiyang Yu, Zhao Xu, Jingtun Zhang, Yi Liu, Keqiang Yan, Haoran Liu, Cong Fu, Bora M Oztekin, Xuan Zhang, and Shuiwang Ji. DIG: A turnkey library for diving into graph deep learning research. Journal of Machine Learning Research, 22(240):1 9, 2021. URL http://jmlr.org/papers/v22/21-0343.html. [27] Meng Liu, Keqiang Yan, Bora Oztekin, and Shuiwang Ji. Graph EBM: Molecular graph generation with energy-based models. In Energy Based Models Workshop-ICLR 2021, 2021. [28] Meng Liu, Youzhi Luo, Kanji Uchino, Koji Maruhashi, and Shuiwang Ji. Generating 3D molecules for target protein binding. In Kamalika Chaudhuri, Stefanie Jegelka, Le Song, Csaba Szepesvari, Gang Niu, and Sivan Sabato, editors, Proceedings of the 39th International Conference on Machine Learning, volume 162 of Proceedings of Machine Learning Research, pages 13912 13924. PMLR, 17 23 Jul 2022. URL https://proceedings.mlr.press/v162/liu22m. html. [29] Yi Liu, Limei Wang, Meng Liu, Yuchao Lin, Xuan Zhang, Bora Oztekin, and Shuiwang Ji. Spherical message passing for 3D molecular graphs. In International Conference on Learning Representations, 2022. URL https://openreview.net/forum?id=givs RXs Ot9r. [30] Shitong Luo, Chence Shi, Minkai Xu, and Jian Tang. Predicting molecular conformation via dynamic graph score matching. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, volume 34, pages 19784 19795. Curran Associates, Inc., 2021. URL https://proceedings.neurips.cc/paper/ 2021/file/a45a1d12ee0fb7f1f872ab91da18f899-Paper.pdf. [31] Youzhi Luo and Shuiwang Ji. An autoregressive flow model for 3D molecular geometry generation from scratch. In International Conference on Learning Representations, 2022. URL https://openreview.net/forum?id=C03Ajc-NS5W. [32] Youzhi Luo, Keqiang Yan, and Shuiwang Ji. Graph DF: A discrete flow model for molecular graph generation. In Marina Meila and Tong Zhang, editors, Proceedings of the 38th International Conference on Machine Learning, volume 139 of Proceedings of Machine Learning Research, pages 7192 7203. PMLR, 18 24 Jul 2021. [33] Elman Mansimov, Omar Mahmood, Seokho Kang, and Kyunghyun Cho. Molecular geometry prediction using a deep generative graph neural network. Scientific reports, 9(1):1 13, 2019. [34] Yashaswi Pathak, Karandeep Singh Juneja, Girish Varma, Masahiro Ehara, and U Deva Priyakumar. Deep learning enabled inorganic material generator. Physical Chemistry Chemical Physics, 22(46):26935 26943, 2020. [35] Chris J. Pickard. Airss data for carbon at 10gpa and the c+n+h+o system at 1gpa, 2020. URL https://archive.materialscloud.org/record/2020.0026/v1. [36] Zekun Ren, Siyu Isaac Parker Tian, Juhwan Noh, Felipe Oviedo, Guangzong Xing, Jiali Li, Qiaohao Liang, Ruiming Zhu, Armin G. Aberle, Shijing Sun, Xiaonan Wang, Yi Liu, Qianxiao Li, Senthilnath Jayavelu, Kedar Hippalgaonkar, Yousung Jung, and Tonio Buonassisi. An invertible crystallographic representation for general inverse design of inorganic crystals with targeted properties. Matter, 5(1):314 335, 2022. ISSN 2590-2385. doi: https://doi.org/10.1016/j.matt. 2021.11.032. URL https://www.sciencedirect.com/science/article/pii/S2590238521006251. [37] Robin Rombach, Andreas Blattmann, Dominik Lorenz, Patrick Esser, and Björn Ommer. Highresolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10684 10695, 2022. [38] Chitwan Saharia, William Chan, Saurabh Saxena, Lala Li, Jay Whang, Emily Denton, Seyed Kamyar Seyed Ghasemipour, Raphael Gontijo-Lopes, Burcu Karagol Ayan, Tim Salimans, Jonathan Ho, David J. Fleet, and Mohammad Norouzi. Photorealistic text-to-image diffusion models with deep language understanding. In Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho, editors, Advances in Neural Information Processing Systems, 2022. URL https://openreview.net/forum?id=08Yk-n5l2Al. [39] Victor Garcia Satorras, Emiel Hoogeboom, Fabian Bernd Fuchs, Ingmar Posner, and Max Welling. E(n) equivariant normalizing flows. In A. Beygelzimer, Y. Dauphin, P. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, 2021. URL https://openreview.net/forum?id=N5h QI_Row VA. [40] Jonathan Schmidt, Love Pettersson, Claudio Verdozzi, Silvana Botti, and Miguel A. L. Marques. Crystal graph attention networks for the prediction of stable materials. Science Advances, 7(49): eabi7948, 2021. doi: 10.1126/sciadv.abi7948. [41] Chence Shi*, Minkai Xu*, Zhaocheng Zhu, Weinan Zhang, Ming Zhang, and Jian Tang. Graph AF: a flow-based autoregressive model for molecular graph generation. In 8th International Conference on Learning Representations, 2020. [42] Chence Shi, Shitong Luo, Minkai Xu, and Jian Tang. Learning gradient fields for molecular conformation generation. In Marina Meila and Tong Zhang, editors, Proceedings of the 38th International Conference on Machine Learning, volume 139 of Proceedings of Machine Learning Research, pages 9558 9568. PMLR, 18 24 Jul 2021. [43] Gregor Simm and Jose Miguel Hernandez-Lobato. A generative model for molecular distance geometry. In Hal Daumé III and Aarti Singh, editors, Proceedings of the 37th International Conference on Machine Learning, volume 119 of Proceedings of Machine Learning Research, pages 8949 8958. PMLR, 13 18 Jul 2020. [44] Yang Song and Stefano Ermon. Generative modeling by estimating gradients of the data distribution. In H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 32. Curran Associates, Inc., 2019. URL https://proceedings.neurips.cc/paper/2019/file/ 3001ef257407d5a371a96dcd947c7d93-Paper.pdf. [45] Yang Song, Jascha Sohl-Dickstein, Diederik P Kingma, Abhishek Kumar, Stefano Ermon, and Ben Poole. Score-based generative modeling through stochastic differential equations. In International Conference on Learning Representations, 2021. URL https://openreview.net/ forum?id=Px TIG12RRHS. [46] Pascal Vincent. A connection between score matching and denoising autoencoders. Neural computation, 23(7):1661 1674, 2011. [47] Shiyu Wang, Xiaojie Guo, and Liang Zhao. Deep generative model for periodic graphs. In Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho, editors, Advances in Neural Information Processing Systems, 2022. URL https://openreview.net/forum?id= lg NGDj WRTo-. [48] Joseph L Watson, David Juergens, Nathaniel R Bennett, Brian L Trippe, Jason Yim, Helen E Eisenach, Woody Ahern, Andrew J Borst, Robert J Ragotte, Lukas F Milles, et al. De novo design of protein structure and function with rfdiffusion. Nature, 620(7976):1089 1100, 2023. [49] Max Welling and Yee Whye Teh. Bayesian learning via stochastic gradient Langevin dynamics. In Proceedings of the 28th International Conference on International Conference on Machine Learning, ICML 11, page 681 688, Madison, WI, USA, 2011. Omnipress. ISBN 9781450306195. [50] Lemeng Wu, Chengyue Gong, Xingchao Liu, Mao Ye, and qiang liu. Diffusion-based molecule generation with informative prior bridges. In Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho, editors, Advances in Neural Information Processing Systems, 2022. URL https://openreview.net/forum?id=TJUNti Zi TKE. [51] Tian Xie and Jeffrey C Grossman. Crystal graph convolutional neural networks for an accurate and interpretable prediction of material properties. Physical review letters, 120(14):145301, 2018. [52] Tian Xie, Xiang Fu, Octavian-Eugen Ganea, Regina Barzilay, and Tommi S. Jaakkola. Crystal diffusion variational autoencoder for periodic material generation. In International Conference on Learning Representations, 2022. URL https://openreview.net/forum?id=03RLpj-tc_. [53] Minkai Xu, Shitong Luo, Yoshua Bengio, Jian Peng, and Jian Tang. Learning neural generative dynamics for molecular conformation generation. In International Conference on Learning Representations, 2021. URL https://openreview.net/forum?id=p Abm1qfhe Gk. [54] Minkai Xu, Wujie Wang, Shitong Luo, Chence Shi, Yoshua Bengio, Rafael Gomez-Bombarelli, and Jian Tang. An end-to-end framework for molecular conformation generation via bilevel programming. In Marina Meila and Tong Zhang, editors, Proceedings of the 38th International Conference on Machine Learning, volume 139 of Proceedings of Machine Learning Research, pages 11537 11547. PMLR, 18 24 Jul 2021. [55] Minkai Xu, Lantao Yu, Yang Song, Chence Shi, Stefano Ermon, and Jian Tang. Geo Diff: A geometric diffusion model for molecular conformation generation. In International Conference on Learning Representations, 2022. URL https://openreview.net/forum?id=Pzcvx EMzv QC. [56] Zhao Xu, Youzhi Luo, Xuan Zhang, Xinyi Xu, Yaochen Xie, Meng Liu, Kaleb Dickerson, Cheng Deng, Maho Nakata, and Shuiwang Ji. Molecule3D: A benchmark for predicting 3D geometries from molecular graphs. ar Xiv preprint ar Xiv:2110.01717, 2021. [57] Keqiang Yan, Yi Liu, Yuchao Lin, and Shuiwang Ji. Periodic graph transformers for crystal material property prediction. In Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho, editors, Advances in Neural Information Processing Systems, 2022. URL https://openreview.net/forum?id=pq CT3L-BU9T. [58] Jiaxuan You, Bowen Liu, Zhitao Ying, Vijay Pande, and Jure Leskovec. Graph convolutional policy network for goal-directed molecular graph generation. In S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 31, pages 6410 6421. Curran Associates, Inc., 2018. [59] Anthony Zee. Group theory in a nutshell for physicists, volume 17. Princeton University Press, 2016. [60] Xuan Zhang, Limei Wang, Jacob Helwig, Youzhi Luo, Cong Fu, Yaochen Xie, Meng Liu, Yuchao Lin, Zhao Xu, Keqiang Yan, Keir Adams, Maurice Weiler, Xiner Li, Tianfan Fu, Yucheng Wang, Haiyang Yu, Yu Qing Xie, Xiang Fu, Alex Strasser, Shenglong Xu, Yi Liu, Yuanqi Du, Alexandra Saxton, Hongyi Ling, Hannah Lawrence, Hannes Stärk, Shurui Gui, Carl Edwards, Nicholas Gao, Adriana Ladera, Tailin Wu, Elyssa F. Hofgard, Aria Mansouri Tehrani, Rui Wang, Ameya Daigavane, Montgomery Bohde, Jerry Kurtin, Qian Huang, Tuong Phung, Minkai Xu, Chaitanya K. Joshi, Simon V. Mathis, Kamyar Azizzadenesheli, Ada Fang, Alán Aspuru-Guzik, Erik Bekkers, Michael Bronstein, Marinka Zitnik, Anima Anandkumar, Stefano Ermon, Pietro Liò, Rose Yu, Stephan Günnemann, Jure Leskovec, Heng Ji, Jimeng Sun, Regina Barzilay, Tommi Jaakkola, Connor W. Coley, Xiaoning Qian, Xiaofeng Qian, Tess Smidt, and Shuiwang Ji. Artificial intelligence for science in quantum, atomistic, and continuum systems. ar Xiv preprint ar Xiv:2307.08423, 2023. [61] Yong Zhao, Mohammed Al-Fahdi, Ming Hu, Edirisuriya MD Siriwardane, Yuqi Song, Alireza Nasiri, and Jianjun Hu. High-throughput discovery of novel cubic crystal materials using deep generative neural networks. Advanced Science, 8(20):2100566, 2021. A Proof of Propositions A.1 Proof of Proposition 3.1 We first prove Proposition 3.1, i.e., if the material distribution p(M) is invariant to permutation, rotation, translation, and periodic transformations described in Section 3.1, the score matrix P log p(P |A, L) will be invariant to translation and periodic transformations, and equivariant to permutation and rotation transformations. (1) For M = (A , P , L) where A and P are obtained by permuting the elements of A and column vectors of P with the same order, from p(M) = p(M ), we know that p(P |A , L) = p(P |A, L) holds, so P log p(P |A , L) actually equals to the matrix obtained by permuting the column vectors of P log p(P |A, L) with the same permutation order. In other words, P log p(P |A, L) is permutation-equivariant. (2) For M = (A, P , L ) obtained by rotating M = (A, P , L) with the rotation matrix Q and translating it with the translation vector b, because p(M ) = p(M) and P = QP + b1T , we have p(P |A, L ) = p(P |A, L) and P log p(P |A, L ) = Q P log p(P |A, L). In other words, P log p(P |A, L) is translation-invariant and rotation-equivariant. (3) For M K = (A, P K, L) where P K = P + LK is obtained by periodic transformation, since p(M) = p(M K), we can easily find that p(P K|A, L) = p(P |A, L) and P K log p(P K|A, L) = P log p(P |A, L) hold. In other words, P log p(P |A, L) is invariant to periodic transformations. A.2 Proof of Proposition 3.2 We next prove Proposition 3.2, i.e., if si,j,k is always invariant to permutation, rotation, translation, and periodic transformations described in Section 3.1, the score matrix P log p(P |A, L) computed by Equation (3) will always be invariant to translation and periodic transformations, and equivariant to permutation and rotation transformations. (1) The score matrix P log p(P |A, L) = [s1, ..., sn] is formed by stacking coordinate score vector si according to the atom order, and this stacking operation is clearly equivariant to permutations. Also, from Equation (3), we can easily find that if si,j,k is always invariant to permutation, the computed coordinate score vector si for each coordinate pi will also be invariant to permutations. Hence, P log p(P |A, L) from s1, ..., sn is equivariant to permutations. (2) For any Q R3 3, QT Q = I and b R3, let P = [p 1, ..., p n] = QP + b1T where p i = Qpi + b, i = 1, ..., n, L = QL, and M = (A, P , L ). We can calculate the score vector s i for p i in the way similar to Equation (3) as s i = p i log p(P |A, L ) (j,k) N (i) d i,j,k log p(P |A, L ) p id i,j,k (j,k) N (i) s i,j,k p i + L k p j ||p i + L k p j||2 , (j,k) N (i) s i,j,k Qpi + b + QLk Qpj b ||Qpi + b + QLk Qpj b||2 , (j,k) N(i) si,j,k Q(pi + Lk pj) ||pi + Lk pj||2 , where N (i) = {(j, k) : (i, j, k) E(M )}, N(i) = {(j, k) : (i, j, k) E(M)}. Note that the second to the last equation holds because ||Qx||2 = ||x||2 holds for any x R3 if QT Q = I, and the edge set of multi-graph representation and si,j,k is invariant to any rotation and translation transformations, hence E(M ) = E(M) and s i,j,k = si,j,k. So far, we have shown that P log p(P |A, L) is invariant to translations and equivariant to rotations. (3) For any K Z3 3, let P K = [p 1, ..., p n]P +LK and M K = A, P K, L . Because the multigraph method is strictly invariant to periodic transformations, there must exist a bijective mapping f : E(M) E M K such that for any (i, j, k) E(M), (i, j, k ) = f(i, j, k) satisfies that pi + Lk pj = p i + Lk p j. Also, because si,j,k is invariant to periodic transformations, we can easily find that calculating the score vectors for pi and p i by Equation (3) are actually summing up the same set of vectors, so these two score vectors equal and then P log p(P |A, L) is invariant to periodic transformations. B Additional Implementation Details Sphere Net Model. In Sy Mat, we use Sphere Net model [29] as the backbone GNN architecture for both the encoder model in the VAE for atom type and lattice generation and the score model in the score-based diffusion model for atom coordinate generation. Sphere Net is a powerful 3D GNN model proposed for 3D molecular property prediction, but in our method, we will use the multi-graph representations of materials as inputs to Sphere Net models. Generally, Sphere Net model first initializes the embedding vectors of nodes and edges with a look-up atom type embedding vectors and spherical basis functions, then employ multiple message passing [11] layers to iteratively update the node and edge embeddings. Note that this overall process of initializing and updating embeddings is always invariant to all symmetry transformations described in Section 3.1. The encoder model of VAE model will use a sum pooling over the final node embeddings to obtain a global representation vector, which is passed into an MLP to obtain z A and z L. The score model will use another MLP to predict the edge distance score scalar from the final edge embeddings. Note that we use two independent Sphere Net models whose parameters are not shared for the VAE encoder model and score model. We use the implementations of Sphere Net in DIG package [26]. Obtaining lattice matrix from lattice lengths and angles. From the lattice lengths ℓ= [ℓ1, ℓ2, ℓ3] and lattice angles ϕ = [ϕ12, ϕ23, ϕ13], let γ = arccos cos ϕ23 cos ϕ13 cos ϕ12 sin ϕ23 sin ϕ13 , one lattice matrix L = [ℓ1, ℓ2, ℓ3] can be computed as ℓ1 = [ℓ1 sin ϕ13, 0, ℓ1 cos ϕ13]T , ℓ2 = [ ℓ2 sin ϕ23 cos γ, ℓ2 sin ϕ23 sin γ, ℓ2 cos ϕ23]T , ℓ3 = [0, 0, ℓ3]T . Calculating the denoising score matrix. In the training of score-based diffusion model in Sy Mat, we add Gaussian noise with noise magnitude σt to the coordinate matrix of a material M = (A, P , L) in the dataset to obtain a noisy material f M = A, e P , L , and the score model sθ( ) is optimized to predict the denoising score matrix e P log pσt e P |P , A, L . Then the score model sθ( ) tries to predict the edge distance score si,j,k = di,j,k log pσt e P |P , A, L for every edge (i, j, k) in the multi-graph representation of f M. Motivated from existing methods [30, 42, 55] in molecule generation, we can assume that the distribution of edge distance di,j,k in f M is a Gaussian distribution centered around the edge distance of the same edge in M. Practically, we find that using a new coordinate matrix b P = [ˆp1, ..., ˆpn] obtained by aligning P = [p1, ..., pn] w.r.t. e P = [ p1, ..., pn] to calculate the corresponding edge distance in M can achieve the best performance. Specifically, we follow alignment procedure as in CDVAE [52] and calculate each ˆpi as ˆpi = pi + Lu, where u = arg minv Z3 ||pi+Lv pi||2. The mean ˆdi,j,k of the Gaussian distribution of the edge distance di,j,k is calculated as ˆdi,j,k = ||ˆpi + Lk ˆpj||2. Besides, we assume the Gaussian distributions of ˆdi,j,k for every edge have the same standard deviations σt, which is pre-computed by manually perturbing all materials in the dataset to noisy materials, collecting the edge distances in all noisy materials, and calculating the empirical standard deviation as σt. C More Discussions Limitations. Generally, there are three major limitations in our proposed Sy Mat approach. (1) First, Sy Mat may take as long as an hour to generate atom coordinates of periodic materials with score-based diffusion models, since this process requires to run thousands of Langevin dynamics sampling steps. Actually, this is a common limitation for many other molecule or material generation methods [42, 30, 52] using score-based diffusion models. Nonetheless, we argue that compared with the huge time cost of laborious lab experiments or DFT based calculation, the time cost of Sy Mat is relatively lower and acceptable. We will explore accelerating Sy Mat with SDE based diffusion models [45, 23] in the future. (2) Second, Sy Mat is designed to maintain invariance to periodic transformations, so it cannot be applied to non-periodic materials as this will lead to incorrect probabilistic modeling. Though some materials are non-periodic, such as amorphous solids and polymers, we focus on periodic materials because they occupy a large part of materials used in real-world applications. We believe that developing non-periodic material generation method need to rely on effective periodic material generation methods, but the problem of generating relatively simpler periodic materials still remains challenging and largely unsolved. Hence, we will focus on generating periodic materials in this work and leaves the problem of generating non-periodic materials to the future. (3) Third, our work follows previous material generation studies [52] to evaluate the quality of the generated materials by some basic metrics, such as composition and structure validity. However, they do not use metrics to evaluate the synthesizability of the generated materials, though these metrics are useful for practical applications. In the future, we will collaborate with experts in material science to come up with metrics for evaluating synthesizability of the materials generated by our method. Broader impacts. In this work, we propose a novel method for periodic material generation. Our work can be used to design novel materials with desired properties for many real-world applications, such as batteries or solar cells. Nonetheless, our method may unexpectedly generate materials that produce negative impacts to human life. For example, manufacturing some materials generated by our model may cause environmental pollution. Hence, we believe strict evaluation or validation process need to be done to estimate the related environmental or other social impacts before using the materials generated by our model to real-world scenarios. D Additional Experiment Details and Results D.1 Additional Details in Experimental Settings In our Sy Mat model, the Sphere Net model is composed of 4 message passing layers and the hidden size is set to 128. Also, all MLP models in the VAE decoder is composed of 2 linear layers with a Re LU function between them, and the hidden size is set to 256. During training, we set the learning rate to 0.001, the batch size to 128, and the epoch number to 1,000. The total training needs around 3, 5, and 10 hours on a GTX 2080 GPU for Perov-5, Carbon-24, and MP-20 datasets, separately. We use different weights for different loss terms. Specifically, we set the weights of the reconstruction losses for atom type set size, atom types, atom numbers of each atom type, lattice items as 1.0, 30.0, 1.0, 10.0, respectively, and use a weight of 0.01 for KL-distance loss and a weight of 10.0 for denoising score matching loss. In the generation of atom coordinates with Algorithm 1, we set step size ϵ as 0.0001 and step number q as 100. In addition, we use a geometrically decreasing series between 10 and 0.01 as {σt}T t=1 and set T = 50. In the random generation task, for COV-R and COV-P metrics, we consider one material cover another material if the distances of their chemical element composition fingerprints and 3D structure fingerprints are smaller than the thresholds δc and δs, respectively. For Perov-5 dataset, we use δc = 6, δs = 0.8; for Carbon-24 dataset, we use δc = 4, δs = 1.0; for MP-24 dataset, we use δc = 12, δs = 0.6. D.2 Material Reconstruction In the material reconstruction task, we reconstruct materials in the test sets of Perov-5, Carbon-24 and MP-20 datasets from the encoded latent representations z A and z L. We evaluate the reconstruction performance by several metrics. First, we evaluate the percentage of materials whose atom types can be exactly reconstructed (Atom Type Match Rate), and the root mean square error in lattice lengths and angles between the target and reconstructed materials (Lattice RMSE). Second, to show the score-based diffusion model has the capacity to approximately reconstruct 3D structures, we evaluate the root mean square error in interatomic distances between the target and reconstructed materials (Distance RMSE). We present the experimental results of Sy Mat together with baseline methods in Table 3. Results show that our method performs well in atom types, lattices and interatomic distances reconstruction. Table 3: Reconstruction performance on Perov-5, Carbon-24, and MP-20 datasets. Here means higher metric values lead to better performance, while means the opposite. Note that Cond-DFCVAE can only be used to Perov-5 dataset in which materials have cubic structures. Bold and underline numbers highlight the best and second best performance, respectively. Dataset Method Atom type match rate Lattice RMSE Distance RMSE FTCP 99.67% 0.0786 0.2953 Cond-DFC-VAE 58.92% 0.0765 0.2281 CDVAE 98.16% 0.0231 0.1072 Sy Mat (ours) 98.30% 0.0224 0.0723 FTCP 100.00% 0.1349 0.3759 CDVAE 100.00% 0.0624 0.1745 Sy Mat (ours) 100.00% 0.0632 0.1186 FTCP 71.46% 0.1057 0.2062 CDVAE 68.32% 0.0532 0.0975 Sy Mat (ours) 72.10% 0.0510 0.0876 D.3 Diversity Evaluation We evaluate the diversity of the randomly generated atom types by uniqueness percentage, i.e., the percentage of unique atom types among all the generated atom types. We present the uniqueness percentages of Sy Mat and baseline methods on Perov-5 and MP-20 datasets in Table 4. Note that Cond-DFC-VAE can only be used to Perov-5 dataset in which materials have cubic structures, and Carbon-24 is not used here because all materials in Carbon-24 are composed of only carbon atoms, so the uniqueness is expected to be low. Results show that our method performs well in generating materials with diverse atom types. Table 4: Uniqueness percentages on Perov-5 and MP-20 datasets. Dataset FTCP Cond-DFC-VAE G-Sch Net P-G-Sch Net CDVAE Sy Mat Perov-5 72.33% 80.36% 98.74% 98.57% 98.61% 99.43% MP-20 79.04% 99.23% 99.03% 99.84% 99.98% D.4 Ablation Study We conduct an ablation study experiment to show that the use of distance score matching in our coordinate generation module is significant. We implement a variant of Sy Mat in which the score model in coordinate generation module is trained by applying the denoising score matching loss directly to coordinates, instead of using our proposed distance score matching loss. We evaluate its performance in the random generation task by the structural validity, COV-R, and COV-P metrics as these metrics evaluate the quality of the generated atom coordinates. We summarize the results in Table 5. From the results, we can clearly find that using coordinate score matching always achieves lower COV-R and COV-P than using distance score matching. This demonstrates that distance score matching can more accurately capture the distribution of material structures, so the generated periodic materials are more similar to periodic materials in datasets, thereby achieving higher COV-R and COV-P. Table 5: Comparison of Sy Mat with coordinate score matching and distance score matching. Here means higher metric values lead to better performance. Bold numbers highlight the best best performance. Dataset Method Structure validity COV-R COV-P Perov-5 Sy Mat with coordinate score matching 100.00% 97.46% 95.32% Sy Mat with distance score matching 100.00% 99.68% 98.64% Carbon-24 Sy Mat with coordinate score matching 100.00% 96.72% 93.44% Sy Mat with distance score matching 100.00% 100.00% 97.59% MP-20 Sy Mat with coordinate score matching 100.00% 94.25% 95.61% Sy Mat with distance score matching 100.00% 98.97% 99.97% E Visualization Results Figure 3: Illustrations of some periodic materials generated by Sy Mat models trained on Perov-5, Carbon-24, and MP-20 datasets. -1.671 -0.237 -0.355 -0.521 Figure 4: Illustrations of some periodic materials before and after being optimized by Sy Mat models in the property optimization task. Numbers show how energies (e V/atom) changes in optimization. 2 0 2 4 6 Energies Before Optimization After Optimization (a) Perov-5 Dataset. 155.0 154.8 154.6 154.4 154.2 154.0 Energies Before Optimization After Optimization (b) Carbon-24 Dataset. 8 6 4 2 0 2 4 Energies Before Optimization After Optimization (c) MP-20 Dataset. Figure 5: Illustrations of material energy (e V/atom) distribution before and after being optimized by Sy Mat models in the property optimization task.