# variational_graph_recurrent_neural_networks__d59d53ae.pdf Variational Graph Recurrent Neural Networks Ehsan Hajiramezanali , Arman Hasanzadeh , Nick Duffield , Krishna Narayanan , Mingyuan Zhou , Xiaoning Qian Department of Electrical and Computer Engineering, Texas A&M University {ehsanr, armanihm, duffieldng, krn, xqian}@tamu.edu Mc Combs School of Business, The University of Texas at Austin mingyuan.zhou@mccombs.utexas.edu Representation learning over graph structured data has been mostly studied in static graph settings while efforts for modeling dynamic graphs are still scant. In this paper, we develop a novel hierarchical variational model that introduces additional latent random variables to jointly model the hidden states of a graph recurrent neural network (GRNN) to capture both topology and node attribute changes in dynamic graphs. We argue that the use of high-level latent random variables in this variational GRNN (VGRNN) can better capture potential variability observed in dynamic graphs as well as the uncertainty of node latent representation. With semi-implicit variational inference developed for this new VGRNN architecture (SIVGRNN), we show that flexible non-Gaussian latent representations can further help dynamic graph analytic tasks. Our experiments with multiple real-world dynamic graph datasets demonstrate that SI-VGRNN and VGRNN consistently outperform the existing baseline and state-of-the-art methods by a significant margin in dynamic link prediction. 1 Introduction Node embedding maps each node in a graph to a vector in a low-dimensional latent space, in which classical feature vector-based machine learning formulations can be adopted [5]. Most of the existing node embedding techniques assume that the graph is static and that learning tasks are performed on fixed sets of nodes and edges [19, 23, 12, 20, 14, 1]. However, many real-world problems are modeled by dynamic graphs, where graphs are constantly evolving over time. Such graphs have been typically observed in social networks, citation networks, and financial transaction networks. A naive solution to node embedding for dynamic graphs is simply applying static methods to each snapshot of dynamic graphs. Among many potential problems of such a naive solution, it is clear that it ignores the temporal dependencies between snapshots. Several node embedding methods have been proposed to capture the temporal graph evolution for both networks without attributes [10, 26] and attributed networks [24, 16]. However, all of the existing dynamic graph embedding approaches represent each node by a deterministic vector in a low-dimensional space [2]. Such deterministic representations lack the capability of modeling uncertainty of node embedding, which is a natural consideration when having multiple information sources, i.e. node attributes and graph structure. In this paper, we propose a novel node embedding method for dynamic graphs that maps each node to a random vector in the latent space. More specifically, we first introduce a dynamic graph autoencoder model, namely graph recurrent neural network (GRNN), by extending the use of graph convolutional Both authors contributed equally. 33rd Conference on Neural Information Processing Systems (Neur IPS 2019), Vancouver, Canada. neural networks (GCRN) [21] to dynamic graphs. Then, we argue that GRNN lacks the expressive power for fully capturing the complex dependencies between topological evolution and time-varying node attributes because the output probability in standard RNNs is limited to either a simple unimodal distribution or a mixture of unimodal distributions [3, 22, 6, 8]. Next, to increase the expressive power of GRNN in addition to modeling the uncertainty of node latent representations, we propose variational graph recurrent neural network (VGRNN) by adopting high-level latent random variables in GRNN. Our proposed VGRNN is capable of learning interpretable latent representation as well as better modeling of very sparse dynamic graphs. To further boost the expressive power and interpretability of our new VGRNN method, we integrate semi-implicit variational inference [25] with VGRNN. We show that semi-implicit variational graph recurrent neural network (SI-VGRNN) is capable of inferring more flexible and complex posteriors. Our experiments demonstrate the superior performance of VGRNN and SI-VGRNN in dynamic link prediction tasks in several real-world dynamic graph datasets compared to baseline and state-of-the-art methods. 2 Background Graph convolutional recurrent networks (GCRN). GCRN was introduced by Seo et al. [21] to model time series data defined over nodes of a static graph. Series of frames in videos and spatio-temporal measurements on a network of sensors are two examples of such datasets. GCRN combines graph convolutional networks (GCN) [4] with recurrent neural networks (RNN) to capture spatial and temporal patterns in data. More precisely, given a graph G with N nodes, whose topology is determined by the adjacency matrix A RN N, and a sequence of node attributes X = {X(1), X(2), . . . , X(T )}, GCRN reads M-dimensional node attributes X(t) RN M and updates its hidden state ht Rp at each time step t: ht = f A, X(t), ht 1 . (1) Here f is a non-probabilistic deep neural network. It can be any recursive network including gated activation functions such as long short-term memory (LSTM) or gated recurrent units (GRU), where the deep layers inside them are replaced by graph convolutional layers. GCRN models node attribute sequences by parameterizing a factorization of the joint probability distribution as a product of conditional probabilities such that p X(1), X(2), . . . , X(T ) | A = t=1 p X(t) | X(