# rethinking_graph_lottery_tickets_graph_sparsity_matters__723b2244.pdf Published as a conference paper at ICLR 2023 RETHINKING GRAPH LOTTERY TICKETS: GRAPH SPARSITY MATTERS Bo Hui1, Da Yan2, Xiaolong Ma3, Wei-Shinn Ku1 1 Auburn University 2 The University of Alabama at Birmingham 3 Clemson University Lottery Ticket Hypothesis (LTH) claims the existence of a winning ticket (i.e., a properly pruned sub-network together with original weight initialization) that can achieve competitive performance to the original dense network. A recent work, called UGS, extended LTH to prune graph neural networks (GNNs) for effectively accelerating GNN inference. UGS simultaneously prunes the graph adjacency matrix and the model weights using the same masking mechanism, but since the roles of the graph adjacency matrix and the weight matrices are very different, we find that their sparsifications lead to different performance characteristics. Specifically, we find that the performance of a sparsified GNN degrades significantly when the graph sparsity goes beyond a certain extent. Therefore, we propose two techniques to improve GNN performance when the graph sparsity is high. First, UGS prunes the adjacency matrix using a loss formulation which, however, does not properly involve all elements of the adjacency matrix; in contrast, we add a new auxiliary loss head to better guide the edge pruning by involving the entire adjacency matrix. Second, by regarding unfavorable graph sparsification as adversarial data perturbations, we formulate the pruning process as a min-max optimization problem to gain the robustness of lottery tickets when the graph sparsity is high. We further investigate the question: Can the retrainable winning ticket of a GNN be also effective for graph transferring learning? We call it the transferable graph lottery ticket (GLT) hypothesis. Extensive experiments were conducted which demonstrate the superiority of our proposed sparsification method over UGS, and which empirically verified our transferable GLT hypothesis. 1 INTRODUCTION Graph Neural Networks (GNNs) (Kipf & Welling, 2017; Hamilton et al., 2017) have demonstrated state-of-the-art performance on various graph-based learning tasks. However, large graph size and over-parameterized network layers are factors that limit the scalability of GNNs, causing high training cost, slow inference speed, and large memory consumption. Recently, Lottery Ticket Hypothesis (LTH) (Frankle & Carbin, 2019) claims that there exists properly pruned sub-networks together with original weight initialization that can be retrained to achieve comparable performance to the original large deep neural networks. LTH has recently been extended to GNNs by Chen et al. (2021b), which proposes a unified GNN sparsification (UGS) framework that simultaneously prunes the graph adjacency matrix and the model weights to accelerate GNN inference on large graphs. Specifically, two differentiable masks mg and mθ are applied to the adjacency matrix A and the model weights Θ, respectively, during end-to-end training by element-wise product. After training, lowest-magnitude elements in mg and mθ are set to zero w.r.t. pre-defined ratios pg and pθ, which basically eliminates low-scored edges and weights, respectively. The weight parameters are then rewound to their original initialization, and this pruning process is repeated until pre-defined sparsity levels are reached, i.e., graph sparsity 1 mg 0 A 0 sg and weight sparsity 1 mθ 0 where . 0 is the L0 norm counting the number of non-zero elements. Intuitively, UGS simply extends the basic parameter-masking algorithm of Frankle & Carbin (2019) for identifying winning tickets to also mask and remove graph edges. However, our empirical study Published as a conference paper at ICLR 2023 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 Weight sparsity Accuracy(%) Graph sparsity=0.264 Weight sparsity=0.744 UGS Only weight 80% edges (a) Performance Study of UGS Variants Nodes in training set Neighbors within 2-hop ? Neighbors out of 2-hop Edges related to training loss Edges with zero grad (b) Types of Edges and Nodes in GNN Training Figure 1: UGS Analysis finds that the performance of a sparsified GNN degrades significantly when the graph sparsity goes beyond a certain level, while it is relatively insensitive to weight sparsification. Specifically, we compare UGS with its two variants: (1) Only weight, which does not conduct graph sparsification, and (2) 80% edges, which stops pruning edges as soon as the graph sparsity is increased to 20% or above. Figure 1(a) shows the performance comparison of UGS and the two variants on the Cora dataset (Chen et al., 2021b), where we can see that the accuracy of UGS (the red line) collapses when the graph sparsity becomes larger than 25%, while the two variants do not suffer from such a significant performance degradation since neither of them sparsifies the edges beyond 20%. Clearly, the performance of GNNs is vulnerable to graph sparsification: removing certain edges tends to undermine the underlying structure of the graph, hampering message passing along edges. In this paper, we propose two techniques to improve GNN performance when the graph sparsity is high. The first technique is based on the observation that in UGS, only a fraction of the adjacency matrix elements (i.e., graph edges) are involved in loss calculation. As an illustration, consider the semi-supervised node classification task shown in Figure 1(b) where the nodes in yellow are labeled, and we assume that a GNN with 2 graph convolution layers is used so only nodes within two hops from the yellow nodes are involved in the training process, which are highlighted in green. Note that the gray edges in Figure 1(b) are not involved in loss calculation, i.e., no message passing happens along the dashed edges so their corresponding mask elements get zero gradients during backpropagation. On the Cora dataset, we find that around 50% edges are in such a situation, leaving the values of their corresponding mask elements unchanged throughout the entire training process. After checking the source code of UGS, we find that it initializes these mask elements by adding a random noise, so the ordering of the edge mask-scores is totally determined by this initial mask randomization rather than the graph topology. As a result, the removal of low-scored edges tends to be random in later iterations of UGS, causing performance collapse as some important dashed edges are removed. To address this problem, we add a new auxiliary loss head to better guide the edge pruning by involving the entire adjacency matrix. Specifically, this loss head uses a novel loss function that measures the inter-class separateness of nodes with Wasserstein distance (WD). For each class, we calculate the WD between (1) the set of nodes that are predicted to be in the class and (2) the set of other nodes. By minimizing WD for all classes, we maximize the difference between the extracted node features of different classes. Now that this loss function involves all nodes, all elements in the graph mask mg will now have gradient during backpropagation. Our second technique is based on adversarial perturbation, which is widely used to improve the robustness of deep neural networks (Wong et al., 2020). To improve the robustness of the graph lottery tickets (i.e., the pruned GNN subnetworks) when graph sparsity is high, we regard unfavorable graph sparsification as an adversarial data perturbation and formulate the pruning process as a min-max optimization problem. Specifically, a minimizer seeks to update both the weight parameters Θ and its mask mθ against a maximizer that aims to perturb the graph mask mg. By performing projected gradient ascent on the graph mask, we are essentially using adversarial perturbations to significantly improve the robustness of our graph lottery tickets against the graph sparsity. We further investigate the question: Can we use the obtained winning ticket of a GNN for graph transfer learning? Studying this problem is particularly interesting since the retrainability of a winning ticket (i.e., a pruned sub-network together with original weight initialization) on the same task is the most distinctive property of Lottery Ticket Hypothesis (LTH): many works (Liu Published as a conference paper at ICLR 2023 et al., 2019; Frankle & Carbin, 2019) have verified the importance of winning ticket initialization and stated that the winning ticket initialization might land in a region of the loss landscape that is particularly amenable to optimization. Here, we move one step further in the transfer learning setting, to investigate if the winning ticket identified on a source task also works on a target task. The answer is affirmative through our empirical study, and we call it the transferable graph LTH. The contributions of this paper are summarized as follows: We design a new auxiliary loss function to guide the edges pruning for identifying graph lottery tickets. Our new auxiliary loss is able to address the issue of random edge pruning in UGS. We formalize our sparsification algorithm as a min-max optimization process to gain the robustness of lottery tickets to the graph sparsity. We empirically investigated and verified the transferable graph lottery ticket hypotheses. 2 PRELIMINARY This section defines our notations, reviews the concept of GNNs, defines graph lottery tickets, and reports our empirical study of UGS s random edge pruning issue. Notations and GNN Formulation. We consider an undirected graph G = (V, E) where V and E are the sets of nodes and edges of G, respectively. Each node vi V has a feature vector xi RF , where F is the number of node features. We use X RN F to denote the feature matrix of the whole graph G, where N = |V| is the number of nodes, and X stacks xi of all nodes vi as its rows. Let A RN N be the adjacency matrix of G, i.e., Ai,j = 1 if there is an edge ei,j E between nodes vi and vj , and Ai,j = 0 otherwise. Graph Convolutional Networks (GCN) (Kipf & Welling, 2017) is the most popular and widely adopted GNN model. Without loss of generality, we consider a GCN with two graph convolution layers, which is formulated as follows: Z = softmax ˆA σ ˆAXW(0) W(1) RN C. (1) Here, W(0) is an input-to-hidden weight matrix and W(1) is a hidden-to-output weight matrix. The softmax activation function is applied row-wise and σ( ) = max(0, ) is the Re LU activation function. The total number of classes is C. The adjacency matrix with self-connections A = A + IN is normalized by ˆA = D 1 2 where D is the degree matrix of A. For the semi-supervised node classification tasks, the cross-entropy error over labeled nodes is given by: j=1 Ylj log(Zlj), (2) where YL is the set of node indices for those nodes with labels, Yl is the one-hot encoding (row) vector of node vl s label. Graph Lottery Tickets. If a sub-network of a GNN with the original initialization trained on a sparsified graph has a comparable performance to the original GNN trained on the full graph in terms of test accuracy, then the GNN subnetwork along with the sparsified graph is defined as a graph lottery ticket (GLT) (Chen et al., 2021b). Different from the general LTH literature, GLT consists of three elements: (1) a sparsified graph obtained by pruning some edges in G, (2) a GNN sub-network and (3) the initialization of its learnable parameters. Without loss of generality, let us consider a 2-layer GCN denoted by f(X, Θ0) where Θ0 = {W(0) 0 , W(1) 0 } is the initial weight parameters. Then, the task is to find two masks mg and mθ such that the sub-network f(X, mθ Θ0) along with the sparsified graph {mg A, X} can be trained to a comparable accuracy as f(X, Θ0). UGS Analysis. LTH algorithms including UGS (Chen et al., 2021b) sort the elements in a mask by magnitude and zero out the smallest ones for sparsification. To find and remove the insignificant connections and weights in the graph and the GNN model, UGS updates two masks mg and mθ by backpropagation. However, recall from Figure 1(b) that UGS suffers from the issue that only a fraction of edges in G are related to the training loss as given by Eq (2) for semi-supervised node Published as a conference paper at ICLR 2023 Table 1: Statistic of Edges with Gradients Dataset # of nodes (training/all) Percentage of edges related to loss Cora 140 / 2,708 51% (2,702 / 5,429) Citeseer 120 / 3,327 34% (1,593 / 4,732) Pub Med 60 / 44,338 9% (3,712 / 44,338) classification. Table 1 shows (1) the number of nodes in the training v.s. entire dataset and (2) the percentage of edges with gradients during backpropagation for three graph datasets: Cora, Citeseer and Pub Med. We can see that a significant percentage of edges (up to 91% as in Pub Med) are not related to the training loss. We carefully examined the source code of UGS and found that the pruning process for these edges tends to be random since each element in the graph mask is initialized by adding a random noise to its magnitude. As we have seen in Figure 1(a), the accuracy of UGS collapses when the graph sparsity becomes high. It is crucial to properly guide the mask learning for those edges not related to the training loss in Eq (2) to alleviate the performance degradation. 3 METHODOLOGY Auxiliary Loss Function. Since L0 as given by Eq (2) does not involve all edges for backpropagation, we hereby design another auxiliary loss function L1 to better guide the edge pruning together with L0. Given GNN output Z and a class c, we separate potential nodes of class c and the other nodes by: Zc = {zi rows(Z) | argmax(zi) = c} and Z c = {zi rows(Z) | argmax(zi) = c}, (3) for c {1, 2, , C}. Since zi of those i YL are properly guided by L0, we can use argmax(zi) = c to filter out the potential nodes that belong to class c. For each class c, we then calculate the Wasserstein distance (WD) between Zc and Z c: WD(Zc, Z c) = inf π Π(Zc,Z c) E(zi,zj) π[ zi zj ], (4) where Π(Zc, Z c) denotes the set of all joint distributions π(zi, zj) whose marginals are, respectively, Zc and Z c: Π(Zc, Z c) = {P : P1 = Zc, PT 1 = Z c}. (5) Wasserstein distance (WD) (Villani, 2009) is widely used to measure the distance between two distributions, especially under the vanishing overlap of two distributions. Intuitively, it is the cost of the optimal plan for transporting one distribution to match another distribution: in Eq (5), Π(Zc, Z c) is the set of valid transport plans, where P is a coupling matrix with each element Pij representing how much probability mass from point zi Zc is assigned to a point zj Z c; and 1 is an all-one vector. We consider Zc and Z c as the distribution of class c and other classes, respectively. Equation (4) can be approximated by a smooth convex optimization problem with an entropic regularization term: min P D, P ε X i,j Pij log(Pij), (6) where D is the pairwise Euclidean distance matrix between points of Zc and Z c, and , is the Frobenius inner product. Minimizing Eq (6) is solved using Sinkhorn iterations (Cuturi, 2013), which form a sequence of linear operations straightforward for backpropagation. Ideally, we want the node representations in one class to be very different from those in other classes so that the classes can be easily distinguished. By maximizing WD(Zc, Z c) for all classes c, we maximize the difference between GNN outputs for different classes. Formally, the auxiliary loss function to minimize is given by: c C WD(Zc, Z c). (7) To explore the relationship between unsupervised loss L1 with the supervised loss L0 (i.e., Eq (2)), we train a 2-layer GCN on Cora with L0 as the only loss function, and Figure 2 plots the value of L0 as well as the corresponding value of P c C WD(Zc, Z c) (abbr. WD) during the training process. We can see that as training loss L0 decreases, the value of WD increases. The curves of both L0 and WD Published as a conference paper at ICLR 2023 0 50 100 150 200 Epochs Figure 2: Supervised Loss L0 v.s. WD. Note that as training loss L0 decreases, the value of WD increases. Both L0 and WD start to bounce back after 100 epochs. start to bounce back after 100 epochs. Therefore, maximizing WD is consistent with improving classification accuracy. Note that P c C WD(Zc, Z c) is unsupervised and related to all nodes in the training set. It is natural to use this new metric to guide the pruning of edges that are not directly related to L0: if masking an edge decreases WD, this edge tends to be important and its removal will decrease the test accuracy of node classification. Our sparsification method uses a semi-supervised loss that combines L0 and L1 (or, WD): L = L0 + λL1. Sparsification. To prune the edges and weights while retaining robustness, we formulate GNN sparsification as a min-max problem given by: min Θ,mθ max mg (L0 + λL1), (8) where the outer minimization seeks to train the model weights and their masks to minimize the classification error, and the inner maximization aims to adversarially perturb the graph mask. Let us use L(Θ, mθ, mg) to denote (L0 + λL1), then the optimization problem in Eq (8) can be solved by alternating the following two steps as specified by (i) Eq (9) and (ii) Eqs (10) & (11). Specifically, the inner maximization perturbs the graph mask at (t + 1)th iteration as follows: m(t+1) g P[0,1]N N h m(t) g + η1 mg L(Θ(t), m(t) θ , m(t) g ) i , (9) where PC[m] = arg minm C m m 2 F is a differentiable projection operation that projects the mask values onto [0, 1], and η1 > 0 is a proper learning rate. While the second step for outer minimization is given by: m(t+1) θ m(t) θ η2 mθL(Θ(t), m(t) θ , m(t+1) g ) + α d h m(t+1) g i T dm(t) θ mg L(Θ(t), m(t) θ , m(t+1) g ) , Θ(t+1) Θ(t) η2 ΘL(Θ(t), m(t) θ , m(t+1) g ) + α d h m(t+1) g i T dΘ(t) mg L(Θ(t), m(t) θ , m(t+1) g ) , (11) where the last term is an implicit gradient obtained by chain rule over m(t+1) g as given by Eq (9). Algorithm 1 Iterative pruning process Input:Initial masks mg = A, mθ = 1 R||Θ|| Output: Sparsified masks mg and mθ 1: while 1 mg 0 A 0 < sg and 1 mθ 0 Θ 0 < sθ do 2: m(0) g = mg, m(0) θ = mθ, Θ(0) = {W(0) 0 , W(1) 0 } 3: for iteration t = 0, 1, 2, , T 1 do 4: Compute m(t+1) g with Eq (9) 5: Compute m(t+1) θ with Eq (10) 6: Compute Θ(t+1) with Eq (11) 7: end for 8: mg = m(T 1) g , mθ = m(T 1) θ 9: Set pg% of the lowest-scored values in mg to 0 and set the others to 1 10: Set pθ% of the lowest-scored values in mθ to 0 and set the others to 1 11: end while We follow UGS s framework to repeatedly train, prune, and reset a GNN with initial weights after each round until the desired sparsity levels sg and sθ are reached. Algorithm 1 illustrates our iterative Published as a conference paper at ICLR 2023 pruning process. At each round once the model is trained and masks are computed, we set pg fraction of the lowest-scored values in mg to 0 and set the others to 1. Likewise, (1 pθ) fraction of the weights in Θ will survive to the next round. Finally, we retrain the obtained sub-network using only L0 to report the test accuracy. 4 TRANSFER LEARNING WITH GRAPH LOTTERY TICKETS This section explores whether a graph lottery ticket (GLT) can be used for graph transfer learning. Specifically, we ask two questions: (1) Can we use the graph winning ticket of a GNN for transfer learning? Here, the graph winning ticket is associated with the originally initialized weight parameters. (2) Can we use the post-trained winning ticket of a GNN for transfer learning? Here, post-trained means that the winning ticket directly uses the trained weight parameters on the source task to bootstrap the training on the target task. 0 200 400 600 800 1000 Epochs Accuracy on target task Post-trained GLT Re-initialized GLT Randomly initialized GNN Post-trained GNN Randomly initialized GLT Figure 3: Training Curves on MAG. Notice that the post-trained winning ticket achieves the similar performance as the post-trained GNN. Also, the winning ticket achieves even better performance than a randomly initialized GNN. To answer these two questions, we consider the node classification tasks on two citation networks ar Xiv and MAG (Hu et al., 2020). In both tasks, each node represents an author and is associated with a 128-d feature vector obtained by averaging the word embeddings of papers that are published by the author. We train a 3-layer GCN model on ar Xiv and transfer the model to MAG by replacing the last dense layer. Since the graphs of ar Xiv and MAG are different, we only transfer mθ and Θ from ar Xiv, and use the original graph of MAG without sparsification. Figure 3 shows the training accuracy curves on MAG with winning tickets obtained from ar Xiv. Here, Post-trained GLT (red line) starts training with the pretrained winning ticket on ar Xiv, while "Re-initialized GLT" (green line) starts training with the winning ticket re-initialized using original initial weight parameters. We can see that the sub-network obtained from ar Xiv transfers well on MAG in both settings, so the answers to our two questions are affirmative. Also, the blue (resp. black) line starts training with the post-trained GNN from ar Xiv without any edge or weight pruning; while the yellow line starts training with the winning ticket but with randomly initialized weight parameters. We can observe that (1) the post-trained winning ticket (red) achieves the same performance as the post-trained GNN (blue), showing that the transferred performance is not sensitive to weight pruning; and that (2) the winning ticket identified on the source task (green) achieves even better performance on the target task than a randomly initialized GNN (black). Note that gradient descent could get trapped in undesirable stationary points if it starts from arbitrary points, and our GLT identified from the source task allows the initialization to land in a region of the loss landscape that is particularly amenable to optimization (Ganea et al., 2018). 5 EXPERIMENT Experimental Setups. Following Chen et al. (2021b), we evaluate our sparsification method with three popular GNN models: GCN (Kipf & Welling, 2017), GAT (Velickovic et al., 2018) and GIN (Xu et al., 2019), on three widely used graph datasets from Chen et al. (2021b) (Cora, Citeseer and Pub Med) and two OGB datasets from Hu et al. (2020) (Arxiv and MAG) for semi-supervised node classification. We compare our method with two baselines: UGS (Chen et al., 2021b) and random pruning (Chen et al., 2021b). For fair comparison, we follow UGS to use the default setting: pg = 5 and pθ = 20 unless otherwise stated. The value of λ is configured as 0.1 by default. More details on the dataset statistics and model configurations can be found in the appendix. Evaluation of Our Sparsification Method. Figures 4, 5 and 6 compare our method with UGS and random pruning in terms of the test accuracy on three 2-layer GNN models: GCN, GIN and GAT, respectively. In each figure, the first (resp. second) row shows how the test accuracy changes with graph sparsity (resp. weight sparsity) during the iterative pruning process, on the three datasets: Cora, Citeseer and Pub Med. We can observe that the accuracy of our method (green line) is much higher than UGS (yellow line) when the graph sparsity is high and when the weight sparsity is high. Published as a conference paper at ICLR 2023 Table 2: Pair-wise comparison Cora (17 iterations) Model GCN GIN GAT Methods/Metrics GS(%) WS(%) Accuracy(%) GS(%) WS(%) Accuracy(%) GS(%) WS(%) Accuracy(%) UGS 56.28 97.52 72.5 57.62 97.25 69.6 57.42 97.38 78.7 Our 58.02 97.80 75.1 57.88 98.26 72.8 58.12 97.89 80.0 Citeseer (17 iterations) Model GCN GIN GAT Methods/Metrics GS(%) WS(%) Accuracy(%) GS(%) WS(%) Accuracy(%) GS(%) WS(%) Accuracy(%) UGS 54.33 96.53 66.7 56.05 96.58 63.2 54.99 97.21 70.1 Our 55.78 97.38 70.3 55.95 97.25 66.8 56.03 97.30 70.6 Pub Med (17 iterations) Model GCN GIN GAT Methods/Metrics GS(%) WS(%) Accuracy(%) GS(%) WS(%) Accuracy(%) GS(%) WS(%) Accuracy(%) UGS 58.21 97.76 77.7 58.23 97.19 76.3 57.64 97.32 78.8 Our 59.43 97.81 79.5 58.13 97.76 77.2 58.22 98.18 80.0 To clearly see this accuracy difference, we show the test accuracy, graph sparsity (GS) and weight sparsity (WS) after the 17th pruning iteration in Table 2 for both UGS and our method. Note that each graph sparsity is paired with a weight sparsity. We use and to represent UGS and our method, respectively. Given the same sparcification iteration number, the sub-networks found by our method have a higher accuracy than those found by UGS. For example, when GS = 56.28% and WS = 97.52%, the accuracy of GCN on Cora is 72.5%, whereas our method improves the accuracy to 75.1% with a comparable setting: GS = 57.95% and WS = 97.31%. Note that although we use the same pg and pθ for both our method and UGS, their sparsity values in the same pruning iteration are comparable but not exactly equal. This is because multiple mask entries can have the same value, and they span across the pruning threshold pg or pθ; we prune them all in an iteration such that slightly more than pg fraction of edges or pθ fraction of weights may be removed. The higher accuracy of our method over UGS demonstrates that our approach finds lottery tickets that are more robust against the sparsity, and better prevents a performance collapse when the graph sparsity is high. In addition, we use and to indicate the last GLT with an accuracy higher than the original model in the sparsification process of UGS and our method, respectively. We can observe that the GLTs identified by our method have higher sparsity than those located by UGS. It further verifies that our method can improve the robustness of GLT against the sparsity. Note that in Figure 4, the third subfigure, the accuracy of our method (green line) is very stable even when graph sparsity is high. This is because Pub Med is a dense graph where some edges can result from noise during data collection, so graph sparsification functions as denoising, which retains or even improves model accuracy by alleviating the oversmoothing problem of GCN. 0.0 0.1 0.2 0.3 0.4 0.5 0.6 Graph sparsity Accuracy (Cora) UGS Random Ours 0.0 0.1 0.2 0.3 0.4 0.5 0.6 Graph sparsity Accuracy (Citeseer) UGS Random Ours 0.0 0.1 0.2 0.3 0.4 0.5 0.6 Graph sparsity Accuracy (Pub Med) UGS Random Ours 0.0 0.2 0.4 0.6 0.8 1.0 Weight sparsity Accuracy (Cora) UGS Random Ours 0.0 0.2 0.4 0.6 0.8 1.0 Weight sparsity Accuracy (Citeseer) UGS Random Ours 0.0 0.2 0.4 0.6 0.8 1.0 Weight sparsity Accuracy (Pub Med) UGS Random Ours Figure 4: Performance on GCN. Marker and denote the spasified GCN (after 17 iterations) with UGS and our method, respectively. Marker and indicate the last GLT that reaches higher accuracy than the original model in the sparsification process of UGS and our method, respectively. Published as a conference paper at ICLR 2023 0.0 0.1 0.2 0.3 0.4 0.5 0.6 Graph sparsity Accuracy (Cora) UGS Random Ours 0.0 0.1 0.2 0.3 0.4 0.5 0.6 Graph sparsity Accuracy (Citeseer) UGS Random Ours 0.0 0.1 0.2 0.3 0.4 0.5 0.6 Graph sparsity Accuracy (Pub Med) UGS Random Ours 0.0 0.2 0.4 0.6 0.8 1.0 Weight sparsity Accuracy (Cora) UGS Random Ours 0.0 0.2 0.4 0.6 0.8 1.0 Weight sparsity Accuracy (Citeseer) UGS Random Ours 0.0 0.2 0.4 0.6 0.8 1.0 Weight sparsity Accuracy (Pub Med) UGS Random Ours Figure 5: Performance on GIN 0.0 0.1 0.2 0.3 0.4 0.5 0.6 Graph sparsity Accuracy (Cora) UGS Random Ours 0.0 0.1 0.2 0.3 0.4 0.5 0.6 Graph sparsity Accuracy (Citeseer) UGS Random Ours 0.0 0.1 0.2 0.3 0.4 0.5 0.6 Graph sparsity Accuracy (Pub Med) UGS Random Ours 0.0 0.2 0.4 0.6 0.8 1.0 Weight sparsity Accuracy (Cora) UGS Random Ours 0.0 0.2 0.4 0.6 0.8 1.0 Weight sparsity Accuracy (Citeseer) UGS Random Ours 0.0 0.2 0.4 0.6 0.8 1.0 Weight sparsity Accuracy (Pub Med) UGS Random Ours Figure 6: Performance on GAT Also note from Figure 6 that the performance of GAT is less sensitive to the sparsity as compared with GCN and GIN. This is because the attention mechanism in GAT leads to many more parameters than the other two models, allowing for pruning more parameters; moreover, GAT itself is able to identify the importance of edges through the attention mechanism, which better guides the edge pruning. 0.54 0.56 0.58 0.60 0.62 0.64 Graph sparsity Accuracy (Cora) = 0 = 0.001 = 0.1 = 1 0.965 0.970 0.975 0.980 0.985 0.990 Weight sparsity Accuracy (Cora) = 0 = 0.001 = 0.1 = 1 Figure 7: Ablation study Ablation Study. To verify the effectiveness of each design component in our sparsification algorithm, we compare our method with several variants. First, we remove the Wassersteindistance based auxiliary loss head by setting λ = 0 during sparsification. As shown in Figure 7, removing the auxiliary loss function leads to performance degradation. However, there exists a tradeoff to weight the importance of L1. For example, if we increase λ from 0.1 to 1, the performance backfires especially when the sparisty is high. We further investigate if our adversarial perturbation strategy can improve the robustness of the pruned GNNs, by replacing our min-max optimization process with the algorithm in UGS. Specifically, the variant UGS+WD uses the sparsification algorithm in UGS to minimize L(θ, mθ, mg). Compared with this variant, our method Published as a conference paper at ICLR 2023 can alleviate the performance degradation when graph sparsity is high, which verifies that adversarial perturbation improves robustness. 0.05 0.10 0.15 0.20 0.25 pg 79.3 75.2 70.7 69.4 68.3 79.1 75.0 71.8 69.1 68.2 79.7 75.3 74.4 70.2 69.2 79.3 76.9 72.1 70.0 70.2 79.3 75.9 72.6 70.9 68.7 Figure 8: Varying pg and pθ Pairwise Sparsity. In Figure 8, we investigate the accuracy of the pruned GNNs after 10 iterations with different combinations of pg and pθ. Note that larger pg and pθ result in higher sparsity of the adjacent matrix and the weight matrices, respectively. We can see that the performance of the pruned GNN is more sensitive to the graph sparsity than the weight sparsity. It further verifies the necessity of finding robust GLTs when the graph sparsity is high. Since the pruned GNN is less sensitive to the weight sparsity, we can obtain GLTs with higher overall sparsity by setting pθ to be larger pg. At the same time, a larger pg can result in the pruning of more noisy and unimportant edges, which may help prevent GCN oversmoothing. Therefore, a tradeoff exists when choosing a proper combination of pg and pθ. While users may try different combinations of pg and pθ to find a proper combination, our approach is superior to UGS in all combinations of pg and pθ, so should be a better algorithm choice. 6 RELATED WORK In the last few years, a series of GNN models (Kipf & Welling, 2017; Velickovic et al., 2018; Xu et al., 2019; Hamilton et al., 2017; Ying et al., 2018; Klicpera et al., 2019; Hui et al., 2020; Tong et al., 2020; Yun et al., 2019) have been designed to utilize the graph structure in spatial domain and spectral domain for effective feature extraction. However, the computational cost of GNNs on large-scale graphs can be very high especially when the graph is dense with many edges. Many models have been proposed to reduce this computational cost, such as Fast GCN (Chen et al., 2018) and SGC (Wu et al., 2019). However, their accuracy tends to be worse than GCN, while the speedup achieved is still limited. So, it is desirable to develop a pruning method to sparsify both the edges and the weight parameters for any given GNN model (including Fast GCN and SGC) without accuracy compromise. The lottery ticket hypothesis is first proposed in Frankle & Carbin (2019) to find an identically initialized sub-network that can be trained to achieve a similar performance as the original neural network. Many works (Frankle et al., 2019; 2020; Diffenderfer & Kailkhura, 2021; Malach et al., 2020; Ma et al., 2021b; Savarese et al., 2020; Zhang et al., 2021a;b; Zhou et al., 2019; Chen et al., 2021c; Su et al., 2020; Chen et al., 2022) have verified that the identified winning tickets are retrainable to reduce the high memory cost and long inference time of the original neural networks. LTH has been extended to different kinds of neural networks such as GANs (Chen et al., 2021e;a; Kalibhat et al., 2021), Transformers (Brix et al., 2020; Prasanna et al., 2020; Chen et al., 2020; Behnke & Heafield, 2020) and even deep reinforcement learning models (Vischer et al., 2021; Yu et al., 2020). It has also been studied on GNNs by UGS (Chen et al., 2021b). Despite the competitive performance of winning tickets, the iterative train-prune-retrain process and even the subsequent fine-tuning (Liu et al., 2021) are costly to compute. To address this issue, Pr AC (Zhang et al., 2021c) proposes to find lottery tickets efficiently by using a selected subset of data rather than using the full training set. Other works (Wang et al., 2020; Tanaka et al., 2020) attempt to find winning tickets at initialization without training. The new trend of dynamic sparse training shows that any random initialized sparse neural networks can achieves comparable accuracy to the dense neural networks (Ye et al., 2020; Evci et al., 2020; Hou et al., 2022; Ma et al., 2021a; Yuan et al., 2021). Customized hardware architectures have also been used to accelerate sparse training (Goli & Aamodt, 2020; Raihan & Aamodt, 2020). The Elastic Lottery Ticket Hypothesis Chen et al. (2021d) studies the transferability of winning tickets between different network architectures within the same design family. For object recognition, Mehta (2019) shows that winning tickets from VGG19 do not transfer well to down-stream tasks. In the natural image domain, Morcos et al. (2019) finds that the winning ticket initializations generalize across a variety of vision benchmark datasets. 7 CONCLUSION In this paper, we empirically observe that the performance of GLTs collapses when the graph sparsity is high. We design a new auxiliary loss to address this limitation. Moreover, we formulate the GNN sparsification process as a min-max optimization problem which adopts adversarial perturbation to improve the robustness of graph lottery tickets. Our experiments verified the superiority of our method over UGS. We further empirically studied transfer learning with GLTs and confirmed the transferability that gives rise to the transferable graph lottery ticket hypotheses. Published as a conference paper at ICLR 2023 Maximiliana Behnke and Kenneth Heafield. Losing heads in the lottery: Pruning transformer attention in neural machine translation. In Bonnie Webber, Trevor Cohn, Yulan He, and Yang Liu (eds.), Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, EMNLP 2020, Online, November 16-20, 2020, pp. 2664 2674. Association for Computational Linguistics, 2020. doi: 10.18653/v1/2020.emnlp-main.211. URL https://doi.org/10. 18653/v1/2020.emnlp-main.211. Christopher Brix, Parnia Bahar, and Hermann Ney. Successfully applying the stabilized lottery ticket hypothesis to the transformer architecture. In Dan Jurafsky, Joyce Chai, Natalie Schluter, and Joel R. Tetreault (eds.), Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020, Online, July 5-10, 2020, pp. 3909 3915. Association for Computational Linguistics, 2020. doi: 10.18653/v1/2020.acl-main.360. URL https://doi. org/10.18653/v1/2020.acl-main.360. Jie Chen, Tengfei Ma, and Cao Xiao. Fastgcn: Fast learning with graph convolutional networks via importance sampling. In 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings. Open Review.net, 2018. URL https://openreview.net/forum?id=rytstx WAW. Tianlong Chen, Jonathan Frankle, Shiyu Chang, Sijia Liu, Yang Zhang, Zhangyang Wang, and Michael Carbin. The lottery ticket hypothesis for pre-trained BERT networks. In Hugo Larochelle, Marc Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan Tien Lin (eds.), Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, Neur IPS 2020, December 6-12, 2020, virtual, 2020. URL https://proceedings.neurips.cc/paper/2020/hash/ b6af2c9703f203a2794be03d443af2e3-Abstract.html. Tianlong Chen, Yu Cheng, Zhe Gan, Jingjing Liu, and Zhangyang Wang. Data-efficient gan training beyond (just) augmentations: A lottery ticket perspective. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan (eds.), Advances in Neural Information Processing Systems, volume 34, pp. 20941 20955. Curran Associates, Inc., 2021a. URL https://proceedings.neurips.cc/paper/2021/file/ af4f00ca48321fb026865c5a1772dafd-Paper.pdf. Tianlong Chen, Yongduo Sui, Xuxi Chen, Aston Zhang, and Zhangyang Wang. A unified lottery ticket hypothesis for graph neural networks. In Marina Meila and Tong Zhang (eds.), Proceedings of the 38th International Conference on Machine Learning, ICML 2021, 18-24 July 2021, Virtual Event, volume 139 of Proceedings of Machine Learning Research, pp. 1695 1706. PMLR, 2021b. URL http://proceedings.mlr.press/v139/chen21p.html. Tianlong Chen, Zhenyu Zhang, Sijia Liu, Shiyu Chang, and Zhangyang Wang. Long live the lottery: The existence of winning tickets in lifelong learning. In 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3-7, 2021. Open Review.net, 2021c. URL https://openreview.net/forum?id=LXMSv Pmsm0g. Tianlong Chen, Xuxi Chen, Xiaolong Ma, Yanzhi Wang, and Zhangyang Wang. Coarsening the granularity: Towards structurally sparse lottery tickets. In International Conference on Machine Learning (ICML), 2022. Xiaohan Chen, Yu Cheng, Shuohang Wang, Zhe Gan, Jingjing Liu, and Zhangyang Wang. The elastic lottery ticket hypothesis. In Marc Aurelio Ranzato, Alina Beygelzimer, Yann N. Dauphin, Percy Liang, and Jennifer Wortman Vaughan (eds.), Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, Neur IPS 2021, December 6-14, 2021, virtual, pp. 26609 26621, 2021d. URL https://proceedings.neurips.cc/ paper/2021/hash/dfccdb8b1cc7e4dab6d33db0fef12b88-Abstract.html. Xuxi Chen, Zhenyu Zhang, Yongduo Sui, and Tianlong Chen. Gans can play lottery tickets too. In 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3-7, 2021. Open Review.net, 2021e. URL https://openreview.net/forum?id= 1Ao Mhc_9j ER. Published as a conference paper at ICLR 2023 Marco Cuturi. Sinkhorn distances: Lightspeed computation of optimal transport. In Christopher J. C. Burges, Léon Bottou, Zoubin Ghahramani, and Kilian Q. Weinberger (eds.), Advances in Neural Information Processing Systems 26: 27th Annual Conference on Neural Information Processing Systems 2013. Proceedings of a meeting held December 5-8, 2013, Lake Tahoe, Nevada, United States, pp. 2292 2300, 2013. URL https://proceedings.neurips.cc/paper/ 2013/hash/af21d0c97db2e27e13572cbf59eb343d-Abstract.html. James Diffenderfer and Bhavya Kailkhura. Multi-prize lottery ticket hypothesis: Finding accurate binary neural networks by pruning A randomly weighted network. In 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3-7, 2021. Open Review.net, 2021. URL https://openreview.net/forum?id=U_mat0b9iv. Utku Evci, Trevor Gale, Jacob Menick, Pablo Samuel Castro, and Erich Elsen. Rigging the lottery: Making all tickets winners. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13-18 July 2020, Virtual Event, volume 119 of Proceedings of Machine Learning Research, pp. 2943 2952. PMLR, 2020. URL http://proceedings.mlr.press/v119/ evci20a.html. Jonathan Frankle and Michael Carbin. The lottery ticket hypothesis: Finding sparse, trainable neural networks. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019. Open Review.net, 2019. URL https://openreview. net/forum?id=r Jl-b3Rc F7. Jonathan Frankle, Gintare Karolina Dziugaite, Daniel M. Roy, and Michael Carbin. The lottery ticket hypothesis at scale. Co RR, abs/1903.01611, 2019. URL http://arxiv.org/abs/1903. 01611. Jonathan Frankle, Gintare Karolina Dziugaite, Daniel M. Roy, and Michael Carbin. Linear mode connectivity and the lottery ticket hypothesis. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13-18 July 2020, Virtual Event, volume 119 of Proceedings of Machine Learning Research, pp. 3259 3269. PMLR, 2020. URL http://proceedings. mlr.press/v119/frankle20a.html. Octavian-Eugen Ganea, Gary Bécigneul, and Thomas Hofmann. Hyperbolic entailment cones for learning hierarchical embeddings. In Jennifer G. Dy and Andreas Krause (eds.), ICML, 2018. Negar Goli and Tor M. Aamodt. Resprop: Reuse sparsified backpropagation. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, June 13-19, 2020, pp. 1545 1555. Computer Vision Foundation / IEEE, 2020. doi: 10.1109/CVPR42600.2020.00162. URL https://openaccess.thecvf.com/content_ CVPR_2020/html/Goli_Re Sprop_Reuse_Sparsified_Backpropagation_ CVPR_2020_paper.html. William L. Hamilton, Zhitao Ying, and Jure Leskovec. Inductive representation learning on large graphs. In Isabelle Guyon, Ulrike von Luxburg, Samy Bengio, Hanna M. Wallach, Rob Fergus, S. V. N. Vishwanathan, and Roman Garnett (eds.), Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4-9, 2017, Long Beach, CA, USA, pp. 1024 1034, 2017. URL https://proceedings.neurips.cc/ paper/2017/hash/5dd9db5e033da9c6fb5ba83c7a7ebea9-Abstract.html. Zejiang Hou, Minghai Qin, Fei Sun, Xiaolong Ma, Kun Yuan, Yi Xu, Yen-Kuang Chen, Rong Jin, Yuan Xie, and Sun-Yuan Kung. Chex: Channel exploration for cnn model compression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 12287 12298, 2022. Weihua Hu, Matthias Fey, Marinka Zitnik, Yuxiao Dong, Hongyu Ren, Bowen Liu, Michele Catasta, and Jure Leskovec. Open graph benchmark: Datasets for machine learning on graphs. In Hugo Larochelle, Marc Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin (eds.), Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, Neur IPS 2020, December 6-12, 2020, virtual, 2020. URL https://proceedings.neurips.cc/paper/2020/hash/ fb60d411a5c5b72b2e7d3527cfc84fd0-Abstract.html. Published as a conference paper at ICLR 2023 Bo Hui, Da Yan, Wei-Shinn Ku, and Wenlu Wang. Predicting economic growth by region embedding: A multigraph convolutional network approach. In Mathieu d Aquin, Stefan Dietze, Claudia Hauff, Edward Curry, and Philippe Cudré-Mauroux (eds.), CIKM 20: The 29th ACM International Conference on Information and Knowledge Management, Virtual Event, Ireland, October 19-23, 2020, pp. 555 564. ACM, 2020. doi: 10.1145/3340531.3411882. URL https://doi.org/ 10.1145/3340531.3411882. Neha Mukund Kalibhat, Yogesh Balaji, and Soheil Feizi. Winning lottery tickets in deep generative models. In Thirty-Fifth AAAI Conference on Artificial Intelligence, AAAI 2021, Thirty-Third Conference on Innovative Applications of Artificial Intelligence, IAAI 2021, The Eleventh Symposium on Educational Advances in Artificial Intelligence, EAAI 2021, Virtual Event, February 2-9, 2021, pp. 8038 8046. AAAI Press, 2021. URL https://ojs.aaai.org/index.php/AAAI/ article/view/16980. Thomas N. Kipf and Max Welling. Semi-supervised classification with graph convolutional networks. In 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 2426, 2017, Conference Track Proceedings. Open Review.net, 2017. URL https://openreview. net/forum?id=SJU4ay Ygl. Johannes Klicpera, Stefan Weißenberger, and Stephan Günnemann. Diffusion improves graph learning. In Hanna M. Wallach, Hugo Larochelle, Alina Beygelzimer, Florence d Alché-Buc, Emily B. Fox, and Roman Garnett (eds.), Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, Neur IPS 2019, December 8-14, 2019, Vancouver, BC, Canada, pp. 13333 13345, 2019. URL https://proceedings.neurips.cc/paper/2019/hash/ 23c894276a2c5a16470e6a31f4618d73-Abstract.html. Nishai Kooverjee, Steven James, and Terence L. van Zyl. Investigating transfer learning in graph neural networks. Co RR, abs/2202.00740, 2022. URL https://arxiv.org/abs/2202. 00740. Ning Liu, Geng Yuan, Zhengping Che, Xuan Shen, Xiaolong Ma, Qing Jin, Jian Ren, Jian Tang, Sijia Liu, and Yanzhi Wang. Lottery ticket preserves weight correlation: Is it desirable or not? In Marina Meila and Tong Zhang (eds.), Proceedings of the 38th International Conference on Machine Learning, ICML 2021, 18-24 July 2021, Virtual Event, volume 139 of Proceedings of Machine Learning Research, pp. 7011 7020. PMLR, 2021. URL http://proceedings. mlr.press/v139/liu21aa.html. Zhuang Liu, Mingjie Sun, Tinghui Zhou, Gao Huang, and Trevor Darrell. Rethinking the value of network pruning. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019. Open Review.net, 2019. URL https://openreview. net/forum?id=r Jln B3C5Ym. Xiaolong Ma, Minghai Qin, Fei Sun, Zejiang Hou, Kun Yuan, Yi Xu, Yanzhi Wang, Yen-Kuang Chen, Rong Jin, and Yuan Xie. Effective model sparsification by scheduled grow-and-prune methods. In International Conference on Learning Representations (ICLR), 2021a. Xiaolong Ma, Geng Yuan, Xuan Shen, Tianlong Chen, Xuxi Chen, Xiaohan Chen, Ning Liu, Minghai Qin, Sijia Liu, Zhangyang Wang, and Yanzhi Wang. Sanity checks for lottery tickets: Does your winning ticket really win the jackpot? Co RR, abs/2107.00166, 2021b. URL https: //arxiv.org/abs/2107.00166. Eran Malach, Gilad Yehudai, Shai Shalev-Shwartz, and Ohad Shamir. Proving the lottery ticket hypothesis: Pruning is all you need. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13-18 July 2020, Virtual Event, volume 119 of Proceedings of Machine Learning Research, pp. 6682 6691. PMLR, 2020. URL http://proceedings. mlr.press/v119/malach20a.html. Rahul Mehta. Sparse transfer learning via winning lottery tickets. Co RR, abs/1905.07785, 2019. URL http://arxiv.org/abs/1905.07785. Published as a conference paper at ICLR 2023 Ari S. Morcos, Haonan Yu, Michela Paganini, and Yuandong Tian. One ticket to win them all: generalizing lottery ticket initializations across datasets and optimizers. In Hanna M. Wallach, Hugo Larochelle, Alina Beygelzimer, Florence d Alché-Buc, Emily B. Fox, and Roman Garnett (eds.), Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, Neur IPS 2019, December 8-14, 2019, Vancouver, BC, Canada, pp. 4933 4943, 2019. URL https://proceedings.neurips.cc/paper/2019/hash/ a4613e8d72a61b3b69b32d040f89ad81-Abstract.html. Sai Prasanna, Anna Rogers, and Anna Rumshisky. When BERT plays the lottery, all tickets are winning. In Bonnie Webber, Trevor Cohn, Yulan He, and Yang Liu (eds.), Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, EMNLP 2020, Online, November 16-20, 2020, pp. 3208 3229. Association for Computational Linguistics, 2020. doi: 10.18653/v1/2020.emnlp-main.259. URL https://doi.org/10.18653/v1/2020. emnlp-main.259. Md Aamir Raihan and Tor M. Aamodt. Sparse weight activation training. In Hugo Larochelle, Marc Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin (eds.), Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, Neur IPS 2020, December 6-12, 2020, virtual, 2020. URL https://proceedings.neurips.cc/paper/2020/hash/ b44182379bf9fae976e6ae5996e13cd8-Abstract.html. Pedro Savarese, Hugo Silva, and Michael Maire. Winning the lottery with continuous sparsification. In Hugo Larochelle, Marc Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin (eds.), Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, Neur IPS 2020, December 6-12, 2020, virtual, 2020. URL https://proceedings.neurips.cc/paper/2020/hash/ 83004190b1793d7aa15f8d0d49a13eba-Abstract.html. Jingtong Su, Yihang Chen, Tianle Cai, Tianhao Wu, Ruiqi Gao, Liwei Wang, and Jason D. Lee. Sanity-checking pruning methods: Random tickets can win the jackpot. In Hugo Larochelle, Marc Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan Tien Lin (eds.), Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, Neur IPS 2020, December 6-12, 2020, virtual, 2020. URL https://proceedings.neurips.cc/paper/2020/hash/ eae27d77ca20db309e056e3d2dcd7d69-Abstract.html. Hidenori Tanaka, Daniel Kunin, Daniel L. K. Yamins, and Surya Ganguli. Pruning neural networks without any data by iteratively conserving synaptic flow. In Hugo Larochelle, Marc Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin (eds.), Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, Neur IPS 2020, December 6-12, 2020, virtual, 2020. URL https://proceedings.neurips.cc/paper/2020/hash/ 46a4378f835dc8040c8057beb6a2da52-Abstract.html. Zekun Tong, Yuxuan Liang, Changsheng Sun, Xinke Li, David S. Rosenblum, and Andrew Lim. Digraph inception convolutional networks. In Hugo Larochelle, Marc Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin (eds.), Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, Neur IPS 2020, December 6-12, 2020, virtual, 2020. URL https://proceedings.neurips.cc/paper/ 2020/hash/cffb6e2288a630c2a787a64ccc67097c-Abstract.html. Petar Velickovic, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Liò, and Yoshua Bengio. Graph attention networks. In 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings. Open Review.net, 2018. URL https://openreview.net/forum?id=r JXMpik CZ. Cédric Villani. Optimal transport: old and new, volume 338. Springer, 2009. Marc Aurel Vischer, Robert Tjarko Lange, and Henning Sprekeler. On lottery tickets and minimal task representations in deep reinforcement learning. Co RR, abs/2105.01648, 2021. URL https: //arxiv.org/abs/2105.01648. Published as a conference paper at ICLR 2023 Chaoqi Wang, Guodong Zhang, and Roger B. Grosse. Picking winning tickets before training by preserving gradient flow. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020. Open Review.net, 2020. URL https:// openreview.net/forum?id=Skgs ACVKPH. Eric Wong, Leslie Rice, and J. Zico Kolter. Fast is better than free: Revisiting adversarial training. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020. Open Review.net, 2020. URL https://openreview.net/forum?id= BJx040EFv H. Felix Wu, Amauri H. Souza Jr., Tianyi Zhang, Christopher Fifty, Tao Yu, and Kilian Q. Weinberger. Simplifying graph convolutional networks. In Kamalika Chaudhuri and Ruslan Salakhutdinov (eds.), Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 915 June 2019, Long Beach, California, USA, volume 97 of Proceedings of Machine Learning Research, pp. 6861 6871. PMLR, 2019. URL http://proceedings.mlr.press/v97/ wu19e.html. Keyulu Xu, Weihua Hu, Jure Leskovec, and Stefanie Jegelka. How powerful are graph neural networks? In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019. Open Review.net, 2019. URL https://openreview.net/forum? id=ry Gs6i A5Km. Mao Ye, Lemeng Wu, and Qiang Liu. Greedy optimization provably wins the lottery: Logarithmic number of winning tickets is enough. In Hugo Larochelle, Marc Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin (eds.), Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, Neur IPS 2020, December 6-12, 2020, virtual, 2020. URL https://proceedings.neurips.cc/paper/ 2020/hash/be23c41621390a448779ee72409e5f49-Abstract.html. Rex Ying, Ruining He, Kaifeng Chen, Pong Eksombatchai, William L. Hamilton, and Jure Leskovec. Graph convolutional neural networks for web-scale recommender systems. In Yike Guo and Faisal Farooq (eds.), Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, KDD 2018, London, UK, August 19-23, 2018, pp. 974 983. ACM, 2018. doi: 10.1145/3219819.3219890. URL https://doi.org/10.1145/3219819.3219890. Haonan Yu, Sergey Edunov, Yuandong Tian, and Ari S. Morcos. Playing the lottery with rewards and multiple languages: lottery tickets in RL and NLP. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020. Open Review.net, 2020. URL https://openreview.net/forum?id=S1xn XRVFw H. Geng Yuan, Xiaolong Ma, Wei Niu, Zhengang Li, Zhenglun Kong, Ning Liu, Yifan Gong, Zheng Zhan, Chaoyang He, Qing Jin, et al. Mest: Accurate and fast memory-economic sparse training framework on the edge. Advances in Neural Information Processing Systems, 34:20838 20850, 2021. Seongjun Yun, Minbyul Jeong, Raehyun Kim, Jaewoo Kang, and Hyunwoo J. Kim. Graph transformer networks. In Hanna M. Wallach, Hugo Larochelle, Alina Beygelzimer, Florence d Alché-Buc, Emily B. Fox, and Roman Garnett (eds.), Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, Neur IPS 2019, December 8-14, 2019, Vancouver, BC, Canada, pp. 11960 11970, 2019. URL https://proceedings.neurips.cc/paper/2019/hash/ 9d63484abb477c97640154d40595a3bb-Abstract.html. Shuai Zhang, Meng Wang, Sijia Liu, Pin-Yu Chen, and Jinjun Xiong. Why lottery ticket wins? a theoretical perspective of sample complexity on sparse neural networks. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan (eds.), Advances in Neural Information Processing Systems, volume 34, pp. 2707 2720. Curran Associates, Inc., 2021a. URL https://proceedings.neurips.cc/paper/2021/file/ 15f99f2165aa8c86c9dface16fefd281-Paper.pdf. Zeru Zhang, Jiayin Jin, Zijie Zhang, Yang Zhou, Xin Zhao, Jiaxiang Ren, Ji Liu, Lingfei Wu, Ruoming Jin, and Dejing Dou. Validating the lottery ticket hypothesis with inertial manifold Published as a conference paper at ICLR 2023 theory. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan (eds.), Advances in Neural Information Processing Systems, volume 34, pp. 30196 30210. Curran Associates, Inc., 2021b. URL https://proceedings.neurips.cc/paper/2021/file/ fdc42b6b0ee16a2f866281508ef56730-Paper.pdf. Zhenyu Zhang, Xuxi Chen, Tianlong Chen, and Zhangyang Wang. Efficient lottery ticket finding: Less data is more. In Marina Meila and Tong Zhang (eds.), Proceedings of the 38th International Conference on Machine Learning, ICML 2021, 18-24 July 2021, Virtual Event, volume 139 of Proceedings of Machine Learning Research, pp. 12380 12390. PMLR, 2021c. URL http: //proceedings.mlr.press/v139/zhang21c.html. Hattie Zhou, Janice Lan, Rosanne Liu, and Jason Yosinski. Deconstructing lottery tickets: Zeros, signs, and the supermask. In Hanna M. Wallach, Hugo Larochelle, Alina Beygelzimer, Florence d Alché-Buc, Emily B. Fox, and Roman Garnett (eds.), Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, Neur IPS 2019, December 8-14, 2019, Vancouver, BC, Canada, pp. 3592 3602, 2019. URL https://proceedings.neurips.cc/paper/2019/hash/ 1113d7a76ffceca1bb350bfe145467c6-Abstract.html. A.1 DATASET AND CONFIGURATION The statistics of benchmark datasets in the experiment are summarized in Table 3. For three citation networks: Cora, Cite Seer and Pub Med, the feature of each node corresponds to the bag-of-words representation of the document. For ar Xiv and MAG, each node is associated with a 128-dimensional feature vector obtained by averaging the word embeddings of papers. Three models (GCN, GIN and GAT) are configured with the default parameters in their corresponding source code, including number of layers, dimension of hidden state and learning rate. In our sparsification process, the value of η1, η2 and α are configured as 1e-2, 1e-2 and 1e-1 by default, respectively. In each pruning round, the number of epochs to update masks is configured as the default number of epochs in the training process of the original GNN model. Table 3: Statistics of datasets Dataset #Nodes #Edges Average Deg. Split ratio Features Classes Cora 2,708 5,429 3.88 120/500/1000 1,433 7 Citeseer 3,327 4,732 2.84 140/500/1000 3,703 6 Pub Med 19,717 44,338 4.50 60/500/1000 500 3 ar Xiv 169,343 1,166,243 13.7 54%18%/28% 128 40 MAG 1,939,743 21,111,007 21.7 85%/9%/6% 128 349 Table 4: GLT comparison Cora Model GCN GIN GAT Methods/Metrics GS(%) WS(%) Accuracy(%) GS(%) WS(%) Accuracy(%) GS(%) WS(%) Accuracy(%) UGS 9.76 36.42 81.3 43.96 91.88 80.1 Ours 18.52 59.12 82.2 4.86 20.07 79.6 63.82 98.93 80.7 Citeseer Model GCN GIN GAT Methods/Metrics GS(%) WS(%) Accuracy(%) GS(%) WS(%) Accuracy(%) GS(%) WS(%) Accuracy(%) UGS 22.65 67.34 71.7 19.03 59.43 70.3 64.48 99.41 70.2 Ours 22.63 67.76 72.1 33.89 83.57 69.3 62.29 98.63 70.8 Pub Med Model GCN GIN GAT Methods/Metrics GS(%) WS(%) Accuracy(%) GS(%) WS(%) Accuracy(%) GS(%) WS(%) Accuracy(%) UGS 48.69 94.53 79.5 35.32 83.27 78.1 57.64 97.32 78.8 Ours 65.35 98.91 79.6 48.84 94.52 78.5 64.26 99.07 80.4 Published as a conference paper at ICLR 2023 A.2 GLT COMPARISON In Table 4, we show the test accuracy, graph sparsity (GS) and weight sparsity (WS) of the last GLT that reaches higher accuracy than the original model in the sparsification process of UGS and our method. Note that GLT of GIN cannot be found on Cora in the sparsification process. We can observe that the last GLT identified by our method either has higher sparsity or achieves higher accuracy with comparable sparsity. In Table 5, we evaluate the test accuracy of sparsified GCN on two large-scale graph datasets: ar Xiv and MAG. We can observe that our method can result in higher accuracy than UGS with comparable graph sparsity and weight sparsity after 5, 15, and 20 iterations. It further verifies the effectiveness of the proposed method on large-scale datasets. Table 5: Performance of GCN on large-scale graph datasets ogbn-arxiv Iterations 5 15 20 Methods/Metrics GS(%) WS(%) Accuracy(%) GS(%) WS(%) Accuracy(%) GS(%) WS(%) Accuracy(%) UGS 22.65 67.23 71.92 53.69 96.49 71.15 64.27 98.86 69.87 Ours 23.51 67.24 71.94 52.81 96.52 71.89 63.82 98.93 70.79 ogbn-mag Iterations 5 15 20 Methods/Metrics GS(%) WS(%) Accuracy(%) GS(%) WS(%) Accuracy(%) GS(%) WS(%) Accuracy(%) UGS 23.66 67.31 33.43 55.99 96.50 30.71 67.42 98.77 30.12 Ours 22.65 67.31 34.11 54.81 96.49 32.23 63.59 98.85 31.76 A.3 PERFORMANCE STANDARD AND INFERENCE TIME To combat randomness, we visualize the standard deviation of accuracy on Cora with a shaded color. Following UGS, we also show the inference MACs (multiply accumulate operations) of pruned models in the sparsification process. We can see that our method can significantly reduce the inference cost of GNNs with less performance drop. Inference MACs (M) Inference MACs (M) Inference MACs (M) Figure 9: Performance over inference MACs of GCN, GIN, and GAT on Cora A.4 TRANSFER LEARNING WITH GLTS To verify the transferable graph lottery ticket hypothesis, we identify the GLT on ar Xiv and investigate the transfer ratio Kooverjee et al. (2022) of this GLT on MAG. Since the number of classes on ar Xiv 0 200 400 600 800 1000 Epochs Accuracy on target task Post-trained GLT Re-initialized GLT Randomly initialized GNN Post-trained GNN Randomly initialized GLT (a) Replacing Last Dense Layer 0 200 400 600 800 1000 Epochs Accuracy on target task Post-trained GLT Re-initialized GLT Randomly initialized GNN Post-trained GNN Randomly initialized GLT (b) Add a New Dense Layer 0 200 400 600 800 1000 Epochs Accuracy on target task WS=67.76% WS=89.26% WS=96.48% WS=98.85% (c) Different GLTs Figure 10: Transfer Learning from Arxiv to MAG Published as a conference paper at ICLR 2023 is different from that on MAG, we need to either add a new dense layer for class mapping or replace the last dense layer to match the output classes in MAG. Figure 10(a) shows the target-task test accuracy of 3-layer GLTs (or GNNs) where the last dense layer is replaced. We can see that the post-trained GLT and post-trained GNN identified on the source task have a similar transfer ratio on the target task. Both the post-trained GLT and post-trained GNN have significant improvement over the randomly initialized GNNs. Also, we can observe improvement of re-initialized GLT over randomly initialized GLT. As shown in Figure 10(b), similar results can be observed when adding a new dense layer. Lastly, Figure 10(c) investigates the relation between sparsity and transfer ratio (recall that we only transfer WS rather than GS), where we can see that the performance of the target task decreases as the sparsity of GLT increases.