# s2fgl_spatial_spectral_federated_graph_learning__4de11dc7.pdf S2FGL: Spatial Spectral Federated Graph Learning Zihan Tan * 1 Suyuan Huang * 1 Guancheng Wan 1 Wenke Huang 1 He Li 1 Mang Ye 1 Federated Graph Learning (FGL) combines the privacy-preserving capabilities of Federated Learning (FL) with the strong graph modeling capability of Graph Neural Networks (GNNs). Current research addresses subgraph-FL from the structural perspective, neglecting the propagation of graph signals on the spatial and spectral domains of the structure. From a spatial perspective, subgraph-FL introduces edge disconnections between clients, leading to disruptions in label signals and a degradation in the semantic knowledge of the global GNN. From a spectral perspective, spectral heterogeneity causes inconsistencies in signal frequencies across subgraphs, which makes local GNNs overfit the local signal propagation schemes. As a result, spectral client drift occurs, undermining global generalizability. To tackle the challenges, we propose a global knowledge repository to mitigate the challenge of poor semantic knowledge caused by label signal disruption. Furthermore, we design a frequency alignment to address spectral client drift. The combination of Spatial and Spectral strategies forms our framework S2FGL. Extensive experiments on multiple datasets demonstrate the superiority of S2FGL. The code is available at https://github. com/Wonder7racer/S2FGL.git 1. Introduction Graph Neural Networks (GNNs) have demonstrated remarkable efficacy in modeling graph-structured data (Wan et al., 2025a; Fang et al., 2025), thereby finding applications across various domains, such as social networks (Fan et al., 2020; Zhang et al., 2022b), epidemiology (Liu et al., 2024), and *Equal contribution 1National Engineering Research Center for Multimedia Software, School of Computer Science, Wuhan University, Wuhan, China. Correspondence to: Mang Ye . Proceedings of the 42 nd International Conference on Machine Learning, Vancouver, Canada. PMLR 267, 2025. Copyright 2025 by the author(s). Client Scale Structure Inertia Score Spectral Heterogeneity Poor Semantic Figure 1. In the first place, compared with centralized GNN training, subgraph-FL is encountering label signal disruption challenge, leading to decreased structure inertia score and poor semantic knowledge for GNNs. Moreover, we demonstrate the heat map of the Kullback-Leibler divergence of eigenvalue distributions across clients. Inconsistency in subgraph signal frequency caused by spectral heterogeneity leads to spectral client drift. fraud detection (Wang et al., 2019; Tang et al., 2022). However, in real-world scenarios, graph data is often generated at the edge devices rather than in centralized systems (Zhang et al., 2021a). To address this, Federated Graph Learning (FGL) has emerged (Fu et al., 2022; Liu & Yu, 2022; Tan et al., 2025; 2024; Huang et al., 2022; Wan et al., 2024b; 2025b), leveraging the data privacy-preserving capabilities of Federated Learning (FL) (Huang et al., 2024; 2023b;c; 2022) to enable the efficient distributed training of GNNs (Huang et al., 2024). A prominent application of FGL is subgraph-FL, in which each participant holds a subgraph derived from the same overall graph data. Although numerous FGL methods have attempted to provide solutions based on structure to enhance effectiveness, including identifying structurally similar collaborators (Baek et al., 2023; Xie et al., 2021; Li et al., 2024), enhancing structural knowledge exchange (Tan et al., 2023; Huang et al., 2023a; Tan et al., 2025), and retrieving generic information under structural shifts (Wan et al., 2024a; Tan et al., 2024). Nevertheless, these approaches overlooked the propagation S2FGL: Spatial Spectral Federated Graph Learning Centralized Learning Client i Client j : Label Signal Disruption : Labeled Node : Signal Propagation Client i Client j Spectra Spectra 0 1 2 0 1 2 (a) Label Signal Disruption (b) Spectral Client Drifts Figure 2. Problem Illustration. (a) From the spatial perspective, nodes in subgraph-FL lose label signals from originally nearby labeled nodes due to edge loss, namely label signal disruption. Correspondingly, GNNs suffer from poor semantic knowledge, leading to a deteriorated global GNN. (b) From the spectral perspective, spectral heterogeneity induces inconsistencies in signal frequencies across subgraphs, leading to spectral client drift in the signal propagation paradigms of GNNs and degraded global generalizability. of graph signals within the structure. Specifically, graph signal propagation can be analyzed from two perspectives: the spatial and the spectral domain. Specifically, the spatial domain governs the explicit transmission of signals among linked nodes, while the spectral domain characterizes signal diffusion across varying frequency spectra. From the spatial perspective, due to edge loss, we hypothesize that nodes in subgraph-FL lose label signals from originally nearby labeled nodes. This degradation hampers the ability of GNNs to learn comprehensive semantic knowledge, resulting in poor global performance and reduced generalizability. Correspondingly, we define this phenomenon as label signal disruption, which naturally exists in subgraph-FL. For verifying, inspired by graph active learning research (Han et al., 2023), we investigate how the Structure Inertia Score (SIS) varies in subgraph-FL. Specifically, SIS evaluates the influence and significance of labels on graphs. In Fig. 1, we empirically demonstrate that the SIS decreases in subgraph-FL compared with centralized training. Correspondingly, existing methods suffer from poor semantic knowledge. Based on our empirical analysis, we pose the question: I) How can we address the challenge of poor semantic knowledge under label signal disruption? From the spectral perspective, inconsistencies in signal frequencies across clients caused by spectral heterogeneity induce spectral client drift in the signal transmission schemes of GNNs, thereby undermining the collaboration. To verify this phenomenon, we examine graph spectra across clients and demonstrate the heterogeneity in Fig. 1. It reveals inconsistent eigenvalue distribution across clients. As a result, GNNs learn distinct signal propagation schemes of sub- graphs and optimize in different spectral directions, leading to spectral client drift and degraded generalizability. Based on our analysis, we pose the question: II) How can we alleviate spectral client drift under spectral heterogeneity? To address the challenge of poor semantic knowledge under label signal disruption in Question I), we propose Node Label Information Reinforcement (NLIR). Specifically, our strategy leverages structurally and semantically representative nodes to construct a prototype-based global repository of semantic knowledge. During training, NLIR calculates the similarity distribution between all representative prototypes with node features, which provides multidimensional semantic localization of nodes. Consequently, our strategy injects semantic knowledge from the repository into the local GNN during training, effectively mitigating the issue of poor semantic knowledge under label signal disruption. Considering the spectral client drift posed by spectral heterogeneity in II), we propose Frequency-aware Graph Modeling Alignment (FGMA). Our method utilizes the similarity relationship of the node feature of the frozen global GNN and the local GNN to reconstruct spectra that incorporates GNNs adjacency awareness. FGMA then projects the highfrequency and low-frequency components of the features onto this spectrum. Subsequently, by aligning the local projections with the global one, we encourage the GNNs to learn a globally generic frequency processing scheme, thereby mitigating spectral client drift. In conclusion, our key contributions are: First, we identify and empirically reveal the issue of S2FGL: Spatial Spectral Federated Graph Learning poor semantic knowledge under label signal disruption. In addition, we reveal the spectral client drift under spectral heterogeneity in subgraph-FL. We design our framework S2FGL including strategy Node Label Information Reinforcement and Frequency-aware Graph Modeling Alignment, effectively addressing the challenges of poor semantic knowledge and spectral client drift in subgraph-FL. We conduct extensive experiments on various datasets, validating the superiority of our proposed S2FGL. 2. Related Work Federated Graph Learning. Federated graph learning leverages the powerful graph modeling capabilities of GNNs along with the privacy-preserving attributes of federated learning, thus gaining increasing attention these days (He et al., 2021a; Fu et al., 2022; Liu & Yu, 2022; Wan et al., 2025b). Current FGL research can generally be categorized into two types: intra-graph FGL and inter-graph FGL. Intra-graph FGL research primarily focuses on subgraph-FL scenarios, where each client participates in the collaboration with a part of the whole graph (Zhang et al., 2021b). Correspondingly, the training targets include missing link prediction (Chen et al., 2021; Baek et al., 2023), node classification (Huang et al., 2023a; Li et al., 2024; Wan et al., 2024a; Zhu et al., 2024), and so on. On the other hand, clients in inter-graph FGL own independent local graph data, such as multiple graphs from different domains (Tan et al., 2023; Xie et al., 2021). In this paper, we focus on subgraph-FL scenarios of intra-graph FGL. Specifically, we are the first to empirically reveal and address the challenge of poor semantic knowledge under label signal disruption and client drift under spectral heterogeneity among subgraphs, while existing methods inevitably fail spatially and spectrally due to the lack of targeted solutions. Federated Learning. Federated learning (Huang et al., 2023c; 2024; Yang et al., 2023; Wan et al., 2024a) has gained increasing attention in recent years as it addresses the issue of data silos while ensuring data privacy. Several research directions have emerged from FL, including robustness (Xu et al., 2022; Hong et al., 2023; Zhu et al., 2023; Fang & Ye, 2022), fairness (Chen et al., 2024; Ezzeldin et al., 2023; Ray Chaudhury et al., 2022), and asynchronous federated learning (Xu et al., 2023; Zhang et al., 2023d). Generally, FL can be categorized into two main types by their optimization objective: traditional FL (t FL) and personalized FL (Hu et al., 2024; Shang et al., 2022; Lv et al., 2024; Smith et al., 2017). Research of t FL aims at aggregating a highly generalizable global model (Mc Mahan et al., 2017; Li et al., 2020; Acar et al., 2021; Zhang et al., 2022a). For instance, Fed NTD (Lee et al., 2022) preserves the global perspective on local data for the not-true classes, FEDGEN (Zhu et al., 2021) ensembles user information in a data-free manner to regulate local training, and SCAFFOLD (Karimireddy et al., 2020) uses variance reduction for the client drift phenomenon. Instead, strategies of personalized FL (p FL) aim to customize models that perform optimally for each client (Wu et al., 2023; Zhou & Konukoglu, 2023; Li et al., 2021; Zhang et al., 2023b). Specifically, Fed ALA (Zhang et al., 2023c) proposed adaptive masks to achieve personalized aggregation, DBE (Zhang et al., 2023a) stores domain biases for elimination, and Fed Ro D (Chen & Chao, 2022) leverages two heads for global and personalized tasks. Graph Spectrum Being related closely to graph connectivity, signal propagation, and structure, graph spectra have proven essential in performing various tasks on graphstructured data. For instance, it plays an essential role in anatomy detection, (Gao et al., 2023; Tang et al., 2022), graph condensation (Kreuzer et al., 2021; Liu et al., 2023), and graph contrastive learning (Bo et al., 2023a; Liu et al., 2022). Additionally, spectral GNNs (Wu et al., 2020) based on spectral filters are showing powerful ability in modeling graph data and attracting more attention. Specifically, existing research either (He et al., 2021b; Defferrard et al., 2016; He et al., 2022; Wang & Zhang, 2023) leverages various orthogonal polynomials to approximate arbitrary filters, or utilizes neural networks to parameterize the filters (Liao et al., 2019; Bo et al., 2023b). Although the potential of graph spectrum has been explored in various scenarios and tasks, the spectral domain in generalizable subgraph-FL has remained unexplored. Consequently, current methods suffer from optimization diverging on spectra and are trapped in suboptimal learning. Instead, our approach remarkably mitigates the challenge by targeted alignment on spectra. Graph Signal Propagation: Graph signal propagation describes how node signals diffuse on graph structures. In the spatial domain, propagation occurs through explicit signal passing along edges. In the spectral domain, propagation is characterized by how signals distribute across different frequency components. Label Signal Disruption: As subgraphs experience edge loss, nodes lose critical label signals containing class knowledge from their formerly adjacent labeled neighbors. Consequently, it limits the ability of GNNs to capture class distinctions accurately, leading to poor semantic knowledge under label signal disruption. Spectral Client Drift: Inconsistencies in signal frequencies on graph spectra across subgraphs lead to spectral heterogeneity and diverging signal propagation schemes, causing spectral client drift and degrading the generalizability of the global model. S2FGL: Spatial Spectral Federated Graph Learning Client Objective: Similarity Distribution Similarity Distribution Feature Similarity Low High Feature Similarity Low High Global Model Local Model Global Repository Representative Node (a) Node Label Information Reinforcement (b) Frequency-aware Graph Modeling Alignment Figure 3. Framework Illustration. (a) Node Label Information Reinforcement (NLIR) leverages a structurally and semantically representative global prototype repository. It provides multidimensional semantic localization of nodes through similarity distribution and allows LFKD to inject the semantic knowledge during training. (b) Frequency-aware Graph Modeling Alignment (FGMA) aligns local high and low spectral adjacency awareness with the global GNN for a generic signal propagation scheme, mitigating spectral drifts. 3. Problem Statement Notation. Let the graph data be represented as G = (V, E), where V is the set of nodes with |V| = N vertices, and E V V denotes the set of edges connecting these nodes. The adjacency matrix is represented by A RN N, where Auv = 1 indicates the presence of an edge euv E, and Auv = 0 otherwise. Moreover, X the feature vector matrix of the graph G. The Laplacian matrix is given by L = D A, where D is the degree matrix. The unitary matrix U is composed of the eigenvectors of L. To distinguish between local and global properties, we introduce the following notation: the symbol i represents local properties or entities, whereas g denotes global properties or entities. Definition 3.1. Personalized Page Rank (PPR): The PPR matrix quantifies the influence each node has on every other node within the graph and is defined as: P = α I (1 α)D 1A 1 . (1) Here, α (0, 1) is the teleportation probability, representing the probability of the random walk restarting from the source node. A typical value is 0.15, which makes the continuation probability (1 α) = 0.85. I is the identity matrix, A is the adjacency matrix of the graph, and D is the degree matrix with Dii denoting the degree of node i. Definition 3.2. Structure Inertia Score (SIS): The SIS quantifies the cumulative influence of the training nodes on the entire graph and is defined as: SIS(P, t) = i=1 max j (Pi,j tj) . (2) Here, P is the PPR matrix, and t {0, 1}n is a binary vector indicating the training nodes, where tj = 1 if node j is part of the training set, and tj = 0 otherwise. The SIS aggregates the maximum personalized Page Rank values from each node to any labeled node, effectively measuring the strongest influence each node in the graph receives from the training set. A higher SIS indicates greater structural inertia, suggesting that the labels have a significant influence over the network overall graph structure. 4. Methodology 4.1. Motivation Signal propagation over graph structures fundamentally shapes the signal transmission paradigm of GNNs. Therefore, rather than focusing solely on challenges arising from static graph structures in subgraph-FL, it is crucial to consider the dynamics of signal propagation. Correspondingly, S2FGL: Spatial Spectral Federated Graph Learning we conduct our analysis from both the spatial and spectral domains from the perspective of the graph signal. Specifically, the spatial domain governs the explicit passing of signals between connected nodes, while the spectral domain captures signal diffusion across different frequencies. Accordingly, we empirically validate the presence of two major challenges from the spatial and spectral perspectives: label signal disruption and spectral client drift. These phenomena respectively pose challenges of poor semantic knowledge and spectral client drift, which severely constrain the potential of subgraph-FL collaboration. Motivation of NLIR. Graph data is fragmented across clients in subgraph-FL, which inevitably disrupts semantic signals from labeled nodes across clients. Label signal disruption undermines key pathways for propagating semantic information. This results in biased local feature representations, which ultimately degrade the performance of GNNs. For validation, we investigate the relationship between the decrease in SIS and the client scale. Specifically, the SIS exhibits a downward trend as the client scale increases, with notably lower scores in the range emphasized by mainstream subgraph-FL studies, thereby highlighting the label signal disruption phenomenon. We aim to mitigate its negative impact by preserving valuable semantic knowledge across fragmented subgraphs. Correspondingly, we propose NLIR. By selecting nodes with both structural representativeness and rich semantic information for the construction of a global repository and injecting it during local training, NLIR reduces the information loss inherent in the subgraph scenarios. Subsequently, NLIR assesses the similarity distributions between node features and all representative prototypes for both local and global GNNs during local training, thereby enabling multidimensional semantic localization of nodes. By aligning the two similarity distributions, it effectively injects semantic knowledge and enhances feature modeling semantically. Motivation of FGMA. We reveal the challenge of spectral client drift in subgraph-FL in Fig. 1, where GNNs across different clients capture frequency information inconsistently due to graph spectral heterogeneity. This further leads to overfitting local signal propagation frequencies, which brings spectral conflicts during collaboration and compromises the generalizability of the global GNN. To address spectral client drift, we propose the Frequency-aware Graph Modeling Alignment. Specifically, FGMA reconstructs the local graph spectra with the GNN adjacency awareness by calculating the node similarity matrix. Subsequently, by projecting feature representations separately onto highand low-frequency components of the reconstructed spectra and aligning the local and global projections, FGMA promotes the learning of a generalizable spectral signal propagation paradigm across clients, thereby reducing frequency-based discrepancies during collaboration and mitigating spectral client drift. Consequently, our strategy effectively enhances the global generalizability spectrally. 4.2. Node Label Information Reinforcement First of all, we introduce the Structure-Aware Label Centrality (SALC) metric, denoted as ΛSALC u for node u. It is defined as the combination of the label influence centrality Λl u and the structural prominence score Λs u: ΛSALC u = Λs u + Λl u, (3) where Λs u assesses the structural representativeness of node u, while Λl u quantifies the influence propagation of labels. Λs u = max Pu,v τv , Λl u = X P (L) v,u , (4) where Pu,v is the (u, v)-th element of the standard PPR matrix P, and τv represents the prior importance score of node v, typically initialized to 1 for all nodes. The structural prominence score captures the maximum influence exerted by any node on node u, weighted by its importance. As clarified, the PPR matrix used for computing label influence centrality is denoted as P(L), and defined as: P(L) = α I (1 α)D 1A 1 . (5) Here, VL denotes the set of labeled nodes, and P (L) v,u represents the influence of node v on node u as captured by P(L). To accurately capture the influence of labeled nodes, the inclusion of self-loops in A ensures that each labeled node s own label contributes to its Λl u. Compared to the intuitive approach of directly selecting labeled nodes, the SALC metric ΛSALC u considers both structural representativeness of the nodes and diffusion of label signals, thus avoiding biases caused by isolated labeled nodes. It is also capable of selecting unlabeled nodes that still possess rich label signals and structural advantages. This improves knowledge quality, enriches the repository, and mitigates the label signal disruption problem. After computing the SALC scores, we rank the nodes based on their ΛSALC u values and select the top K nodes, where the default value of K is 1/3 of the total number of nodes. Subsequently, for each class c, the local prototype Hi c at each client is computed as the mean feature vector of the selected nodes belonging to class c: Hi c = 1 |Vic| u Vic hi u, (6) where Vi c represents the set of selected nodes categorized as class c on client i, and hi u is the feature vector of node u on client i. Once the local prototypes are computed, clients upload their prototypes to the server along with the node count. For each class c, the server aggregates the S2FGL: Spatial Spectral Federated Graph Learning prototypes from α percent of the clients by weighting each local prototype according to its sample size. Four global anchor prototypes are constructed for each class. Each global prototype Hg,k c is computed as: Hg,k c = 1 P i N k c |Vic| i N k c |Vi c|Hi c, H = Hg,1 1 Hg,2 1... Hg,4 C where Hg,k c represents the k-th global prototype for class c, N k c is the set of clients randomly selected to contribute to the k-th global prototype for class c, and C is the number of classes. The global repository H contains all the global prototypes and will be broadcast back to clients. After the global knowledge repository is constructed, it is distributed to the local clients along with the model parameters. The global features hg u used in the following loss formulation refer to the frozen inference features extracted locally using the distributed global model. To regulate local training, we propose a federated knowledge distillation loss function aimed at harmonizing the semantic feature localization of the local GNN with its global counterpart, namely by aligning their similarity distributions for the representative prototypes stored in the global repository: LFKD = 1 |Vi| u Vi KL σ φ(hi u, H) , σ (φ(hg u, H)) , (8) where KL( , ) represents the Kullback-Leibler divergence, and σ( ) is the softmax function applied to similarity scores computed by the function φ(h, H), which returns a vector of cosine similarities between the feature and all prototypes. 4.3. Frequency-aware Graph Modeling Alignment To emphasize the GNN spectral adjacency awareness and more accurately capture similarities between nodes, we leverage the feature matrix h to compute node similarity matrices. In the construction of the following similarity matrix, h denotes operations applied to both hi and hg. Specifically, for each node u, we identify its ksim most similar neighbors based on the cosine similarity of their feature vectors hu. We then construct a sparse self-similarity matrix S as: ( hu hv hu 2 hv 2 v among the top ksim 0 otherwise . (9) Subsequently, we calculate the graph laplacian matrix L based on this sparse similarity matrix S as: L = D S , (10) where D is the diagonal degree matrix of S . Moreover, we perform eigendecomposition on the laplacians L i and L g. For L i, let {ulow,i m }keig m=1 be the eigenvectors corresponding to the smallest eigenvalues, which represents low-frequency. While {uhigh,i m }keig m=1 denotes the largest eigenvalues, which represents high-frequency. Similarly, for L g, we obtain {ulow,g m }keig m=1 and {uhigh,g m }keig m=1. The feature matrix h is then projected onto each of these eigenvectors. For instance: Zlow m = (ulow m ulow T m )h, Zhigh m = (uhigh m uhigh T m )h. (11) Applying these projections for each m {1, . . . , keig}, we obtain several sets of projected feature matrices used in the loss computation. Specifically, local features hi are projected onto low/high-frequency eigenvectors of the local graph, yielding Zi,low m and Zi,high m , respectively. Similarly, frozen global inference features hg are projected onto corresponding eigenvectors from L g, yielding Zg,low m and Zg,high m . Conequently, loss LFGMA is then defined as the sum of MSE over all eigenvector-projected pairs: MSE(Zi,low m , Zg,low m ) + MSE(Zi,high m , Zg,high m ) . (12) This loss addresses spectral heterogeneity by aligning client and global signal characteristics in both spectral domains. Combining strategies Node Label Information Reinforcementand Frequency-aware Graph Modeling Alignment, our framework S2FGL reinforces semantic knowledge during local modeling and mitigates spectral client drift. Ultimately, the loss for local training is: L = LCE + λ1LFKD + λ2LFGMA, (13) where LCE denotes the standard cross-entropy loss for node classification, while λ1 and λ2 are balancing hyperparameters for the proposed methods NLIR and FGMA. 5. Experiments 5.1. Experimental Setup Datasets. We conducted experiments on various datasets to validate the superiority of our framework S2FGL. The homophilic graph datasets include Cora, Citeseer, and Pubmed, while the heterophilic graph datasets comprise Texas, Wisconsin, and Minesweeper. The following provides a description of each dataset. Cora (Mc Callum et al., 2000) dataset consists of 2708 scientific publications classified into one of seven classes. There are 5429 edges in the network of citations. 1433 distinct words make up the dictionary. Citeseer (Giles et al., 1998) dataset consists of 3312 scientific publications classified into one of six classes and 4732 edges. The dictionary contains 3703 unique words. Pubmed (Sen et al., 2008) dataset consists of 19717 scientific papers on diabetes that have been categorized into one of three categories in the Pub Med database. The citation network has 44338 edges. S2FGL: Spatial Spectral Federated Graph Learning Table 1. Performance Comparison with the state-of-the-art methods on homophilic and heterophilic graph datasets. We report the node classification accuracies with the performance improvement over Fed Avg. The best results are highlighted in bold. Methods Cora Cite Seer Pub Med Texas Wisconsin Minesweeper Fed Avg [ASTAT17] 81.9 0.7 74.3 0.4 87.3 0.3 72.8 2.2 77.6 2.7 79.6 0.1 Fed Prox [ar Xiv18] 82.1 0.5 0.2 74.4 0.3 0.1 87.9 0.4 0.6 73.5 3.7 0.7 77.3 3.4 0.3 79.7 0.1 0.1 Fed Nova [Neur IPS20] 81.6 1.2 0.3 74.4 0.4 0.1 88.2 0.5 0.9 73.0 4.4 0.2 77.4 4.2 0.2 79.9 0.4 0.3 Fed Fa [ICLR23] 82.7 0.5 0.8 74.9 0.6 0.6 87.8 0.5 0.5 73.9 3.6 1.1 78.1 4.6 0.5 80.1 0.3 0.5 Fed Sage+ [Neur IPS19] 82.3 0.7 0.4 75.2 0.3 0.9 88.2 0.7 0.9 73.7 4.0 0.9 79.0 3.3 1.4 79.9 0.2 0.3 Fed Star [AAAI23] 82.6 0.5 0.7 74.5 0.3 0.2 88.1 0.6 0.8 74.3 2.7 1.5 78.3 4.7 0.7 79.8 0.1 0.2 Fed Pub [ICML23] 82.3 0.8 0.4 74.8 0.7 0.5 88.0 0.4 0.7 73.4 3.5 0.6 77.8 3.1 0.2 79.9 0.2 0.3 FGSSL [IJCAI23] 82.6 0.4 0.7 74.9 0.2 0.6 87.6 0.7 0.3 73.6 4.6 0.8 77.8 3.8 0.2 79.9 0.2 0.3 Fed GTA [VLDB24] 82.4 0.8 0.5 75.1 0.5 0.8 87.7 0.9 0.4 72.6 4.2 0.2 77.8 4.1 0.2 80.2 0.3 0.6 FGGP [AAAI24] 82.5 0.4 0.6 74.7 0.5 0.4 87.5 0.4 0.2 73.6 2.8 0.8 78.2 3.4 0.6 80.4 0.3 0.8 S2FGL (ours) 83.4 0.2 1.5 76.0 0.3 1.7 88.6 0.2 1.3 74.8 2.3 2.0 79.0 1.0 1.4 80.5 0.1 0.9 Texas and Wisconsin datasets are subsets of the Web KB dataset (Craven et al., 1998). The Web KB dataset was introduced in 1998, comprising web pages from the computer science departments of various universities, including the University of Texas and the University of Wisconsin. The dataset is commonly used for tasks such as webpage classification and link prediction, serving as a benchmark for evaluating machine learning models in graph-based learning scenarios. Minesweeper (Baranovskiy et al., 2023) dataset is a synthetic graph dataset inspired by the Minesweeper game. In this dataset, the graph is structured as a regular 100x100 grid, where each node represents a cell connected to its neighboring nodes, except for edge nodes, which have fewer neighbors. The primary task is to predict which nodes contain mines. This dataset is commonly used to evaluate the performance of GNNs under heterophily. Evaluation Metric. Following mainstream FGL research experimantal practices, we utilize the accuracy of the node classification task as the evaluation metric. Baselines. We compare S2FGL with several state-of-theart approaches, including traditional federated learning methods such as Fed Avg (Mc Mahan et al., 2017), Fed Prox (Li et al., 2020), Fed Nova (Wang et al., 2020), and Fed Fa (Zhou & Konukoglu, 2023); federated graph learning approaches including FGSSL (Huang et al., 2023a) and FGGP (Wan et al., 2024a); as well as personalized federated graph learning methods such as Fed Sage+ (Zhang et al., 2021b), Fed Star (Tan et al., 2023), Fed Pub (Baek et al., 2023), and Fed GTA (Li et al., 2024). This comprehensive set of baseline methods spans various FL and FGL paradigms, allowing us to evaluate the performance and advantages of our proposed S2FGL across diverse scenarios. Implement Details. Following prevalent methodologies in FGL research, we employ the Louvain community detection algorithm to partition the graph into subgraphs assigned to different clients. For each dataset, we divide the nodes into training, validation, and testing sets with ratios of 60%, 20%, and 20%, respectively. Additionally, we simulate various collaborative scenarios by configuring the number of clients to 10 for Cora, Citeseer, Pubmed, and Minesweeper datasets, and 3 for the Texas and Wisconsin datasets. The primary evaluation metric is the node classification accuracy on the clients test sets. We conduct each experiment five times and report the average accuracy from the last five communication epochs as the final performance. We conduct experiments with the ACM-GCN (Luan et al., 2022), which achieves a strong ability on both homophilic and heterophilic graph datasets. 5.2. Experiment Results In this section, we comprehensively evaluate the proposed S2FGL by addressing the following questions: Q1: How does S2FGL perform compared to existing FL and FGL methods in subgraph-FL? Q2: What is the impact of each component of S2FGL on the overall performance? Q3: Does S2FGL maintain consistent performance across different hyperparameter and client scales? Q4: Do NLIR and FGMA mitigate the effects of label signal disruption and spectral heterogeneity? Q1: How does S2FGL perform compared to existing FL and FGL methods in subgraph-FL? We present the results of node classification tasks across various FGL scenarios using multiple graph datasets, and S2FGL: Spatial Spectral Federated Graph Learning we summarize the final average test accuracy in Tab. 1. It demonstrates that our proposed S2FGL consistently outperforms all baseline approaches across all six datasets. This superiority highlights the effectiveness of S2FGL. Q2: What is the impact of each component of S2FGL on the overall performance? To evaluate the individual contributions of NLIR and FGMA strategies within the S2FGL framework, we conducted ablation experiments on the Cora and Citeseer datasets. In this study, we removed each component to assess its impact on the overall performance. The results are presented in Tab. 2, which demonstrate that both NLIR and FGMA independently contribute to the overall performance. Dataset NLIR FGMA Cora Citeseer 81.9 0.7 74.3 0.4 83.2 0.4 75.6 0.3 82.6 0.3 75.0 0.2 83.4 0.2 76.0 0.3 Table 2. Ablation study on key components of S2FGL. Q3: Does S2FGL maintain consistent performance across different hyperparameter and client scales? We evaluated the stability and adaptability of our proposed framework S2FGL under varying hyperparameter configurations of and different client scales. Varying Hyperparameters of NLIR and FGMA. For NLIR, we test hyperparameter sets of 100, 50, 10, and 1. For FGMA, the settings are 0.01, 0.05, 0.5, and 1. The results in Fig. 4 indicate that our method maintains consistent performance across varying hyperparameter configurations. (b) Citeseer Figure 4. Analysis of the performance growth between S2FGL and Fed Avg under different hyperparameters of NLIR and FGMA Varying Client Scales. We assessed performance with different client scales: 5, 10 and 20. Specifically, we compared S2FGL with other FL and FGL baselines, including Fed Avg, Fed Prox, and FGSSL. The results in Fig. 5 demonstrate that S2FGL consistently delivers reliable results regardless of the client scales. Overall, S2FGL exhibits strong stability (b) Citeseer Figure 5. Analysis of performance under different client scales. and adaptability across varying hyperparameter settings and client partition configurations on both the Cora and Citeseer datasets. These findings confirm that S2FGL not only sustains its effectiveness under diverse conditions but also adapts seamlessly to varying client scales, demonstrating its suitability for real-world subgraph-FL scenarios. Q4: Do NLIR and FGMA mitigate the effects of label signal disruption and spectral heterogeneity? In Fig. 1, our experiments demonstrate that the SIS score in subgraph scenarios declines compared to centralized training, while spectral heterogeneity exists among clients. Here, we further verify the targeted effectiveness of NILR and FGMA with respect to these two issues, respectively. First, in Fig. 6 (a), we investigate the relationship between the performance gain brought by NILR to Fed Avg and the change in the SIS score. The results show that this method achieves higher performance when semantic signals are more limited, thereby confirming its effectiveness and targeted nature. Second, in Fig. 6 (b), we examine the relationship between the performance improvement of FGMA for Fed Avg and spectral heterogeneity, measured by the average KL divergence between the eigenvalue distributions of different clients. Experimental results show that greater spectral heterogeneity corresponds to larger performance gains, confirming the effectiveness of our method. Specifically, green indicates the performance gain of the proposed method relative to Fed Avg, blue denotes variations in SIS, and orange corresponds to spectral heterogeneity. Client Scale Client Scale Heterogeneity Figure 6. Analysis of the targeted effectiveness of NILR and FGMA. (a) The performance gain from NILR increases as the semantic signal becomes limited. (b) The performance improvement from FGMA grows with higher spectral heterogeneity. S2FGL: Spatial Spectral Federated Graph Learning 6. Conclusion In this paper, we identify and empirically demonstrate two phenomena in subgraph-FL from both spatial and spectral perspectives of graph signal propagation: label signal disruption and spectral heterogeneity. These phenomena pose challenges of poor semantic knowledge and spectral client drift. To address these challenges, we propose two key strategies: NLIR and FGMA. NLIR selects structurally and semantically representative nodes and constructs a global repository accordingly. By injecting semantic information from the repository into local training, it alleviates the poor semantic knowledge caused by label signal disruption. In addition, FGMA aligns and feature projections in both the highand low-frequency reconstructed graph spectra, thereby promoting a generic signal propagation paradigm and mitigating client drifts under spectral heterogeneity. By integrating these strategies, S2FGL effectively tackles both spatial and spectral challenges in subgraph-FL. Extensive experiments on multiple datasets demonstrate that S2FGL significantly enhances global generalizability. Acknowledgement This work is supported by the National Key Research and Development Program of China (2024YFC3308400), and National Natural Science Foundation of China under Grant (62361166629, 62176188, 62225113, 623B2080), the Wuhan University Undergraduate Innovation Research Fund Project. The supercomputing system at the Supercomputing Center of Wuhan University supported the numerical calculations in this paper. Impact Statement This paper presents work whose goal is to advance the field of Machine Learning. There are many potential societal consequences of our work, none of which we feel must be specifically highlighted here. Acar, D. A. E., Zhao, Y., Matas, R., Mattina, M., Whatmough, P., and Saligrama, V. Federated learning based on dynamic regularization. In ICLR, 2021. Baek, J., Jeong, W., Jin, J., Yoon, J., and Hwang, S. J. Personalized subgraph federated learning. In ICML, pp. 1396 1415, 2023. Baranovskiy, D., Oseledets, I., and Babenko, A. A critical look at the evaluation of gnns under heterophily: Are we really making progress? ar Xiv preprint ar Xiv:2302.11640, 2023. Bo, D., Fang, Y., Liu, Y., and Shi, C. Graph contrastive learn- ing with stable and scalable spectral encoding. Neur IPS, 36:45516 45532, 2023a. Bo, D., Shi, C., Wang, L., and Liao, R. Specformer: Spectral graph neural networks meet transformers. ar Xiv preprint ar Xiv:2303.01028, 2023b. Chen, H.-Y. and Chao, W.-L. On bridging generic and personalized federated learning for image classification. In ICLR, 2022. Chen, M., Zhang, W., Yuan, Z., Jia, Y., and Chen, H. Fede: Embedding knowledge graphs in federated setting. In IJCKG, pp. 80 88, 2021. Chen, Y., Huang, W., and Ye, M. Fair federated learning under domain skew with local consistency and domain diversity. In CVPR, pp. 12077 12086, 2024. Craven, M., Di Pasquo, D., Freitag, D., Mc Callum, A., Mitchell, T., Nigam, K., and Slattery, S. Learning to extract symbolic knowledge from the world wide web. AAAI/IAAI, pp. 2, 1998. Defferrard, M., Bresson, X., and Vandergheynst, P. Convolutional neural networks on graphs with fast localized spectral filtering. In Neur IPS, 2016. Ezzeldin, Y. H., Yan, S., He, C., Ferrara, E., and Avestimehr, A. S. Fairfed: Enabling group fairness in federated learning. In AAAI, 2023. Fan, W., Ma, Y., Li, Q., Wang, J., Cai, G., Tang, J., and Yin, D. A graph neural network framework for social recommendations. TKDE, 2020. Fang, X. and Ye, M. Robust federated learning with noisy and heterogeneous clients. In CVPR, pp. 10072 10081, 2022. Fang, X., Easwaran, A., Genest, B., and Suganthan, P. N. Adaptive hierarchical graph cut for multi-granularity outof-distribution detection. IEEE TAI, 2025. Fu, X., Zhang, B., Dong, Y., Chen, C., and Li, J. Federated graph machine learning: A survey of concepts, techniques, and applications. ar Xiv preprint ar Xiv:2207.11812, 2022. Gao, Y., Wang, X., He, X., Liu, Z., Feng, H., and Zhang, Y. Addressing heterophily in graph anomaly detection: A perspective of graph spectrum. In Proceedings of the ACM Web Conference 2023, pp. 1528 1538, 2023. Giles, C. L., Bollacker, K. D., and Lawrence, S. Citeseer: An automatic citation indexing system. In Proceedings of the third ACM conference on Digital libraries, pp. 89 98, 1998. S2FGL: Spatial Spectral Federated Graph Learning Han, H., Liu, X., Ma, L., Torkamani, M., Liu, H., Tang, J., and Yamada, M. Structural fairness-aware active learning for graph neural networks. In ICLR, 2023. He, C., Balasubramanian, K., Ceyani, E., Yang, C., Xie, H., Sun, L., He, L., Yang, L., Yu, P. S., Rong, Y., et al. Fedgraphnn: A federated learning system and benchmark for graph neural networks. In ICLR, 2021a. He, M., Wei, Z., Xu, H., et al. Bernnet: Learning arbitrary graph spectral filters via bernstein approximation. In Neur IPS, pp. 14239 14251, 2021b. He, M., Wei, Z., and Wen, J.-R. Convolutional neural networks on graphs with chebyshev approximation, revisited. In Neur IPS, pp. 7264 7276, 2022. Hong, J., Wang, H., Wang, Z., and Zhou, J. Federated robustness propagation: sharing adversarial robustness in heterogeneous federated learning. In AAAI, pp. 7893 7901, 2023. Hu, M., Yue, Z., Xie, X., Chen, C., Huang, Y., Wei, X., Lian, X., Liu, Y., and Chen, M. Is aggregation the only choice? federated learning via layer-wise model recombination. In SIGKDD, pp. 1096 1107, 2024. Huang, W., Ye, M., and Du, B. Learn from others and be yourself in heterogeneous federated learning. In CVPR, pp. 10143 10153, 2022. Huang, W., Wan, G., Ye, M., and Du, B. Federated graph semantic and structural learning. In Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence, pp. 3830 3838, 2023a. Huang, W., Ye, M., Shi, Z., and Du, B. Generalizable heterogeneous federated cross-correlation and instance similarity learning. TPAMI, pp. 712 728, 2023b. Huang, W., Ye, M., Shi, Z., Li, H., and Du, B. Rethinking federated learning with domain shift: A prototype view. In CVPR, pp. 16312 16322. IEEE, 2023c. Huang, W., Ye, M., Shi, Z., Wan, G., Li, H., Du, B., and Yang, Q. A federated learning for generalization, robustness, fairness: A survey and benchmark. TPAMI, 2024. Karimireddy, S. P., Kale, S., Mohri, M., Reddi, S. J., Stich, S. U., and Suresh, A. T. Scaffold: Stochastic controlled averaging for on-device federated learning. In ICML, pp. 5132 5143, 2020. Kreuzer, D., Beaini, D., Hamilton, W., L etourneau, V., and Tossou, P. Rethinking graph transformers with spectral attention. In Neru IPS, pp. 21618 21629, 2021. Lee, G., Jeong, M., Shin, Y., Bae, S., and Yun, S.-Y. Preservation of the global knowledge by not-true distillation in federated learning. In Koyejo, S., Mohamed, S., Agarwal, A., Belgrave, D., Cho, K., and Oh, A. (eds.), Neur IPS, pp. 38461 38474, 2022. Li, T., Sahu, A. K., Zaheer, M., Sanjabi, M., Talwalkar, A., and Smith, V. Federated optimization in heterogeneous networks. MLSys, 2:429 450, 2020. Li, X., Wu, Z., Zhang, W., Zhu, Y., Li, R.-H., and Wang, G. Fedgta: Topology-aware averaging for federated graph learning. ar Xiv preprint ar Xiv:2401.11755, 2024. Li, X.-C., Zhan, D.-C., Shao, Y., Li, B., and Song, S. Fedphp: Federated personalization with inherited private models. In ECML, pp. 587 602, 2021. Liao, R., Zhao, Z., Urtasun, R., and Zemel, R. S. Lanczosnet: Multi-scale deep graph convolutional networks. In ICLR, 2019. Liu, N., Wang, X., Bo, D., Shi, C., and Pei, J. Revisiting graph contrastive learning from the perspective of graph spectrum. In Neur IPS, pp. 2972 2983, 2022. Liu, R. and Yu, H. Federated graph neural networks: Overview, techniques and challenges. ar Xiv preprint ar Xiv:2202.07256, 2022. Liu, Y., Bo, D., and Shi, C. Graph condensation via eigenbasis matching. ar Xiv preprint ar Xiv:2310.09202, 2023. Liu, Z., Wan, G., Prakash, B. A., Lau, M. S., and Jin, W. A review of graph neural networks in epidemic modeling. ar Xiv preprint ar Xiv:2403.19852, 2024. Luan, S., Hua, C., Lu, Q., Zhu, J., Zhao, M., Zhang, S., Chang, X.-W., and Precup, D. Revisiting heterophily for graph neural networks. In Neur IPS22, pp. 1362 1375, 2022. Lv, F., Shang, X., Zhou, Y., Zhang, Y., Li, M., and Lu, Y. Personalized federated learning on heterogeneous and long-tailed data via expert collaborative learning. ar Xiv preprint ar Xiv:2408.02019, 2024. Mc Callum, A. K., Nigam, K., Rennie, J., and Seymore, K. Automating the construction of internet portals with machine learning. Information Retrieval, 3(2):127 163, 2000. Mc Mahan, B., Moore, E., Ramage, D., Hampson, S., and y Arcas, B. A. Communication-efficient learning of deep networks from decentralized data. In AISTATS, pp. 1273 1282, 2017. S2FGL: Spatial Spectral Federated Graph Learning Ray Chaudhury, B., Li, L., Kang, M., Li, B., and Mehta, R. Fairness in federated learning via core-stability. In Neur IPS, pp. 5738 5750, 2022. Sen, P., Namata, G., Bilgic, M., Getoor, L., Galligher, B., and Eliassi-Rad, T. Collective classification in network data. AI magazine, 29(3):93 93, 2008. Shang, X., Lu, Y., Huang, G., and Wang, H. Federated learning on heterogeneous and long-tailed data via classifier re-training with federated features. In IJCAI, 2022. Smith, V., Chiang, C.-K., Sanjabi, M., and Talwalkar, A. S. Federated multi-task learning. In Neur IPS, 2017. Tan, Y., Liu, Y., Long, G., Jiang, J., Lu, Q., and Zhang, C. Federated learning on non-iid graphs via structural knowledge sharing. In AAAI, pp. 9953 9961, 2023. Tan, Z., Wan, G., Huang, W., and Ye, M. Fedssp: Federated graph learning with spectral knowledge and personalized preference. In Neur IPS, pp. 34561 34581, 2024. Tan, Z., Wan, G., Huang, W., Li, H., Zhang, G., Yang, C., and Ye, M. Fedspa: Generalizable federated graph learning under homophily heterogeneity. In CVPR, pp. 15464 15475, 2025. Tang, J., Li, J., Gao, Z., and Li, J. Rethinking graph neural networks for anomaly detection. In ICML, pp. 21076 21089, 2022. Wan, G., Huang, W., and Ye, M. Federated graph learning under domain shift with generalizable prototypes. In AAAI, pp. 15429 15437, 2024a. Wan, G., Tian, Y., Huang, W., Chawla, N. V., and Ye, M. S3gcl: Spectral, swift, spatial graph contrastive learning. In ICML, pp. 49973 49990, 2024b. Wan, G., Huang, Z., Zhao, W., Luo, X., Sun, Y., and Wang, W. Rethink graphode generalization within coupled dynamical system. In ICML, 2025a. Wan, G., Shi, Z., Huang, W., Zhang, G., Tao, D., and Ye, M. Energy-based backdoor defense against federated graph learning. In ICLR, 2025b. Wang, D., Lin, J., Cui, P., Jia, Q., Wang, Z., Fang, Y., Yu, Q., Zhou, J., Yang, S., and Qi, Y. A semi-supervised graph attentive network for financial fraud detection. In ICDM, pp. 598 607, 2019. Wang, J., Liu, Q., Liang, H., Joshi, G., and Poor, H. V. Tackling the objective inconsistency problem in heterogeneous federated optimization. In Neur IPS, pp. 7611 7623, 2020. Wang, X. and Zhang, M. How powerful are spectral graph neural networks. In ICML, pp. 23341 23362, 2023. Wu, X., Liu, X., Niu, J., Zhu, G., and Tang, S. Bold but cautious: Unlocking the potential of personalized federated learning through cautiously aggressive collaboration. In ICCV, pp. 19375 19384, 2023. Wu, Z., Pan, S., Chen, F., Long, G., Zhang, C., and Philip, S. Y. A comprehensive survey on graph neural networks. TNNLS, pp. 4 24, 2020. Xie, H., Ma, J., Xiong, L., and Yang, C. Federated graph classification over non-iid graphs. In Neru IPS, pp. 18839 18852, 2021. Xu, C., Qu, Y., Xiang, Y., and Gao, L. Asynchronous federated learning on heterogeneous devices: A survey. Computer Science Review, 50:100595, 2023. Xu, J., Chen, Z., Quek, T. Q., and Chong, K. F. E. Fedcorr: Multi-stage federated learning for label noise correction. In CVPR, pp. 10184 10193, 2022. Yang, X., Huang, W., and Ye, M. Dynamic personalized federated learning with adaptive differential privacy. Advances in Neural Information Processing Systems, 36: 72181 72192, 2023. Zhang, H., Shen, T., Wu, F., Yin, M., Yang, H., and Wu, C. Federated graph learning a position paper. In ar Xiv preprint ar Xiv:2105.11099, 2021a. Zhang, J., Li, Z., Li, B., Xu, J., Wu, S., Ding, S., and Wu, C. Federated learning with label distribution skew via logits calibration. In ICML, pp. 26311 26329, 2022a. Zhang, J., Hua, Y., Cao, J., Wang, H., Song, T., XUE, Z., Ma, R., and Guan, H. Eliminating domain bias for federated learning in representation space. In Neur IPS, pp. 14204 14227, 2023a. Zhang, J., Hua, Y., Wang, H., Song, T., Xue, Z., Ma, R., Cao, J., and Guan, H. Gpfl: Simultaneously learning global and personalized feature information for personalized federated learning. In CVPR, pp. 5041 5051, 2023b. Zhang, J., Hua, Y., Wang, H., Song, T., Xue, Z., Ma, R., and Guan, H. Fedala: Adaptive local aggregation for personalized federated learning. In AAAI, pp. 11237 11244, 2023c. Zhang, K., Yang, C., Li, X., Sun, L., and Yiu, S. M. Subgraph federated learning with missing neighbor generation. In Neur IPS, pp. 6671 6682, 2021b. Zhang, T., Gao, L., Lee, S., Zhang, M., and Avestimehr, S. Timelyfl: Heterogeneity-aware asynchronous federated learning with adaptive partial training. In CVPR, pp. 5064 5073, 2023d. S2FGL: Spatial Spectral Federated Graph Learning Zhang, Y., Gao, S., Pei, J., and Huang, H. Improving social network embedding via new second-order continuous graph neural networks. In KDD, pp. 2515 2523, 2022b. Zhou, T. and Konukoglu, E. Fedfa: Federated feature augmentation. In ICLR, 2023. Zhu, B., Wang, L., Pang, Q., Wang, S., Jiao, J., Song, D., and Jordan, M. I. Byzantine-robust federated learning with optimal statistical rates. In AISTATS, pp. 3151 3178, 2023. Zhu, Y., Li, X., Wu, Z., Wu, D., Hu, M., and Li, R.-H. Fedtad: Topology-aware data-free knowledge distillation for subgraph federated learning. ar Xiv preprint ar Xiv:2404.14061, 2024. Zhu, Z., Hong, J., and Zhou, J. Data-free knowledge distillation for heterogeneous federated learning. In ICML, pp. 12878 12889, 2021.