# subgraph_federated_learning_for_local_generalization__9b73578e.pdf Published as a conference paper at ICLR 2025 SUBGRAPH FEDERATED LEARNING FOR LOCAL GENERALIZATION Sungwon Kim1, , Yoonho Lee1, , Yunhak Oh1, , Namkyeong Lee1, , Sukwon Yun2, , Junseok Lee1, , Sein Kim1, , Carl Yang3, & Chanyoung Park1, 1KAIST, 2UNC Chapel Hill, 3Emory University Federated Learning (FL) on graphs enables collaborative model training to enhance performance without compromising the privacy of each client. However, existing methods often overlook the mutable nature of graph data, which frequently introduces new nodes and leads to shifts in label distribution. Since they focus solely on performing well on each client s local data, they are prone to overfitting to their local distributions (i.e., local overfitting), which hinders their ability to generalize to unseen data with diverse label distributions. In contrast, our proposed method, Fed Lo G, effectively tackles this issue by mitigating local overfitting. Our model generates global synthetic data by condensing the reliable information from each class representation and its structural information across clients. Using these synthetic data as a training set, we alleviate the local overfitting problem by adaptively generalizing the absent knowledge within each local dataset. This enhances the generalization capabilities of local models, enabling them to handle unseen data effectively. Our model outperforms baselines in our proposed experimental settings, which are designed to measure generalization power to unseen data in practical scenarios. Our code is available at https://github.com/sung-won-kim/Fed Lo G 1 INTRODUCTION In the realm of Graph Neural Networks (GNNs) (Hamilton, 2020), most systems are designed for a unified, centralized graph. However, real-world applications (Zhang et al., 2021a) frequently involve individual users or institutions maintaining private graphs, isolated due to privacy concerns. Graph Federated Learning (GFL) (Liu et al., 2024) provides a solution by enabling clients to independently train local GNNs on their data. This decentralized training approach allows a central server to aggregate the locally updated weights from multiple clients, creating a unified model that respects privacy constraints. In this paper, among the various settings in GFL, we focus on one of the most challenging aspects distributed subgraphs (subgraph-FL), where clients manage largely disjoint sets of nodes and their edges. In real-world scenarios, graph data frequently changes, particularly in social, citation, and e-commerce networks (Sen et al., 2008; Mc Auley et al., 2015; Shchur et al., 2018). While these changes often result in new label distribution patterns that are distinct from the existing local label distribution within each client, existing subgraph-FL methods (Zhang et al., 2021a; Wu et al., 2021a; Yao et al., 2024; Baek et al., 2023) primarily focus on optimizing models based on the current label distribution within each client (i.e., local optimization). On the other hand, some studies (Zhang et al., 2022; Lee et al., 2022; Guo et al., 2024) demonstrate that client models are particularly prone to local overfitting after local updates, resulting in a significant decrease in the accuracy of minority classes (i.e., tail classes) within the local data. Given these limitations, current approaches face substantial practical challenges, particularly in adapting to new nodes added to the original local graph. This is especially difficult for nodes belonging to tail classes or unseen classes that are missing from the local graph but exist in other graphs (i.e., missing classes). These nodes, which form new connections with existing nodes, often have structural patterns unfamiliar to local clients, leading to substantial discrepancies in both label and structural distributions of the local graphs. Existing methods in FL (Li & Zhan, 2021; Zhang et al., 2022; Lee et al., 2022; Guo et al., 2024) aim to ensure that local models can make predictions for all classes without bias by mitigating local Corresponding author (cy.park@kaist.ac.kr) Published as a conference paper at ICLR 2025 overfitting caused by the local label distribution. Specifically, they propose regularizing the logits of each class in the local models to align more closely with those of the global model. While these methods effectively address local overfitting and manage tail or missing classes, increasing the logits of local tail data risks amplifying noisy data, which is harmful for the class representation of the global model. Beyond FL, another common approach to mitigating the problem of overfitting on training data involves addressing class imbalance. Techniques such as down-sampling (He & Ma, 2013), over-sampling (Chawla et al., 2002; Zhao et al., 2021; He & Bai, 2008), calibration (Niculescu-Mizil & Caruana, 2005; Zadrozny & Elkan, 2001), or constructing expert models for tail data (Menon et al., 2020; Yun et al., 2022) are commonly used. Despite their effectiveness, they require at least one data point to be present for each class, facing challenges when a class is missing in a local client while present in others. In this paper, we propose to address the local overfitting issue of subgraph-FL by introducing reliable global synthetic data that mitigate class imbalance while addressing missing classes. Specifically, we aggregate knowledge from local data across all clients for each class, and integrate it into the global synthetic data. Subsequently, each client adaptively utilizes the global synthetic data as additional training data to ensure effective learning for all classes, including those that are underrepresented or missing in each client. This strategy helps prevent local overfitting even after local updates and enables accurate class representation (i.e., local generalization). However, there exist two crucial challenges that need consideration: C1. Which data across all clients should be aggregated to ensure reliability? Since clients in FL heavily rely on knowledge from other clients to learn locally absent information, it is crucial for each client to share the most reliable knowledge within its own local graph. Here, data reliability refers to the accuracy and consistency of information sourced from decentralized nodes. C2. How can data from other clients be utilized without compromising privacy? While direct sharing of the data between clients prevents local overfitting, it raises severe privacy concerns. Furthermore, directly using training data from all clients incurs high communication costs. Solution to C1. Knowledge from head degree and head class nodes. Our findings indicate that nodes with a high number of connections (i.e., head degree) and those belonging to the majority class (i.e., head class) provide reliable structural and class-representative information, respectively, which significantly enhances the model s ability to generalize to unseen data. Building on these insights, we aggregate knowledge from clients and filter it based on the headness of both their degree and class. Solution to C2. Data condensation. We propose to condense only reliable knowledge into synthetic data to share across the clients. This avoids the direct use of individual client data, mitigating the privacy concerns, while also minimizing the amount of data transferred between the server and local clients, thereby lowering communication costs. In summary, we propose a subgraph Federated Learning framework for Local Generalization, Fed Lo G, that generates global synthetic data with a novel reliable knowledge condensation strategy. This approach reduces the risk of noise in class representations, and enables each client to compensate for locally absent knowledge without compromising privacy. By doing so, Fed Lo G prevents local overfitting and ensures a well-generalized representation of all classes, enabling successful handling of unseen data in our three proposed evaluation settings; 1) Unseen Node: New nodes with seen classes are added, and they introduce structural changes in local graphs. 2) Missing Class: New nodes with classes previously absent in the client s graph are added. 3) New Client: A completely new graph with distinct label distribution and structure is added. In this paper, we make the following contributions: We introduce Fed Lo G, the first work in subgraph-FL that focuses on preventing local overfitting, including the issue of missing classes, to address the mutable nature of the graph domain. This approach enhances the performance of the global model and improves the generalization power of local models, allowing them to effectively handle unseen data. We analyze what constitutes reliable data in graph-based federated learning and propose a method to condense and share this knowledge across clients. This approach not only leverages reliable data effectively but also protects privacy by using condensed synthetic data. We propose practical and important evaluation settings on unseen data for subgraph-FL (i.e., Unseen Node, Missing Class, and New Client), enabling measurement of the model s generalization on future data, assessing robustness in mutable graph domains, and demonstrating consistent outperformance over other baselines. Published as a conference paper at ICLR 2025 2 RELATED WORK 2.1 SUBGRAPH FEDERATED LEARNING Recent works (He et al., 2021; Liu et al., 2024) have introduced FL frameworks that enable collaborative GNN training without sharing graph data. Subgraph-FL aims to leverage disjoint graphs from each local client to collaboratively train a global model for solving downstream tasks. Existing studies (Zhang et al., 2021a; Yao et al., 2024; Wu et al., 2021a; Liu et al., 2024) have attempted to supplement the local absent knowledge among local graphs that each client currently holds. For instance, Fed SAGE+ (Zhang et al., 2021a), Fed GNN (Wu et al., 2021a), and Fed GCN (Yao et al., 2024) request node information from other clients to recover missing neighborhood nodes and compensate for potential edges. Fed PUB (Baek et al., 2023) and Fed Star (Tan et al., 2023) aim to personalize local models by adapting the global model to specialize in the local data of each client. However, due to the mutable properties of graph domains, subgraph-FL must generalize well not only to the current label distribution but also to new nodes that will emerge in the future. Unlike these approaches (Zhang et al., 2021a; Wu et al., 2021a; Yao et al., 2024; Baek et al., 2023) that only focus on finding missing knowledge relevant to the current state, our model learns representations for all classes and their connection patterns, ensuring better generalization across various future scenarios. 2.2 LOCAL OVERFITTING IN FEDERATED LEARNING Imbalanced data distribution is common in real-world scenarios, and significant efforts (Cui et al., 2019; Menon et al., 2020; Tan et al., 2020; Yun et al., 2022; Li et al., 2022; Ma et al., 2023) have been made to address the resulting deterioration in model performance. Federated learning, the task at hand, inherently faces the data imbalance problem as well. Specifically, the involvement of multiple clients means that each client has its own imbalanced dataset, making local models prone to overfitting to their local data (Zhang et al., 2022; Lee et al., 2022; Guo et al., 2024). Recent works (Chen et al., 2023; Lee et al., 2022; Zhang et al., 2022; Guo et al., 2024) aim to alleviate local overfitting in FL by regularizing local models to be similar to the global model. Fed HKD (Chen et al., 2023), Fed LC (Zhang et al., 2022), and Fed ED (Guo et al., 2024) introduce logit calibration, which aligns the logits of each class in the local models more closely with those of the global model. While Fed ED (Guo et al., 2024) addresses the missing class problem in FL, it does not consider the noisy properties of tail data (Subramonian et al., 2024; Liu et al., 2021; Wu et al., 2021b; Xiao et al., 2021). Our method, Fed Lo G, addresses local overfitting and ensures reliable representation of all classes by leveraging class-specific knowledge across clients and considering their structural properties. To the best of our knowledge, this is the first work to tackle local overfitting with missing classes in subgraph-FL. 3 PRELIMINARIES Notations. We use G = (V, E) to denote a graph with the set of nodes V and the set of edges E V V. The dataset D = (G, Y ) includes labels Y for the nodes that belong to one of |CV| distinct classes, and XV R|V| d is the feature matrix with d as the feature dimension, where each node v V is associated with a feature vector xv Rd. In subgraph-FL, a server S and K clients manage disjoint subgraphs Gk = (Vk, Ek) for each client k. The global set of nodes is V = SK k=1 Vk with Vi Vj = for all i = j. The local dataset for client k is Dk = (Gk, Yk), and the combined local dataset is Dlocal = SK k=1 Dk. Additionally, we generate a global synthetic set Dglobal = (Gglobal, Y global), where Gglobal = (Vglobal, E ) consists of isolated nodes vg Vglobal with no edges E . Vglobal includes s nodes per class, totaling s |CV| nodes. Problem Statement. We aim to develop a distributed learning framework for collaborative training of a node classifier. Specifically, the classifier F uses optimized parameters ϕ to minimize a predefined task loss. The objective is to find global parameters ϕ that minimizes the aggregated local empirical risk R, defined as: ϕ = arg minϕ R(F(ϕ)) = 1 K PK k=1 Rk(Fk(ϕ)), where Rk(Fk(ϕ)) := E(Gk,Yk) Dlocal [Lk(Fk(ϕ; Gk), Yk)] and the task-specific loss Lk is defined as: Lk := 1 |Vk| P vk Vk l(ϕ; Gk(vk), yvk) + 1 |Vglobal| P vg Vglobal l(ϕ; vg, yvg). To allow each client to generalize across all classes, including missing classes, we generate global synthetic data Dglobal and introduce an additional loss term to take into account this data to prevent local overfitting. 3.1 WHICH DATA ARE RELIABLE? Data reliability refers to the accuracy and consistency of information from decentralized nodes, crucial for training models across varied environments (i.e., clients). Inspired by the robust performance of Published as a conference paper at ICLR 2025 GNNs on head class and head degree nodes (Yun et al., 2022; Park et al., 2021; Zhao et al., 2021; Liu et al., 2021), we found that data reliability largely depends on 1) the extent of data connections (i.e., degree headness) and 2) the predominance of certain classes (i.e., class headness). To corroborate our arguments, we measured the target class accuracy of a client (receiver) receiving information (i.e., weights) from other clients (contributors). To check how degree/class headness of the contributors impacts the receiver, we varied the contributors training sets, adjusting the degree/class headness of the training nodes, while keeping the receiver s training set constant. The global model was constructed by averaging client weights and evaluated on the receiver s local graph. Receiver s Accuracy Receiver s Accuracy Figure 1: Data Reliability Analysis1(Pub Med used). Figure 1(top) shows receiver accuracy with contributor training sets composed of 1) Only head degree nodes , 2) Only tail degree nodes , and 3) Balanced (head+tail) degree nodes . The receiver s performance is better when knowledge is received from head degree nodes, indicating the reliability of knowledge from head degree nodes over that from tail degree nodes. Moreover, as the number of clients increases, the performance gap widens, highlighting the accumulated negative impact of noise within tail degree nodes. Figure 1(bottom) manipulates label distribution within each contributor s local graph, transforming the target class into a tail or head class by varying the number of training nodes in other classes while keeping the target class constant. When the class headness of the target class within contributors is high (i.e., negative imbalance rates), contributors enhance the receiver s performance. Conversely, with low headness (i.e., positive imbalance rates), the receiver s performance deteriorates as contributors struggle to represent the target class (Yun et al., 2022). This negative impact is magnified as the number of clients increases. In summary, data with headness in both degree and class from other clients (i.e., contributors) helps the target client (i.e., receiver) learn reliable representations, while tailness data negatively impacts the model training due to insufficient or noisy information. Building on our observations, our method, Fed Lo G, collects knowledge from head degree and head class data across all clients to alleviate locally absent knowledge. 4 PROPOSED METHODOLOGY: FEDLOG Our proposed subgraph-FL framework, Fed Lo G, works as follows. Figure 2 shows the overall framework of Fed Lo G. Step 1 Local Fitting (Section 4.1): The server initializes the local model parameters of K clients with the parameters of the global model ϕglobal. Each local model is then trained using local data Dk. Concurrently, head degree and tail degree knowledge are condensed into synthetic nodes within each client, denoted as Vk,head and Vk,tail. Step 2 Global Aggregation and Global Synthetic Data Generation (Section 4.2): After local training, the server aggregates the local models to create the global model ϕglobal and generates global synthetic data Dglobal by aggregating Vk,head for all k, weighted according to the head classes within each client k. Step 3 Local Fitting (Section 4.1) & Local Generalization (Section 4.3): At the start of each round, local fitting is performed first. After local fitting, local models are generalized using Dglobal which possess both head degree and head class knowledge, adaptively learning the locally absent knowledge. While the framework starts with Step 1 (r = 0), it continues to alternate between Steps 2 and 3 until the final round R is reached (r 1). In summary, our method extracts head degree knowledge at the client level and head class knowledge at the server level, then condense them into the global synthetic data, which is utilized to train the local model during Local Generalization to adaptively compensate for the locally absent knowledge within each client. Algorithm 1 outlines the algorithm of Fed Lo G. 1Please refer to Appendix A.8 for detailed description of experimental settings. Published as a conference paper at ICLR 2025 Head-degree Branch 𝒉!! Tail-degree Branch 𝒑 (b2) Local Generalization (b3) Local model 𝜙! 0.8 0.2 0.0 0.1 0.1 0.8 𝜙"#$%&# (a1) Global Aggregation (a2) Global Synthetic Data Gen. (𝑟'-: cls rate) (𝑟,-: cls rate) Local 𝜙' Local 𝜙, ℙ.,)*&+ 𝒉!!,#$%& : Prompt node PG: Prompt Generator Global PGc 𝒢. Prompt Generation (b1) Local Fitting Init 𝜙. with Local dataset 𝒟. = (𝒢., 𝑌.) (b) Client 𝒌 Figure 2: Overview of Fed Lo G with 2 Clients and 3 Classes. 4.1 LOCAL FITTING The local model for each client k consists of one GNN embedder (φk) and two classifiers (θk,H and θk,T ), for the head and tail degree branches, respectively, as shown in Figure 2(b3). Each branch forms a prototypical network (Snell et al., 2017)-based architecture for prediction. While a prototypical network is originally designed for few-shot learning, we repurpose it by designating learnable nodes as prototypes, enabling both prediction and the condensation of class information into these prototypes. Specifically, each branch contains |CV| s learnable nodes, each with features of dimension d, allocating s nodes per class. Thus, each client k has learnable node sets Vk,head and Vk,tail with features XVk,head R(|CV| s) d and XVk,tail R(|CV| s) d, respectively. At the client level, we condense knowledge from locally observed nodes into these learnable nodes. Head degree nodes are condensed into Vk,head and tail degree nodes into Vk,tail, integrating condensation and prediction into a single process. The detailed processes are described as follows (See Figure 2(b1) and (b3)): (Initialization) In the initial round (r = 0), we initialize the local model weights for each client k, denoted as ϕk = {φk, θk,H, θk,T }, with the global set of parameters ϕglobal = {φglobal, θglobal H , θglobal T }. (Embedding) Given client k s local graph Gk = (Vk, Ek) where the feature of each node vk Vk is initialized with h(0) vk = xvk, a shared Graph SAGE (Hamilton et al., 2017) GNN encoder φk is employed to embed each local node vk Vk and learnable nodes vk,head Vk,head and vk,tail Vk,tail: hvk = GNNφk(vk, Gk), hvk,head = GNNφk(vk,head, Gk,head), hvk,tail = GNNφk(vk,tail, Gk,tail) (1) where hvk, hvk,head, and hvk,tail are the representations of nodes vk, vk,head, and vk,tail, respectively. Gk,head = (Vk,head, E ) and Gk,tail = (Vk,tail, E ) denote synthetic graphs constructed by learnable nodes from head and tail degree branches, respectively. It is worth noting that the learnable nodes do not adhere to a specific graph structure but share the same GNN encoder with the local graph, allowing us to condense structural information into the features of the learnable nodes. After acquiring node representations, we generate model predictions in each branch using class prototypes, which are the representations of learnable nodes. Since the process of generating model predictions is identical in both branches, we only explain the head branch here. In the head branch, prototypes are defined as Pk,head = {hv(1,1) k,head, . . . , hv(1,s) k,head, . . . , hv (|CV |,1) k,head , . . . , hv (|CV |,s) k,head }, with s proto- types per class. To ensure all class information contributes to the final prediction, the target node representations are further updated based on feature differences with all prototypes assigned to each class as follows: h vk = θk,H(hvk, {(hvk hv(1,1) k,head), . . . , (hvk hv (|CV |,s) k,head )}). Please refer to Appendix A.4 for more details on θk,H. Then, the class probability for target node vk is given as follows: p(c|h vk) = exp( d(h vk, h Vc k,head)) P|CV | c =1 exp( d(h vi, h Vc (k,head))) , (2) where d( , ) is the squared Euclidean distance and h Vc k,head indicates the average of prototype representations of class c, i.e., h Vc k,head = 1 s Ps i=1 hv(c,i) k,head. By updating the target node s representation based on its relationships with all prototypes and minimizing the distance to its correct class prototype, we aim to guide the synthetic learnable node to effectively learn a representation that reflects both the correct class and the broader relationships among all classes. To obtain final prediction p , we combine the class probabilities from both branches phead and ptail by weighting them based on the degree value of the target node vk (i.e., deg(vk)) as follows: p = α phead + (1 α) ptail, (3) Published as a conference paper at ICLR 2025 where α = 1/(1+e (deg(vk) (λ+1))), and λ is the tail degree threshold outlined in the Appendix A.9. Note that the hyperparameter α balances the influence between the head and tail branches based on the node s degree. Specifically, α acts as a weight for the prediction loss contributed by each branch for the target node. For instance, when a target node has a high degree, α becomes large, increasing the influence of the head degree branch on the final predictions. Consequently, knowledge from high-degree nodes is condensed into Vk,head, as the model significantly updates the learnable features XVk,head of Vk,head to minimize prediction loss. Conversely, when a node has a low degree, its knowledge is condensed into XVk,tail of Vk,tail instead. Additionally, the sigmoid-based formulation of α prevents nodes with extremely high degrees from dominating the learnable features, ensuring a balanced contribution across all nodes. Training objective for each client k. The prediction loss for each client k is calculated as Lk,cls = P c CV I(yvk = c) log(p[c]). Furthermore, to ensure the stability of the condensation process, we minimize the L2 norm of the learnable features, denoted as Lk,norm = P v Vk,head Vk,tail xv 2. Thus, the total loss for model parameters is Lk(ϕk, XVk,head, XVk,tail) = Lk,cls + β Lk,norm, where β adjusts the extent of regularization. It is important to note that besides updating the local model weights ϕk = {φk, θk,H, θk,T } for each client k, the learnable features XVk,head and XVk,tail are also updated. 4.2 GLOBAL AGGREGATION AND GLOBAL SYNTHETIC DATA GENERATION Global Aggregation. As shown in Figure 2(a1), after training the K local clients, the server aggregates the local model weights for round r using the weighted average ϕglobal 1 K PK k=1 ϕ(r) k . Global Synthetic Data Generation. In addition, as shown in Figure 2(a2), the server generates global synthetic data Dglobal, which will be employed during the Local Generalization phase (Section 4.3) to help mitigate the issue of local overfitting. More specifically, we first generate node features in the global synthetic data Dglobal by merging Vk,head from all K clients. For example, when merging Vk,head for all K clients regarding a class c, the server gives more weight to the synthetic data from client k for whom class c belongs to the head classes within the local data Dk. This is to take into account the expert knowledge of each client regarding the dominant classes within its local data, which is supported by our empirical analysis in Section 3.1. More formally, for each class c CV, the feature vector of the i-th global synthetic node for class c, i.e., xv(c,i) g Rd, is generated as follows: xv(c,i) g = 1 PK k=1 rc k k=1 rc kxv(c,i) k,head, (4) where xv(c,i) k,head XVk,head and rc k = |Vc k| |Vk| represents the proportion of nodes belonging to class c in the k-th client s dataset. By giving more weight to clients with expertise in each head class, the server effectively combines the most reliable knowledge from all clients to create the global synthetic nodes. This process results in |CV| s global synthetic nodes Vglobal with features XVglobal = {xv(1,1) g , . . . , xv(1,s) g , . . . , xv (|CV |,1) g , . . . , xv (|CV |,s) g }. The final global synthetic data is represented as Dglobal = (Gglobal, Y global), where Gglobal = (Vglobal, E ), with node features XVglobal. At the end of each round r, the server distributes the aggregated model weights ϕglobal and the generated global synthetic data Dglobal. In summary, considering both the degree headness through the learnable features XVk,head and the class headness through rc k allows the global features XVglobal to contain knowledge about both head degree and head class across all clients. 4.2.1 DISCUSSIONS ON THE GRAPH STRUCTURES OF GLOBAL SYNTHETIC NODES Note that even though each global synthetic node contains only features without explicit graph structures, these features still implicitly capture the original graph structure. This is because these features are learned using a graph embedder φk that is shared across both synthetic and original nodes. As a result, the structural information of the original graph is condensed into the features of the synthetic nodes. 4.3 LOCAL GENERALIZATION At the beginning of each round (r 1), each client k initializes its local model ϕk with the distributed global model parameters ϕglobal followed by the local update of ϕk based on its local data Dk (i.e., Published as a conference paper at ICLR 2025 Local Fitting described in Section 4.1). Then, as shown in Figure 2(b2) and (b3), we additionally train the local model with the global synthetic data Dglobal, enabling the model to generalize to locally absent knowledge (i.e., Local Generalization), such as tail and missing classes. Since each client has different locally absent knowledge, we first adaptively customize the global synthetic data for the current state of the local model through two strategies, i.e., 1) feature scaling and 2) prompt generation, and then train the local model with the customized data. Strategy 1) Feature Scaling. When the local model is strongly biased toward the local distribution, it tends to assign high logits to the dominant class, making it difficult to predict tail or missing classes. To address this issue, we apply strong perturbation to the training data of the dominant class to help balance predictions across all classes, allowing the model to effectively learn absent knowledge within the local data. To achieve this, we use feature scaling for the perturbation on the global synthetic data XVglobal = {xv(1,1) g , . . . , xv(1,s) g , . . . , xv (|CV |,1) g , . . . , xv (|CV |,s) g } as follows: ˆxv(c,i) g = xv(c,i) g + γk[c] ( x Vglobal xv(c,i) g ), where x Vglobal = 1 Vglobal X xvg XVglobal xvg, (5) where γk R|CV| is the class-wise adaptive factor that adjusts the strength of the perturbation by moving the global synthetic data of class c to the average of global synthetic data, making it harder to predict. Note that when γk[c] = 1, the global synthetic data for class c is completely replaced with the average of the global synthetic data, whereas the global synthetic data for class c remains unchanged when γk[c] = 0. During training, we dynamically modify the factor by incrementing it by 0.001 whenever the local model s accuracy for class c exceeds the threshold at the end of the round, thereby increasing the perturbation of the corresponding class. Strategy 2) Prompt Generation. While we have focused on condensing knowledge into the global synthetic data, we have not yet addressed how to train our GNN encoder on this data. Recall that the GNN encoder φk is shared across both the synthetic and original nodes. However, since global synthetic nodes only contain features without an explicit graph structure while the original nodes involve a graph structure, the parameters of the shared GNN encoder would be differently affected by them, leading to discrepancies in the gradient matrices. Our main idea to solve this issue is to ensure that synthetic nodes are trained in the same environment as original nodes within the local graph. Specifically, the objective of the prompt generator is to make the synthetic graph which consists of a target node and its corresponding prompt node derived from the target node s features produce the same gradient matrix as when the target node is predicted using its true h-hop subgraph within the local graph. Details on pretraining the prompt generators are in Appendix A.7. More formally, we generate a prompt node for each feature-scaled global synthetic node using the class-specific prompt generator PGc corresponding to class c (Figure 2(b2)) as follows: xvp vg = PGc=yvg (ˆxvg), where xvp vg is the feature of the prompt node vp vg corresponding to the node vg. We then construct a synthetic graph Gsyn k = S vg Vglobal Gsyn k,vg for client k, where each Gsyn k,vg consists of a global synthetic node vg and its prompt node vp vg, connected to each other. Training with Customized Data. Using the synthetic graph Gsyn k , the prediction for the target synthetic nodes within Vglobal follows Eqs. 1-3 using the same local model ϕk as for normal nodes Vk, with α set to 0.5. The prediction loss for Dglobal is Lk,g = P vg Vglobal P c CV I(yvg = c) log(p[c]). At the end of each round, we adjust the adaptive factor γk[c] c based on the class prediction accuracy. In summary, the final loss of the local model is: Lk = Lk,cls + Lk,g + β Lk,norm. 5 EXPERIMENTS 5.1 EXPERIMENTAL SETTINGS Datasets. We conduct experiments on five real-world graph datasets. Distributed subgraphs are constructed by dividing each dataset into a certain number of clients using the METIS graph partitioning algorithm (Karypis & Kumar, 1997). The datasets used are Cora (Sen et al., 2008), Cite Seer (Sen et al., 2008), Pub Med (Sen et al., 2008), Amazon Computer (Mc Auley et al., 2015), and Amazon Photo (Shchur et al., 2018). For more details, see Appendix A.12. Baseline Methods. 1) Local: Refers to local training without any weight sharing. 2) Fed Avg (Mc Mahan et al., 2017): The most widely-used FL baseline. 3) Fed SAGE+ (Zhang et al., 2021a), 4) Fed GCN (Yao et al., 2024) and 5) Fed PUB (Baek et al., 2023): subgraph-FL baselines that primarily Published as a conference paper at ICLR 2025 Table 1: Model performance on Seen Graph settings. Mean accuracy with std. over 3 runs. Cora Cite Seer Pub Med Amazon Photo Amazon Computers Methods 3 Clients 5 Clients 10 Clients 3 Clients 5 Clients 10 Clients 3 Clients 5 Clients 10 Clients 3 Clients 5 Clients 10 Clients 3 Clients 5 Clients 10 Clients Local 0.7357 (0.0030) 0.7325 (0.0066) 0.8039 (0.0008) 0.6674 (0.0069) 0.6647 (0.0045) 0.7128 (0.0035) 0.8445 (0.0003) 0.8108 (0.0000) 0.8024 (0.0011) 0.6724 (0.0003) 0.7959 (0.0106) 0.7562 (0.0137) 0.6523 (0.0221) 0.5764 (0.0001) 0.6645 (0.0051) Fed Avg 0.8416 (0.0044) 0.6332 (0.0166) 0.7162 (0.0382) 0.7426 (0.0024) 0.7498 (0.0049) 0.7252 (0.0035) 0.7126 (0.0000) 0.8640 (0.0024) 0.8586 (0.0010) 0.7668 (0.0414) 0.5695 (0.0483) 0.5669 (0.0974) 0.5626 (0.0715) 0.4195 (0.0173) 0.4858 (0.0187) Fed SAGE+ 0.7560 (0.0237) 0.4156 (0.0034) 0.3522 (0.1196) 0.7505 (0.0150) 0.5167 (0.0389) 0.4929 (0.0075) 0.8980 (0.0001) 0.9091 (0.0025) 0.9041 (0.0012) 0.9239 (0.0083) 0.6670 (0.0206) 0.6246 (0.0585) 0.7539 (0.0062) 0.6934 (0.0006) 0.6656 (0.0082) Fed GCN 0.8226 (0.0062) 0.8124 (0.0158) 0.7243 (0.0172) 0.7376 (0.0111) 0.7649 (0.0010) 0.7123 (0.0122) 0.7127 (0.0000) 0.8504 (0.0011) 0.8441 (0.0070) 0.7398 (0.0036) 0.5717 (0.0583) 0.5627 (0.0957) 0.5782 (0.0623) 0.4217 (0.0243) 0.4908 (0.0183) Fed PUB 0.8476 (0.0021) 0.8448 (0.0009) 0.8622 (0.0059) 0.7455 (0.0065) 0.7694 (0.0074) 0.7505 (0.0081) 0.9064 (0.0016) 0.9069 (0.0019) 0.9092 (0.0019) 0.9399 (0.0020) 0.9122 (0.0016) 0.8983 (0.0052) 0.8339 (0.0142) 0.8202 (0.0141) 0.8181 (0.0124) Fed NTD 0.8452 (0.0067) 0.8526 (0.0024) 0.6984 (0.0030) 0.7455 (0.0069) 0.7826 (0.0047) 0.7146 (0.0079) 0.9049 (0.0002) 0.9065 (0.0009) 0.9061 (0.0012) 0.9378 (0.0029) 0.9166 (0.0021) 0.9119 (0.0036) 0.8492 (0.0107) 0.8619 (0.0034) 0.8707 (0.0055) Fed ED 0.8542 (0.0084) 0.8398 (0.0024) 0.6779 (0.0343) 0.7305 (0.0086) 0.7624 (0.0050) 0.6251 (0.0149) 0.9080 (0.0006) 0.9086 (0.0027) 0.8985 (0.0025) 0.9463 (0.0014) 0.9101 (0.0027) 0.8950 (0.0059) 0.8623 (0.0136) 0.8722 (0.0035) 0.8356 (0.0158) Fed Lo G 0.8601 (0.0118) 0.8575 (0.0074) 0.8451 (0.0103) 0.7663 (0.0086) 0.7728 (0.0049) 0.7624 (0.0063) 0.9180 (0.0005) 0.9129 (0.0015) 0.9115 (0.0043) 0.9653 (0.0020) 0.9496 (0.0037) 0.9305 (0.0049) 0.9073 (0.0012) 0.8986 (0.0014) 0.8742 address missing knowledge within the current local label distribution. To ensure a fair comparison, we also evaluate our method against 6) Fed NTD (Lee et al., 2022) and 7) Fed ED (Guo et al., 2024), which address local overfitting in FL. For more details, see Appendix A.13. Evaluation Protocol. We perform FL for 100 rounds. Node classification accuracy is measured on the client side and averaged across all clients over three runs. More details are in Appendix A.14. 5.2 EXPERIMENT RESULTS Q1. How does Fed Lo G perform in conventional FL settings? Table 1 presents the evaluation of models on graphs that were used for training. The label distributions of the test nodes match the training label distribution of each client. We refer to this conventional setting as Seen Graph, where models are evaluated on test nodes within the same graph structure as the training nodes (i.e., transductive setting (Kipf & Welling, 2016)). The overall performance of Fed Lo G on the Seen Graph outperforms that of other baselines, demonstrating its strong performance in conventional settings. Q2. Does Fed Lo G generalize to unseen data after local updates? In this section, we introduce practical and novel test settings for subgraph-FL, aiming to assess the model s ability to generalize to potential unseen data in real-world scenarios. By proposing these new settings, we emphasize the importance of evaluating models beyond conventional FL settings, and demonstrate that our method consistently achieves superior performance compared to other baselines. We introduce three practical scenarios for unseen data: 1) Unseen Node (Table 2(a)). Each client has new nodes with seen classes added to its local graph, which introduce structural changes. We perform evaluations on the new nodes to assess how well the FL framework adapts to these structural changes. 2) Missing Class (Table 2(b)). Each client has new nodes with missing classes added to its local graph. We evaluate the performance on new nodes representing missing classes for each client, assessing how effectively the FL framework enables the local model to learn previously absent knowledge. 3) New Client (Table 2(c)). A new client that has never participated in the FL framework emerges. This client has a distinct label distribution and graph structure. We assess how well the FL framework generalizes to accommodate this new client, ensuring robust performance across diverse scenarios. To do so, we perform evaluations on the unseen graph of the new client using each trained local model without the new client being involved in the training, and then report the mean accuracy over all clients. These approaches help simulate real-world scenarios where clients have incomplete information. Detailed settings are described in Appendix A.10 and A.16. In Table 2, we observe that Fed Lo G outperforms baselines by preventing local overfitting and effectively addressing unseen data. Specifically, in the Unseen Node and Missing Class settings, Fed Lo G shows superior performance on the added nodes, even with missing classes. For the Missing Class, which requires extensive knowledge from other clients, Local and personalized FL models like Fed PUB (Baek et al., 2023) fail to predict the missing classes as they optimize for the training label distribution. Although Fed SAGE+ (Zhang et al., 2021a) and Fed GCN (Yao et al., 2024) attempt to compensate for missing neighbors, they are not always effective because the missing class is not always within the neighbors. Moreover, Fed NTD (Lee et al., 2022) and Fed ED (Guo et al., 2024) address local overfitting and achieve relatively high performance in missing class prediction. However, they regularize the local model logits to match the global model, risking noisy information from tail data and resulting in inconsistent performance across different settings. In contrast, Fed Lo G alleviates local overfitting by using reliable class representations and structural information across clients, reducing the emphasis on noisy information. Thus, Fed Lo G successfully addresses unseen data, ensuring robust performance even with unseen graph structures (New Client) due to its generalization ability across all classes and structural features. Published as a conference paper at ICLR 2025 Table 2: Model performance in FL settings on unseen data. Mean accuracy with std. over 3 runs. (a) Unseen Node Cora Cite Seer Pub Med Amazon Photo Amazon Computers Methods 3 Clients 5 Clients 10 Clients 3 Clients 5 Clients 10 Clients 3 Clients 5 Clients 10 Clients 3 Clients 5 Clients 10 Clients 3 Clients 5 Clients 10 Clients Local 0.1250 (0.0030) 0.2957 (0.0079) 0.2854 (0.0263) 0.4443 (0.0131) 0.3471 (0.0020) 0.5177 (0.0052) 0.7510 (0.0010) 0.7292 (0.0000) 0.7489 (0.0013) 0.1333 (0.0000) 0.1900 (0.0392) 0.3958 (0.0211) 0.1687 (0.0000) 0.2488 (0.0000) 0.3890 (0.0043) Fed Avg 0.5403 (0.0797) 0.5198 (0.0179) 0.4139 (0.1308) 0.6585 (0.0220) 0.6098 (0.0301) 0.6199 (0.0084) 0.6154 (0.0000) 0.8189 (0.0120) 0.8070 (0.0043) 0.1782 (0.0419) 0.2125 (0.0136) 0.3727 (0.0221) 0.2275 (0.1017) 0.3095 (0.0179) 0.4177 (0.0335) Fed SAGE+ 0.5653 (0.0546) 0.4265 (0.0062) 0.3836 (0.0705) 0.6572 (0.0093) 0.4023 (0.0339) 0.6154 (0.0034) 0.8944 (0.0048) 0.8921 (0.0089) 0.8926 (0.0051) 0.4781 (0.0093) 0.2298 (0.0394) 0.4607 (0.0391) 0.3462 (0.0325) 0.3555 (0.0196) 0.3596 (0.0024) Fed GCN 0.3689 (0.0646) 0.5877 (0.0018) 0.5075 (0.0001) 0.6232 (0.0243) 0.6530 (0.1095) 0.6139 (0.0119) 0.6154 (0.0000) 0.7759 (0.0058) 0.8188 (0.0049) 0.2565 (0.0067) 0.2604 (0.0077) 0.3708 (0.0304) 0.2086 (0.0597) 0.3084 (0.0140) 0.4103 (0.0342) Fed PUB 0.5529 (0.0246) 0.5192 (0.0064) 0.4767 (0.0286) 0.6798 (0.0334) 0.6691 (0.0057) 0.6938 (0.0245) 0.8878 (0.0003) 0.8822 (0.0056) 0.8836 (0.0043) 0.4085 (0.0118) 0.3890 (0.0404) 0.5033 (0.0155) 0.4414 (0.0225) 0.5025 (0.0466) 0.5253 (0.0471) Fed NTD 0.6355 (0.0195) 0.5880 (0.0041) 0.3913 (0.1235) 0.7057 (0.0173) 0.7014 (0.0614) 0.6151 (0.0155) 0.8939 (0.0068) 0.8852 (0.0044) 0.8816 (0.0031) 0.4042 (0.0155) 0.5833 (0.0043) 0.5286 (0.0030) 0.5056 (0.0440) 0.6034 (0.0309) 0.6482 (0.0226) Fed ED 0.7338 (0.0294) 0.5514 (0.0117) 0.3916 (0.1184) 0.6646 (0.0658) 0.6148 (0.0097) 0.5381 (0.0781) 0.9008 (0.0027) 0.8884 (0.0036) 0.8730 (0.0077) 0.6227 (0.0429) 0.4265 (0.0675) 0.4629 (0.0137) 0.4582 (0.0176) 0.5408 (0.0223) 0.4940 (0.0275) Fed Lo G 0.7341 (0.0273) 0.7413 (0.0316) 0.7406 (0.0527) 0.7624 (0.0522) 0.7415 (0.0142) 0.8044 (0.0078) 0.9044 (0.0021) 0.8956 (0.0033) 0.8965 (0.0061) 0.7065 (0.0715) 0.7077 (0.0571) 0.7176 (0.0277) 0.7677 (0.0237) 0.8156 (0.0326) 0.6735 (b) Missing Class Cora Cite Seer Pub Med Amazon Photo Amazon Computers Methods 3 Clients 5 Clients 10 Clients 3 Clients 5 Clients 10 Clients 3 Clients 5 Clients 10 Clients 3 Clients 5 Clients 10 Clients 3 Clients 5 Clients 10 Clients Local 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) Fed Avg 0.3900 (0.1104) 0.1119 (0.0202) 0.0652 (0.0568) 0.2022 (0.0751) 0.1914 (0.0140) 0.3189 (0.0218) 0.0000 (0.0000) 0.0013 (0.0013) 0.0020 (0.0010) 0.0000 (0.0000) 0.0000 (0.0000) 0.0085 (0.0148) 0.0000 (0.0000) 0.0000 (0.0000) 0.0073 (0.0127) Fed SAGE+ 0.5000 (0.0457) 0.1393 (0.0317) 0.0287 (0.0111) 0.5581 (0.0524) 0.1622 (0.0470) 0.3701 (0.0528) 0.0000 (0.0000) 0.0015 (0.0013) 0.0034 (0.0004) 0.0000 (0.0000) 0.0000 (0.0000) 0.0036 (0.0051) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) Fed GCN 0.0702 (0.0713) 0.2123 (0.0197) 0.0549 (0.0091) 0.1648 (0.0187) 0.1702 (0.0833) 0.0833 (0.0584) 0.0000 (0.0000) 0.0000 (0.0000) 0.0006 (0.0006) 0.0000 (0.0000) 0.0156 (0.0271) 0.0097 (0.0169) 0.0000 (0.0000) 0.0000 (0.0000) 0.0085 (0.0148) Fed PUB 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0012 (0.0021) 0.0053 (0.0026) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0002 (0.0003) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) Fed NTD 0.3714 (0.1273) 0.1895 (0.0098) 0.0336 (0.0317) 0.4257 (0.0077) 0.2438 (0.0476) 0.2878 (0.0459) 0.0003 (0.0002) 0.0256 (0.0256) 0.0512 (0.0050) 0.0000 (0.0000) 0.0019 (0.0009) 0.0061 (0.0064) 0.0038 (0.0015) 0.0008 (0.0008) 0.0104 (0.0024) Fed ED 0.5305 (0.1078) 0.1080 (0.0158) 0.0350 (0.0305) 0.3184 (0.2660) 0.1534 (0.0470) 0.1039 (0.1785) 0.0056 (0.0045) 0.0097 (0.0097) 0.0050 (0.0022) 0.2192 (0.0395) 0.1796 (0.0634) 0.0004 (0.0007) 0.1162 (0.0297) 0.0005 (0.0005) 0.0004 (0.0007) Fed Lo G 0.6472 (0.0811) 0.4948 (0.0930) 0.4037 (0.0619) 0.6142 (0.0292) 0.5922 (0.1037) 0.5958 (0.0608) 0.1070 (0.0623) 0.1700 (0.1236) 0.2290 (0.0308) 0.4795 (0.0949) 0.5525 (0.1464) 0.1328 (0.0562) 0.3580 (0.1256) 0.2175 (0.0638) 0.0412 (c) New Client Cora Cite Seer Pub Med Amazon Photo Amazon Computers Methods 3 Clients 5 Clients 10 Clients 3 Clients 5 Clients 10 Clients 3 Clients 5 Clients 10 Clients 3 Clients 5 Clients 10 Clients 3 Clients 5 Clients 10 Clients Local 0.0995 (0.0084) 0.1488 (0.0059) 0.1778 (0.0284) 0.1435 (0.0113) 0.1968 (0.0006) 0.1337 (0.0000) 0.3570 (0.0000) 0.3947 (0.0000) 0.3936 (0.0001) 0.0313 (0.0001) 0.0965 (0.0004) 0.1169 (0.0089) 0.1571 (0.0000) 0.1974 (0.0000) 0.2816 (0.0111) Fed Avg 0.3583 (0.0206) 0.2713 (0.0057) 0.3924 (0.1880) 0.2572 (0.0222) 0.2859 (0.0348) 0.2976 (0.0138) 0.3333 (0.0000) 0.5243 (0.0138) 0.5175 (0.0240) 0.2597 (0.0118) 0.0853 (0.0044) 0.1661 (0.0172) 0.1978 (0.0000) 0.2924 (0.0637) 0.4799 (0.0904) Fed SAGE+ 0.2411 (0.0109) 0.3250 (0.0226) 0.4129 (0.1052) 0.3630 (0.0385) 0.0646 (0.0099) 0.1048 (0.0322) 0.3834 (0.0022) 0.5616 (0.0126) 0.5279 (0.0000) 0.2782 (0.0086) 0.0900 (0.0021) 0.1472 (0.0549) 0.2132 (0.3683) 0.3718 (0.1044) 0.2756 (0.0062) Fed GCN 0.3449 (0.0494) 0.3320 (0.0052) 0.4825 (0.0189) 0.2572 (0.0072) 0.4548 (0.1900) 0.2144 (0.0566) 0.3333 (0.0000) 0.4830 (0.0048) 0.4890 (0.0067) 0.2597 (0.0550) 0.0659 (0.0226) 0.1605 (0.0169) 0.2066 (0.0781) 0.3009 (0.0590) 0.4841 (0.0920) Fed PUB 0.3990 (0.0239) 0.2258 (0.0153) 0.4031 (0.0087) 0.3929 (0.0485) 0.3408 (0.0113) 0.3930 (0.0296) 0.4112 (0.0007) 0.6036 (0.0060) 0.5743 (0.0008) 0.3171 (0.0120) 0.1075 (0.0102) 0.2540 (0.0078) 0.5244 (0.0276) 0.6764 (0.0289) 0.5543 (0.0586) Fed NTD 0.3805 (0.0328) 0.3169 (0.0010) 0.3705 (0.1879) 0.4321 (0.0479) 0.5288 (0.0940) 0.3057 (0.0166) 0.4153 (0.0039) 0.6321 (0.0093) 0.6026 (0.0106) 0.4617 (0.1349) 0.1473 (0.0036) 0.2980 (0.0361) 0.0038 (0.0602) 0.7146 (0.0101) 0.7873 (0.0232) Fed ED 0.4527 (0.0353) 0.2537 (0.0165) 0.3194 (0.1364) 0.3303 (0.1068) 0.4053 (0.0240) 0.1346 (0.0136) 0.4842 (0.0226) 0.6352 (0.0019) 0.5969 (0.0131) 0.5451 (0.0586) 0.1563 (0.0098) 0.2622 (0.0027) 0.1162 (0.0712) 0.7147 (0.0067) 0.5228 (0.0610) Fed Lo G 0.5047 (0.0884) 0.4439 (0.0455) 0.6055 (0.0914) 0.5973 (0.1623) 0.5647 (0.0179) 0.6487 (0.1143) 0.6053 (0.1293) 0.7091 (0.0557) 0.7546 (0.0107) 0.5605 (0.1052) 0.5083 (0.1794) 0.5574 (0.0368) 0.7386 (0.0190) 0.9164 (0.0071) 0.8029 (a) Seen Graph (b) Unseen Node (c) Missing Class (d) New Client Figure 3: Impact of headness of class/degree for various scenarios (Amazon Clothing - 3 Clients). Q3: Do the headness of degree and class really help other clients? We evaluate the importance of the headness of degree/class under various scenarios, both of which are expected to enhance data reliability. As Fed Lo G additionally trains the clients using global synthetic data, we measure the impact by varying the knowledge condensed into global synthetic data. Specifically, we compare four different test settings for constructing global synthetic data, i.e., using 1) head class & head degree nodes (HH), 2) head class & tail degree nodes (HT), 3) tail class & head degree nodes (TH), and 4) tail class & tail degree nodes (TT). Detailed descriptions are provided in Appendix A.15. Figure 3 shows test accuracy curves illustrating the impact of each test setting on performance and stability. Data reliability varies with global synthetic data knowledge; HH knowledge is the most reliable. Class headness significantly affects reliability, evident in the performance gap between head and tail classes. Degree headness impacts stability; tail degree settings show more fluctuations. Thus, using HH knowledge is crucial for maintaining reliability and stable outcomes. w/o PG w/o FS w/o LG Figure 4: Ablation Studies (Cite Seer - 3 Clients). Q4. Does each module effectively address the local overfitting problem? Figure 4 shows the results of ablation studies under the Missing Class setting: w/o LG denotes excluding the Local Generalization (LG) phase, w/o PG denotes without Prompt Generation, and w/o FS denotes without Feature Scaling. We validate that the Local Generalization phase is crucial for addressing the absent knowledge, while other modules (i.e., PG and FS) impact performance and stability. Refer to Appendix A.11.4 for details. Further experiments, including an assessment of Feature Scaling s adaptive effects, hyperparameter analysis, and evaluation on unseen data in an open set, are provided in Appendix A.11. Communication overhead and time complexity analysis are described Appendix A.3 and A.5, respectively. Published as a conference paper at ICLR 2025 6 PRIVACY ANALYSIS Q1. Does utilizing the class distribution of the clients pose a privacy problem? The class distribution does not include individual data but merely represents the proportion of each class, which is far less sensitive than the raw data. As the class distribution is general statistical information that indicates the trends within a group rather than specific data about individual users, we argue that it is very difficult for an attacker to infer specific data of individual nodes from the class distribution. However, in case privacy concerns remain, we can add noise to the class rate to make it difficult to determine the exact class distribution. We experimented with two methods of adding noise to the class rate: 1) adding class-wise Gaussian noise with µ as 0 and σ as a rc k, where a is chosen from [0.01, 0.1, 0.5] and rc k = |Vc k| |Vk| , and 2) performing random permutation of the elements in the class rate vector. To maintain the trend while applying random permutation, we permuted only the elements within the head classes and within the tail classes. Table 3: Class rate with noises (Cora - 3 clients). No Noise GN (a = 0.01) GN (a = 0.1) GN (a = 0.5) RP Seen Graph 0.8601 (0.0118) 0.8530 (0.0089) 0.8542 (0.0080) 0.8560 (0.0010) 0.8631 (0.0010) Unseen Node 0.7341 (0.0273) 0.7127 (0.0191) 0.7217 (0.0318) 0.7057 (0.0123) 0.7351 (0.0054) Missing Class 0.6472 (0.0811) 0.6244 (0.05275) 0.6032 (0.0457) 0.6277 (0.0148) 0.6328 (0.0586) New Client 0.5047 (0.0884) 0.5278 (0.0326) 0.5297 (0.0523) 0.4883 (0.0232) 0.5199 (0.0421) In Table 3, both Gaussian Noise (GN) and Random Permutation (RP) methods, which result in class rates roughly similar to the original, showed no significant difference in performance, except for the GN (a = 0.5) setting that highly deteriorates the trend of the class rate. This indicates that Fed Lo G does not require an exact class distribution as long as the general trend is maintained, allowing us to protect privacy more rigorously. Class 0 Class 1 Class 2 Synthetic feature Original feature Figure 5: 2D PCA visualization of feature distributions for the same class in the Cite Seer dataset. Q2. Can synthetic data be specified to match the original data s features? Since the synthetic data is generated by condensing the original nodes within each client s graph, there may be a potential privacy risk. One way to assess this risk is to compare the feature distributions of the original nodes (i.e., Vk) and the synthetic nodes (i.e., Vk,head and Vk,tail). If these distributions overlap, it suggests that the original features can be reconstructed from the synthetic data, which entails privacy risk. In Figure 5, we present a 2-dimensional PCA visualization of both the original feature matrix (blue) and the synthetic feature matrix (orange) for the same class in the Cite Seer dataset. The clear difference between the two distributions shows that sharing synthetic nodes poses minimal privacy risk. Moreover, privacy risk is further reduced because the synthetic data represents an aggregation of all training nodes, without being tied to any specific node (i.e., condensation). This aggregation also incorporates structural information into features, resulting in a distribution that diverges from the original feature space. Since the synthetic nodes lack an explicit graph structure, their feature space distills structural information differently from the original nodes, leading to a distinct distribution. Q3. How does Fed Lo G provide protection against gradient inversion attacks? Fed Lo G enhances protection against gradient inversion attacks (Zhu et al., 2019), where adversaries attempt to reconstruct the original data from gradients uploaded to the server. This protection is primarily due to each client being trained not only on its local data but also on global synthetic data. The inclusion of this synthetic data introduces noise into the gradients, making it difficult for adversaries to extract information solely from the original data. Additionally, each client applies different levels of feature scaling to the global synthetic data (Section 4.3), and these scaling factors are never shared. This variation further obscures the gradients derived from the synthetic data, making it even harder for adversaries to accurately invert the gradients and reconstruct the original data. 7 CONCLUSION In this study, we address the challenges of local overfitting and unseen data (i.e., unseen node, missing class, and new client) in subgraph-FL with our proposed method, Fed Lo G. Our model generates global synthetic data by condensing reliable information from each class representation and its structural information across clients, enabling adaptive generalization of absent knowledge within local datasets without directly using data from other clients. This approach enhances the generalization capabilities of local models, allowing them to handle unseen data effectively while also mitigating privacy concerns. Our experimental results demonstrate that Fed Lo G outperforms existing baselines, proving its efficacy in novel practical scenarios for generalizing to unseen data. Published as a conference paper at ICLR 2025 ACKNOWLEDGEMENTS This work was supported by Institute of Information & Communications Technology Planning & Evaluation (IITP) and National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (RS-2022-II220157 and RS-2024-00406985) and National Research Foundation of Korea (NRF) funded by Ministry of Science and ICT (NRF-2022M3J6A1063021). Jinheon Baek, Wonyong Jeong, Jiongdao Jin, Jaehong Yoon, and Sung Ju Hwang. Personalized subgraph federated learning. In International Conference on Machine Learning, pp. 1396 1415. PMLR, 2023. Nitesh V Chawla, Kevin W Bowyer, Lawrence O Hall, and W Philip Kegelmeyer. Smote: synthetic minority over-sampling technique. Journal of artificial intelligence research, 16:321 357, 2002. Hong-You Chen and Wei-Lun Chao. On bridging generic and personalized federated learning for image classification. ar Xiv preprint ar Xiv:2107.00778, 2021. Huancheng Chen, Haris Vikalo, et al. The best of both worlds: Accurate global and personalized models through federated learning with data-free hyper-knowledge distillation. ar Xiv preprint ar Xiv:2301.08968, 2023. Yin Cui, Menglin Jia, Tsung-Yi Lin, Yang Song, and Serge Belongie. Class-balanced loss based on effective number of samples. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 9268 9277, 2019. Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial networks. Communications of the ACM, 63(11):139 144, 2020. Kuangpu Guo, Yuhe Ding, Jian Liang, Ran He, Zilei Wang, and Tieniu Tan. Not all minorities are equal: Empty-class-aware distillation for heterogeneous federated learning. ar Xiv preprint ar Xiv:2401.02329, 2024. Will Hamilton, Zhitao Ying, and Jure Leskovec. Inductive representation learning on large graphs. Advances in neural information processing systems, 30, 2017. William L Hamilton. Graph representation learning. Morgan & Claypool Publishers, 2020. Chaoyang He, Keshav Balasubramanian, Emir Ceyani, Carl Yang, Han Xie, Lichao Sun, Lifang He, Liangwei Yang, Philip S Yu, Yu Rong, et al. Fedgraphnn: A federated learning system and benchmark for graph neural networks. ar Xiv preprint ar Xiv:2104.07145, 2021. Haibo He and Yang Bai. Adasyn: Adaptive synthetic sampling approach for imbalanced learning. 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence), pp. 1322 1328, 2008. Haibo He and Yunqian Ma. Imbalanced learning: foundations, algorithms, and applications. 2013. Wei Jin, Lingxiao Zhao, Shichang Zhang, Yozen Liu, Jiliang Tang, and Neil Shah. Graph condensation for graph neural networks. ar Xiv preprint ar Xiv:2110.07580, 2021. Wei Jin, Xianfeng Tang, Haoming Jiang, Zheng Li, Danqing Zhang, Jiliang Tang, and Bing Yin. Condensing graphs via one-step gradient matching. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pp. 720 730, 2022. George Karypis and Vipin Kumar. Metis: A software package for partitioning unstructured graphs, partitioning meshes, and computing fill-reducing orderings of sparse matrices. 1997. Seong-Woong Kim and Dong-Wan Choi. Stable federated learning with dataset condensation. J. Comput. Sci. Eng., 16(1):52 62, 2022. Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. ar Xiv preprint ar Xiv:1412.6980, 2014. Published as a conference paper at ICLR 2025 Thomas N Kipf and Max Welling. Semi-supervised classification with graph convolutional networks. ar Xiv preprint ar Xiv:1609.02907, 2016. Gihun Lee, Minchan Jeong, Yongjin Shin, Sangmin Bae, and Se-Young Yun. Preservation of the global knowledge by not-true distillation in federated learning. Advances in Neural Information Processing Systems, 35:38461 38474, 2022. Qinbin Li, Yiqun Diao, Quan Chen, and Bingsheng He. Federated learning on non-iid data silos: An experimental study. In 2022 IEEE 38th international conference on data engineering (ICDE), pp. 965 978. IEEE, 2022. Xin-Chun Li and De-Chuan Zhan. Fedrs: Federated learning with restricted softmax for label distribution non-iid data. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, pp. 995 1005, 2021. Zexi Li, Tao Lin, Xinyi Shang, and Chao Wu. Revisiting weighted aggregation in federated learning with neural networks. In International Conference on Machine Learning, pp. 19767 19788. PMLR, 2023a. Zexi Li, Xinyi Shang, Rui He, Tao Lin, and Chao Wu. No fear of classifier biases: Neural collapse inspired federated learning with synthetic and fixed classifier. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 5319 5329, 2023b. Ping Liu, Xin Yu, and Joey Tianyi Zhou. Meta knowledge condensation for federated learning. ar Xiv preprint ar Xiv:2209.14851, 2022. Rui Liu, Pengwei Xing, Zichao Deng, Anran Li, Cuntai Guan, and Han Yu. Federated graph neural networks: Overview, techniques, and challenges. IEEE Transactions on Neural Networks and Learning Systems, 2024. Zemin Liu, Trung-Kien Nguyen, and Yuan Fang. Tail-gnn: Tail-node graph neural networks. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, pp. 1109 1119, 2021. Yanbiao Ma, Licheng Jiao, Fang Liu, Shuyuan Yang, Xu Liu, and Lingling Li. Curvature-balanced feature manifold learning for long-tailed classification. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 15824 15835, 2023. Julian Mc Auley, Christopher Targett, Qinfeng Shi, and Anton Van Den Hengel. Image-based recommendations on styles and substitutes. In Proceedings of the 38th international ACM SIGIR conference on research and development in information retrieval, pp. 43 52, 2015. Brendan Mc Mahan, Eider Moore, Daniel Ramage, Seth Hampson, and Blaise Aguera y Arcas. Communication-efficient learning of deep networks from decentralized data. In Artificial intelligence and statistics, pp. 1273 1282. PMLR, 2017. Aditya Krishna Menon, Sadeep Jayasumana, Ankit Singh Rawat, Himanshu Jain, Andreas Veit, and Sanjiv Kumar. Long-tail learning via logit adjustment. ar Xiv preprint ar Xiv:2007.07314, 2020. Alexandru Niculescu-Mizil and Rich Caruana. Predicting good probabilities with supervised learning. In Proceedings of the 22nd international conference on Machine learning, pp. 625 632, 2005. Seungeun Oh, Jihong Park, Eunjeong Jeong, Hyesung Kim, Mehdi Bennis, and Seong-Lyun Kim. Mix2fld: Downlink federated learning after uplink federated distillation with two-way mixup. IEEE Communications Letters, 24(10):2211 2215, 2020. Joonhyung Park, Jaeyun Song, and Eunho Yang. Graphens: Neighbor-aware ego network synthesis for class-imbalanced node classification. In International conference on learning representations, 2021. Mohammad Rasouli, Tao Sun, and Ram Rajagopal. Fedgan: Federated generative adversarial networks for distributed data. ar Xiv preprint ar Xiv:2006.07228, 2020. Prithviraj Sen, Galileo Namata, Mustafa Bilgic, Lise Getoor, Brian Galligher, and Tina Eliassi-Rad. Collective classification in network data. AI magazine, 29(3):93 93, 2008. Published as a conference paper at ICLR 2025 Oleksandr Shchur, Maximilian Mumme, Aleksandar Bojchevski, and Stephan Günnemann. Pitfalls of graph neural network evaluation. ar Xiv preprint ar Xiv:1811.05868, 2018. Myung Jae Shin, Chihoon Hwang, Joongheon Kim, Jihong Park, Mehdi Bennis, and Seong-Lyun Kim. Xor mixup: Privacy-preserving data augmentation for one-shot federated learning. ar Xiv preprint ar Xiv:2006.05148, 2020. Jake Snell, Kevin Swersky, and Richard Zemel. Prototypical networks for few-shot learning. Advances in neural information processing systems, 30, 2017. Arjun Subramonian, Jian Kang, and Yizhou Sun. Theoretical and empirical insights into the origins of degree bias in graph neural networks. ar Xiv preprint ar Xiv:2404.03139, 2024. Jingru Tan, Changbao Wang, Buyu Li, Quanquan Li, Wanli Ouyang, Changqing Yin, and Junjie Yan. Equalization loss for long-tailed object recognition. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 11662 11671, 2020. Yue Tan, Yixin Liu, Guodong Long, Jing Jiang, Qinghua Lu, and Chengqi Zhang. Federated learning on non-iid graphs via structural knowledge sharing. In Proceedings of the AAAI conference on artificial intelligence, volume 37, pp. 9953 9961, 2023. Yuan Wang, Huazhu Fu, Renuga Kanagavelu, Qingsong Wei, Yong Liu, and Rick Siow Mong Goh. An aggregation-free federated learning for tackling data heterogeneity. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 26233 26242, 2024. Chuhan Wu, Fangzhao Wu, Yang Cao, Yongfeng Huang, and Xing Xie. Fedgnn: Federated graph neural network for privacy-preserving recommendation. ar Xiv preprint ar Xiv:2102.04925, 2021a. Jiancan Wu, Xiang Wang, Fuli Feng, Xiangnan He, Liang Chen, Jianxun Lian, and Xing Xie. Selfsupervised graph learning for recommendation. In Proceedings of the 44th international ACM SIGIR conference on research and development in information retrieval, pp. 726 735, 2021b. Teng Xiao, Zhengyu Chen, Donglin Wang, and Suhang Wang. Learning how to propagate messages in graph neural networks. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, pp. 1894 1903, 2021. Yuhang Yao, Weizhao Jin, Srivatsan Ravi, and Carlee Joe-Wong. Fedgcn: Convergencecommunication tradeoffs in federated training of graph convolutional networks. Advances in Neural Information Processing Systems, 36, 2024. Rui Ye, Mingkai Xu, Jianyu Wang, Chenxin Xu, Siheng Chen, and Yanfeng Wang. Feddisco: Federated learning with discrepancy-aware collaboration. In International Conference on Machine Learning, pp. 39879 39902. PMLR, 2023. Tehrim Yoon, Sumin Shin, Sung Ju Hwang, and Eunho Yang. Fedmix: Approximation of mixup under mean augmented federated learning. ar Xiv preprint ar Xiv:2107.00233, 2021. Xianyao You, Caiyun Liu, Jun Li, Yan Sun, and Ximeng Liu. Fedmdo: Privacy-preserving federated learning via mixup differential objective. IEEE Transactions on Circuits and Systems for Video Technology, 2024. Sukwon Yun, Kibum Kim, Kanghoon Yoon, and Chanyoung Park. Lte4g: Long-tail experts for graph neural networks. In Proceedings of the 31st ACM International Conference on Information & Knowledge Management, pp. 2434 2443, 2022. Bianca Zadrozny and Charles Elkan. Calibrating probability with undersampling for unbalanced data. In Proceedings of the First SIAM International Conference on Data Mining, 2001. Jie Zhang, Zhiqi Li, Bo Li, Jianghe Xu, Shuang Wu, Shouhong Ding, and Chao Wu. Federated learning with label distribution skew via logits calibration. In International Conference on Machine Learning, pp. 26311 26329. PMLR, 2022. Ke Zhang, Carl Yang, Xiaoxiao Li, Lichao Sun, and Siu Ming Yiu. Subgraph federated learning with missing neighbor generation. Advances in Neural Information Processing Systems, 34:6671 6682, 2021a. Published as a conference paper at ICLR 2025 Longling Zhang, Bochen Shen, Ahmed Barnawi, Shan Xi, Neeraj Kumar, and Yi Wu. Feddpgan: federated differentially private generative adversarial networks framework for the detection of covid-19 pneumonia. Information Systems Frontiers, 23(6):1403 1415, 2021b. Tianxiang Zhao, Xiang Zhang, and Suhang Wang. Graphsmote: Imbalanced node classification on graphs with graph neural networks. In Proceedings of the 14th ACM international conference on web search and data mining, pp. 833 841, 2021. Ligeng Zhu, Zhijian Liu, and Song Han. Deep leakage from gradients. Advances in neural information processing systems, 32, 2019. Published as a conference paper at ICLR 2025 Supplementary Material for Subgraph Federated Learning for Local Generalization A Appendix 16 A.1 Algorithm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16 A.2 Notations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17 A.3 Communication Overhead . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18 A.4 Detailed Process of the Classifiers θk,H and θk,T . . . . . . . . . . . . . . . . . . 18 A.5 Time Complexities . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19 A.6 Related Works . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19 A.6.1 Improving Generalization in Federated Learning . . . . . . . . . . . . . . 19 A.6.2 Synthetic-based Federated Learning . . . . . . . . . . . . . . . . . . . . . 20 A.7 Detailed Process of Pretraining the Prompt Generator . . . . . . . . . . . . . . . . 21 A.8 Detailed Process of Evaluating Data Reliability . . . . . . . . . . . . . . . . . . . 22 A.9 Criteria for Threshold Degree Value for Tail-Degree Nodes . . . . . . . . . . . . . 23 A.10 Detailed Process of Evaluating Unseen Data . . . . . . . . . . . . . . . . . . . . . 23 A.10.1 Closed Set . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23 A.10.2 Open Set . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24 A.11 Additional Experiments . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24 A.11.1 Impact of the Hyperparameters . . . . . . . . . . . . . . . . . . . . . . . . 24 A.11.2 Assessing the Adaptive Impact of Feature Scaling on Local Clients . . . . 26 A.11.3 Experimental Results on the Open Set . . . . . . . . . . . . . . . . . . . . 26 A.11.4 Ablation Study . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26 A.12 Datasets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27 A.12.1 Dataset Statistics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28 A.13 Baselines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28 A.14 Implementation Details . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29 A.15 Detailed Process of Generating HH/HT/TH/TT Global Synthetic Data . . . . . . . 29 A.16 Experimental Dataset Statistics . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30 Published as a conference paper at ICLR 2025 A.1 ALGORITHM Algorithm 1 Fed Lo G: The Overall Algorithm Server S 1: if round == 0 then 2: Initialize global model ϕglobal 3: else 4: Aggregate local models: ϕglobal 1 K P k ϕk 5: Generate global synthetic data Dglobal: 6: Combine Vk,head based on class proportions Equation 4 7: end if Local Client k 8: if round == 0 then 9: Initialize local model: ϕk ϕglobal 10: Local Fitting: Update ϕk, Vk,head, Vk,tail using Dk 11: else 12: Update local model: ϕk ϕglobal 13: Local Fitting: Update ϕk, Vk,head, Vk,tail using Dk 14: Local Generalization: 15: 1. Download Dglobal, PGc c from server 16: 2. Feature Scaling: Adapt Dglobal features locally Equation 5 17: 3. Prompt Generation: Generate prompt nodes of Dglobal via PGc 18: 4. Update ϕk, Vk,head, Vk,tail using adapted Dglobal Pretrain Prompt Generators (PGs) 20: Local Client k: Pretrain PGk using Dk 21: Server S: Aggregate PGc 1 P k rc k PGk for each class c Weighted by rc k = |Vc k| |Vk| Published as a conference paper at ICLR 2025 A.2 NOTATIONS Table 4: Summary of the notations. For simplicity, we describe the notation based on a head-branch. Notation Description General Notations S Server K Number of clients G Graph V Set of nodes E Set of edges D Dataset consisting of G and Y Y Label set for the nodes XV Feature matrix of a set of nodes V xv Feature vector of a node v V ϕ Set of parameters of a global model |CV| Number of classes within a set of nodes V r Current round hv Representation of a node v α Weight of prediction between head and tail branches Local Client k Notations Gk Local graph for client k Vk Set of nodes within a local graph for client k Ek Set of edges within a local graph for client k Dk Local dataset for client k Dlocal Combined local datasets, Dlocal = SK k=1 Dk Yk Label set for the nodes within a local graph for client k ϕk Set of parameters of a local model for client k φk Parameters of a GNN embedder for client k θk,H Parameters of a head-branch classifier for client k θk,T Parameters of a tail-branch classifier for client k Vk,head Set of synthetic nodes within a head-degree branch of client k v(c,s) k,head s-th synthetic node for class c in a head-branch for client k XVk,head Feature matrix of a set of nodes Vk,head x(c,s) k,head Feature of the s-th synthetic node for class c in a head-branch for client k h(c,s) k,head Representation of the s-th synthetic node for class c in a head-branch for client k Pk,head Set of prototype representations (i.e., representations of synthetic nodes) in a headbranch for client k h Vc k,head Average of prototype representations of class c in a head-branch for client k rc k Proportion of nodes labeled c in client k s dataset Dk PGk Pretrained prompt generator for client k (regardless of specific class) γk Class-wise adaptive factor for client k Gsyn k Synthetic graph set consisting of graphs, each containing the global synthetic nodes (which are adapted locally) neighboring with their generated prompt nodes. vp k Generated prompt node of node vk ˆxvp k Generated feature of generated prompt node vp k for the input feature xvk of node vk x Nvk Average of features of the h-hop neighbors of node vk within Gk Global Synthetic Node Notations Dglobal Global synthetic dataset Gglobal Global synthetic graph Vglobal Set of global synthetic nodes Y global Label set for the global synthetic nodes XVglobal Feature matrix of the global synthetic nodes Vglobal xv(c,s) g Feature of the s-th global synthetic node vg for class c PGc Class-specific prompt generator for class c (by aggregating NGk for all k in a class-wise manner) Hyperparameter Notations s Hyperparameter for assigning the number of synthetic nodes per class λ Hyperparameter for adjusting tail degree threshold β Hyperparameter for adjusting the extent of regularization of the features of synthetic nodes Published as a conference paper at ICLR 2025 A.3 COMMUNICATION OVERHEAD In this section, we provide comparisons of communication overhead across different baselines. Fed Lo G uploads and downloads both the synthetic data and the model parameters at the end of each round. For the Cora dataset with a setting of 3 clients and s = 20, our model has 1,081,926 parameters to share with the server, resulting in 4 bytes 1,081,926=4.32 MB (excluding the prompt generator, which is only trained once at the first round). Additionally, the synthetic data has 182,000 parameters (s |CV| d, where |CV| denotes the number of classes and d denotes the dimension of the features), amounting to 0.72 MB. In summary, our model requires 10.08 MB (2 (4.32 + 0.72)) for upload and download each round. Below are comparisons of communication overhead between models over 100 rounds: Fed Avg Fed SAGE+ Fed GCN Fed PUB Fed NTD Fed ED Fed Lo G Cost (MB) 393.11 1543.58 393.11 786.03 393.11 393.11 1011.14 Table 5: Comparison of Communication Cost (MB) Across Different Models Although Fed Lo G relatively requires higher communication overhead compared to other baselines, it shows faster convergence due to its utilization of reliable class representation, leading to a stable training process. Below are comparisons of communication overhead until each model reaches the same accuracy (i.e., 0.8 on the Cora dataset with 3 clients). Model Fed Avg Fed SAGE+ Fed GCN Fed PUB Fed NTD Fed ED Fed Lo G Rounds to Reach 0.8 58 100 (Fails to reach) 57 29 19 39 10 Cost (MB) 228.00 1543.58 224.07 227.95 72.79 149.41 101.11 Table 6: Rounds to reach 0.8 accuracy and corresponding communication cost in MB across different models. Despite Fed Lo G s higher communication overhead per round, its faster convergence results in a lower overall communication overhead to achieve the same accuracy compared to other baselines. This demonstrates the efficiency and stability of Fed Lo G s training process, making it an effective approach despite the initially higher communication cost per round. A.4 DETAILED PROCESS OF THE CLASSIFIERS θk,H AND θk,T In this section, we provide the detailed process of the classifiers θk,H and θk,T . Specifically, we describe Eq. 2: h vk = θk,H hvk, hvk hv(1,1) k,head , . . . , hvk hv (|CV |,s) k,head which corresponds to the head-degree classifier in the local model. Since both the head-degree and tail-degree classifiers have the same architecture, we focus on describing the head-degree classifier. The primary objective of the classifier is to ensure that all prototypes in Pk,head contribute to the final prediction of the target node, allowing the prediction loss to be influenced by all prototypes. To achieve this, we first generate a message mkj from each prototype node vj Pk,head to the target node vk, based on the distance dkj between them. This is computed using MLPmsg as follows: mkj = MLPmsg([hvk nvk dkj]), (6) where the distance dkj is calculated as: dkj = rkj 2, rkj = [hvk nvk] [hvj hvj], (7) and denotes concatenation. Here, nvk = 1 |Nvk | P vo Nvk hvo is the average embedding of the 1-hop neighbors of the target node vk within the graph Gk. The neighbor information is used because the embeddings of neighbors from head-degree and taildegree nodes differ, enabling each branch to leverage degree-specific knowledge. The target node Published as a conference paper at ICLR 2025 embedding hvk is then updated by applying a learned transformation to the representation differences rkj, followed by aggregation: tkj = [hvk hvj] MLPtrans(mkj), (8) t k = 1 |Pk,head| h vk = hvk + t k, (10) where MLPtrans transforms the message mkj into a scalar value. In summary, the classifiers θk,H and θk,T update the embedding of the target node hk by reflecting the interactions between all different prototypes in Pk,head, ensuring that the final prediction and its loss are influenced by all prototypes (i.e., learnable synthetic nodes). A.5 TIME COMPLEXITIES In this section, we assess the time complexity of Fed Lo G and demonstrate its efficiency in computational requirements. Notably, each classifier, including the head and tail branches, shares the same time complexity as a multi-layer perceptron (MLP), specifically O(d2), where d represents the feature dimension of its input. This ensures that the branches have minimal computational overhead, even when processing high-dimensional features. We provide detailed time complexity calculations for each module as follows: Classifier. Pairwise Distance between Prototypes (Eq. 7). The naive time complexity is O(PF), where P is the number of prototypes (i.e., |Cv| s) and F is the dimension of the input (2 d, where d denotes the dimension of its inputs). Since P is small enough to be negligible, the complexity reduces to O(d). Distance-Based Message Generation (Eq. 6). The naive time complexity is O(PF 2), where F is the input dimension (i.e., 2 d + 1). With P being negligible, this results in a complexity of O(d2), which is the same as that of an MLP. Updating the Target Node s Representation (Eqs. 810). Eq. 8 includes an MLP and elementwise operations with subtraction, giving a total complexity of O(d2). Eqs. 9 and 10 involve only simple additions and are therefore negligible. Thus, the total time complexity for this update is O(d2). Prompt Generator. In the pretraining phase, the prompt generator requires O(|V|d) complexity for Eq. 11, and O(|E|d + |V|d2) for Eq. 12. Therefore, the total time complexity of the prompt generator is O(|E|d + |V|d2), which is the same as the GNN encoder. However, it is worth noting that, during inference, it requires only O(d2), which has the same time complexity as the MLP. Feature Scaling. For feature scaling in Eq. 5, the time complexity is O(d) since the operation involves only simple element-wise additions. Consequently, the total time complexity of the classifiers, including the Prompt Generator and Feature Scaling, is O(d2). This is significantly lighter than the complexity of the GNN encoder, O(|E|d + |V|d2), where |E| and |V| denote the number of edges and nodes, respectively. Graph Encoder. As we utilize Graph SAGE for the GNN encoder, it requires O(|E|d + |V|d2) for both forward and backward passes. Overall Model Complexity. To sum up, our model requires O(|E|d + |V|d2) complexity, which is the same as a GNN encoder. Importantly, each classifier has the same time complexity as an MLP (i.e., O(d2)), which has little influence on the total complexity of our architecture. This highlights the efficiency of Fed Lo G in handling computational demands, even for large-scale graphs with high-dimensional data. A.6 RELATED WORKS A.6.1 IMPROVING GENERALIZATION IN FEDERATED LEARNING One of the core challenges in Federated Learning (FL) is achieving strong generalization across heterogeneous and biased client datasets. Clients often have non-i.i.d. data distributions or class Published as a conference paper at ICLR 2025 imbalances, which make it difficult to train robust and generalized models. To address these challenges, various approaches (Chen & Chao, 2021; Li et al., 2023b;a; Ye et al., 2023) have been proposed, focusing on improving the generalization capabilities of both global and local models. Fed Ro D (Chen & Chao, 2021) bridges the gap between generic FL and personalized FL by leveraging a class-balanced loss and empirical risk minimization. While this approach improves generic FL, it depends on the presence of at least one data point for each class within each client. This reliance makes it less effective in scenarios where certain classes are entirely absent in some clients, a common challenge in federated learning (i.e., the Missing Class setting). Fed ETF (Li et al., 2023b) addresses classifier biases by enhancing the generalization of the global model and enabling personalized adaptation through local fine-tuning. To improve generalization, Fed ETF employs a balanced feature loss weighted by the number of samples in each class. However, its generalization phase does not adequately handle the Missing Class scenario, where certain classes have no samples at all. Furthermore, its reliance on local fine-tuning exacerbates the local overfitting problem, making local models more prone to overfitting their biased data distributions and struggling to generalize to unseen data, such as missing classes. Fed LAW (Li et al., 2023a) enhances the generalization of global models by introducing a learnable weighted aggregation mechanism, where the L1 norm of the aggregation weights is constrained to be less than 1. Additionally, it incorporates the concept of client coherence to identify clients that positively contribute to generalization. Similarly, Fed Disco (Ye et al., 2023) proposes a weighted aggregation method based on the discrepancy between local and global category distributions, further improving the performance of the global model. While both Fed LAW and Fed Disco primarily focus on enhancing the generalization of the global model, our work takes a different approach by addressing the local overfitting problem. Specifically, we aim to improve the generalization of local models, which are prone to overfitting their local data distributions after a few local updates from the global model, even when the global model itself is well generalized. A.6.2 SYNTHETIC-BASED FEDERATED LEARNING Generating synthetic data using aggregated knowledge from clients has emerged as a promising approach to compensate for the limitations of local training data. This method facilitates data augmentation while addressing challenges such as class imbalance and limited data availability. Mix Up-based synthetic data generation methods (Yoon et al., 2021; You et al., 2024; Oh et al., 2020; Shin et al., 2020) augment training datasets by mixing data samples with privacy-preserving techniques. However, these approaches operate in the raw feature space, which poses significant risks of privacy leakage, particularly when the local data size is small. Alternatively, GAN-based (Goodfellow et al., 2020) methods, such as Fed GAN (Rasouli et al., 2020) and Fed DPGAN (Zhang et al., 2021b), leverage generative models to create synthetic data. These methods aim to generalize a global generator to produce synthetic data that can mitigate data imbalance while preserving privacy. However, they incur high computational costs, which limits their practical applicability. Recently, condensation-based methods (Kim & Choi, 2022; Liu et al., 2022; Wang et al., 2024) have been proposed to alleviate the impact of data heterogeneity. Fed DC (Kim & Choi, 2022) condenses synthetic data based on local data and fine-tunes the global model at the server level to ensure stable convergence. Fed MK (Liu et al., 2022) generates synthetic data by condensing private data into meta-knowledge, which is used as an additional training set to accelerate convergence. Fed AF (Wang et al., 2024) introduces an aggregation-free paradigm, where the server directly trains the global model using condensed synthetic data. Key distinctions of our approach compared to existing synthetic-based methods are as follows: No reliance on raw features. Mix Up-based (Yoon et al., 2021; You et al., 2024; Oh et al., 2020; Shin et al., 2020) and condensation-based methods (Kim & Choi, 2022; Liu et al., 2022; Wang et al., 2024) generate synthetic data by augmenting or condensing data at the raw feature level of the input, which can lead to privacy leakage, particularly when the original data is limited. In contrast, our synthetic data has a distinct feature distribution from the original data, arising from differences in the embedding approach used for synthetic and original data, particularly due to the presence or absence of the explicit structure. Furthermore, our method leverages not only the original data but also global synthetic data as an additional training set for condensation. This design significantly Published as a conference paper at ICLR 2025 reduces privacy risks, especially when the local data size is small. Moreover, we only share a subset of the synthetic data (i.e., synthetic data within the head-degree branch), which not only excludes complete information about the local data to enhance privacy but is also specifically designed to capture reliable information relevant to the graph domain. Handling the local overfitting problem. Our method effectively addresses the local overfitting problem, which is one of the most challenging issues in federated learning. Local overfitting occurs after a few local updates with the distributed global model, causing the local model to severely struggle in predicting unseen data that involves unseen distributions, particularly for missing classes. Mix Up-based approaches (Yoon et al., 2021; You et al., 2024; Oh et al., 2020; Shin et al., 2020) still depend on local data for augmenting the training set, which limits their ability to generate data for missing classes. In contrast, our method generates global synthetic data even in scenarios where local data for certain classes is completely absent. This is achieved without relying on raw feature-based Mix Up, ensuring both privacy and flexibility. Optimizing training with synthetic data. We extend beyond the generation of synthetic data by investigating how to train it effectively. Since our synthetic data has a different feature distribution from the original data but is utilized as training data (i.e., local generalization), it is essential to explore how to optimize the model training process with global synthetic data. To address this, we propose the Feature Scaling and Prompt Generator phases, as detailed in Section 4.3, to minimize the training-effect gap between original nodes and synthetic nodes. A.7 DETAILED PROCESS OF PRETRAINING THE PROMPT GENERATOR In this section, we explain the process of pretraining local prompt generators and how they are aggregated on the server to produce unbiased prompt nodes for each class. Specifically, the primary goal of the prompt generator is to ensure that the synthetic graph comprising a target node and its corresponding prompt node derived from the target node s features produces a similar gradient matrix as when the target node is predicted using its true h-hop subgraph within the local graph. Synthetic graph, 𝒢!,#$ True subgraph, 𝒢!,#$ Local graph Local (Client 𝒌). Global 𝐏𝐆𝒄s Feature Matching Gradient Matching Figure 6: Overview of Pretraining the Prompt Generator. Training Local Prompt Generators. Each client k trains its own prompt generator PGk, aiming to generate a synthetic prompt node that optimizes the GNN s training effectiveness on feature-only data (Figure 6(left)). Let Gsyn k,vk be the generated synthetic graph consisting of the target node vk and its generated prompt node vp k. The objective is to approximate the true h-hop subgraph around vk within the local graph Gk (denoted as Gk,vk = (Vk,vk, Ek,vk) Gk) in a compact form within Gsyn k,vk (Jin et al., 2021; 2022). To achieve this, the prompt generator applies feature matching to ensure that the generated prompt node has features similar to those of the true neighbors of the target node. In essence, the synthetic prompt node acts as a "compressed summary" of the surrounding structural information derived from the target node s features. The input to the generator PGk is the feature vector of the target node xvk, and it outputs a synthetic prompt feature ˆxvp k = PGk(xvk) Rd. To align the synthetic prompt node with the average features of real h-hop neighbors within the local graph, we minimize the following Published as a conference paper at ICLR 2025 feature-matching loss: Lk,feat = 1 |Vk| vk Vk ˆxvp k x Nvk 2 2, where x Nvk = 1 |Vk,vk| 1 v Vk,vk \vk xv. (11) To ensure that the training effect on the synthetic graph Gsyn k,vk resembles that on the true h-hop graph Gk,vk, gradient matching is applied. This approach minimizes the difference between gradients of the GNN when trained on the true h-hop graph versus the synthetic graph, aligning the parameter updates and thus making the learning process similar. The gradient-matching loss is defined as follows: Lk,grad = 1 ϕ(n) rand l(ϕ(n) rand; Gsyn k,vk(vk), yvk) ϕ(n) rand l(ϕ(n) rand; Gk,vk(vk), yvk) 2 where N is the number of randomly initialized weights ϕ(n) rand (for n = 1, 2, . . . , N) used to optimize for the target-node classification task. The combined loss for optimizing the local prompt generator is: LPGk = Lk,feat + Lk,grad. (13) Thus, the generator produces a single synthetic prompt node feature ˆxvp k Rd, which serves two main purposes: (1) it captures essential structural information from the target node s features, and (2) it enhances the GNN s learning effect on feature-only data. We pretrain PGk for all k {1, . . . , K} over P (i.e., 100) epochs using the training sets within each local dataset, resulting in a collection PG = {PG1, . . . , PGK}. We set N to 20. A.8 DETAILED PROCESS OF EVALUATING DATA RELIABILITY In this section, we detail the process of evaluating data reliability as outlined in Section 3.1. We define data reliability as the accuracy and consistency of information from decentralized nodes. Specifically, we assess which data within the local dataset positively or negatively impacts other clients in the FL framework. Inspired by the robust performance of GNNs for head degree and head degree data (Yun et al., 2022; Park et al., 2021; Zhao et al., 2021; Liu et al., 2021), we design experiments to evaluate data reliability from two perspectives: 1) headness of degree and 2) headness of class. We use the Pub Med dataset for validation. We set the base settings for both perspectives. In the FL framework, we assign two roles to each client. The receiver is the client who receives information about the target class from other clients. This client is trained using the same training data across all settings for this section, ensuring a fair comparison to validate the impact from other clients. Contributors are the clients who share knowledge from their own data with the receiver . Their training sets (i.e., information shared through the FL framework) vary for each setting, such as adjusting the proportion of head/tail degree nodes or class imbalance rate. In a global setting with K clients in FL, we assign one client as the receiver and the others as contributors (i.e., K 1 clients). To assess how degree or class headness affects data reliability, we measure the target class accuracy of the receiver when varying the training sets of contributors . This helps identify whether headness or tailness of data positively or negatively impacts the receiver . We construct the global model by averaging the weights from each client and then evaluate the global model on the receiver s local graph following Fed Avg (Mc Mahan et al., 2017). Detailed Process for Headness of Degree Perspective We divide head degree and tail degree using the tail degree threshold λ set to 3, as justified in Appendix A.9. Nodes with degrees less than or equal to 3 are considered tail degree nodes, while those with degrees greater than 3 are head degree nodes. We only vary the training dataset of the contributors . We create three different training sets for each contributor : 1) Head degree nodes only (Head degree), 2) Tail degree nodes only (Tail degree), and 3) Balanced degree nodes (Balanced degree). Each training set contains the same number of nodes, but their headness differs according to the setting. The Head degree setting includes only Published as a conference paper at ICLR 2025 head degree nodes with the target class, the Tail degree setting includes only tail degree nodes, and the Balanced degree setting includes an equal mix of head and tail degree nodes. We use the Fed Avg (Mc Mahan et al., 2017) framework for the federated learning setting with 100 rounds. At the final round, we evaluate the accuracy of the target class within the receiver s local data using the global model. We average the performances across all classes and report the mean of three seeds results. Impact of Class Headness on the Data Reliability We define the imbalance rate using the proportion of the number of the target class within each contributor s local data. We fix the training nodes of the target class for each contributor , and varies the number of training nodes for other classes which are not the target class. Let nc be the number of nodes per class, and let the number of training nodes of the target class be nt. We then assign nk number of training nodes for each non-target class: nk = nt + rimb 10 min(nc c C) (14) where min(nc c C) is the minimum number of nodes across all other classes c within the set C, and rimb is the imbalance rate can be defined as: rimb = 10 nk nt min(nc c C). Thus, if the rimb has a negative value (i.e., nt > nk), it means the target class becomes a head class within the local data. Conversely, when the rimb has a positive value, the target class becomes a tail class. As the value of rimb increases, the tailness of the target class gets higher. We set rimb in the range from -5 to +5, and we average the performances of the receiver at the final round across all classes and report the mean of three seed results. A.9 CRITERIA FOR THRESHOLD DEGREE VALUE FOR TAIL-DEGREE NODES Figure 7: The number of nodes for HH/HT/TH/TT at threshold λ (Cora dataset used). Recent methods (Yun et al., 2022; Liu et al., 2021) addressing the degree long-tail problem consider nodes with degrees less than or equal to 5 as tail degree nodes, while those with degrees greater than 5 are considered head degree nodes. As shown in Figure 7, we illustrate the number of nodes belonging to 1) head class & head degree (HH), 2) head class & tail degree (HT), 3) tail class & head degree (TH), and 4) tail class & tail degree (TT) as we vary the threshold value λ within the global graph. We use the Cora dataset for validation. When the threshold λ increases, the number of HH nodes significantly decreases, reducing the amount of knowledge that can be condensed into the global synthetic data. In this work, we set λ to 3 to utilize a sufficient amount of HH knowledge while filtering out noisy information from tail degree nodes. A.10 DETAILED PROCESS OF EVALUATING UNSEEN DATA In this section, we provide a detailed description of our proposed Unseen Data test settings (i.e., Unseen Node , Missing Class , and New Client ). To evaluate realistic scenarios, we define two different settings for evaluating unseen data: 1) Closed set nodes setting (Closed set) and 2) Open set nodes setting (Open set). The results in Table 2 are evaluated on the closed set nodes setting. A.10.1 CLOSED SET Published as a conference paper at ICLR 2025 New Client 𝒱# (a) Closed set (b) Open set Figure 8: Overview of Unseen Data settings (K = 3). Following recent work (Baek et al., 2023), we partition the global graph into several subgraphs using the Metis graph partitioning algorithm (Karypis & Kumar, 1997). For the New Client setting, we generate an additional subgraph, resulting in the partitioning of the global graph into k + 1 subgraphs, where k denotes the number of clients. Due to the properties of the Metis algorithm, the extra subgraph has a distinct label distribution, as the algorithm minimizes the number of edges between partitions, leading to the formation of distinct communities. The closed set setting includes unseen data for the Unseen Node and Missing Class settings from other clients. Specifically, in Figure 8(a), the global set of nodes is V = SK k=1 Vk, with Vi Vj = for all i = j. We construct the Unseen Node and Missing Class nodes for client k by expanding the h-hop subgraph from the local graph Gk. Since we allocate all nodes within the global node set V to the clients, the nodes within the h-hop subgraph (i.e., Vu k ) inevitably overlap with those of other clients. Although nodes may overlap, no edges are shared between different clients. Unseen nodes from other clients establish new connections with the local data. For the Missing Class setting, we select the missing classes for each client and then exclude the nodes corresponding to those classes (i.e., Vuc k ) within each local graph Gk. To maintain the overall context of the local graph, we select the missing class from tail classes, which have the smallest portion within each local graph. If the number of nodes corresponding to the missing classes is insufficient, we add additional missing classes for those clients. Excluded nodes Vuc k are included in Vu k . When evaluating the Missing Class at test time, we expand the local graph Gk to the range of h-hop, and within the evolved graph structure, the local model predicts the labels of nodes in Vuc k . For Unseen Node , the local model predicts the labels of nodes in Vu k \ Vuc k . For real-world case for the closed set setting, consider Store-A, which uses a model tailored to the purchasing habits of its regular customers. This model may struggle to adapt to the distinct buying patterns of customers from Store-B. These new patterns could create unfamiliar also-bought connections between products within Store-A, especially if they involve new products that Store A has never sold before. However, these customers can visit Store-A at any time, forming new relationships with existing nodes, reflecting a real-world scenario. This complexity increases the difficulty in effectively integrating and addressing new nodes in the model. In addition, we provide the data statistics for each setting in Appendix A.16. A.10.2 OPEN SET In real-world scenarios, unseen data outside the global nodes V in the FL system can emerge and form new relationships with existing nodes. We define this setting as Open Set, where the unseen nodes are Vu k V = . To create this setting, we randomly crop 20% of the global graph before partitioning it into k + 1 subgraphs, denoting the cropped node set as Vcrop. Similar to the Closed Set, we exclude nodes corresponding to locally assigned missing classes within each local graph. At test time, for the Unseen Node and Missing Class settings, we reconstruct the structure between cropped nodes Vcrop and local nodes Vk. Within the reconstructed graph, we evaluate the nodes in Vcrop that belong to the missing classes for the Missing Class setting and those having locally trained classes for the Unseen Node setting. In Table 9, we provide the experimental results on the open set in Appendix A.11.3. A.11 ADDITIONAL EXPERIMENTS A.11.1 IMPACT OF THE HYPERPARAMETERS In Figure 9, we analyze the impact of hyperparameters such as the number of synthetic data for each class (s) and the tail degree threshold (λ). Published as a conference paper at ICLR 2025 (a) The number of synthetic data, 𝑠 (b) Tail-degree Threshold 𝜆 Seen Graph Unseen Node Missing Class New Client Seen Graph Unseen Node Missing Class New Client Figure 9: Hyperparameter analysis. The Number of Synthetic Data, s For generating the global synthetic data, sets of learnable nodes Vk,head and Vk,tail are constructed during the Local Fitting phase within each client. We assign s learnable synthetic nodes per class and vary s to assess its impact on global synthetic nodes. As shown in Figure 9(a), we vary s within the range [1, 5, 10, 20, 50] and evaluate the model s performance on the same test data using the Cora dataset with 3 clients. Notably, s significantly impacts the Unseen Data settings, particularly the Missing Class setting, which relies heavily on global synthetic data. A larger number of synthetic data condenses diverse knowledge expressions. However, too many synthetic data points complicate modeling the interaction between the target node and each synthetic nodes (i.e., prototypes), as all prototypes participate in the final prediction described in Section 4.1. Consequently, accuracy for Unseen Data including Unseen Node , Missing Class , and New Client improves with more synthetic data, but an excessive number (e.g., s = 50) can reduce performance. Conversely, the performance of the Seen Graph settings shows robustness to the number of synthetic data compared to the Unseen Data settings because the dependency on knowledge from other clients is lower for test data with the same distribution as the training data. Tail-Degree Threshold, λ We evaluate the impact of the tail-degree threshold λ on performance. Varying λ within the range [0, 3, 5, 10, 20], we use the Cite Seer dataset with 3 clients for the evaluation. As shown in Figure 9(b), the tail-degree threshold λ significantly impacts the Unseen Data settings as it directly influences the knowledge condensed into the global synthetic data. Increasing λ filters out more knowledge from tail-degree nodes, condensing primarily head-degree node knowledge. However, as illustrated in Figure 7 in Section A.9, the number of HH nodes significantly decreases with a higher λ, reducing the amount of knowledge to be condensed into the global synthetic data. Thus, setting λ to 3 yields the best performance, effectively filtering out tail-degree knowledge while ensuring a sufficient amount of HH nodes. Degree-Based Branch Weight for Prediction, α The primary objective of each branch is to distill knowledge from the input data into learnable synthetic data. To achieve this, we designed a prototypical network-based branch that uses learnable synthetic data to represent class-specific knowledge. By adjusting the weight (i.e., alpha) of each branch s final prediction based on the target node s degree, we guide the gradient flow from head degree nodes primarily towards the head branch. This approach enables head degree knowledge to be distilled within the head branch s synthetic data, and similarly, tail-degree knowledge within the tail branch. As shown in Section 3.1, tail-degree knowledge negatively affects the performance of other clients. This effect is illustrated in Figure 3, where we see that synthetic data generated from head-branch knowledge (i.e., HH: head class/head degree and TH: tail class/head degree) outperforms that generated from tail-branch knowledge (i.e., HT: head class/tail degree and TT: tail class/tail degree). Specifically, the performance hierarchy (HH > HT and TH > TT) suggests that the head branch holds more reliable knowledge from the input data, particularly head degree knowledge, as discussed in Section 3.1. This demonstrates that each branch effectively captures distinct types of knowledge, successfully separating head and tail degree information from the input data. Published as a conference paper at ICLR 2025 Table 7: Impact of degree-based branch weight α on performance (Cora dataset used). 3 Clients 5 Clients 10 Clients SG UN MC NC SG UN MC NC SG UN MC NC Fed Lo G (α = 0.5) 0.8613 (0.0108) 0.7154 (0.0239) 0.5668 (0.0330) 0.4852 (0.0329) 0.8519 (0.0065) 0.7397 (0.0016) 0.4837 (0.0331) 0.3982 (0.0055) 0.8377 (0.0089) 0.7160 (0.0770) 0.3769 (0.1721) 0.5285 (0.0986) Fed Lo G 0.8601 (0.0118) 0.7341 (0.0273) 0.6472 (0.0811) 0.5047 (0.0884) 0.8575 (0.0074) 0.7413 (0.0316) 0.4948 (0.0930) 0.4439 (0.0455) 0.8451 (0.0103) 0.7406 (0.0527) 0.4037 (0.0619) 0.6055 (0.0914) Improvement (%p) 0.12 +1.87 +8.04 +1.95 +0.56 +0.16 +1.11 +4.57 +0.74 +2.46 +2.68 +7.70 SG: Seen Graph, UN: Unseen Node, MC: Missing Class, NC: New Client To provide additional clarity, we conducted an ablation study to evaluate the effectiveness of weight averaging based on the target node s degree. In this ablation, we fixed the α value in Eq. 3 at 0.5, preventing degree-specific knowledge from being divided across branches. This configuration results in a simple ensemble of the two branches without considering the degree. The results of this ablation study are shown in Table 7. The performance on the unseen data settings (i.e., UN, MC, and NC) differs significantly from that on the SG (Seen Graph) setting. Specifically, eliminating degree-based weight averaging leads to a significant performance decrease in unseen data settings. This difference arises because, in unseen data settings, the model relies more heavily on the reliability of global synthetic data. Consequently, weighting predictions from each branch based on the target node s degree effectively extracts reliable knowledge into the head branch while preserving tail-specific knowledge within the tail branch. A.11.2 ASSESSING THE ADAPTIVE IMPACT OF FEATURE SCALING ON LOCAL CLIENTS Table 8: γ[c] values at the final round R (Cite Seer - 3 Clients). Class 1 2 3 4 5 6 # Nodes 5 27 129 16 0 25 γ[c] 0.288 0.289 0.297 0.282 0.282 0.286 Fed Lo G shares the same global synthetic data with clients at the end of each round, but clients have distinct absent knowledge due to different label distributions. Clients in Fed Lo G adaptively utilize the global synthetic data by adjusting its perturbation strength in a class-wise manner, as described in Section 4.3. We verify that the adaptive factor optimally adjusts the perturbation strength for local clients. Table 8 shows the value of the adaptive factor for class c (i.e., γ[c]) in Client 1 at the last round R. The adaptive factor for the head class (i.e., class 3) is higher than that for tail (i.e., class 1) and missing classes (i.e., class 5), showing that it effectively adjusts the perturbation strength based on each client s current learning status. A.11.3 EXPERIMENTAL RESULTS ON THE OPEN SET We evaluate the Unseen Node and Missing Class in the Open Set settings to validate the model s ability to generalize to nodes never seen at the global level. The results are provided in Table 9. Similar to the Closed Set, our method, Fed Lo G, outperforms the baselines across most settings. However, in the Unseen Node setting on the Pub Med dataset, some baselines show better performance than our method. We attribute this to the Pub Med dataset providing a sufficient number of training data for each class, allowing methods to generalize well within each class s local data. Conversely, in the Missing Class setting, the baselines fail to generalize due to the absence of local data for the missing classes. In contrast, our model effectively generalizes to all classes, including missing classes, demonstrating its robustness on various real-world scenarios. A.11.4 ABLATION STUDY Unseen Node Missing Class New Client w/o LG w/o PG w/o FS Figure 10: Ablation studies (Cite Seer - 3 Clients). Published as a conference paper at ICLR 2025 Table 9: Performance on Unseen Node and Missing Class in the Open Set setting. (a) Unseen Node Cora Cite Seer Pub Med Amazon Photo Amazon Computers Methods 3 Clients 5 Clients 10 Clients 3 Clients 5 Clients 10 Clients 3 Clients 5 Clients 10 Clients 3 Clients 5 Clients 10 Clients 3 Clients 5 Clients 10 Clients Local 0.1250 (0.0030) 0.2957 (0.0077) 0.2854 (0.0263) 0.4443 (0.0131) 0.3471 (0.0020) 0.5177 (0.0052) 0.7510 (0.0010) 0.7292 (0.0000) 0.7489 (0.0013) 0.1333 (0.0000) 0.1900 (0.0039) 0.3958 (0.0211) 0.1687 (0.0001) 0.2891 (0.0000) 0.3890 (0.0043) Fed Avg 0.6696 (0.0232) 0.5939 (0.0215) 0.4243 (0.1304) 0.6055 (0.0033) 0.7126 (0.0210) 0.5255 (0.0119) 0.8679 (0.0059) 0.7192 (0.0170) 0.6793 (0.0127) 0.2481 (0.0455) 0.2491 (0.0671) 0.2692 (0.0304) 0.3480 (0.0428) 0.2980 (0.0198) 0.2617 (0.0074) Fed SAGE+ 0.6362 (0.0764) 0.5050 (0.0047) 0.3953 (0.0527) 0.4090 (0.0155) 0.2667 (0.0160) 0.3945 (0.0571) 0.9035 (0.0028) 0.8820 (0.0015) 0.8312 (0.0124) 0.3117 (0.0071) 0.2529 (0.0198) 0.3651 (0.0183) 0.4205 (0.0073) 0.6028 (0.0055) 0.3404 (0.0189) Fed GCN 0.6840 (0.0083) 0.6299 (0.0022) 0.4389 (0.1433) 0.6148 (0.0124) 0.6500 (0.0319) 0.5767 (0.0254) 0.8571 (0.0027) 0.7138 (0.0114) 0.6558 (0.0044) 0.2329 (0.0448) 0.2411 (0.0512) 0.2617 (0.0238) 0.3519 (0.0506) 0.2923 (0.0235) 0.2621 (0.0041) Fed PUB 0.6772 (0.0039) 0.5971 (0.0117) 0.4717 (0.0114) 0.6097 (0.0264) 0.7222 (0.0087) 0.5958 (0.0049) 0.8842 (0.0114) 0.8954 (0.0022) 0.8864 (0.0023) 0.4842 (0.0204) 0.5109 (0.0331) 0.3790 (0.0359) 0.4886 (0.0123) 0.5068 (0.0286) 0.4574 (0.0125) Fed NTD 0.7066 (0.0241) 0.6402 (0.0076) 0.4245 (0.1274) 0.6443 (0.0105) 0.7639 (0.0082) 0.5664 (0.0168) 0.8953 (0.0052) 0.8769 (0.0033) 0.8789 (0.0013) 0.5516 (0.0283) 0.6196 (0.0098) 0.4903 (0.0165) 0.4183 (0.0033) 0.6707 (0.0252) 0.6778 (0.0102) Fed ED 0.6904 (0.0163) 0.5453 (0.0185) 0.3024 (0.0038) 0.5985 (0.0330) 0.6568 (0.0060) 0.4448 (0.0232) 0.8978 (0.0047) 0.8771 (0.0043) 0.8805 (0.0028) 0.6491 (0.0346) 0.5872 (0.0395) 0.2581 (0.0460) 0.4326 (0.0119) 0.7420 (0.0291) 0.5751 (0.0374) Fed Lo G 0.7224 (0.0102) 0.7163 (0.0216) 0.6203 (0.0089) 0.6363 (0.0153) 0.7645 (0.0141) 0.6634 (0.0235) 0.8627 (0.0078) 0.8622 (0.0058) 0.8627 (0.0062) 0.8754 (0.0049) 0.8275 (0.0340) 0.6576 (0.0202) 0.7759 (0.0475) 0.8625 (0.0180) 0.7163 (b) Missing Class Cora Cite Seer Pub Med Amazon Photo Amazon Computers Methods 3 Clients 5 Clients 10 Clients 3 Clients 5 Clients 10 Clients 3 Clients 5 Clients 10 Clients 3 Clients 5 Clients 10 Clients 3 Clients 5 Clients 10 Clients Local 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0001 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) Fed Avg 0.0000 (0.0000) 0.2091 (0.0291) 0.0000 (0.0317) 0.1801 (0.0405) 0.4269 (0.0517) 0.1490 (0.0387) 0.2771 (0.0207) 0.0499 (0.0133) 0.0166 (0.0064) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0246) Fed SAGE+ 0.2030 (0.0883) 0.2774 (0.0528) 0.0244 (0.0326) 0.3243 (0.2832) 0.4155 (0.0220) 0.2007 (0.1161) 0.0495 (0.0116) 0.0733 (0.0160) 0.1166 (0.0217) 0.0000 (0.0000) 0.0000 (0.0000) 0.0175 (0.0304) 0.0000 (0.0000) 0.0000 (0.0000) 0.0189 (0.0327) Fed GCN 0.0000 (0.0000) 0.2940 (0.0280) 0.0579 (0.0520) 0.0961 (0.0364) 0.3562 (0.1246) 0.1831 (0.0253) 0.2035 (0.0165) 0.0478 (0.0058) 0.0049 (0.0012) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0177 (0.0085) Fed PUB 0.0000 (0.0000) 0.0082 (0.0094) 0.0352 (0.0000) 0.0000 (0.0000) 0.0251 (0.0220) 0.0070 (0.0060) 0.0318 (0.0125) 0.0002 (0.0004) 0.0100 (0.0087) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) 0.0000 (0.0000) Fed NTD 0.1182 (0.0686) 0.2650 (0.0189) 0.0457 (0.0402) 0.4054 (0.0375) 0.5822 (0.0504) 0.2054 (0.0166) 0.0733 (0.0638) 0.2688 (0.0659) 0.3895 (0.0263) 0.0292 (0.0479) 0.0385 (0.0376) 0.1558 (0.0305) 0.1708 (0.0137) 0.0088 (0.0047) 0.0017 (0.0014) Fed ED 0.1333 (0.0844) 0.1449 (0.0143) 0.0370 (0.0339) 0.2523 (0.0364) 0.2922 (0.0412) 0.1197 (0.0219) 0.1487 (0.0460) 0.1118 (0.0696) 0.1604 (0.0067) 0.0503 (0.0820) 0.0000 (0.0000) 0.0029 (0.0050) 0.0613 (0.0434) 0.1232 (0.0973) 0.0091 (0.0063) Fed Lo G 0.4273 (0.0567) 0.5528 (0.0569) 0.2649 (0.0174) 0.4234 (0.0324) 0.5342 (0.0449) 0.4484 (0.0203) 0.5697 (0.2118) 0.4758 (0.1292) 0.5697 (0.2169) 0.3333 (0.0142) 0.5423 (0.1803) 0.4397 (0.1365) 0.7648 (0.0985) 0.2548 (0.0204) 0.2929 We perform an ablation study on 1) Local Generalization (w/o LG), 2) Prompt Generation (w/o PG), and 3) Feature Scaling (w/o FS). As these modules are all directly related to addressing unseen data, we depict the test accuracy curves in Unseen Data settings to easily verify the effectiveness of each module. Local Generalization Local Generalization is an essential phase to prevent local overfitting after the local updates of each client within the FL framework. The Local Generalization phase enables clients to learn locally absent knowledge from the global synthetic data, allowing them to generalize all classes even if they don t have any data for certain classes within their local data (i.e., missing class). As shown in Figure 10, our method without the Local Generalization phase fails to generalize the missing class, which means Local Generalization is crucial for addressing the absent knowledge. Furthermore, for the Unseen Node and New Client settings, the performance deteriorates when we omit the Local Generalization phase. Prompt Generation We evaluate the effectiveness of the prompt generators PGc [c]. The prompt generators generate the prompt nodes of the global synthetic data Dg, which contain the h-hop neighbor information for the target nodes and also contribute to training by mimicking the true h-hop subgraphs gradient. We perform the ablation study for the prompt generators by omitting the generation of prompt nodes for the global synthetic data, which means we train them without any generated prompts. In Figure 10, without prompt nodes, there is a discrepancy in the training mechanism of the GNN between isolated nodes and nodes within the graph structure, leading to a performance decrease for all settings. Furthermore, the learning curves fluctuate when training the global synthetic data without prompt generation, indicating that using only the features of synthetic nodes negatively affects stability. Feature Scaling Feature Scaling helps each client learn all classes adaptively. Feature Scaling adjusts the strength of the perturbation of the global synthetic data for each client depending on the class prediction ability for all classes at the current round. Thus, Feature Scaling affects the stability of learning for each client. In Figure 10, we can verify the effectiveness of Feature Scaling, as the learning curves are more fluctuating than the original Fed Lo G method, and the performance is decreased. A.12 DATASETS Cora (Sen et al., 2008): The Cora dataset consists of 2,708 scientific publications classified into one of seven classes. The citation network contains 5,429 links. Each publication in the dataset is described by a 1,433-dimensional binary vector, indicating the absence/presence of a word from a dictionary. Published as a conference paper at ICLR 2025 Cite Seer (Sen et al., 2008): The Cite Seer dataset comprises 3,327 scientific publications classified into one of six classes. The citation network consists of 4,732 links. Each publication is described by a 3,703-dimensional binary vector. Pub Med (Sen et al., 2008): The Pub Med dataset includes 19,717 scientific publications from the Pub Med database pertaining to diabetes, classified into one of three classes. The citation network comprises 44,338 links. Each publication is described by a TF/IDF-weighted word vector from a dictionary with a size of 500. Amazon Computers (Mc Auley et al., 2015): The Amazon Computers dataset is a subset of the Amazon co-purchase graph. It consists of 13,752 nodes (products) and 245,861 edges (co-purchase relationships). Each product is described by a 767-dimensional feature vector, and the task is to classify products into 10 classes. Amazon Photos (Shchur et al., 2018): The Amazon Photos dataset is another subset of the Amazon co-purchase graph. It consists of 7,650 nodes (products) and 143,663 edges (co-purchase relationships). Each product is described by a 745-dimensional feature vector, and the task is to classify products into 8 classes. A.12.1 DATASET STATISTICS Table 10: Dataset Statistics Dataset Nodes Edges Features Classes Description Cora 2,708 5,429 1,433 7 Scientific publications Cite Seer 3,327 4,732 3,703 6 Scientific publications Pub Med 19,717 44,338 500 3 Scientific publications Amazon Computers 13,752 245,861 767 10 Amazon co-purchase Amazon Photos 7,650 143,663 745 8 Amazon co-purchase A.13 BASELINES In this section, we provide details for the baselines and the URLs of the official codes, where available. Local. This is a non-FL baseline where each local model is trained independently using the GCN embedder without any weight sharing. Fed Avg. (Mc Mahan et al., 2017) This FL baseline involves clients sending their local model weights to the server, which then averages these weights based on the number of training samples at each client. The aggregated model is then distributed back to the clients. In our implementation, we use GCN as the graph embedder. Fed SAGE+. (Zhang et al., 2021a) This subgraph-FL baseline involves clients using Graph SAGE as an embedder and a missing neighbor generator, trained using a graph mending technique. The neighbor generator creates missing neighbors based on their number and features. With the neighbor generator, local models are trained with compensated neighbors and then their weights are aggregated on the server using Fed Avg-based FL aggregation. Fed GCN. (Yao et al., 2024) This subgraph-FL baseline involves clients who collect h-hop averaged neighbor node features from other clients at the beginning of training to address missing information. The server then collects local model weights for Fed Avg-based FL aggregation. Fed PUB. (Baek et al., 2023) This subgraph-FL baseline proposes weight aggregation based on the similarity between clients. It identifies highly correlated clients with similar community graph structures by using the functional embeddings of local GNNs, which are computed using random graphs as inputs to determine similarities. Published as a conference paper at ICLR 2025 Fed NTD. (Lee et al., 2022) This FL baseline is designed to tackle the challenge of overfitting in local models due to non-IID data across clients. It performs local-side distillation only for non-true classes to prevent forgetting global knowledge corresponding to regions outside the local distribution. In our implementation, we use GCN as the graph embedder. Fed ED. (Guo et al., 2024) This FL baseline is designed to tackle the challenge of overfitting in local models and addresses the issue of local missing classes. Similar to our task, it addresses the missing class problem in FL by adding a loss term that regularizes the logits of missing classes to be similar to those of the global model. In our implementation, we use GCN as the graph embedder. Table 11: Baselines and their corresponding code repositories. * We utilized the Fed Avg code implemented in the official Fed Pub code. Baseline URL / Note Fed Avg* https://github.com/Jinheon Baek/FED-PUB/ Fed SAGE https://github.com/zkhku/fedsage Fed GCN https://github.com/yh-yao/Fed GCN Fed Pub https://github.com/Jinheon Baek/FED-PUB/ Fed NTD https://github.com/Lee-Gihun/Fed NTD Fed ED Self-implemented due to absence of official code. A.14 IMPLEMENTATION DETAILS In this section, we provide implementation details of Fed Lo G. Model Architecture. In our experiments, we use a 2-layer Graph SAGE (Hamilton et al., 2017) implementation (φE) with a dropout rate of 0.5, a hidden dimension of 128, and an output dimension of 64. The model parameters with learnable features XVk,head and XVk,tail are optimized with Adam (Kingma & Ba, 2014) using a learning rate of 0.001. The classifiers θH and θT consist of 2 main learnable functions (i.e., MLPmsg and MLPtrans) as follows: Message generating function (MLPmsg): Two linear layers with Si LU activation (Inputs Linear (2 64 64) Si LU Linear (64 64) Si LU Outputs). Message embedding function (MLPtrans): Three linear layers with Si LU activation (Inputs Linear (64 64) Si LU Linear (64 64) Linear (64 1) Outputs). In all experiments, we utilize 2-layer classifiers. Training Details. Our method is implemented on Python 3.10, Py Torch 2.0.1, and Torch-geometric 2.4.0. All experiments are conducted using four 24GB NVIDIA Ge Force RTX 4090 GPUs. For all experiments, we set the number of rounds (R) to 100 and the number of local epochs to 1. This setting is applied consistently across all baselines. Evaluation Details. For the evaluation under the Seen Node setting, we assess the test nodes using the model that achieves the best validation performance across all rounds (R) in the Seen Node setting. This model is then used to evaluate performance in the other unseen data settings: Unseen Node, Missing Class, and New Client, by testing on the corresponding test nodes for each setting. Hyperparameters. We set the number of learnable nodes s to 20, the tail-degree threshold γ to 3, and select the regularization parameter β to values in the range of [0.01, 0.1, 1]. A.15 DETAILED PROCESS OF GENERATING HH/HT/TH/TT GLOBAL SYNTHETIC DATA In this section, we describe the process of generating global synthetic data using 1) head class & head degree nodes (HH), 2) head class & tail degree nodes (HT), 3) tail class & head degree nodes (TH), and 4) tail class & tail degree nodes (TT). Fed Lo G has two branches, each generating Vk,head and Vk,tail, which contain knowledge from head degree nodes and tail degree nodes, respectively. Published as a conference paper at ICLR 2025 HH. As described in Section 4.2, we generate HH global synthetic data by merging the head degree condensed nodes Vk,head from all clients, weighted by the proportion of head classes for each client. In Figure 2(d), for each class c C, the feature vector of the i-th global synthetic node for class c, xv(c,i) g , is defined as: xv(c,i) g = 1 PK k=1 rc k k=1 rc kxv(c,i) k,head, where rc k = |Vc k| |Vk| represents the proportion of nodes labeled c in the k-th client s dataset. HT. In generating HT global synthetic data, we substitute Vk,head with Vk,tail. Thus, for each class c C, the feature vector of the i-th global synthetic node for class c, xv(c,i) g , is defined as: xv(c,i) g = 1 PK k=1 rc k k=1 rc kxv(c,i) k,tail , TH. For generating TH global synthetic data, we aim to give more weight to the tail classes. To achieve this, we adjust the weights inversely proportional to rc k, ensuring that tail classes (with lower rc k) receive higher weights. The new equation is given by: xv(c,i) g = 1 PK k=1 αc k k=1 αc kxv(c,i) k,head, where αc k = PK j=1 rc j rc k + ϵ (15) Here, ϵ is a very small positive value added to prevent division by zero. In this revised equation, αc k assigns higher weights to classes with smaller rc k values, thereby giving more importance to the tail classes. TT. Finally, we generate TT global synthetic data using: xv(c,i) g = 1 PK k=1 αc k k=1 αc kxv(c,i) k,tail , where αc k = PK j=1 rc j rc k + ϵ (16) A.16 EXPERIMENTAL DATASET STATISTICS In this section, we provide the experimental dataset statistics for all testing settings for three clients, allowing for an easy verification of the data distribution of each client and the New Client. In the Global row, we sum up the statistics from all local clients. Published as a conference paper at ICLR 2025 Table 12: Cora Dataset Statistics (Closed Set) Seen Graph Unseen Node Missing Class Dataset Class Train Valid Test Test Test 0 49 37 32 268 86 1 82 45 75 258 0 2 140 110 133 217 220 3 242 206 171 605 0 4 120 82 90 268 0 5 45 41 32 120 29 6 53 42 27 9 59 0 8 12 4 121 0 1 7 4 3 124 0 2 4 2 3 190 0 3 208 168 131 125 0 4 33 22 22 96 0 5 0 0 0 0 29 6 0 1 0 1 23 0 0 0 0 0 86 1 7 1 3 117 0 2 136 108 130 27 0 3 6 15 14 202 0 4 74 49 60 74 0 5 3 4 2 44 0 6 0 0 0 0 36 0 41 25 28 147 0 1 68 40 69 17 0 2 0 0 0 0 220 3 28 23 26 278 0 4 13 11 8 98 0 5 42 37 30 76 0 6 53 41 27 8 0 0 - - 222 - - 1 - - 12 - - 2 - - 2 - - 3 - - 107 - - 4 - - 87 - - 5 - - 166 - - 6 - - 9 - - Published as a conference paper at ICLR 2025 Table 13: Cite Seer Dataset Statistics (Closed Set) Seen Graph Unseen Node Missing Class Dataset Class Train Valid Test Test Test 0 41 24 25 72 0 1 90 75 90 185 0 2 196 163 137 424 93 3 116 85 77 210 0 4 154 96 110 132 121 5 37 25 26 88 53 0 19 5 11 26 0 1 52 50 58 59 0 2 67 52 32 296 0 3 73 52 44 76 0 4 7 1 7 60 0 5 0 0 0 0 53 0 5 5 2 18 0 1 27 16 26 79 0 2 129 111 105 128 0 3 16 15 14 66 0 4 0 0 0 0 121 5 25 13 17 44 0 0 17 14 12 28 0 1 11 9 6 47 0 2 0 0 0 0 93 3 27 18 19 68 0 4 147 95 103 72 0 5 12 12 9 44 0 0 - - 35 - - 1 - - 53 - - 2 - - 12 - - 3 - - 110 - - 4 - - 93 - - 5 - - 211 - - Published as a conference paper at ICLR 2025 Table 14: Pub Med Dataset Statistics (Closed Set) Seen Graph Unseen Node Missing Class Dataset Class Train Valid Test Test Test Global 0 1271 983 949 221 168 1 1176 897 951 572 1003 2 2972 2171 2141 658 0 Client 0 0 277 222 209 93 0 1 0 0 0 0 346 2 1642 1227 1189 193 0 Client 1 0 994 761 740 128 0 1 0 0 0 0 657 2 594 408 414 213 0 Client 2 0 0 0 0 0 168 1 1176 897 951 572 0 2 736 536 538 252 0 New Client 0 - - 787 - - 1 - - 3500 - - 2 - - 591 - - Published as a conference paper at ICLR 2025 Table 15: Photos Dataset Statistics (Closed Set) Seen Graph Unseen Node Missing Class Dataset Class Train Valid Test Test Test 0 150 102 108 61 28 1 592 502 468 881 0 2 266 219 197 61 20 3 358 241 258 329 0 4 300 258 250 327 0 5 332 217 247 0 25 6 214 146 128 907 0 7 39 17 27 307 0 0 0 0 0 0 28 1 71 50 49 345 0 2 261 215 196 3 0 3 49 26 33 135 0 4 5 6 10 186 0 5 332 217 247 0 0 6 9 5 7 96 0 7 9 5 9 59 0 0 146 100 107 7 0 1 394 369 333 226 0 2 0 0 0 0 20 3 11 8 8 101 0 4 3 1 0 67 0 5 0 0 0 0 18 6 197 132 113 69 0 7 1 0 3 40 0 0 4 2 1 54 0 1 127 83 86 310 0 2 5 4 1 58 0 3 298 207 217 93 0 4 292 251 240 74 0 5 0 0 0 0 7 6 8 9 8 742 0 7 29 12 15 208 0 0 - - 0 - - 1 - - 72 - - 2 - - 3 - - 3 - - 43 - - 4 - - 64 - - 5 - - 1 - - 6 - - 1412 - - 7 - - 248 - - Published as a conference paper at ICLR 2025 Table 16: Computers Dataset Statistics (Closed Set) Seen Graph Unseen Node Missing Class Dataset Class Train Valid Test Test Test 0 168 120 115 146 140 1 297 221 201 1604 0 2 558 442 410 2 479 3 91 63 66 765 0 4 1399 1094 1103 3397 0 5 129 69 98 0 60 6 182 134 164 132 93 7 343 220 232 4 167 8 748 597 580 1321 0 9 119 80 79 105 22 0 164 118 114 32 0 1 98 65 71 398 0 2 558 442 410 2 0 3 66 42 46 206 0 4 41 25 30 813 0 5 0 0 0 0 46 6 0 0 0 0 93 7 343 220 232 4 0 8 12 10 8 447 0 9 108 59 61 25 0 0 4 2 1 114 0 1 125 84 80 573 0 2 0 0 0 0 262 3 7 6 5 235 0 4 139 123 117 1504 0 5 129 69 98 0 0 6 181 133 164 4 0 7 0 0 0 0 143 8 707 561 552 178 0 9 11 21 18 80 0 0 0 0 0 0 140 1 74 72 50 633 0 2 0 0 0 0 217 3 18 15 15 324 0 4 1219 946 956 1080 0 5 0 0 0 0 14 6 1 1 0 128 0 7 0 0 0 0 24 8 29 26 20 696 0 9 0 0 0 0 22 0 - - 30 - - 1 - - 1374 - - 2 - - 2 - - 3 - - 302 - - 4 - - 1360 - - 5 - - 1 - - 6 - - 1 - - 7 - - 1 - - 8 - - 167 - - 9 - - 5 - -