# towards_understanding_adversarial_transferability_in_federated_learning__a4f8c375.pdf Published in Transactions on Machine Learning Research (11/2024) Towards Understanding Adversarial Transferability in Federated Learning Yijiang Li yijiangli@ucsd.edu University of California San Diego Ying Gao gaoying@scut.edu.cn South China University of Technology Haohan Wang haohanw@illinois.edu University of Illinois Urbana-Champaign Reviewed on Open Review: https: // openreview. net/ forum? id= hafn Y2Pi Tn We investigate a specific security risk in FL: a group of malicious clients has impacted the model during training by disguising their identities and acting as benign clients but later switching to an adversarial role. They use their data, which was part of the training set, to train a substitute model and conduct transferable adversarial attacks against the federated model. This type of attack is subtle and hard to detect because these clients initially appear to be benign. The key question we address is: How robust is the FL system to such covert attacks, especially compared to traditional centralized learning systems? We empirically show that the proposed attack imposes a high security risk to current FL systems. By using only 3% of the client s data, we achieve the highest attack rate of over 80%. To further offer a full understanding of the challenges the FL system faces in transferable attacks, we provide a comprehensive analysis over the transfer robustness of FL across a spectrum of configurations. Surprisingly, FL systems show a higher level of robustness than their centralized counterparts, especially when both systems are equally good at handling regular, non-malicious data. We attribute this increased robustness to two main factors: 1) Decentralized Data Training: Each client trains the model on its own data, reducing the overall impact of any single malicious client. 2) Model Update Averaging: The updates from each client are averaged together, further diluting any malicious alterations. Both practical experiments and theoretical analysis support our conclusions. This research not only sheds light on the resilience of FL systems against hidden attacks but also raises important considerations for their future application and development. 1 Introduction Although Federated Learning (FL) provides a promising solution to collaboratively train models without exchanging data, especially in privacy-concerned areas Li et al. (2022), it is still susceptible to attacks such as data poisoning Huang et al. (2011), model poisoning Bhagoji et al. (2019); Bagdasaryan et al. (2020); Huang et al. (2023), free-riders attack Lin et al. (2019), and reconstruction attacks Geiping et al. (2020); Zhu et al. (2019). It is also vulnerable to adversarial attacks during inference Biggio et al. (2013); Szegedy et al. (2013), including adversarial examples designed to deceive the model Zizzo et al. (2020). Research on robust FL methods against adversarial examples has primarily focused on a white-box setting where attackers have full model access Zhou et al. (2020); Reisizadeh et al. (2020a); Hong et al. (2021); Qiao et al. (2024). However, real-world FL applications, like Gboard Hard et al. (2018), usually restrict such access. Published in Transactions on Machine Learning Research (11/2024) We observe a distinct FL security challenge: malicious clients may pose as benign contributors, only revealing their adversarial intent post-training. This setting raises a new security challenge because these clients have access to a subset of the training data, potentially leading to a better surrogate model for transferable attacks. Current FL applications lack effective mechanisms to eliminate such hostile participants Hard et al. (2018), even if selection mechanisms exist, such as Krum Fang et al. (2020); Li et al. (2020a); Bagdasaryan et al. (2020), as attackers do not exhibit hostile behavior during training. After obtaining data, an attacker could train a surrogate model for transfer-based black-box attacks. In this paper, we pioneer an exploration of this practical perspective of FL robustness. Stemming from the above scenarios, we propose a simple yet practical assumption: the attacker possesses a limited amount of the users data but no knowledge about the target model or the full training set. To assess current FL system robustness and guide future research in this regard, we investigate the adversarial transferability under a spectrum of practical FL settings. Adversarial transferability refers to the ability of adversarial examples generated from the source model to successfully attack the target model, which measures the amount of threat white-box attack poses to the model and system. We first evaluate the transferability of adversarial examples generated from different source models to attack a federated-trained model. Then a comprehensive evaluation of practical configurations is conducted to assess the feasibility of our attack. We further investigate two properties of FL: decentralized training and the averaging operation and their correlation with federated robustness. To provide a comprehensive evaluation in a practical aspect, we consider the attack timing, the architecture, and different aggregation methods in our experiments. We have the following findings: We find that, while there are indeed security challenges from the novel attack setting, the federated model is more robust under white-box attack compared with its centralized-trained counterpart when their accuracy on clean images is comparable. We investigate the transferability of adversarial examples generated from models trained by various numbers of users data. We observe that without any elaborated techniques such as dataset synthesis Papernot et al. (2017) or attention Wu et al. (2020b), a regularly trained source model with only limited users data can perform the transfer attack. With Res Net50 on CIFAR10, we achieve a transfer rate of 70% and 80% with only 5% and 7% of users with augmentations and further improve this number to 81% and 85% with Auto Attack Croce & Hein (2020). We investigate two intrinsic properties of the FL: the property of distributed training and the averaging operation and discover that both heterogeneity and dispersion degree of the decentralized data as well as the averaging operations can significantly decrease the transfer rate of transfer-based black-box attack. To further understand the phenomenon, We provide theoretical analysis to further explain the observations. 2 Background 2.1 Adversarial Robustness The adversarial robustness of a model is usually defined as the model s ability to predict consistently in the presence of small changes in the input. Intuitively, if the changes to the image are so tiny that they are imperceptible to humans, these perturbations will not alter the prediction. Formally, given a well trained classifier f and image-label pairs (x, y) on which the model correctly classifies f(x) = y, f is defined to be ϵ-robust with respect to distance metric d( , ) if E(x,y) min x :d(x ,x) ϵ α(f(x ; θ), y) = E(x,y)α(f(x; θ), y) (1) which is usually optimized through maximizing: E(x,y) min x :d(x ,x) ϵ α(f(x ; θ), y) (2) where α denotes the function evaluating prediction accuracy. In the case of classification, α(f(x; θ), y) yields 1 if the prediction f(x; θ) equals ground-truth label y, 0 otherwise. The distance metric d( , ) is usually Published in Transactions on Machine Learning Research (11/2024) approximated by L0, L2 or L to measure the visual dissimilarity of original image x and the perturbed image x . Despite the change to the input is small, the community have found a class of methods that can easily manipulate model s predictions by introducing visually imperceptible perturbations in images Szegedy et al. (2013); Goodfellow et al. (2014); Moosavi-Dezfooli et al. (2016). From the optimization standpoint, it is achieved by maximizing the loss of the model on the input Madry et al. (2017): max δ l(f(x + δ; θ), y) s.t. d(x + δ, x) < ϵ (3) where l( , ) denotes the loss function (e.g., cross-entropy loss) for training the model f parameterized by θ. While these attack methods are powerful, they usually require some degrees of knowledge about the target model f (e.g., the gradient). Arguably, for many real-world settings, such knowledge is not available, and we can only expect less availability of such knowledge on FL applications trained and deployed by service providers. On the other hand, the hostile attacker having access to some but limited amount of users data is a much more realistic scenario. Thus, we propose the following assumption for practical attack in FL: given the data of n malicious users Dm = Sn i=1 Di where Di = {(xk, yk)|k = 1, , mi}(i) contains mi data points, we aim to acquire a transferable perturbation δ by maximizing the same objective as in Equation 3 but with a surrogate model f parameterized by θ trained by Dm: δ = arg max δ l(f (x + δ; θ ), y) s.t. d(x + δ, x) < ϵ (4) We hope to test whether this δ can be used to deceive the target model f as well. 2.2 The Security and Robustness of Federated Learning Poisoning and Backdoor Attack. Poisoning attacks Biggio et al. (2012); Fang et al. (2020) aim to disrupt the global model by injecting malicious data or manipulating local model parameters on compromised devices. In contrast, backdoor attacks Bagdasaryan et al. (2020); Sun et al. (2019); Wang et al. (2023) infuse a malicious task into the existing model without impacting its primary task accuracy Zhang et al. (2023); Huang et al. (2024), including model replacement Chen et al. (2017), label-flipping Fung et al. (2018), fixed-trigger backdoor Dai & Li (2023); Liu et al. (2024) and trigger-optimization Huang (2020); Li et al. (2023a); Nguyen et al. (2024). Defenses against these attacks often involve anomaly detection methods, such as Byzantine-tolerant aggregation Shejwalkar & Houmansadr (2021) (e.g., Krum, Multi Krum Blanchard et al. (2017), Bulyan Guerraoui et al. (2018), Trimmed-mean and Median Yin et al. (2018)), focusing on the geometric distance between hostile and benign gradients. More advanced defenses take detection, aggregation, detection, and differential privacy into consideration Huang et al. (2023); Nguyen et al. (2022). The robustness and attack performance of backdoor attacks is significantly influenced by FL data heterogeneity Zawad et al. (2021). Transfer Attack. Transfer-based adversarial attacks employ the full training set of the target model to train a surrogate model Zhou et al. (2018), a challenging condition to meet in practice, especially in FL where data privacy is paramount. Another line of inquiry delves into the mechanisms of black-box attacks, which exploit the high transferability of adversarial examples even between different model architectures Szegedy et al. (2013); Goodfellow et al. (2014). This transferability is partially attributed to the similarity between source and target models Goodfellow et al. (2014); Liu et al. (2016); Li et al. (2023b), as adversarial perturbations align closely with a model s weight vectors, and different models learn similar decision boundaries for the same task. Tramèr et al. (2017) found that adversarial examples span a large, contiguous subspace, facilitating transferability. Meanwhile, Ilyas et al. (2019) posits that adversarial perturbations are non-robust features captured by the model, and Waseda et al. (2023) utilizes this theory to explain differing mistakes in transfer attacks. Additionally, Demontis et al. (2019); Zhang et al. (2024) reveals that similar gradients in source and target models and lower variance in loss landscapes increase transfer attack probability. Despite transfer attacks being more realistic and practical, not much attention has been focused on this aspect of the safety and robustness of FL. We take the initiative to investigate this area and underscore our setting s distinctiveness and importance compared to others. Key Difference 1: Different from query-based or transfer-based black-box attack, we assume the malicious clients possess the data themselves, impacting the target model during training and attack during inference. Published in Transactions on Machine Learning Research (11/2024) We also present a comparison of our attack setting and the query-based attack in Section 4.3. Note that our attack setting doesn t contradict the query-based attack. In fact, we can perform with both if the FL system allows a certain number of queries, which we leave to future works to explore. Key Difference 2: Poisoning attack or backdoor attack manipulates the parameters update during target model training which can be defended by anomaly detection. Moreover, in practice, despite clients preserving the training data locally, the training procedure and communication with the server are highly encapsulated and encrypted with secret keys, which is even more unrealistic and laborsome to manipulate. Our attack setting circumvents this risk since no hostile action is performed during the training but successfully boost the attack possibility during inference time. Significance: Besides the potential data leakage by malicious participants, we also emphasize that despite, ideally, each participant having access to the global model, in real-world applications (e.g. Gboard Hard et al. (2018)), the infrastructure provider will impose additional protection such as encryption or encapsulation over the local training. For instance, Google s Gboard provides next-word prediction with FL, which requires users to install an app to participate. For an adversary, it s impractical to obtain the global model without breaking or hacking the app or hijacking and decrypting the communication, despite all the things happening "locally". We believe this is much more difficult and laborsome than our setting which significantly boosts the transferability by simply acting as a benign. 3 Investigation Setup and Research Goals GOAL 1: We aim to investigate the possibility of a transfer attack with limited data and validate whether it is possible and practical for the attacker to lay benign during the training process and leverage the obtained data to perform the adversarial attack. GOAL 2: We aim to explore how different degrees of decentralization, the heterogeneity of data and the aggregation, i.e. average affect the transferability of the adversarial examples against the FL model in a practical configuration. 3.1 Experiment Setup Threat Model. Following (Zizzo et al., 2020), we use PGD (Madry et al., 2017) with 10 iterations, no random restart, and an epsilon of 8 / 255 over L norm on CIFAR10. For experiments on Image Net200, we use PGD (Madry et al., 2017) of the same setup but with an epsilon of 2 / 255. Settings. We first build up the basic FL setting. We split the datatset into 100 partitions of equal size in an iid fashion. We adopt two models for the experiments: CNN from (Mc Mahan et al., 2017) since it is commonly used in the FL literature and the widely used Res Net50 (He et al., 2016) which represents a more realistic evaluation. We conduct training in three paradigms: the centralized model, the federated model and the source model with a limited number of clients data. For the federated model, we use SGD without momentum, weight decay of 1e-3, learning rate of 0.1, and local batch size of 50 following (Acar et al., 2021). We follow the cross-device setting and use a subset of clients in each round of training. We use a 10% as default where not specified. We train locally 5 epochs on Res Net50 and 1 epoch on CNN. For centralized and source model training, we leverage SGD with a momentum of 0.9, weight decay of 1e-3, a learning rate of 0.01 and batch size of 64. For adversarial training, we use the same setting as centralized and leverage PGD to perform the adversarial training. We refer to (Zizzo et al., 2020) for the details of adversarial training. All experiments are conducted with one RTX 2080Ti GPU. Metrics. We report Accuracy (Acc) and adversarial accuracy (Adv.Acc) for the performance and the robustness of white-box attack, and for adversarial transferability, we report transfer accuracy (T.Acc) and transfer success rate (T.Rate) as detailed in 3.2. Published in Transactions on Machine Learning Research (11/2024) 3.2 Adversarial Transferability in Federated Learning To define the transferability of adversarial examples, we first introduce the definition of the source model, target model and adversarial example. The source model is the surrogate model used to generate adversarial examples while the target model is the target aimed to attack. Given the validation set x = {(xi, yi)}, source model f , target model f and adversarial perturbation function adv( , ) (e.g., PGD), we first define the following sets: s1 = {xi|f (xi) = yi}, s2 = {xi|f (adv(xi, f )) = yi}, s3 = {xi|f(xi) = yi}, s4 = {xi|f(adv(xi, f )) = yi} Adversarial examples are defined as those samples that are originally correctly classified by model f but are misclassified when the adversarial perturbation is added, i.e., s1 s2. Adversarial transferability against the target model refers to the ability of adversarial examples generated from the source model to attack the target model (become an adversarial example of the target model). We define transfer rate (T.Rate) and transfer accuracy (T.Acc) to measure the adversarial transferability: T.Rate = ||s1 s2 s3 s4|| ||s1 s2 s3|| , T.Acc = 1 ||s4|| ||x|| where || || denotes the cardinality of a set. We are the first one to propose and use the Transfer Rate metric to measure the transferability of adversarial examples which measures the transferability of the surrogate model by measuring the portion of transferable examples. This serves as a complimentary to accuracy as plain accuracy fails to accurately measure robustness. 4 Experiments 4.1 Robustness with Comparable Accuracy Table 1: Centralized and federated model under white-box attack. We can see that, with comparable clean accuracy, the FL model shows greater robustness against white-box attacks compared with the centralized model. This observation is consistent across different datasets and model architectures. Paradigm Architecture same-acc regular Acc Adv.A Acc Adv.A CIFAR10 centralized R50 90.20 0.01 95.24 0.40 R50 (adv) 81.23 23.27 89.46 46.09 CNN 75.06 1.24 82.41 0.35 CNN (adv) 73.15 20.89 76.78 28.92 R50 90.29 0.05 92.31 0.02 R50 (adv) 80.05 36.44 81.05 39.11 CNN 75.09 3.68 76.83 3.98 CNN (adv) 72.85 25.5 72.87 24.35 Image Net200 centralized R50 55.05 3.68 65.79 8.41 R50 (adv) 50.04 31.20 55.59 32.84 federated R50 55.03 13.42 60.59 15.68 R50 (adv) 50.18 38.31 54.92 41.13 In order to provide a preliminary understanding about the robustness of the FL model, we train the centralized model for 200 epochs and the federated model for 400 rounds resulting in a decent accuracy of over 90% (see the regular column of Tab. 1). For the CNN model, we train 200 epochs for the centralized model and 600 rounds for the federated model to achieve an accuracy of over 75%. We can observe that the federated model s clean and adversarial accuracy is lower than its centralized counterpart, aligned as the result in (Zizzo et al., 2020). However, we conjecture that such an increase in adversarial accuracy is not attributed to the intrinsic robustness of the centralized model but largely due to its high clean accuracy. To validate this hypothesis and facilitate a fair comparison between the two paradigms, we early-stop both models when their clean accuracy reaches 90% (75% for CNN) and report the results in the same-acc column of Tab. 1. We early stop at 80% for adversarial training (72% for CNN). We can see that when both models reach a comparable clean accuracy, the FL model shows greater robustness against white-box attacks compared with the centralized model. To further validate our hypothesis, we perform the experiment on a much larger and more realistic dataset, i.e. Image Net200. We early stopped both models at 55% accuracy. Results can be seen in the bottom two rows of Tab. 1. We can see that FL models demonstrate superior robustness against white-box attacks compared with the centralized model on both same-acc and regular settings. Published in Transactions on Machine Learning Research (11/2024) 4.2 Robustness Against Transfer Attack Table 2: T.Rate and transfer accuracy of PGD attack between pairs of models using various training paradigms. The row and column denote the source and target model respectively. For each cell, the left is the transfer accuracy and the right is the T.Rate. federated centralized R50 federated 0.15 / 99.83 2.01 / 97.67 centralized 24.28 / 71.94 7.41 / 91.48 CNN federated 19.31 / 76.32 21.59 / 71.84 centralized 30.57 / 56.59 22.62 / 68.19 Image Net200 R50 federated 2.29 / 95.60 6.52 / 86.74 centralized 22.72 / 54.00 8.27 / 83.13 We turn to the black-box attack which is more practical and realistic in real-world applications. We explore the examples generated by two different training paradigms and their transferability to different models. Since the similarity of decision boundary and clean accuracy influences and reflects the transferability between models (Goodfellow et al., 2014; Liu et al., 2016; Demontis et al., 2019), we early stop both federated and centralized models. For CIFAR10, we early stop both models at 90% of accuracy (75% for CNN). For Image Net200, we follow Section 4.1 and early stop at 55%. We follow this training setting for the rest of this paper. Tab. 2 shows that the adversarial examples generated by the federated model are highly transferable to both the federated and centralized model while adversarial examples generated by the centralized model exhibit less transferability. The T.Rate of federatedto-centralized attack is even larger than centralized-to-centralized attack. Secondly, T.Rate of adversarial examples between models trained under same paradigms is larger than models trained under different paradigms, which can be attributed to the difference of the two training paradigms, e.g., the discrepancy in the decision boundary (Goodfellow et al., 2014; Liu et al., 2016) or different sub-space (Tramèr et al., 2017). 4.3 Transfer Attack with limited data In this section, we comprehensively evaluate the practicality and plausibility of our proposed attack setting, i.e. transfer attack with limited data. We present the overview of our attack setup in Algorithm 1. To simulate this scenario, we fix the generated partition used in the federated training and randomly select a specified number of users as malicious clients whose data is available for performing the attack. To perform the transfer attack, we train a surrogate model in a centralized manner with the collected data. Training details are specified in Section 3.1. Algorithm 1: Our Attack as Benign Setting Input: A set of N clients {C1, C2, . . . , CN} where M clients {C1, C2, . . . , CM} are corrupted by attacker, datasets Xi for each client Ci, sampling rate K, total communication round T, local iteration number E and surrogate model training iteration number Ts. Output: Adversarial perturbation δ on example x /* Normal Federated Training with corrupted clients acting as benign */ initialize FL model f0; for each round t = 1, 2, , T do Server samples K devices St and distributes the global model ft for each client Ci in St do in parallel f i t+1 Client Update(ft, Xi) end ft+1 Aggregate(f 1 t+1, , f K t+1) end /* Train surrogate model */ Adversary collects data from corrupted clients {C1, C2, . . . , CM}; initialize surrogate model f 0; for each iter t = 1, 2, , Ts do f t Client Update(f t 1, S i=1, ,M Xi) end δ Adv Perturb(f , x); One of the key differences between the proposed setting and the conventional transfer-based attack is the amount of available data to train the substitute (source) model, which is a key factor for a successful attack since one would reason that more data will lead to a higher success rate. This is measured by the number of Published in Transactions on Machine Learning Research (11/2024) clients used to train the source model. To provide an overview of the transferability of adversarial examples generated by the source model trained with different numbers of clients, we plot their relation in Fig. 1. We have the following observations: Observation 1. T.Rate increases as the number of users increases, which is consistently observed in both centralized and federated models. Observation 2. With only 20% of clients the source model achieves an T.Rate of 90% and 50% with Res Net50 and CNN respectively. We notice that with Res Net50, the T.Rate of 20% clients is even larger than a transfer attack with full training data (71.94% T.Rate). With CNN, the source model trained with 20% clients can achieve 50% T.Rate which only lags behind the transfer attack by 6% (56.59%). From observation 2, we can see that the proposed attack can achieve comparable or even better T.Rate or lower T.Acc. Consequently, we can conclude that the proposed attack setting with limited data is likely to cause significant security breaches in the current and future FL applications. To further explain observation 1 and an intriguing phenomenon that the T.Rate of Res Net50 model rises to the peak and then decreases, we provide the following hypothesis: When the number of clients used to train the source model is small, the clean accuracy of the source model is also low, leading to a large discrepancy in the decision boundary. Increasing the number of users used in the source model minimizes such discrepancy until the amount of data is sufficient to train a source model with similar accuracy. At this point, the difference between the federated and centralized (Caldarola et al., 2022) becomes the dominant factor affecting the transferability since the source model is trained in the centralized paradigm. 0 10 20 30 40 50 60 70 80 90 100 110 transfer rate number of user (a)Res Net50 on CIFAR10 (c)boosting transfer rate (d)transfer rate in different training stage 0 10 20 30 40 50 60 70 80 90 100 110 number of user 0 10 20 30 40 50 60 70 80 90 100 110 federated centralized (b)CNN on CIFAR10 transfer acc 0 10 20 30 40 50 60 70 80 90 100 110 centralized Resnet50->Res Net50 VGG16->Res Net50 Res Net50->VGG16 40 60 80 100 100 40e 80e 120e 160e 200e 2 4 6 8 10 0 (e)Res Net50on CIFAR10 with 10 and 30 users 0 5 10 15 20 25 30 0 10 20 30 40 50 60 70 80 90 transfer acc transfer rate 60 80 100 120 0 20 regular auto augment pretrain pretrain+auto augment federated auto attack skip 0 20 40 60 80 100 120 number of user number of user number of user transfer rate Figure 1: Attack with data from a limited number of users. (a) We show the transfer rate of our attack with Res Net50 on the CIFAR10 dataset. We additionally experiment with attackers of different architectures. (b) experiment with CNN on CIFAR10 dataset. (c) we boost the performance of our attack with standard augmentation and pretraining techniques. (d) we perform our attack on different training stages. (e) we provide experiments on easier scenarios with 10 and 30 users, which further demonstrate the threat our attack poses. Training surrogate model in a federated manner. To validate the conjecture above, we train the surrogate model in a federated manner. Specifically, since we know each sample s client, we partition the collected data from malicious clients as the federated model and train the source model in a federated manner. Results can be seen in (c) of Fig. 1: with a federated source model, the T.Rate can be slightly boosted at the beginning (limited number of clients) but continue to increase as the number of users increases and finally contributes to a significantly high T.Rate of 99%. This demonstrates our conjecture and also shows that if the hostile party trains the surrogate model in a federated manner, the T.Rate can be further increased. We further evaluate whether the partition information is crucial to the higher transferability in Appendix A. Boosting transfer rate. Following the conjectures above, we propose several effective approaches to enhance the transferability of the surrogate model trained by the malicious data only. We first boost the T.Rate of the source model trained with limited data with data augmentation and model pretraining. Without loss of generality, we leverage Auto Augment (Cubuk et al., 2018) and Image Net pretrain. We believe other forms of augmentation and pretrained weights will exhibit similar effects. From (c) of Fig. 1, we can see, with these techniques we can successfully increase the T.Rate of 1% and 2% of clients from around 3% to 36% and 48% respectively. With 7% or 10% of clients data, the proposed attack setting achieves a high T.Rate of more Published in Transactions on Machine Learning Research (11/2024) than 80% (10% higher than the transfer-based attack). With simple training techniques, malicious clients can attack with more than 40% success rate with one or more clients and 80% with 7 to 10 clients. Advanced Transfer Attack. To further show the threat posed by our attack, we leverage more advanced attacks to replace the default PGD attack as shown in (c) of Fig. 1: both Auto Attack (Croce & Hein, 2020) and skip attack (Wu et al., 2020a) achieve a transfer rate of over 80% with only 3 and 10 users respectively. Noticeably, Auto Attack achieves the highest transfer rate of 85% with 10% of the users. Comparison with query-based attack As elaborated in Sec. 2, our transfer attack with limited data is similar but different from the query-based black-box attack, as we assume the malicious clients possess a portion of the original training data. Query-based black-box attack, on the other hand, aims to attack the target model with limited queries to the target model. Through these queries, the source model manages to optimize the decision boundary towards the target model. However, most APIs and FL-based systems require charges to access the service or equip anomaly detection that detects multiple or malicious queries. We provide a comparison of our proposed attack setting and the query-based black-box attack. To facilitate a fair comparison, we follow the same experiment setting to train the query-based model. We plot the comparison in (a) of Fig. 2. Despite query-based outperforming the default of our proposed attack by a slight margin at the cost of expensive queries, our attack combined with federated training outperforms the query-based attack consistently at all configurations. It is also pertinent to note that facilitating the queries for data from only one client necessitates the execution of 500 requests to the target model (in a 100-client partition). Further, the query cost demonstrates a proportional escalation as the number of required queries increases, which may be regarded as impracticable in tangible operational contexts. Moreover, we underscore that our attack does not counteract the query-based attack. In fact, we can perform both if the FL system allows a certain number of queries. We believe that, combined with a limited number of queries, our attack strategy will be augmented to attain a higher transfer rate since the queries from the target model can help the surrogate model learn a decision boundary that more congruently aligns with that of the target model. 2 0 4 0 6 0 8 0 1 0 0 2 0 4 0 6 0 8 0 1 0 0 q u e r y - b a s e d o u r s ( r e g u l a r ) o u r s ( f e d e r a t e d ) o u r s ( f e d e r a t e d + A u t o A u g m e n t ) t r a n s f e r r a t e n u m b e r o f u s e r s 1 0 2 0 3 0 4 0 5 0 t r a n s f e r r a t e n u m b e r o f u s e r s o u r s s a m e d i s t r i b u t i o n ( n o p a r t i c i p a t i o n ) s a m e d i s t r i b u t i o n n o n - i i d d i r = 0 . 5 ( n o p a r t i c i p a t i o n ) s a m e d i s t r i b u t i o n n o n - i i d s h a r d s = 5 ( n o p a r t i c i p a t i o n ) s o u r c e m o d e l t r a i n e d o n C I F A R 1 0 0 t r a n s f e r r a t e n u m b e r o f u s e r s Figure 2: (a) Comparison with query-based black-box attack. (b) Attack with the surrogate model trained with the same distribution (no participation in the FL training) (c) Attack with the surrogate model trained with similar knowledge, i.e. CIFAR100. Practical evaluation. To practically evaluate our attack setting, we consider two cases: 1. the malicious clients can only participate in some periods during the FL training and 2. the attacker has no knowledge of the model deployed (i.e. unknown architecture). We simulate these situations by attacking in different training stages and with different architectures, as shown in (d) of Fig. 1. Is Data Distribution or Similar Knowledge Sufficient? We emphasize the significance of the process that malicious clients stay benign during training and affect the training with their own data. We demonstrate in this section that this obtained training data is crucial to a successful attack. We first show that using a similar distribution is not as good as the actual training data to train a surrogate model, as shown in (b) of Fig. 2. We note that the inherent heterogeneity of FL training will further impose challenges. The distinct distribution of each client s data makes it nearly impossible to approximate an overall training distribution due to the scale and variability of the data. Published in Transactions on Machine Learning Research (11/2024) Training the surrogate model with a dataset of similar knowledge or characteristics is also ineffective. We simulate this by using the CIFAR100 dataset to train a surrogate model and perform the attack against the FL model trained on the CIFAR10 dataset, as shown in (c) of Fig. 2. We also want to emphasize that our approach to obtaining the training data by acting benignly during training is practically impossible to defend against as there is no way to distinguish a benign client and a malicious client if they stay benign. More Configurations. We also conduct experiments with 10 and 30 users as shown in (e) of Fig. 1 which is a relatively simpler setting. Specifically, we emphasize that when the federated model is trained on 10 and 30 uses, our attack achieves the highest 90% transfer rate with only 2 and 5 malicious clients respectively. When there are only 1 malicious client in the 10-user setting and 3 malicious clients in the 30-user setting, we can achieve over 60% transfer rate and over 70% transfer rate. We further perform experiments on CIFAR100, SVHN and a much larger and more realistic dataset Image Net200, as demonstrated in (a) of Fig. 5 where similar trends are demonstrated. These results emphasize the threat of our attack setting. 5 Two intrinsic properties contributing to transfer robustness To fully understand how adversarial examples transfer between centralized and federated models, we study two intrinsic properties of FL and its relation with transfer robustness. To protect the privacy of clients and leverage the massive data from user-end, FL utilizes distributed data to train a global model through local updates and aggregation at the server (Mc Mahan et al., 2017). As a consequence, the heterogeneity of the distributed data and the aggregation operation is the core component of an FL method. In this section, we study how these two properties affect the transfer robustness of the FL model. Figure 3: T.Rate vs. data of different heterogeneity and dispersion degree. (a): top 3 are results of Res Net50; (b): bottom 3 are results of CNN; Left: T.Rate as a function of the number of users in federated training; Middle: T.Rate as a function of dirichlet alpha; Right:T.Rate as a function of unbalanced sgm. 5.1 Decentralized training and data Heterogeneity This section aims to explore the relationship between the degree of heterogeneity and transfer robustness. Control the degree of dispersion and heterogeneity. To explore the impact of distributed data on adversarial transferability, we control decentralization and heterogeneity through four indexes. By varying the number of clients in the partition, we alter the degree of dispersion of distributed data. We provide two approaches to control the heterogeneity of the distributed data. 1. Change the number of maximum classes per client (Mc Mahan et al., 2017). 2. Change the alpha value of Dirichlet distributions (α) (Wang et al., 2020; Yurochkin et al., 2019) (smaller α means a more non-iid partition) and the log-normal variance (sgm) of the Log-Normal distribution (larger variance denotes more unbalanced partitions) used in unbalanced partition (Zeng et al., 2021). We leverage the Fed Lab framework to generate the different partitions (Zeng et al., 2021). For simplicity, we leverage the centralized trained model as the source model for the rest of the experiments. Published in Transactions on Machine Learning Research (11/2024) Degree of decentralization reduces transferability. We first explore the relation between decentralization and transferability. To control the degree of decentralization, we generate partitions with different numbers of clients and train target federated models on these partitions. Then we perform the transfer attack using the centralized model as the source model. As seen in Fig. 3 (left of (a) and (b)), we can observe that despite some fluctuations, T.Rate drops with an increasing number of users, which demonstrates that more decentralized data leads to lower transferability. We provide statistical testing for the correlation coefficient in Appendix C to further validate the above observation. With the Spearman correlation coefficient, we report a significant negative correlation on both Res Net50 and CNN between the degree of decentralization and T.Rate under a significance level of 0.1 with p-value=0.03 and 0.01 respectively for Res Net50 and CNN. We also visualize linear regression to fit the negative correlation between the degree of decentralization and T.Rate the in Appendix H. a x i m u m n u m b e r o f c l a s s e s p e r u s e r t r a n s f e r r a t e ( a ) R e s N e t 5 0 o n C I F A R 1 0 a x i m u m n u m b e r o f c l a s s e s p e r u s e r 5 5 2 5 % ~ 7 5 % ( 1 0 u s e r s ) 2 5 % ~ 7 5 % ( 1 0 0 u s e r s ) R a n g e w i t h i n 1 . 5 I Q R M e d i a n L i n e M e a n O u t l i e r s ( b ) C N N o n C I F A R 1 0 Figure 4: Transfer rate v.s. maximum number of classes per client; (a): Results of Res Net50 on CIFAR10 dataset; (b): Results of CNN on CIFAR10 dataset; It is noteworthy to observe that the transfer rate in the 10-user setting persistently surpasses that in the 100-user setting. This further substantiates the proposition that more decentralized training leads to lower adversarial transferability for the federated model. Data heterogeneity affects transfer attack. As discussed in Sec. 5.1, we provide two approaches for controlling the heterogeneity of the data. First, we alter the alpha values of the Dirichlet distributions to generate heterogeneous data of different degrees. As per the middle plot of (a) and (b) in Fig. 3, we can see that T.Rate increases as α increases. We also alter the maximum number of classes per client to generate heterogeneous data of different degrees. Results are reported in Fig. 4. We can observe from Fig. 4 a clear increase trend in both the 10user partition and the 100-user partition setting with Res Net50 and CNN. As the degree of heterogeneity decreases, the transferability increases. Experiments in both settings can illustrate our findings. This proves that our observations hold to wider circumstances and scenarios. We further explore whether unbalanced data (i.e. different clients possess different numbers of samples) leads to a decrease in transferability in Fig. 3 (right of (a) and (b)). We can observe that a larger variance leads to a lower transfer rate, meaning that unbalanced data also contribute to higher robustness. To validate the above observation, we provide statistical testing for the correlation coefficient in Appendix C. With the Spearman correlation coefficient, we report a significant negative correlation on Res Net50 between log-normal variance and T.Rate under a significance level of 0.1 (p-value=0.05). We report a significant correlation on all results with a level of 0.05 except the CNN experiments with different Dirichlet α and unbalanced sgm. Visualization of linear fitting is in Appendix H. To provide a more practical evaluation, we follow the setting in practical evaluation of Sec. 4.3 and evaluate the above findings in different training stages and with different architectures. As per (c) and (d) of Fig. 3, we can observe a similar correlation and trends between Dirichlet alpha and transfer rate. This further illustrates that our findings hold in various settings and configurations. 5.2 Averaging Leads to Robustness We explore the other core property, averaging operation, of FL and its correlation with transfer robustness. To change the degree of averaging in FL, we alter the number of clients selected to average at each round. To comprehensively illustrate this finding, we use two source models to attack. The first one is a centralized model trained with all the data. The Second surrogate model is trained with 30% of all the clients data, simulating the attack in Sec. 4.3. We plot the relation of T.Rate and #averaged users in (a) to (d) of Fig. 6. Both CNN and Res Net50 exhibit a decreasing trend as more users participate in the averaging operation. This demonstrates that the averaging operation contributes to the robustness of the FL model and more clients to average per round leads to higher transfer robustness. We provide statistical testing to validate Published in Transactions on Machine Learning Research (11/2024) the correlation in Appendix C. With the Spearman correlation coefficient, we report a significant negative correlation in all four experiments (all p-values are less than .001). To provide a more practical evaluation, we follow the setting in practical evaluation of Sec. 4.3 and evaluate the above findings in different training stages and with different architectures. As shown in (e), (f) of Fig. 6, we can observe a similar trend as mentioned above. This demonstrates that this phenomena hold to more practical settings and wider configurations. We further illustrate that this observation generalizes to different aggregation methods in Appendix B. Res Net18 on SVHN Res Net50 on CIFAR100 (a) our attack (c) heterogeneity (b) decentralization (d) averaging 0.2 0.4 0.6 0.8 transfer rate 20 50 100 200 300 400 500 70 number of user transfer rate 1 5 10 30 50 70 100 transfer rate user to average 0.2 0.4 0.6 0.8 1 32 transfer rate 1 5 10 30 50 70 100 35 transfer rate user to average 1 10 100 transfer rate number of users source traind with fed centralized federated 1 10 100 55 transfer rate number of users source traind with fed centralized federated 20 50 100 200 300 36 number of user transfer rate Res Net50 on Image Net200 transfer rate transfer rate Figure 5: Additional experiments on SVHN, CIFAR100 and Image Net200. (a) Our attack with Res Net18 on SVHN (first row), Res Net50 on CIFAR100 (second row) and Res Net50 on Image Net200 (third row). (b) How decentralization relates to the transfer robustness of FL model on SVHN (first row), CIFAR100 (second row) and Image Net200 (third row) datasets. (c) How heterogeneity relates to the transfer robustness of FL model on SVHN (first row), CIFAR100 (second row) and Image Net200 (third row) datasets. (d) How averaging operation relates to the transfer robustness of FL model on SVHN dataset (first row), CIFAR100 (second row) and Image Net200 (third row) datasets. 5.3 Discussion We summarize the above investigations as the below take-home messages and provide implications for understanding adversarial transferability in FL and secure FL applications: The heterogeneous data and a large degree of decentralization both result in lower transferability of adversarial examples from the surrogate model. The attacker can benefit from closing the discrepancy between the surrogate model and the target model (e.g., train the surrogate model in a federated manner). With more clients to average at each round, the federated model becomes increasingly robust to black-box attacks. Defenders can benefit from increasing the number of clients selected at each round to average. In addition, we also identify a different, simpler, but practical attack evaluation for FL, which can serve as the standard robustness evaluation for future FL applications. Published in Transactions on Machine Learning Research (11/2024) 1 5 10 30 50 70 100 56 58 60 62 64 66 68 70 72 74 76 78 80 82 84 86 88 (a) Res Net50(centralized) (b) Res Net50(30users) (d) CNN(30users) (c) CNN(centralized) number of user to average per round number of user to average per round (e) transfer rate v.s. #user to average number of user to average per round number of user to average per round (f) transfer rate v.s. #user to average 1 5 10 30 50 70 100 70 72 74 76 78 80 82 84 86 88 90 92 94 96 1 5 10 30 50 70 100 1 5 10 30 50 70 100 transfer rate training stage 1 5 10 30 50 70 100 number of user to average per round transfer rate transfer rate 20% 40% 60% 80% 100% r-r r-v v-r r-r r-v v-r r-r r-v v-r r-r r-v v-r r-r r-v v-r r-r r-v v-r r-r r-v v-r 1 5 10 30 50 70 100 40 45 50 55 60 65 70 75 80 85 90 transfer rate Res Net50-Res Net50 Res Net50->VGG16 VGG16->Res Net50 Figure 6: Transfer rate v.s. different number of clients selected to average in each round; (a) Res Net50 results with source model trained in centralized manner with full data; (b) Res Net50 results with source model trained with 30 users; (c) CNN results with source model trained in centralized manner with full data; (d) CNN results with source model trained trained with 30 users; 6 Supporting Theoretical Evidence Notation. We use (X, Y ) to denote a dataset, where X Rn p and Y Rn. We use (x, y) to denote a sample following Section 2. We consider the problem of federated learning optimization model: minθ{l(θ) PN k=1 pklk(θ)}, where lk(θ) = 1 nk Pnk j=1 l(f(x; θ), y), N is the total number devices and nk is the number of samples possessed by device k. pk is the weight of device k with the constraint that P pk = 1. Following Li et al. (2020b). we assume the algorithm performs a total of T stochastic gradient descent (SGD) iterations and each round of local training possesses E iterations. T E is thus the total communication times. At each communication, a maximal number of K devices will participate in the process. We quantify the degree of non-iid using Γ = l PN k=1 pklk where l and lk is the minimum value of l(θ) and lk(θ) respectively. We use the relative increase of adversarial loss Demontis et al. (2018) to measure the transferability for easier derivation, which is defined as the loss of the target model of an adversarial example, simplified through a linear approximation of the loss function: T = l(f(x + ˆδ; θ), y) l(f(x; θ), y) + ˆδT xl(f(x; θ), y), where ˆδ is some perturbation generated by the surrogate model, which corresponds to the maximization of an inner product over an ϵ-sized ball under the above linear approximation: ˆδ arg max δ p<ϵ l(f (x + δ; θ ), y), max δ p<ϵ δT xl(f (x; θ ), y) = ϵ xl(f (x; θ ), y) q where q and p are dual norm. Without loss of generality, we take p = 2 and gives optimal ˆδ = ϵ xl(f (x; θ ), y)/ xl(f (x; θ ), y) 2 from the surrogate model. Substituting it back to the Equation 6 we have the loss increment, bounded by the loss of the white-box attack: l = ϵ xl(f (x; θ )T xl(f(x; θ), y) xl(f (x; θ ), y)T 2 ϵ xl(f(x; θ), y) 2 We define the relative increase in loss in the black-box case compared to the white box as R(x, y), which we show has a lower bound. R(x, y, θ, θ ) = xl(f (x; θ ), y)T xl(f(x; θ), y) xl(f (x; θ ), y) 2 xl(f(x; θ), y) 2 (5) Published in Transactions on Machine Learning Research (11/2024) Then we provide a low bound for R. Theorem 6.1. With Assumptions 2-8, we have: E[R(x, y, θ , θ )] 2µ(γ + T 1)θT θ 4(B + C)κ + µ2γκE θ1 θ 2 (6) where L, µ, σk, G are defined in the assumptions, κ = L µ , γ = max{8κ, E} , B = PN k=1 p2 kσ2 k+6LΓ+8(E 1)2G2 and C = 4 K E2G2. θ1 is the parameter after one step update of SGD. θ is the optimal parameter for the centralized model. We refer to Appendix D for proof. Corollary 6.2. We derive the lower bound of the expectation of R(x, y, θ , θ ) by setting E = 1 and K = 1, where θ represents the centralized source model. E[R(x, y, θ , θ )] 2µ(γ + T 1)θT θ 4(PN k=1 p2 kσ2 k + 6LΓ + 4G2)κ + µ2γκE θ1 θ 2 Remark 6.3. The difference between the lower bound of FL (Lemma D.1) and the centralized model (Corollary 6.2) lies in the denominator. With FL model, k=1 p2 kσ2 k + 6LΓ + 8(E 1)2G2 + 4 while centralized gives a smaller k=1 p2 kσ2 k + 6LΓ + 4G2, leading to a larger lower bound. Thus, the transferability of adversarial examples generated by the surrogate centralized model to attack the federated model is less than that when attacking a centralized model. Remark 6.3 supports the empirical findings in Section 4.2. Remark 6.4. With Theorem 6.1, we can see the degree of non-iid Γ lies in the denominator of the lower bound, meaning that the larger the degree of non-iid among devices, the less the transferability of examples generated by centralized surrogate model. Remark 6.4 aligns with our empirical findings in Section 5.1, which will provide more insights to future research on federated adversarial robustness. Theorem 6.5. Let T denote the train set T = (X, Y ) sampled from some data distribution and x denote the adversarial example following some adversarial distribution. Let f( , T ) be the model trained with dataset T and f( ) = ET [f( ; T )] is the expectation of the model trained on dataset T . We denote the an average of n models as fn = 1 n Pn i=1 fi( ; Ti). Then we have: Ex ,y ET [ y fn(x ) 2 2] = Ex ,y[ y f(x ) 2 2] + 1 n Ex ,T [ f(x ) f(x , T ) 2 2] Remark 6.6. In Theorem 6.5, Ex ,y[ y f(x ) 2 2] and Ex ,T [ f(x ) f(x , T ) 2 2] only depends on f and are fixed with respect to n. Thus, as n becomes larger, the expected error decreases. Remark 6.6 supports the observation in Section Section 5.2. 7 Conclusion We explore the potential for malicious clients to masquerade as benign entities in Federated Learning, then exploit this position to launch transferable attacks. We provide a thorough investigation of the proposed Published in Transactions on Machine Learning Research (11/2024) attack with limited data setting. Our evaluation shows that limited data can yield a comparable transfer rate to a full-dataset attack. To fully understand how adversarial examples transfer between centralized and federated models, we further study two intrinsic properties of FL and its relation with transfer robustness. We discover that decentralized training, heterogeneous data, and averaging operations enhance transfer robustness and reduce the transferability of adversarial examples. We provide evidence from both the perspective of empirical experiments and theoretical analysis. Our findings have implications for understanding the robustness of federated learning systems and poses a practical question for federated learning applications. Durmus Alp Emre Acar, Yue Zhao, Ramon Matas Navarro, Matthew Mattina, Paul N Whatmough, and Venkatesh Saligrama. Federated learning based on dynamic regularization. ar Xiv preprint ar Xiv:2111.04263, 2021. Eugene Bagdasaryan, Andreas Veit, Yiqing Hua, Deborah Estrin, and Vitaly Shmatikov. How to backdoor federated learning. In International Conference on Artificial Intelligence and Statistics, pp. 2938 2948. PMLR, 2020. Arjun Nitin Bhagoji, Supriyo Chakraborty, Prateek Mittal, and Seraphin Calo. Analyzing federated learning through an adversarial lens. In International Conference on Machine Learning, pp. 634 643. PMLR, 2019. Battista Biggio, Blaine Nelson, and Pavel Laskov. Poisoning attacks against support vector machines. ar Xiv preprint ar Xiv:1206.6389, 2012. Battista Biggio, Igino Corona, Davide Maiorca, Blaine Nelson, Nedim Šrndić, Pavel Laskov, Giorgio Giacinto, and Fabio Roli. Evasion attacks against machine learning at test time. In Joint European conference on machine learning and knowledge discovery in databases, pp. 387 402. Springer, 2013. Peva Blanchard, El Mahdi El Mhamdi, Rachid Guerraoui, and Julien Stainer. Machine learning with adversaries: Byzantine tolerant gradient descent. Advances in Neural Information Processing Systems, 30, 2017. Debora Caldarola, Barbara Caputo, and Marco Ciccone. Improving generalization in federated learning by seeking flat minima. ar Xiv preprint ar Xiv:2203.11834, 2022. Xinyun Chen, Chang Liu, Bo Li, Kimberly Lu, and Dawn Song. Targeted backdoor attacks on deep learning systems using data poisoning. ar Xiv preprint ar Xiv:1712.05526, 2017. Francesco Croce and Matthias Hein. Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks. In International conference on machine learning, pp. 2206 2216. PMLR, 2020. Ekin D Cubuk, Barret Zoph, Dandelion Mane, Vijay Vasudevan, and Quoc V Le. Autoaugment: Learning augmentation policies from data. ar Xiv preprint ar Xiv:1805.09501, 2018. Yanbo Dai and Songze Li. Chameleon: Adapting to peer images for planting durable backdoors in federated learning. In International Conference on Machine Learning, pp. 6712 6725. PMLR, 2023. Ambra Demontis, Marco Melis, Maura Pintor, Matthew Jagielski, B. Biggio, Alina Oprea, C. Nita-Rotaru, and F. Roli. Why do adversarial attacks transfer? explaining transferability of evasion and poisoning attacks. Usenix Security Symposium, 2018. Ambra Demontis, Marco Melis, Maura Pintor, Matthew Jagielski, Battista Biggio, Alina Oprea, Cristina Nita-Rotaru, and Fabio Roli. Why do adversarial attacks transfer? explaining transferability of evasion and poisoning attacks. In 28th USENIX security symposium (USENIX security 19), pp. 321 338, 2019. Published in Transactions on Machine Learning Research (11/2024) Minghong Fang, Xiaoyu Cao, Jinyuan Jia, and Neil Gong. Local model poisoning attacks to {Byzantine Robust} federated learning. In 29th USENIX Security Symposium (USENIX Security 20), pp. 1605 1622, 2020. Clement Fung, Chris JM Yoon, and Ivan Beschastnikh. Mitigating sybils in federated learning poisoning. ar Xiv preprint ar Xiv:1808.04866, 2018. Jonas Geiping, Hartmut Bauermeister, Hannah Dröge, and Michael Moeller. Inverting gradients-how easy is it to break privacy in federated learning? Advances in Neural Information Processing Systems, 33: 16937 16947, 2020. Stuart Geman, Elie Bienenstock, and René Doursat. Neural networks and the bias/variance dilemma. Neural computation, 4(1):1 58, 1992. Ian J Goodfellow, Jonathon Shlens, and Christian Szegedy. Explaining and harnessing adversarial examples. ar Xiv preprint ar Xiv:1412.6572, 2014. Rachid Guerraoui, Sébastien Rouault, et al. The hidden vulnerability of distributed learning in byzantium. In International Conference on Machine Learning, pp. 3521 3530. PMLR, 2018. Andrew Hard, Kanishka Rao, Rajiv Mathews, Swaroop Ramaswamy, Françoise Beaufays, Sean Augenstein, Hubert Eichner, Chloé Kiddon, and Daniel Ramage. Federated learning for mobile keyboard prediction. ar Xiv preprint ar Xiv:1811.03604, 2018. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770 778, 2016. Junyuan Hong, Haotao Wang, Zhangyang Wang, and Jiayu Zhou. Federated robustness propagation: Sharing adversarial robustness in federated learning. ar Xiv preprint ar Xiv:2106.10196, 2021. Robert Hönig, Yiren Zhao, and Robert Mullins. Dadaquant: Doubly-adaptive quantization for communicationefficient federated learning. In International Conference on Machine Learning, pp. 8852 8866. PMLR, 2022. Anbu Huang. Dynamic backdoor attacks against federated learning. ar Xiv preprint ar Xiv:2011.07429, 2020. Ling Huang, Anthony D Joseph, Blaine Nelson, Benjamin IP Rubinstein, and J Doug Tygar. Adversarial machine learning. In Proceedings of the 4th ACM workshop on Security and artificial intelligence, pp. 43 58, 2011. Siquan Huang, Yijiang Li, Chong Chen, Leyu Shi, and Ying Gao. Multi-metrics adaptively identifies backdoors in federated learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 4652 4662, 2023. Wenke Huang, Mang Ye, Zekun Shi, Guancheng Wan, He Li, Bo Du, and Qiang Yang. Federated learning for generalization, robustness, fairness: A survey and benchmark. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2024. Andrew Ilyas, Shibani Santurkar, Dimitris Tsipras, Logan Engstrom, Brandon Tran, and Aleksander Madry. Adversarial examples are not bugs, they are features. Advances in neural information processing systems, 32, 2019. Henger Li, Chen Wu, Sencun Zhu, and Zizhan Zheng. Learning to backdoor federated learning. ar Xiv preprint ar Xiv:2303.03320, 2023a. Suyi Li, Yong Cheng, Wei Wang, Yang Liu, and Tianjian Chen. Learning to detect malicious clients for robust federated learning. ar Xiv preprint ar Xiv:2002.00211, 2020a. Published in Transactions on Machine Learning Research (11/2024) Xiang Li, Kaixuan Huang, Wenhao Yang, Shusen Wang, and Zhihua Zhang. On the convergence of fedavg on non-iid data. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020. Open Review.net, 2020b. URL https://openreview.net/forum?id= HJx NAn Vt DS. Yijiang Li, Wentian Cai, Ying Gao, Chengming Li, and Xiping Hu. More than encoder: Introducing transformer decoder to upsample. In 2022 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), pp. 1597 1602. IEEE, 2022. Yijiang Li, Xinjiang Wang, Lihe Yang, Litong Feng, Wayne Zhang, and Ying Gao. Diverse cotraining makes strong semi-supervised segmentor. ar Xiv preprint ar Xiv:2308.09281, 2023b. Jierui Lin, Min Du, and Jian Liu. Free-riders in federated learning: Attacks and defenses. ar Xiv preprint ar Xiv:1911.12560, 2019. Tao Liu, Yuhang Zhang, Zhu Feng, Zhiqin Yang, Chen Xu, Dapeng Man, and Wu Yang. Beyond traditional threats: A persistent backdoor attack on federated learning. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 38, pp. 21359 21367, 2024. Yanpei Liu, Xinyun Chen, Chang Liu, and Dawn Song. Delving into transferable adversarial examples and black-box attacks. ar Xiv preprint ar Xiv:1611.02770, 2016. Aleksander Madry, Aleksandar Makelov, Ludwig Schmidt, Dimitris Tsipras, and Adrian Vladu. Towards deep learning models resistant to adversarial attacks. ar Xiv preprint ar Xiv:1706.06083, 2017. Brendan Mc Mahan, Eider Moore, Daniel Ramage, Seth Hampson, and Blaise Aguera y Arcas. Communicationefficient learning of deep networks from decentralized data. In Artificial intelligence and statistics, pp. 1273 1282. PMLR, 2017. Seyed-Mohsen Moosavi-Dezfooli, Alhussein Fawzi, and Pascal Frossard. Deepfool: a simple and accurate method to fool deep neural networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 2574 2582, 2016. Thien Duc Nguyen, Phillip Rieger, Roberta De Viti, Huili Chen, Björn B Brandenburg, Hossein Yalame, Helen Möllering, Hossein Fereidooni, Samuel Marchal, Markus Miettinen, et al. {FLAME}: Taming backdoors in federated learning. In 31st USENIX Security Symposium (USENIX Security 22), pp. 1415 1432, 2022. Thuy Dung Nguyen, Tuan A Nguyen, Anh Tran, Khoa D Doan, and Kok-Seng Wong. Iba: Towards irreversible backdoor attacks in federated learning. Advances in Neural Information Processing Systems, 36, 2024. Nicolas Papernot, Patrick Mc Daniel, Ian Goodfellow, Somesh Jha, Z Berkay Celik, and Ananthram Swami. Practical black-box attacks against machine learning. In Proceedings of the 2017 ACM on Asia conference on computer and communications security, pp. 506 519, 2017. Yu Qiao, Apurba Adhikary, Chaoning Zhang, and Choong Seon Hong. Towards robust federated learning via logits calibration on non-iid data. ar Xiv preprint ar Xiv:2403.02803, 2024. Amirhossein Reisizadeh, Farzan Farnia, Ramtin Pedarsani, and Ali Jadbabaie. Robust federated learning: The case of affine distribution shifts. Advances in Neural Information Processing Systems, 33:21554 21565, 2020a. Amirhossein Reisizadeh, Aryan Mokhtari, Hamed Hassani, Ali Jadbabaie, and Ramtin Pedarsani. Fedpaq: A communication-efficient federated learning method with periodic averaging and quantization. In International conference on artificial intelligence and statistics, pp. 2021 2031. PMLR, 2020b. Virat Shejwalkar and Amir Houmansadr. Manipulating the byzantine: Optimizing model poisoning attacks and defenses for federated learning. In NDSS, 2021. Ziteng Sun, Peter Kairouz, Ananda Theertha Suresh, and H Brendan Mc Mahan. Can you really backdoor federated learning? ar Xiv preprint ar Xiv:1911.07963, 2019. Published in Transactions on Machine Learning Research (11/2024) Christian Szegedy, Wojciech Zaremba, Ilya Sutskever, Joan Bruna, Dumitru Erhan, Ian Goodfellow, and Rob Fergus. Intriguing properties of neural networks. ar Xiv preprint ar Xiv:1312.6199, 2013. Florian Tramèr, Nicolas Papernot, Ian Goodfellow, Dan Boneh, and Patrick Mc Daniel. The space of transferable adversarial examples. ar Xiv preprint ar Xiv:1704.03453, 2017. Hongyi Wang, Mikhail Yurochkin, Yuekai Sun, Dimitris Papailiopoulos, and Yasaman Khazaeni. Federated learning with matched averaging. ar Xiv preprint ar Xiv:2002.06440, 2020. Xinjiang Wang, Xingyi Yang, Shilong Zhang, Yijiang Li, Litong Feng, Shijie Fang, Chengqi Lyu, Kai Chen, and Wayne Zhang. Consistent-teacher: Towards reducing inconsistent pseudo-targets in semi-supervised object detection. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 3240 3249, 2023. Futa Waseda, Sosuke Nishikawa, Trung-Nghia Le, Huy H Nguyen, and Isao Echizen. Closer look at the transferability of adversarial examples: How they fool different models differently. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 1360 1368, 2023. Dongxian Wu, Yisen Wang, Shu-Tao Xia, James Bailey, and Xingjun Ma. Skip connections matter: On the transferability of adversarial examples generated with resnets. ar Xiv preprint ar Xiv:2002.05990, 2020a. Weibin Wu, Yuxin Su, Xixian Chen, Shenglin Zhao, Irwin King, Michael R Lyu, and Yu-Wing Tai. Boosting the transferability of adversarial samples via attention. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 1161 1170, 2020b. Dong Yin, Yudong Chen, Ramchandran Kannan, and Peter Bartlett. Byzantine-robust distributed learning: Towards optimal statistical rates. In International Conference on Machine Learning, pp. 5650 5659. PMLR, 2018. Mikhail Yurochkin, Mayank Agarwal, Soumya Ghosh, Kristjan Greenewald, Nghia Hoang, and Yasaman Khazaeni. Bayesian nonparametric federated learning of neural networks. In International Conference on Machine Learning, pp. 7252 7261. PMLR, 2019. Syed Zawad, Ahsan Ali, Pin-Yu Chen, Ali Anwar, Yi Zhou, Nathalie Baracaldo, Yuan Tian, and Feng Yan. Curse or redemption? how data heterogeneity affects the robustness of federated learning. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pp. 10807 10814, 2021. Dun Zeng, Siqi Liang, Xiangjing Hu, and Zenglin Xu. Fedlab: A flexible federated learning framework. ar Xiv preprint ar Xiv:2107.11621, 2021. Yechao Zhang, Shengshan Hu, Leo Yu Zhang, Junyu Shi, Minghui Li, Xiaogeng Liu, and Hai Jin. Why does little robustness help? a further step towards understanding adversarial transferability. In Proceedings of the 45th IEEE Symposium on Security and Privacy (S&P 24), volume 2, 2024. Zaixi Zhang, Qi Liu, Zhicai Wang, Zepu Lu, and Qingyong Hu. Backdoor defense via deconfounded representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 12228 12238, 2023. Wen Zhou, Xin Hou, Yongjun Chen, Mengyun Tang, Xiangqi Huang, Xiang Gan, and Yong Yang. Transferable adversarial perturbations. In Proceedings of the European Conference on Computer Vision (ECCV), pp. 452 467, 2018. Yao Zhou, Jun Wu, Haixun Wang, and Jingrui He. Adversarial robustness through bias variance decomposition: A new perspective for federated learning. ar Xiv preprint ar Xiv:2009.09026, 2020. Ligeng Zhu, Zhijian Liu, and Song Han. Deep leakage from gradients. Advances in neural information processing systems, 32, 2019. Giulio Zizzo, Ambrish Rawat, Mathieu Sinn, and Beat Buesser. Fat: Federated adversarial training. ar Xiv preprint ar Xiv:2012.01791, 2020. Published in Transactions on Machine Learning Research (11/2024) A Attack with Same or Different Partition We show in Section 4.3 that training surrogate models in a federated manner can lead to even higher transferability of adversarial examples. However, in the experiment in Section 4.3, we partition the collected data from malicious clients as the federated model and train the source model in a federated manner. In this section, we further evaluate whether this partition information is crucial to the higher transferability. That is, whether different partitions used by the source model affect the transferability of its adversarial examples against the target model. To simulate this setting, we first randomly generate two different partitions with 1 0 2 0 3 0 5 0 7 0 9 0 1 0 0 1 0 1 5 2 0 2 5 3 0 3 5 4 0 4 5 5 0 5 5 6 0 6 5 7 0 7 5 8 0 8 5 9 0 n u m b e r o f u s e r s t r a n s f e r a c c n u m b e r o f u s e r s t r a n s f e r r a t e 1 0 2 0 3 0 5 0 7 0 9 0 1 0 0 2 5 % ~ 7 5 % ( d i f f e r e n t p a r t i t i o n ) 2 5 % ~ 7 5 % ( s a m e p a r t i t i o n ) R a n g e w i t h i n 1 . 5 I Q R M e d i a n L i n e M e a n O u t l i e r s Figure 7: Difference between same partition and different partition; Left: transfer accuracy v.s. number of users data leveraged in source model training; Right: transfer rate v.s. number of users data leveraged in source model training distinct random seeds and then perform the source model training and the target model training on the two different partitions. Then transfer attack is performed with the source model against the target model following the above configuration. We repeat the experiment 4 times with different random seeds and report the averaged results in Figure 7. We observe no significant difference between the same partition and the different partition settings. To further validate this observation, we perform a Hypothesis Test on the obtained results with the Paired Sample T-Test and achieve a p-value of .393 meaning that there is no significant difference. This further demonstrates the possibility of attacking a federated learning system through our proposed attacks and illustrates a higher security risk. B Averaging Leads to Adversarial Robustness with Different Aggregation Methods We show in Section 5.2 that the averaging operation contributes to the robustness of the FL model and more clients to average per round leads to higher transfer robustness. We further validate that this observation generalizes to different aggregation methods, i.e. Krum Blanchard et al. (2017), Geometric Mean Yin et al. (2018) and Trimmed Mean Yin et al. (2018). As per Figure 8, we can see that with all three aggregation methods, there are decreasing trends as the number of averaged users per round increases. median krum trimmed mean median krum trimmed mean median krum trimmed mean median krum trimmed mean median krum trimmed mean median krum trimmed mean median krum trimmed mean 1 10 30 5 50 70 100 25%~75% Range within 1.5IQR Median Line Mean Outliers Figure 8: Transfer rate v.s. #clients selected to average in each round with difference aggregation methods. Published in Transactions on Machine Learning Research (11/2024) C Statistical Hypothesis Testing on Spearman correlation coefficient In Section 5.1, we elucidate the correlation between the degree of decentralization and the heterogeneity of training data with the transferability of adversarial examples. The findings denote that, with more decentralized, heterogeneous data, the federated model is more robust to transfer attack. Furthermore, Section 5.2 portrays a discernible inverse relationship between the number of clients and the average transfer success rate, as depicted through box plots, which illustrate the averaging operation leads to better robustness against adversarial examples. To statistically validate these correlations, we perform two-tailed Hypothesis Testing on Spearman correlation coefficient. To conduct Hypothesis Testing for Spearman correlation coefficient on a specified correlation, we first calculate the Spearman correlation coefficient ρ on the two sets of points (e.g., T.Rate and Dirichlet α): ρ = cov(X, Y ) where cov( , ) denotes the covariance and σ( ) represents the standard deviation. To perform the Hypothesis Test, we first have the Null Hypothesis H0 and Alternate Hypothesis Ha: We choose the significance level to be 0.1, which means we reject H0 is p-value is smaller than 0.1. We report the p-value and Spearman correlation coefficient in Table 3. We can see, as reported in Section 5.1 and 5.2, all experiments except the CNN experiments on dirichlet α and unbalance sgm can deomenstrates significant correlation under a significance level of 0.1. More to the point, we report numerous correlations with p-value less than .001 (significant under level of .001). This Hypothesis Testing validated the findings of our investigation. Table 3: Spearman correlation coefficient and p-value Architecture X p-value (two-tailed) Spearman coefficient dirichlet α .006 .59 unbalance sgm .05 -.44 number of user in partition .003 -.63 maximum number of classes (10 users) < .001 0.80 maximum number of classes (100 users) < .001 .76 number of user to average (30 users) < .001 -.71 number of user to average (centralized) < .001 -.79 dirichlet α .76 .074 unbalance sgm .84 .049 number of user in partition < .001 -.83 maximum number of classes (10 users) .009 .45 maximum number of classes (100 users) .005 .67 number of user to average (30 users) < .001 -.77 number of user to average (centralized) < .001 -.83 D Proof of Theorem 6.1 Following assumptions from Li et al. (2020b) and additional assumptions for adversarial transferability, we provide a low bound for R. Assumption 1 f(x; θ)/ x = θT ρ(x, θ) where ρ(x, θ) is an arbitrary function in the forms of ρ(x, θ) : R1 p Rp 1 R. Published in Transactions on Machine Learning Research (11/2024) There are many functions following our standards (e.g., ℓ2-norm regularized linear regression, logistic regression and softmax classifier). For the rest of our discussion, we define l(f(x; θ), y) = f(x; θ)/ x Lemma D.1. With Assumption 1, we have R(x, y, θ, θ ) = θT θ θ 2 θ 2 1 (7) With Lemma D.1, θ can be directly measured by the cosine similarity of the parameters θ and θ . Furthermore, as we need to offer a discussion regarding multiple aspects of the model, such as the loss, the parameters, and the data, we follow the previous convention Li et al. (2020b) to focus on a narrower scope the model family: Assumption 2 Model f is in the form of f(θ) = P x X(xθ)2 Notice that this is not a significant deviation from previous studies Li et al. (2020b) that focus on f(θ) = θT Aθ for detailed investigation of the parameter behaviors. Assumption 3 The covariance of the samples we study is positive semidefinite, i.e., XT X I Assumption 4: The loss function l is L-smooth: for all v and w, l(v) l(w) + (v w)T l(w) + L Assumption 5: The loss function l is µ-strongly convex: for all v and w, l(v) l(w) + (v w)T l(w) + µ 2 v w 2 2. Assumption 6: Let ξk t be sampled from the k-th device s local data uniformly at random in iteration t. The variance of stochastic gradients in each device is bounded by σ2 k: E l(ξk t ; θk t ) l(θk t ) σ2 k Assumption 7: The expected squared norm of stochastic gradients is uniformly bounded, i.e., E l(ξk t ; θk t ) 2 G2 for all k = 1, , N and t = 1, , T 1. Assumption 8: We assume the federated algorithm is Fed Avg. Assume St contains a subset of K indices randomly selected with replacement according to the sampling probabilities p1, , p N. The aggregation step of Fed Avg performs θt 1 k St θk t . θk t denotes the parameters of device k at iteration t. We use Theorem 2 from Li et al. (2020b) as our lemma to prove our Theorem 6.1. Lemma D.2. With assumption 4 to 8 and L, µ, σk and G be defined therein. Let L denote the minimum loss obtained by optimal estimation from the centralized model and l(θ ) denotes the loss of the federated model. Then E l(θ ) L κ γ + T 1 2(B + C) 2 E θ1 θ (8) where L, µ, σk, G is defined in the assumptions, κ = L µ , γ = max{8κ, E} , B = PN k=1 p2 kσ2 k+6LΓ+8(E 1)2G2 K E2G2. θ1 is the parameter after one step update of SGD. θ is the optimal parameter for centralized model. Now, we prove Theorem 6.1. Proof. We first write out l(θ ) L = θ T XT Xθ θT XT Xθ (9) = (θ θ )T XT X(θ θ ) (10) θ 2 θ 2 (11) when XT X is p.s.d. On the other hand, we can write l(θ ) = L + ϵ where ϵ > 0. Due to the construction of our model, we can write θ = (XT X) 1XT p (Xθ )2 + ϵ (12) Published in Transactions on Machine Learning Research (11/2024) by solving a linear system. Thus, we can get θT θ = θT (XT X) 1XT p (Xθ )2 + ϵ (13) θT (XT X) 1XT Xθ (14) = θT θ (15) Thus, by connecting the above terms, we will have R(x, y, θ , θ ) = θ T θ θ 2 θ 2 θT θ l(θ ) L (16) Finally, by substituting Equation 8 to the denominator, we have E[R(x, y, θ , θ )] (γ + T 1)θT θ 2 E θ1 θ 2) (17) = 2µ(γ + T 1)θT θ 4(B + C)κ + µ2γκE θ1 θ 2 (18) E Proof of Theorem 6.5 Proof. We adopt the Bias-Variance decomposition Geman et al. (1992) to provide theoretical justification for the observation in Section 5.2. Let T denote the train set T = (X, Y ) sampled from some data distribution and x denote the adversarial example following some adversarial distribution. First, we give the Bias-Variance decomposition as follows: Ex ,y ET [ y f(x ; T ) 2 2] = Ex ,y[ y f(x ) 2 2] | {z } Bias2 (19) + Ex ,T [ f(x ) f(x ; T ) 2 2] | {z } Variance where f( ; T ) denotes a model f trained on dataset T . f( ) = ET [f( ; T )] is the expected the model over the data distribution of T . By using an average of multiple model f( ; T ), denoted by fn = 1 n Pn i=1 fi( ; Ti) and with fn( ) = ET [ 1 n Pn i=1 fi( ; T )] = f( ), we can show the following: Ex ,y ET [ y fn(x ) 2 2] = Ex ,y[ y fn(x ) 2 2] | {z } Bias2 (20) + Ex ,T [ fn(x ) fn(x ) 2 2] | {z } Variance = Ex ,y[ y f(x ) 2 2] + 1 n Ex ,T [ f(x ) f(x , T ) 2 2] Inequality 7 holds due to that Var( 1 n Pn i=1 fi( ; Ti)) = 1 n2 Var(Pn i=1 fi( ; Ti)) and since each fi is trained independently, 1 n2 Var(Pn i=1 fi( ; Ti)) = 1 n2 Pn i=1 Var(f( ; T )) = 1 n Var(f( ; T )). F Attacking Communication-efficient FL Methods We also explore whether our attack performs well on communication-efficient federated methods such as DAda Quant Hönig et al. (2022) and Fed PAQ Reisizadeh et al. (2020b). Specifically, we compare the standard Fed Avg, Krum, Trimmed Mean and Fed PAQ Reisizadeh et al. (2020b). Published in Transactions on Machine Learning Research (11/2024) 2 0 4 0 6 0 8 0 1 0 0 t r a n s f e r r a t e ( % # m a l i c i o u s c l i e n t s F e d A v g T r i m e d M e a n K r u m F e d P A Q Figure 9: Attacking communication-efficient FL. G Comparison between transfer attack and attack using traces of training. In this section, we explore using the traces of training (checkpoints during the training of the federated model) to attack and compare with the transfer attack presented in the paper. We must emphasize that the former is a different threat model with the capability of keeping traces of the model at different stages of training. That is to say, the attack is a purely white-box one. We present the attack result using checkpoints of different stages of the training as shown in Tab 4. We early stopped the model when the accuracy reaches 90% which is around 75 epochs. Then we use the checkpoints from 20, 40 and 60 epochs to perform the attack. white-box 20e 40e 60e transfer rate 91.48 89.98 91.45 94.66 1% transfer 10% transfer 30% transfer 100% transfer transfer rate 2.49 57.69 83.91 68.94 Table 4: Comparison between using checkpoint (traces of training) attack and transfer attack. H linear regression to visualize the correlation To better demonstrates the correlation between various factors and adversarial transferability, we perform linear regression with hypothesis testing on the experiment results. We plot scatter graph and linear regression line on each of the correlation and corresponding experiment result as shown in the following figures: Published in Transactions on Machine Learning Research (11/2024) , 0 / % * ) 2 1 ' * / / - ) . $ # - - / # ) / # - ! # + / 3 # . % " 0 ' 0 ( * $ , 0 - # . # - . * ) . - " & , 0 - # (a) Res Net50: transfer rate v.s. number of total users in the partition. , 0 / % * ) 2 1 ' * / / - ) . $ # - - / # ) / # - ! # + / 3 # . % " 0 ' 0 ( * $ , 0 - # . # - . * ) . - " & , 0 - # (b) Res Net50: transfer rate v.s. dirichlet alpha. , 0 / % * ) 2 1 ' * / / - ) . $ # - - / # ) / # - ! # + / 3 # . % " 0 ' 0 ( * $ , 0 - # . # - . * ) . - " & , 0 - # (c) Res Net50: transfer rate v.s. unbalance sgm. , 0 / % * ) 2 1 ' * / / - ) . $ # - - / # ) / # - ! # + / 3 # . % " 0 ' 0 ( * $ , 0 - # . # - . * ) . - " & , 0 - # (d) Res Net50: transfer rate v.s. number of users to average per round (source model trained in centralized manner with full training dataset). Figure 10: Linear regression visualization , 0 / % * ) 2 1 ' * / / - ) . $ # - - / # ) / # - ! # + / 3 # . % " 0 ' 0 ( * $ , 0 - # . # - . * ) . - " & , 0 - # Figure 11: Res Net50: transfer rate v.s. number of users to average per round (source model trained in centralized manner with 30 client s data). Published in Transactions on Machine Learning Research (11/2024) , 0 / % * ) 2 1 ' * / / - ) . $ # - - / # ) / # - ! # + / 3 # . % " 0 ' 0 ( * $ , 0 - # . # - . * ) . - " & , 0 - # Figure 12: Res Net50: transfer rate v.s. maximum number of classes per user (10 users). , 0 / % * ) 2 1 ' * / / - ) . $ # - - / # ) / # - ! # + / 3 # . % " 0 ' 0 ( * $ , 0 - # . # - . * ) . - " & , 0 - # Figure 13: Res Net50: transfer rate v.s. the maximum number of classes per user (100 users). , 0 / % * ) 2 1 ' * / / - ) . $ # - - / # ) / # - ! # + / 3 # . % " 0 ' 0 ( * $ , 0 - # . # - . * ) . - " & , 0 - # Figure 14: CNN: transfer rate v.s. the number of total users in the partition. Published in Transactions on Machine Learning Research (11/2024) , 0 / % * ) 2 1 ' * / / - ) . $ # - - / # ) / # - ! # + / 3 # . % " 0 ' 0 ( * $ , 0 - # . # - . * ) . - " & , 0 - # Figure 15: CNN: transfer rate v.s. dirichlet alpha. , 0 / % * ) 2 1 ' * / / - ) . $ # - - / # ) / # - ! # + / 3 # . % " 0 ' 0 ( * $ , 0 - # . # - . * ) . - " & , 0 - # Figure 16: CNN: transfer rate v.s. unbalance sgm. ) - , " ' & / . $ ' , , * & + ! * * , & , * ( , 0 + " - $ - % ' ! ) - * + * + ' & + * Figure 17: CNN: transfer rate v.s. number of users to average per round (source model trained in centralized manner with full training dataset). Published in Transactions on Machine Learning Research (11/2024) ) - , " ' & / . $ ' , , * & + ! * * , & , * ( , 0 + " - $ - % ' ! ) - * + * + ' & + * Figure 18: CNN: transfer rate v.s. number of users to average per round (source model trained in centralized manner with 30 client s data). , 0 / % * ) 2 1 ' * / / - ) . $ # - - / # ) / # - ! # + / 3 # . % " 0 ' 0 ( * $ , 0 - # . # - . * ) . - " & , 0 - # Figure 19: CNN: transfer rate v.s. maximum number of classes per user (10 users). + / . $ ) ( 1 0 & ) . . , ( - # " , , . " ( . " , " * . 2 " - $ ! / & / ' ) # + / , " - " , - ) ( - , ! % + / , " Figure 20: CNN: transfer rate v.s. the maximum number of classes per user (100 users).