# inverting_gradient_attacks_makes_powerful_data_poisoning__70797622.pdf Published in Transactions on Machine Learning Research (12/2025) Inverting Gradient Attacks Makes Powerful Data Poisoning Wassim (Wes) Bouaziz wassim.s.bouaziz@gmail.com Meta, FAIR & CMAP, École polytechnique Nicolas Usunier Work done at Meta, FAIR El-Mahdi El-Mhamdi CMAP, École polytechnique Reviewed on Open Review: https: // openreview. net/ forum? id= Lvy5Mjy Th3 Gradient attacks and data poisoning tamper with the training of machine learning algorithms to maliciously alter them and have been proven to be equivalent in convex settings. The extent of harm these attacks can produce in non-convex settings is still to be determined. Gradient attacks are practical for fewer systems than data poisoning but have been argued to be more harmful since they can be arbitrary, whereas data poisoning reduces the attacker s power to only being able to inject data points to training sets, via e.g. legitimate participation in a collaborative dataset. This raises the question whether the harm made by gradient attacks can be matched by data poisoning in non-convex settings. In this work, we provide a positive answer and show how data poisoning can mimic gradient attacks to perform an availability attack on (non-convex) neural networks. Through gradient inversion, commonly used to reconstruct data points from actual gradients, we show how reconstructing data points out of malicious gradients can be sufficient to perform a range of attacks. This allows us to show, for the first time, a worst-case availability attack on neural networks through data poisoning, degrading the model s performances to random-level through a minority (as low as 1%) of poisoned points. Code available at https://github.com/wesbz/inverting-gradient. 1 Introduction Security in Machine Learning requires considering various attackers with a wide range of capabilities. Better understanding the capabilities of these attackers is crucial to design effective defenses and evaluations. In training time attacks, an attacker may only need to participate in the training procedure and send strategically chosen, yet legitimate-looking participation. We call gradient attacks the cases where the attacker can send gradients, and data poisoning when they send data points. The severity of attacks range from integrity attacks, where the model s reliability or consistency can be altered, to the most severe, availability attacks, where the model is plainly unusable, e.g. due to performances being too low to allow for normal operations. So far, unless the objective function is convex Biggio et al. (2013), availability attacks have only been possible through gradient attacks Blanchard et al. (2017); El-Mhamdi et al. (2018); Baruch et al. (2019), leaving an open question, whether gradient attacks are fundamentally more powerful than data poisoning attacks outside the convex setting. What can be the damage caused by data poisoning in a worst-case scenario, and can it match the damage of gradient attacks? While data poisoning have been considered weaker than gradient attacks, challenging this perception is crucial to develop comprehensive security measures (Section A illustrates real-world scenarios). In this work, we answer this question by providing empirical evidence that poisoning attacks can lead to availability attacks on (non-convex) neural networks, under a worst-case threat model that allows for comparing both gradient and data poisoning attacks. Published in Transactions on Machine Learning Research (12/2025) Attacker s degree of freedom w.r.t. interaction with the model data poisoning at each iter. & unconstrained data poisoning at each iter. & constrained to data domain one time data poisoning & constrained to data domain Attacker s knowledge established availability attack no known availability attack model weights model architecture gradient attack at each iter. & unconstrained omnipotent attacker embedding attack at each iter. & unconstrained traditional data poisoning Standard Robust Distributed Learning Figure 1: Territory of known availability attacks (in red) within a domain of constraints. The closer to the origin, the more constrained is the setting for the attacker to realize an availability attack. : Geiping et al. (2020b); Zhao & Lao (2022); Ning et al. (2021); Huang et al. (2020), : Blanchard et al. (2017); Baruch et al. (2019), : El-Mhamdi et al. (2018), so far only in convex settings : Farhadkhani et al. (2022), A: Availability attacks (Result A in subsection 5.3), B: Influence of the feasible set (Result B in subsection 5.3). An apple-to-apple comparison of data poisoning attacks and gradient attacks is not straightforward because the literature associates them with different threat models offering different levels of knowledge and interventions. Gradient attacks are only considered in Distributed Learning settings with the Byzantine Machine Learning threat model Blanchard et al. (2017); El-Mhamdi et al. (2018); Baruch et al. (2019), where legitimate workers send actual model updates and malicious workers send arbitrary updates. In contrast, data poisoning attacks are usually considered in Centralized Learning settings Biggio et al. (2013); Muñoz-González et al. (2017); Lu et al. (2022) in a threat model that only allows the attack to be crafted once and for all. In order to remove this confounding factor, we consider a threat model in which both attacks can be executed by allowing an attacker to intervene at each iteration, similarly to Algorithm 1 in Steinhardt et al. (2017), may it be gradients or data poisoning (Figure 1). Once this confounding factor removed, the fundamental difference between gradient and data poisoning attacks comes from the limited expressivity of data poisoning compared to the arbitrary gradient attacks. Figure 2: Images of the gradient operator on different sets. Rd is where an attacker can craft unrestricted gradient attacks. θL(hθ(X), Y) is the set of possible gradients given an unrestricted data poisoning (Result B in Section 5.3), and θL(hθ(FX ), FY) is the set of possible gradients when data poisoning is restricted to a feasible set FX FY X Y (Result A in Section 5.3). Our approach leverages gradient inversion methods, previously used in privacy attacks in distributed learning to reconstruct training data points from actual gradients they induced. Contrary to privacy attacks, malicious gradients might not be achievable by the gradient operator if participants are expected to send legitimate-looking inputs (Figure 2). We reconstruct data points whose induced gradient can replicate as much as possible a malicious gradient, and show that they can constitute a sufficient data poisoning to achieve an availability attack. In our experiments, we exhibit a successful availability attack on neural network architectures, trained on an image classification task with different optimization algorithms, even when protected by a defense mechanism against gradient attacks. We show that: (1) the additional constraints under which data poisoning attacks operate, compared to gradient attacks, make them overall less effective than plain gradient attacks, and (2) the severity of data poisoning attacks covers the same range as gradient attacks, including availability attacks, even on non-convex neural networks. Published in Transactions on Machine Learning Research (12/2025) Our contributions are the following: We leverage gradient inversion mechanism from privacy attacks to reconstruct data poisoning from existing gradient attacks and perform attacks on various settings; We introduce a worst-case threat model that allows for a fair comparison between gradient attacks and data poisoning attacks, by allowing the attacker to intervene at each iteration; We demonstrate that in our worst-case threat model, data poisoning could perform an availability attack on neural networks degrading them to random chance performances. Paper structure. The rest of the paper is organised as follows. Section 2 reviews related work on the broader effort to understand various types of attacks, Section 3 describes the formal setup , Section 4 provides our solution to invert the most important gradient attacks back in the training data domain, and provides a table of correspondence between gradient attacks and data poisoning attacks, Section 5 implements our solution describing empirical outcomes, finally, Section 6 discusses our work and provides concluding remarks. 2 Background Gradient attacks have long been studied in the standard Robust Distributed Learning literature, when an attacker can send arbitrary1 gradients. Lemma 1 in Blanchard et al. (2017) shows that with stochastic gradient descent (SGD), availability attacks are possible with only a single malicious worker. The attacker s freedom allows for a variety of attacks by exploiting the geometry of honest gradients and the inevitable weaknesses El-Mhamdi et al. (2018); Baruch et al. (2019); Xie et al. (2020); Shejwalkar & Houmansadr (2021) of robust gradient aggregators when the model is high-dimensional or when the data is heterogeneous. The attacker s goal is to bring the average gradient to another direction, making a step toward their objective. There is no naive way to determine if a gradient is legitimate or not. In such settings, defending is often considered as a task of robust mean estimation. Robust aggregation of gradients is hence a possible defense mechanisms one can deploy Blanchard et al. (2017); El-Mhamdi et al. (2018); Yin et al. (2018). Still, with better defense mechanisms came better attacks Baruch et al. (2019). Even stronger, Theorem 2 in El-Mhamdi et al. (2023) shows the impossibility of robust mean estimation below a certain threshold that grows with data heterogeneity and model size. Our work relies on these gradient attacks and shows how they can be transferred to data poisoning as successful availability attacks, even in non-convex settings. Data poisoning is the manipulation of training data of ML algorithms with the goal of influencing the algorithms behavior. Several approaches exist to generate poisons: label flipping Shejwalkar et al. (2022), generative methods Muñoz-González et al. (2019); Zhao & Lao (2022), and gradient-based approaches Muñoz-González et al. (2017); Shafahi et al. (2018); Geiping et al. (2020b). The last allows to finely control the resulting gradient on the poisonous points instead of relying on another proxy. Although Shejwalkar et al. (2022); Shejwalkar & Houmansadr (2021) consider data poisoning to be of limited harm and gradient-based approach to be too computationally intensive, clean-label attacks based on gradient matching have shown to be both stealth and effective when performing a targeted integrity attack Geiping et al. (2020b). It was also demonstrated in practical settings such as state of the art image classifiers Bouaziz et al. (2025a), audio classification Bouaziz et al. (2025c), and even language modeling Bouaziz et al. (2025b). To the best of our knowledge, our work is the first to achieve an availability attack on a neural network using data poisoning. Availability attacks have been demonstrated in a range of settings. As shown in Figure 1, we can compare these settings in term of attacker knowledge (from black box to omniscient) and degree of freedom (from a single constrained interaction to an omnipotent attacker). The data poisoning literature on neural networks only allows an attacker to craft a poison once to insert it in the training set, with various levels of knowledge. Geiping et al. (2020b) operate both in black-box setting and with sole model architecture knowledge. Muñoz-González et al. (2019) assume an omniscient attacker and Muñoz-González et al. (2017) add scenarios with attackers not having access to the model weights or to the training data. Farhadkhani 1Often called Byzantine attacks, in reference to the Byzantine faults model in distributed computing Lamport et al. (1982) Published in Transactions on Machine Learning Research (12/2025) et al. (2022) assume an attacker who can interfere at the embedding level of the last layer of a neural network. This particular case is equivalent to attacking a logistic regression which is a convex setting for which an equivalence between data poisoning and gradient attacks holds. Previous works in data poisoning on neural networks were only able to slightly decrease the performance of the attacked algorithm and have yet to demonstrate a complete availability attack that bring a model down to chance-level Zhao & Lao (2022); Lu et al. (2022). Gradient attacks, on the other hand, have established effective ways of making a model utterly useless. Blanchard et al. (2017) show how an omniscient attacker sending unconstrained gradients at each iteration can arbitrarily change the model s weights. Baruch et al. (2019) assume an attacker with access to model s weights and a fraction of the training set (similar to our auxiliary dataset). El-Mhamdi et al. (2020) suppose an omniscient attacker or one that does not know the legitimate gradients. Finally, Liu et al. (2023) claim a data poisoning availability attack, but they not only require a high poisoning ratio (at least 80%) but also the degradation of performances is not as severe as in gradient attacks (Figure 3 in Liu et al. (2023)). In this work, we extend the domain of known availability attacks and demonstrate how they are possible in settings were the attacker knows the model weights and can craft a data poisoning at each iteration in a constrained set or not, similarly to Algorithm 1 in Steinhardt et al. (2017) (Result A and B in subsection 5.3). Defenses against data poisoning have been studied through different approaches such as data sanitization Steinhardt et al. (2017), data augmentation Borgnia et al. (2021), bagging Wang et al. (2022) or pruning and fine-tuning Liu et al. (2018). However, the effectiveness of such defenses rely on strong assumptions such as the learner having access to a clean dataset, on the convexity of the loss w.r.t. the model s parameters or the learner s ability to train a very large number of models. And still, attackers could find ways to break these defenses Koh et al. (2022). Even if a theoretically sound and impenetrable defense mechanism against data poisoning might be impossible El-Mhamdi et al. (2023); Hardt et al. (2023), making the attacker s job harder by adding several imperfect yet constraining defenses in a Swiss cheese 2-like model is still necessary. Inverting gradients has recently been studied in privacy attacks to reconstruct training samples based on the resulting gradient Geiping et al. (2020a); Zhao et al. (2020). Their central recovery mechanism relies on maximizing a similarity measure Sim between a targeted gradient g(tgt) that has been computed on the data which the attacker wants to reconstruct and the gradients G(rec) = { θt L(hθt(x), y)}(x,y) S computed on nrec reconstructed samples S and aggregated through the Agg function: arg max S (X Y)nrec Sim(g(tgt), Agg(G(rec))) (1) Similarly, we try to recover data points that induce the closest gradient possible to a malicious gradient. The existence of a solution for (1) in the privacy attack setting is known, since actual data points have given rise to the targeted gradient g(tgt). More surprisingly, several works Geiping et al. (2020b); Bouaziz et al. (2025a) have shown that it is possible to construct data points whose gradients match a targeted sample s gradient. On the contrary, gradient attacks are not actually calculated from a data point hence there is no guarantee that the inversion can find an existing solution. 3 Framework 3.1 Learning Setting We consider a classification setting where the model is trained on a dataset Dtrain = {(xi, yi)}n i=1 sampled from a distribution D over X Y. The learner trains a neural network hθ parametrized by θ Rd with an iterative optimization algorithm on the (non-convex) loss function L. Its goal is to achieve the lowest test loss on a heldout test set Dtest that is not necessarily sampled from the same distribution D as the training set. We formulate the objective as the following optimization problem: arg min θ Θ 1 ntest (x,y) Dtest L(hθ(x), y) 2a cybersecurity defense model, see https://en.wikipedia.org/wiki/Swiss_cheese_model Published in Transactions on Machine Learning Research (12/2025) We consider that learning occurs through a set of nb Gradient Generation Units {Vi}nb i=1 each of which reports a message in a set Sb t = Message(Dtrain, t) = {vi,t}nb i=1 at each iteration t. Messages are then aggregated through an aggregator Agg and the model weights are updated using the Update algorithm: θt+1 = Update(θt, Agg, Sb t ) This abstraction allows us to represent a large spectrum of learning settings, from centralized learning to fully distributed learning and settings in between (such as federated learning). In the common centralized setting, when training a neural network, the Message operator returns a batch of data points sampled from the training set Sb t = {(xi,t, yi,t)}nb i=1, the aggregator Agg is the Average of their gradients, computed by the learner and the update algorithm is SGD or Adam. In the federated learning setting, the Message operator returns a batch of gradients (or equivalently model updates) that each worker computed separately, the default aggregator is the Average of the messages and the update algorithm is Federated Averaging Mc Mahan et al. (2017) or Local SGD Stich (2018). This allows us to study a common framework for both gradient attacks and data poisoning attacks, normally operating on different settings, characterized by different Message operators. The choice for the Message operator boils down to entrusting the calculation of gradients to the workers or not. This, in turn, gives different degrees of freedom for a malicious Gradient Generation Unit to influence the training: in the scope of this paper, respectively gradient attacks or data poisoning3. In our experiments, we consider an image classification task on the CIFAR10 dataset on which the attacker can tamper with messages (data point or gradient depending on the learning setting). We use the SGD and Adam update algorithms as well as Average and Multi Krum Blanchard et al. (2017) aggregators. Multi Krumf<0.5 is a robust aggregator which can withstand a ratio of malicious elements up to f and is defined for a set of vectors {vi}n i=1 as the average of the n(1 f) 2 vectors minimizing the score function s(i) = P i j vi vj 2, with i j the indices of the n(1 f) 2 closest vectors to vi. The selection of the update algorithm and aggregator is crucial, as both are regarded as defense mechanisms in the Robust Distributed Learning literature El-Mhamdi et al. (2020). 3.2 Threat model In order to study the expressivity of data poisoning in non-convex settings, we consider a worst-case threat model that allows for a fair comparison with gradient attacks. Here, the attacker: has knowledge of the weights of the model θt, the update algorithm Update, the aggregator Agg function and the message operator Message as in, Farhadkhani et al. (2022); Blanchard et al. (2017); Yin et al. (2018); Shafahi et al. (2018); does not have access to the batch Sb t , unlike the stronger standard assumption of omniscient attacker in the robust distributed learning literature Blanchard et al. (2017); El-Mhamdi et al. (2020); Chen et al. (2017); Yin et al. (2018); has access to an auxiliary dataset Da D that is a surrogate to the unobservable training set. This is a standard assumption when not assuming omniscient attackers, as in El-Mhamdi et al. (2018); Farhadkhani et al. (2022); El-Mhamdi et al. (2020); has the control over a ratio α of Gradient Generation Units and the ability to append a set of arbitrarily crafted poisoned messages (gradients or data points) Sp t = {vp i,t}np i=1 to the clean batch Sb t at each iteration t up to a level α of contamination (i.e. |Sp t |/|Sb p t | = α), as in Blanchard et al. 3The notion of Gradient Generation Units can be used as an abstraction for any source of input that is itself a gradient, a model update, or that can be later transformed to a gradient, such as a training data point, a machine in distributed learning or a user account in a social media platform, depending on the level of granularity that is considered. Published in Transactions on Machine Learning Research (12/2025) Figure 3: Threat model. The attacker has access to θ but does not have access to the batch Sb t and uses an auxiliary dataset Da to craft Sp t the set of poisoned messages. Both the batch and the poisons set are gathered into Sb p t . The attacker s goal is either to slow down the training or attack the model s availability. (2017); El-Mhamdi et al. (2018); Yin et al. (2018) for gradients attacks, in Steinhardt et al. (2017) for data poisoning or in Farhadkhani et al. (2022) for both gradients attacks and data poisoning; constrains itself4 to only crafting poison in a feasible domain Sp t Fnp, depending on the task and data structure. For instance, for an image classification task of H W RGB pixels encoded on 3 bytes and labels between 1 and C, we set F to be the set of possible such sized images and labels: F = [0..255]H W 3 [1..C] as in Farhadkhani et al. (2022), but without restricting ourselves to convex settings. The attacker s goal is to perform an availability attack: degrading the learner s performances as much as possible. We also consider slow down attacks, in which the attacker tries to stall the learning procedure as much as possible. This attack is important, given the cost for training large models and the financial repercussion of slowing it down. This threat model is inspired from gradient attacks and assumes much more communication between the attacker and the model than traditional data poisoning threat models. The latter leads to poor availability attacks performance. We hypothesized that once we allowed for the same amount of interactions between gradient attacks and data poisoning, the two methods would display far less discrepancy of destructive power than what is currently thought. At each iteration t, the attacker, controlling p Gradient Generation Units, computes an auxiliary batch Sa t = Message(Da, t) and constructs Sp t such that Update(θt, Agg, Sa t Sp t ) gives poorer performance on Da than Update(θt, Agg, Sa t ). This should, in turn, also decrease the model s performances on Dtrain under the assumption it follows the same distribution as Da. 4.1 Gradient attacks When the Message operator outputs gradients, a malicious Gradient Generation Units can participate in Sb t with a gradient attack in Rd. At each iteration t, the attacker uses the auxiliary dataset Da as a surrogate 4Result A in subsection 5.3 is our strongest result and is obtained under this constraint on the attacker. Result B is 5.3 is obtained without this constraint and serves to further understand the expressivity of data poisoning from inverted gradient attacks. Published in Transactions on Machine Learning Research (12/2025) for the batch Sb t . Let Sa t = {ga t,i}na i=1 = { θt L(hθt(x), y)}(x,y) Da be the set of per-sample auxiliary gradients and ga t = 1 na P g Sa t g the averaged auxiliary gradient. Similarly, gb t = 1 nb P g Sb t g denotes the averaged clean batch gradient. The attacker constructs a set of np attacking gradients Gatk t based on Sa t and sends them as message Sp t . We denote Sb p t = Sb t Sp t the set of poisoned batch of gradients, Sa p t = Sa t Sp t the set of poisoned auxiliary gradients and gb p t and ga p t their respective averages. We consider several gradient attacks to perform an availability attack. Gradient Ascent (GA) Blanchard et al. (2017): the attacker sends a gradient such that the averaged poisoned gradient is anti-collinear with the mean clean gradients. This provokes a gradient ascent step with the SGD update rule, Average, and λ > 0: ESb t [θt+1] = θt ηESb t [gb p t ] = θt + ηλESb t [gb t] Orthogonal Gradient (OG): the attacker sends a gradient such that the averaged poisoned gradient is orthogonal to the mean clean gradient. This should stall the training. Little is Enough (LIE) Baruch et al. (2019); Shejwalkar & Houmansadr (2021): the attacker sends the mean clean gradient deviated by a strategically chosen amount times the coordinate-wise standard deviation of the clean gradients, with σ[j] = q V ar({g[j]}g Sa t ). gatk t = ga t zmaxσ (2) where zmax arg max z R + Aa p t Aa t , and Aa p t = Agg(Sa t {ga t zσ}np i=1) Aa t = Agg(Sa t ) This attack has been shown to be effective against Multi Krum aggregator. Note that contrary to Baruch et al. (2019) and similarly to Shejwalkar & Houmansadr (2021), we use an adaptive approach and choose zmax to maximize the divergence of the poisoned gradients on the aggregation Aa p t . 4.2 Data poisoning When the Message operator outputs data samples, malicious Gradient Generation Units are expected to participate via similarly structured messages, i.e. data poisoning. In our experiments, on CIFAR10 and its 32 32 RGB images, it leaves room for an attacker to participate in the training process via messages crafted in [0, 255]32 32 3. This gives only 32 32 3 = 3, 072 degrees of freedom, far less than commonly used deep learning models number of parameter, hinting that data poisoning should be less expressive than gradient attacks. Models non-linearities further constrain the dimension of the image space of the gradient operator (as illustrated in Figure 2). Data poisoning is allegedly harder as it constitutes a far more constrained problem than gradient attacks (in which the attacker can send an arbitrary gradient from Rd). At each iteration t, the attacker computes a given gradient attack gatk t in the same manner as above and inverts the gradient operator to compute an associated set of data points Sp t = {(xp i,t, yp i,t)}np i=1 such that 1 np Pnp i=1 θt L(hθt(xp i,t), yp i,t) = gatk t . That set of data points is then sent as message Sp t and appended to Sb t , like in the gradient attack case. Similarly to what is done in privacy attacks, we optimize a similarity measure between the gradient attack and the gradients calculated on Sp t . Contrary to privacy attacks where the existence of a solution is known5, there is no guarantee that the inversion can find an existing solution, let alone an effective data poison. We can formulate our data poisoning as an optimization problem where the attacker aims to minimize a poisoning function fp. fp characterizes the dissimilarity of the data poison gradients Gp t = { θt L(hθt(x), y)}(x,y) Sp t 5In privacy attacks, the gradient being inverted was produced by the data point that the attacker is trying to retrieve. Published in Transactions on Machine Learning Research (12/2025) with a gradient attack gatk t calculated from the auxiliary dataset gradients Ga t = { θt L(hθt(x), y)}(x,y) Sa t (since the attacker cannot have access to the batch Sb t ) over a feasible domain F: Sp t arg min S Fnp fp(hθt, Ga t , S) (3) Since characterizing the image space of the gradient operator on the loss function L is a difficult task, we cannot know beforehand if a vector can have an antecedent data point through the gradient operator. Together with constraints on feasibility, finding data poisons which exactly reproduce the gradient attack might be impossible. Since I F = θt L(hθt(FX ), FY) θt L(hθt(X), Y) Rd, the set I F might not cover all the possible candidates for an effective gradient attack, if not any (Figure 2). Thus, achieving an effective data poisoning that performs similarly to a gradient attack is non trivial. We show in our experiments that an attacker can still produce data poisons which have significant impact on the training procedure. Poisons are iteratively updated to minimize the poisoning objective fp using the Adam optimizer and are projected on the feasible set F at each iteration of the poisoning optimization by clipping. Table 1 details the formulas used to determine the gradient attacks and their equivalent poisoning functions in the data poisoning case. Table 1: Gradient attacks formula and their equivalent data poisoning objective functions to be optimized following Equation 3. gp, ga, ga p are respectively the averaged gradients computed on the data poisons, on the auxiliary dataset, and the weighted average between them. cos is the cosine similarity. zmax and σ are defined as in eq. 2. Gradient attack gp Rd s.t. Our data poisoning attack fp Gradient Ascent cos(ga p, ga) = 1 cos (ga p, ga) Orthogonal Gradient ga p, ga = 0 cos (ga p, ga) 2 Little is Enough gp = ga zmax σ gp ga + zmaxσ 2 5 Experiments 5.1 Preliminary experiment To give a better intuition of the capabilities of data poisoning, we present a preliminary experiment on the XOR operator classification task. Sampling a point x in [0, 1]2, its label is y = 1{x[0] > 0.5} 1{x[1] > 0.5} with 1 the indicator function and the XOR operator. We generate a set of possible poisons by regularly sampling on the [0, 1]2 grid and labeling them with XOR then flipping their labels. After training a multilayer perceptron with Gradient Descent (i.e. full batch) on 1000 data points, we perform a single step of gradient descent on the data poisoned with one of the possible poisons, repeated as to reach a contamination level α. The resulting test accuracies obtained for all poisons are illustrated in Figure 4. While a single step at α = 0.005 is not enough to reduce the model accuracy down to random-chance, several steps of poisoned update will do. A 10% level of contamination is already enough to bring the model close to the level of a random guess, which suggests that a single step on poisoned data is enough to obtain an availability attack on neural network for a task such XOR. This proof of concept could however be argued to require a high contamination ratio. We exhibit experimental evidences that lower contamination ratio can actually be effective on, e.g. an image classification setting. 5.2 Experimental setup Model & dataset We demonstrate our poisoning procedure on a custom convolutional neural network (described in Table 5 in Appendix B) and on Vision Transformers models (Vi T-tiny models with patch size 8) trained for 50 epochs on the CIFAR10 dataset partitioned in training, validation, and auxiliary datasets. We use different optimization algorithms and aggregation rules to train the models: SGD & Average, Published in Transactions on Machine Learning Research (12/2025) 0.0 0.5 1.0 x1 0.0 0.5 1.0 x1 0.000 0.025 0.050 0.075 0.100 Minimum accuracy Min. poisoned accuracy Figure 4: Landscape of accuracies after 1 poisoned step with the respective poison with two levels of contamination α = 0.005 (left) and α = 0.1 (center). The black squares represent the border of the feasible domain F. Right: The minimum accuracy in the landscape for different levels of contamination α. Adam & Average, SGD & Multi Krum (with different levels of data truncation f {0.1, 0.2, 0.4}). Since Multi Krum is used as a defense mechanism, we expect it to be robust to the attacks operating within its working assumptions. However, the Little is Enough attack was specifically designed to circumvent these assumptions when using gradient attacks. We thus expect Little is Enough to bypass Multi Krum in the gradient attacks situation. Baseline In every setting, the learning rate is fixed to the value were the learner achieves the best performances without any poisoning to set a baseline for the performances of the model. We then measure the decay in performances caused by an attack. Each setting of this experiment is run 4 times for better statistical significance. Each run has a different model and poisons initialization, and dataset split. Full results can be seen in the Appendix. Attacks In every setting, we perform either one of the considered attacks either via a gradient attack, or a data poisoning attack as specified in section 4.2. The np crafted poisons are added to the batch of size nb at each iteration so that np nb+np = α [0.01, 0.48]. Since the gradients induced by the poisoned data mimic the malicious gradients, we expect our data poisoning is at best as good as the associated gradient attack. Gradient attacks should thus be a topline for our data poisoning attack. 5.3 Results 5.3.1 Gradient attacks Table 2 shows that gradient attacks do perform an availability attack, bringing the models performances down to random-level. We also notice that the Multi Krum aggregation rule does act as a defense mechanism, for levels of contamination lower than its tolerance parameter f. However, as expected, this defense is well circumvented by the Little is Enough attack. 5.3.2 Data poisoning After performing the equivalent data poisoning attacks, the observed effects range from a slowdown of the training procedure to its complete halt and up to degrading the performances down to random levels. Figure 5a compares the different attacks with a random data poisoning (uniformly sampled in [0, 1]) at α = 0.01 with its baseline counterpart. While the Gradient Ascent and Orthogonal attacks behave similarly and at most only slightly slow down the training at most, Little is Enough attack strongly degrades performances, even at such low levels of contamination. Figure 5b shows for the Gradient Ascent and Little is Enough attacks that the higher the level of contamination, the stronger the effect: validation accuracy increases slower or plummets earlier. Therefore, our main result is the following. Availability attacks (Result A). To fairly compare each setting under that attack, each model has been evaluated on the test set with the weights achieving the best validation accuracy. Figure 6 shows that for Published in Transactions on Machine Learning Research (12/2025) Table 2: Best validation accuracy under different attacks with different update rules, different aggregation functions and different levels of contamination α. A high validation accuracy (colored in apricot) indicates a failed attack. A low validation accuracy (colored in pale green) indicates a successful attack. Update; Agg Attack α 0.01 0.05 0.10 0.20 0.30 0.40 0.48 Adam; Average GA 10.0 9.8 10.0 10.0 10.0 10.1 10.0 OG 9.9 9.9 10.1 10.1 9.8 10.0 10.1 LIE 10.2 10.3 10.1 10.2 10.1 10.1 10.0 SGD; Average GA 10.1 10.1 10.1 9.9 10.0 10.0 10.0 OG 10.2 10.2 10.0 9.9 10.1 10.1 10.2 LIE 10.2 9.9 10.2 9.9 9.9 10.0 10.0 SGD; Multi Krumf=0.1 GA 65.1 63.2 33.7 10.1 10.1 10.2 10.2 OG 65.1 65.3 65.9 10.0 10.1 10.2 9.9 LIE 41.7 15.0 10.4 9.9 10.0 9.9 10.1 (a) Comparison of different attacks at α = 0.01. (b) Comparison of different levels of contamination for the Gradient Ascent & Little is Enough attacks. Figure 5: Validation accuracies during training in the SGD & Average setting under different attacks and different level α of contamination. Error bars represent the standard error. the SGD & Average learner under the Little is Enough attack, α = 0.01 is enough to significantly degrade the model s performance. Among the 4 runs, 3 ended up diverging while every poison was in the feasible set F. Figure 9 in the Appendix shows 50 of the first 500 poisons crafted by the attacker in one of the diverging run. Similarly, in the SGD & Multi Krumf=0.1 setting, the Little is Enough attack with contamination level α = 0.05 finds data poisons that circumvent the robust aggregation rule and drastically reduce the model s performance. Table 3: Epoch of best validation accuracy (validation accuracies in parenthesis) for the CNN model with Adam optimizer and Average under different data poisoning attacks. Higher levels of contamination induce slower training and lower performances. α GA OG LIE 0 4 (67.1) 4 (67.1) 4 (67.1) 0.01 5 (66.4) 5 (66.4) 6 (66.3) 0.05 7 (66.5) 7 (65.3) 7 (66.3) 0.1 19 (63.6) 14 (63.9) 10 (65.7) 0.2 34 (61.2) 17 (62.5) 22 (62.1) Published in Transactions on Machine Learning Research (12/2025) Table 4: Poison selection rates in the SGD & Multi Krumf=0.1 averaged over all runs and all epochs. Each row corresponds to a different attack. Standard deviations in parenthesis. α 0.01 0.05 0.1 0.2 GA 0.0 2 10 4 2 10 3 0.91 (0.0) (4 10 4) (2 10 3) (0.08) OG 0.0 3 10 4 0.18 0.85 (0.0) (1 10 3) (0.1) (0.05) LIE 0.92 0.91 0.97 0.87 (0.07) (0.13) (0.02) (0.06) Comparison of update rules. Figure 6 shows that our data poisoning procedure failed at performing an availability attack against the Adam update rule. Because Adam normalizes the aggregated gradients, the training cannot diverge as abruptly as with SGD. However, the slow down attack can still be observed (e.g. in Figure 8 in Appendix B), meaning that the data poisoning procedure could find a solution that perturbs the total gradient enough to slow down convergence, but not enough to completely halt it. Table 3 shows that higher levels of contamination lead to a lower best validation accuracy but most importantly to a higher number of epochs (hence a longer time) before reaching it. Figure 6: Test accuracy of the CNN model which achieved the best validation accuracy under the Little is Enough attack. Each column represents a different setting of update function and aggregation rule. The blue line is the test accuracy obtained with (SGD, Average) setting and no attack. Error bars are the standard errors. Data poisoning against a robust aggregation rule. Table 4 shows that with the Multi Krumf=0.1 aggregation rule, the attacker still manages to have some of its messages not filtered out. For α > f, the aggregation rule does not play a defensive role anymore as per its functioning conditions. On the other hand, for α below this point, the Little is Enough attack displays significantly higher selection rates than the other attacks. This means that the attacker manages to produce data poisons whose gradients deceive Multi Krumf=0.1, similarly to the gradient version of the attack which is particularly designed for this purpose. However Figure 6 shows that a higher selection rate does not necessarily mean success of attack. Even if the attacker sometimes manages to successfully attack the model, Multi Krum overall enhances the robustness of the model while slightly degrading its performances. Influence of the feasible set (Result B). As the feasible set FX changes, the attacker will only be allowed to converge (in case of convergence) towards different poisons. We compare three increasingly restrictive feasible sets to determine their influence on the success of the attack: Constraint-free set: the feasible set is simply the input domain FXfree = RH W 3; Image-encoding set: the feasible set ensures that the poisons respect the same encoding as the clean data FXimg = [0..255]H W 3; and Neighborhood set: this is a subset of the previous one and is composed of all the images that are at a L1 norm of at most ϵ = 32 255 of an actual image FXnei = {x FXimg/ xa Das.t. x xa 1 ϵ}. Published in Transactions on Machine Learning Research (12/2025) Figure 7: Final test accuracies for the SGD; Average setting under the Little is Enough attack for different feasible sets. Error bars are the standard errors. Figure 7 shows that the more constrained the feasible set, the less effective the resulting attack. It is important to note, however, that early stopping helps in limiting the effects of the attack (Figure 10, Appendix B). Neural Network architecture. We perform the same attacks on the same learning pipelines replacing the CNN with a Visual Transformer (Vi T, Dosovitskiy et al. (2020)) tiny with patch size 8. Since Vi Ts are predominantly trained using Adam-like approaches, we report the efficiency of our attacks using it. Figure 11 in the Appendix shows that Vi Ts are also vulnerable to our data poisoning, although with less success and on higher levels of contamination. 6 Discussion Our work shows that even in training settings involving deep neural networks, which are not restricted to convex cases as in Farhadkhani et al. (2022), inverting malicious gradients can result in an effective data poisoning, achieving an availability attack, and mimicking the harm of gradient attacks. To compare both type of attacks, our threat model uses assumptions that differ from the common threat model in data poisoning, making it an impractical threat model for now. These assumptions should be further explored in future work on at least four different fronts. (1) The role of the auxiliary dataset. Further experiments should consider an auxiliary dataset with a distribution different than the training set or no auxiliary dataset at all. (2) Accessing the trained model s weights. Performing an attack when the attacker estimates the victim s model with a surrogate model would make the threat model more practical. (3) The role of the feasible set. Clean label poisoning (Geiping et al., 2020b; Shafahi et al., 2018) shows that it is possible to design data poisons which deceive human annotators by resembling legitimate images. Future work should consider exploring more constraining feasible sets for the attacker to reach such level of stealthiness. (4) Number of interventions. In our threat model, similarly to Farhadkhani et al. (2022) the attacker is allowed to craft and inject malicious data at each iteration, which limits the practicality of our threat model, future work should explore data poisoning availability attacks limiting also the number of times the attacker can craft its poisons (the common data poisoning threat model only allows a single intervention). 7 Conclusion Our work asks whether or not gradient attacks were fundamentally more effective than data poisoning. Our experiments yield a nuanced response. On one hand, inverting malicious gradients sometimes results in a devastating data poisoning, and our results are the first to show the feasibility of a total availability attack on neural network via data poisoning. On the other hand, the success rate of our data poisoning attacks is lower than for their gradient attack counterparts, but the mere possibility of mimicking gradient attacks with feasible data poisoning should motivate further research in defense mechanisms for attacks that evade today s state of the art defense in robust machine learning. Published in Transactions on Machine Learning Research (12/2025) 8 Broader Impact Statement This paper presents work whose goal is to advance the field of Robust Machine Learning. There are many potential societal consequences of our work, all of which fall under the usual considerations to take into account when considering tools that can facilitate data poisoning or make it more potent. In particular, our technique of inverting gradients aims at showing that availability attacks which were believed to be specific to gradient attacks except for convex settings are doable with data poisoning alone, and in small amount. Gilad Baruch, Moran Baruch, and Yoav Goldberg. A little is enough: Circumventing defenses for distributed learning. Advances in Neural Information Processing Systems, 32, 2019. URL https://arxiv.org/abs/ 1902.06156. Battista Biggio, Blaine Nelson, and Pavel Laskov. Poisoning attacks against support vector machines, 2013. Peva Blanchard, El Mahdi El Mhamdi, Rachid Guerraoui, and Julien Stainer. Machine learning with adversaries: Byzantine tolerant gradient descent. Advances in neural information processing systems, 30, 2017. Eitan Borgnia, Valeriia Cherepanova, Liam Fowl, Amin Ghiasi, Jonas Geiping, Micah Goldblum, Tom Goldstein, and Arjun Gupta. Strong data augmentation sanitizes poisoning and backdoor attacks without an accuracy tradeoff. In ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 3855 3859. IEEE, 2021. Wassim Bouaziz, Nicolas Usunier, and El-Mahdi El-Mhamdi. Data taggants: Dataset ownership verification via harmless targeted data poisoning. In ICLR 2025 The Thirteenth International Conference on Learning Representations, 2025a. Wassim Bouaziz, Mathurin Videau, Nicolas Usunier, and El-Mahdi El-Mhamdi. Winter soldier: Backdooring language models at pre-training with indirect data poisoning. ar Xiv preprint ar Xiv:2506.14913, 2025b. Wassim Wes Bouaziz, El-Mahdi El-Mhamdi, and Nicolas Usunier. Targeted data poisoning for black-box audio datasets ownership verification. In ICASSP 2025-2025 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 1 5. IEEE, 2025c. Nicholas Carlini, Matthew Jagielski, Christopher A Choquette-Choo, Daniel Paleka, Will Pearce, Hyrum Anderson, Andreas Terzis, Kurt Thomas, and Florian Tramèr. Poisoning web-scale training datasets is practical. ar Xiv preprint ar Xiv:2302.10149, 2023. URL https://arxiv.org/abs/2302.10149. Paul Charon and Jean-Baptiste Jeangène Vilmer. Chinese influence operations: A machiavellian moment. Report by the Institute for Strategic Research (IRSEM), Paris, Ministry for the Armed Forces, 2021. Yudong Chen, Lili Su, and Jiaming Xu. Distributed statistical machine learning in adversarial settings: Byzantine gradient descent. Proceedings of the ACM on Measurement and Analysis of Computing Systems, 1(2):1 25, 2017. Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. ar Xiv preprint ar Xiv:2010.11929, 2020. El-Mahdi El-Mhamdi, Rachid Guerraoui, and Sébastien Rouault. The hidden vulnerability of distributed learning in byzantium. In International Conference on Machine Learning, pp. 3521 3530. PMLR, 2018. El-Mahdi El-Mhamdi, Rachid Guerraoui, and Sébastien Rouault. Distributed momentum for byzantineresilient learning. ar Xiv preprint ar Xiv:2003.00010, 2020. El-Mahdi El-Mhamdi, Sadegh Farhadkhani, Rachid Guerraoui, Nirupam Gupta, Lê-Nguyên Hoang, Rafael Pinot, Sébastien Rouault, and John Stephan. On the impossible safety of large ai models. 2023. Published in Transactions on Machine Learning Research (12/2025) Sadegh Farhadkhani, Rachid Guerraoui, Oscar Villemaud, et al. An equivalence between data poisoning and byzantine gradient attacks. In International Conference on Machine Learning, pp. 6284 6323. PMLR, 2022. Jonas Geiping, Hartmut Bauermeister, Hannah Dröge, and Michael Moeller. Inverting gradients-how easy is it to break privacy in federated learning? Advances in Neural Information Processing Systems, 33: 16937 16947, 2020a. Jonas Geiping, Liam Fowl, W Ronny Huang, Wojciech Czaja, Gavin Taylor, Michael Moeller, and Tom Goldstein. Witches brew: Industrial scale data poisoning via gradient matching. ar Xiv preprint ar Xiv:2009.02276, 2020b. Moritz Hardt, Eric Mazumdar, Celestine Mendler-Dünner, and Tijana Zrnic. Algorithmic collective action in machine learning. In International Conference on Machine Learning, pp. 12570 12586. PMLR, 2023. Lê-Nguyên Hoang, Louis Faucon, Aidan Jungo, Sergei Volodin, Dalia Papuc, Orfeas Liossatos, Ben Crulis, Mariame Tighanimine, Isabela Constantin, Anastasiia Kucherenko, et al. Tournesol: A quest for a large, secure and trustworthy database of reliable human judgments. ar Xiv preprint ar Xiv:2107.07334, 2021. W Ronny Huang, Jonas Geiping, Liam Fowl, Gavin Taylor, and Tom Goldstein. Metapoison: Practical general-purpose clean-label data poisoning. Advances in Neural Information Processing Systems, 33: 12080 12091, 2020. Pang Wei Koh, Jacob Steinhardt, and Percy Liang. Stronger data poisoning attacks break data sanitization defenses. Machine Learning, pp. 1 47, 2022. Leslie Lamport, Robert Shostak, and Marshall Pease. The byzantine generals problem. ACM Trans. Program. Lang. Syst., 4(3):382 401, jul 1982. ISSN 0164-0925. doi: 10.1145/357172.357176. URL https://doi.org/10.1145/357172.357176. Kang Liu, Brendan Dolan-Gavitt, and Siddharth Garg. Fine-pruning: Defending against backdooring attacks on deep neural networks. In International symposium on research in attacks, intrusions, and defenses, pp. 273 294. Springer, 2018. Yiyong Liu, Michael Backes, and Xiao Zhang. Transferable availability poisoning attacks. ar Xiv preprint ar Xiv:2310.05141, 2023. Yiwei Lu, Gautam Kamath, and Yaoliang Yu. Indiscriminate data poisoning attacks on neural networks. Ar Xiv, abs/2204.09092, 2022. URL https://api.semanticscholar.org/Corpus ID:248266720. Alexandra Luccioni and Joseph Viviano. What s in the box? an analysis of undesirable content in the common crawl corpus. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 2: Short Papers), pp. 182 189, 2021. Brendan Mc Mahan, Eider Moore, Daniel Ramage, Seth Hampson, and Blaise Aguera y Arcas. Communicationefficient learning of deep networks from decentralized data. In Artificial intelligence and statistics, pp. 1273 1282. PMLR, 2017. Luis Muñoz-González, Battista Biggio, Ambra Demontis, Andrea Paudice, Vasin Wongrassamee, Emil C Lupu, and Fabio Roli. Towards poisoning of deep learning algorithms with back-gradient optimization. In Proceedings of the 10th ACM workshop on artificial intelligence and security, pp. 27 38, 2017. Luis Muñoz-González, Bjarne Pfitzner, Matteo Russo, Javier Carnerero-Cano, and Emil C Lupu. Poisoning attacks with generative adversarial nets. ar Xiv preprint ar Xiv:1906.07773, 2019. Rui Ning, Jiang Li, Chunsheng Xin, and Hongyi Wu. Invisible poison: A blackbox clean label backdoor attack to deep neural networks. In IEEE INFOCOM 2021-IEEE Conference on Computer Communications, pp. 1 10. IEEE, 2021. Published in Transactions on Machine Learning Research (12/2025) Christoph Schuhmann, Romain Beaumont, Richard Vencu, Cade Gordon, Ross Wightman, Mehdi Cherti, Theo Coombes, Aarush Katta, Clayton Mullis, Mitchell Wortsman, et al. Laion-5b: An open large-scale dataset for training next generation image-text models. Advances in Neural Information Processing Systems, 35:25278 25294, 2022. Ali Shafahi, W Ronny Huang, Mahyar Najibi, Octavian Suciu, Christoph Studer, Tudor Dumitras, and Tom Goldstein. Poison frogs! targeted clean-label poisoning attacks on neural networks. Advances in neural information processing systems, 31, 2018. Virat Shejwalkar and Amir Houmansadr. Manipulating the byzantine: Optimizing model poisoning attacks and defenses for federated learning. Proceedings 2021 Network and Distributed System Security Symposium, 2021. URL https://api.semanticscholar.org/Corpus ID:231861235. Virat Shejwalkar, Amir Houmansadr, Peter Kairouz, and Daniel Ramage. Back to the drawing board: A critical evaluation of poisoning attacks on production federated learning. In 2022 IEEE Symposium on Security and Privacy (SP), pp. 1354 1371. IEEE, 2022. Jacob Steinhardt, Pang Wei W Koh, and Percy S Liang. Certified defenses for data poisoning attacks. Advances in neural information processing systems, 30, 2017. Sebastian U Stich. Local sgd converges fast and communicates little. ar Xiv preprint ar Xiv:1805.09767, 2018. Onur Varol, Emilio Ferrara, Clayton Davis, Filippo Menczer, and Alessandro Flammini. Online humanbot interactions: Detection, estimation, and characterization. In Proceedings of the international AAAI conference on web and social media, volume 11, pp. 280 289, 2017. Wenxiao Wang, Alexander J Levine, and Soheil Feizi. Improved certified defenses against data poisoning with (deterministic) finite aggregation. In International Conference on Machine Learning, pp. 22769 22783. PMLR, 2022. Cong Xie, Oluwasanmi Koyejo, and Indranil Gupta. Fall of empires: Breaking byzantine-tolerant sgd by inner product manipulation. In Uncertainty in Artificial Intelligence, pp. 261 270. PMLR, 2020. Dong Yin, Yudong Chen, Ramchandran Kannan, and Peter Bartlett. Byzantine-robust distributed learning: Towards optimal statistical rates. In International Conference on Machine Learning, pp. 5650 5659. PMLR, 2018. Bingyin Zhao and Yingjie Lao. CLPA: Clean-Label Poisoning Availability Attacks Using Generative Adversarial Nets. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pp. 9162 9170, 2022. doi: 10.1609/aaai.v36i8.20902. URL https://ojs.aaai.org/index.php/AAAI/article/view/20902. Bo Zhao, Konda Reddy Mopuri, and Hakan Bilen. idlg: Improved deep leakage from gradients. ar Xiv preprint ar Xiv:2001.02610, 2020. Published in Transactions on Machine Learning Research (12/2025) A Discussion on realistic values of α Machine Learning practitioners process data scraped online Luccioni & Viviano (2021); Schuhmann et al. (2022) or sent by users and trust their resulting models Hoang et al. (2021). Since data poisoning has proven to be practical in real case scenario Carlini et al. (2023), we should treat models trained on scrapped or collaborative datasets with as much precaution as untrustworthy data. A few redditors unpurposefully obtained a dedicated token in GPT-3 s tokenizer by artificially inflating their online presence on the platform by massively posting over 160k posts on the r/counting subreddit on which people simply count 6. Even worse, these data can potentially be sent by malevolent agents who can thus influence the models by legitimately participating to these data sources. Estimating a realistic value for α is difficult since the point for attackers is to not be detected. Since bots represent a non-negligible part of online users (between 5% 7 and 15% Varol et al. (2017) on Twitter), we could expect the ratio of stealthy malevolent agents to be somewhat similar (if not higher). Armies as large as 2 million full-time individuals Charon & Vilmer (2021) could be conducing campaigns on social media (each individual manually steering tens of social media accounts to escape automated-activity detection). Collaborative datasets like Wikipedia Carlini et al. (2023) could already be stealthily poisoned. As such, we should not only presume that data we train on might have been poisoned, we should also consider the contamination ratio to be much higher than a fraction of a percent. In this work, we considered a wide range up to the extreme case of α = 0.48 We chose to solve the poisoning optimization problem with gradient-based approaches which are computationally intense, limiting us in experimenting with larger models and datasets. Stronger gradient attacks that only require to steer the model in a direction only slightly dissimilar from honest gradients should also be explored to increase the chances for a stealth data poison to exist and to improve the attack success rate. Stronger or less computationally expensive data poisoning approaches (that may not rely on gradient attacks) can be experimented to enhance the attack success rate. While the tested defense mechanisms for the gradient case appear to generalize and defend against data poisoning, it has been shown that the latter can bypass certain defense mechanisms Koh et al. (2022) and that robust mean estimation only works up to a certain point El-Mhamdi et al. (2023). This leaves room for potentially devastating data poisoning that can bypass defenses while mimicking an unstoppable gradient attack. B Complementary Figures and Tables Table 5: Architecture of the 1.6M parameters convolutional neural network used for our experiments. Layer # of channels kernel stride Conv2d 32 5 5 2 Re LU Conv2d 64 5 5 2 Re LU Linear 512 Re LU Linear 64 Re LU Linear 10 6Solid Gold Magikarp (plus, prompt generation) 7https://twitter.com/paraga/status/1526237585441296385 Published in Transactions on Machine Learning Research (12/2025) Figure 8: Validation accuracies of the CNN during training with the SGD and Adam update rule with the Average aggregation function under the Gradient Ascent attack. This data poisoning manages to slow down the training but not degrade the model s performance to random levels. Error bars represent the standard error. Figure 9: 50 of the first 500 poisons crafted in the (SGD & Average, Little is Enough, α = 0.01) setting. Published in Transactions on Machine Learning Research (12/2025) Figure 10: Best test accuracies for the SGD; Average setting under the Little is Enough attack for different feasible sets. (a) Availability attack. (b) Slow down attack. Figure 11: Visual Transformer (Vi T) tiny with patch size 8 under different attacks. Little is Enough performs a slow down attack whereas Gradient Ascent and Orthogonal Gradient are able to perform an availability attack for high enough contamination levels α.