# genie_generative_hard_negative_images_through_diffusion__db73e393.pdf Published in Transactions on Machine Learning Research (04/2025) Ge NIe: Generative Hard Negative Images Through Diffusion Soroush Abbasi Koohpayegani* 1 Anuj Singh* 2 3 K L Navaneet1 Hamed Pirsiavash1 Hadi Jamali-Rad2 3 1University of California, Davis 2Delft University of Technology, The Netherlands 3Shell Global Solutions International B.V., Amsterdam, The Netherlands {soroush,nkadur,hpirsiav}@ucdavis.edu {a.r.singh,h.jamalirad}@tudelft.nl Reviewed on Open Review: https://openreview.net/forum?id=Vu LEOy Ti PO Data augmentation is crucial in training deep models, preventing them from overfitting to limited data. Recent advances in generative AI, e.g., diffusion models, have enabled more sophisticated augmentation techniques that produce data resembling natural images. We introduce Ge NIe a novel augmentation method which leverages a latent diffusion model conditioned on a text prompt to combine two contrasting data points (an image from the source category and a text prompt from the target category) to generate challenging augmentations. To achieve this, we adjust the noise level (equivalently, number of diffusion iterations) to ensure the generated image retains low-level and background features from the source image while representing the target category, resulting in a hard negative sample for the source category. We further automate and enhance Ge NIe by adaptively adjusting the noise level selection on a per image basis (coined as Ge NIe-Ada), leading to further performance improvements. Our extensive experiments, in both few-shot and long-tail distribution settings, demonstrate the effectiveness of our novel augmentation method and its superior performance over the prior art. Our code is available at: https://github.com/UCDvision/Ge NIe 1 Introduction Data augmentation has become an integral part of training deep learning models, particularly when faced with limited training data. For instance, when it comes to image classification with limited number of samples per class, model generalization ability can be significantly hindered. Simple transformations like rotation, cropping, and adjustments in brightness artificially diversify the training set, offering the model a more comprehensive grasp of potential data variations. Hence, augmentation can serve as a practical strategy to boost the model s learning capacity, minimizing the risk of overfitting and facilitating effective knowledge transfer from limited labelled data to real-world scenarios. Various image augmentation methods, encompassing standard transformations, and learning-based approaches have been proposed (Cubuk et al., 2019b;a; Yun et al., 2019b; Zhang et al., 2018; Trabucco et al., 2024). Some augmentation strategies combine two images possibly from two different categories to generate a new sample image. The simplest ones in this category are Mix Up (Zhang et al., 2018) and Cut Mix (Yun et al., 2019a) where two images are combined in the pixel space. However, the resulting augmentations often do not lie within the manifold of natural images and act as out-of-distribution samples that will not be encountered during testing. Recently, leveraging generative models for data augmentation has gained an upsurge of attention (Trabucco et al., 2024; Roy et al., 2023; Luzi et al., 2022; He et al., 2022b). These interesting studies, either based on finetuning or prompt engineering of diffusion models, are mostly focused on generating generic augmentations * equal contribution Published in Transactions on Machine Learning Research (04/2025) Figure 1: Generative Hard Negative Images Through Diffusion (Ge NIe): generates hard negative images that belong to the target category but are similar to the source image from low-level feature and contextual perspectives. Ge NIe starts from a source image passing it through a partial noise addition process, and conditioning it on a different target category. By controlling the amount of noise, the reverse latent diffusion process generates images that serve as hard negatives for the source category. without considering the impact of other classes and incorporating that information into the generative process for a classification context. We take a different approach to generate challenging augmentations near the decision boundaries of a downstream classifier. Inspired by diffusion-based image editing methods (Meng et al., 2021; Luzi et al., 2022) some of which are previously used for data augmentation, we propose to use conditional latent diffusion models (Rombach et al., 2021a) for generating hard negative images. Our core idea (coined as Ge NIe) is to sample source images from various categories and prompt the diffusion model with a contradictory text corresponding to a different target category. We demonstrate that the choice of noise level (or equivalently number of iterations) for the diffusion process plays a pivotal role in generating images that semantically belong to the target category while retaining low-level features from the source image. We argue that these generated samples serve as hard negatives (Xuan et al., 2021; Mao et al., 2017) for the source category (or from a dual perspective hard positives for the target category). To further enhance Ge NIe, we propose an adaptive noise level selection strategy (dubbed as Ge NIe-Ada) enabling it to adjust noise levels automatically per sample. To establish the impact of Ge NIe, we focus on two challenging scenarios: long-tail and few-shot settings. In real-world applications, data often follows a long-tail distribution, where common scenarios dominate and rare occurrences are underrepresented. For instance, a person jaywalking a highway causes models to struggle with such unusual scenarios. Combating such a bias or lack of sufficient data samples during model training is essential in building robust models for applications such as self-driving cars. Same challenge arises in few-shot learning settings where the model has to learn from only a handful of samples. Our extensive quantitative and qualitative experimentation, on a suite of few-shot and long-tail distribution settings, corroborate the effectiveness of the proposed novel augmentation method (Ge NIe, Ge NIe-Ada) in generating hard negatives, corroborating its significant impact on categories with a limited number of samples. A high-level sketch of Ge NIe is illustrated in Fig. 1. Our main contributions are summarized below: - We introduce Ge NIe, a novel yet elegantly simple diffusion-based augmentation method to create challenging hard-negative augmentations in the manifold of natural images. For the first time, to our best knowledge, Ge NIe achieves this by combining two sources of information (a source image, and a contradictory target prompt) through a noise-level adjustment mechanism in the diffusion-denoising process. - We further extend Ge NIe by automating the noise-level adjustment strategy on a per-sample basis (called Ge NIe-Ada), to enable generating hard negative samples in the context of image classification, leading also to further performance enhancement. Published in Transactions on Machine Learning Research (04/2025) - To substantiate the impact of Ge NIe, we present a suite of quantitative and qualitative results including extensive experimentation on two challenging tasks: few-shot and long tail distribution settings corroborating that Ge NIe (and its extension Ge NIe-Ada) significantly improve the downstream classification performance. 2 Related Work Data Augmentations. Simple flipping, cropping, colour jittering, and blurring are some forms of image augmentations (Shorten & Khoshgoftaar, 2019). These augmentations are commonly adopted in training deep learning models. However, using these data augmentations is not trivial in some domains. For example, using blurring might remove important low-level information from medical images. More advanced approaches, such as Mix Up (Zhang et al., 2018) and Cut Mix (Yun et al., 2019a), mix images and their labels accordingly (Hendrycks et al., 2020; Liu et al., 2022; Kim et al., 2020; Cubuk et al., 2020). However, the resulting augmentations are not natural images anymore, and thus, act as out-of-distribution samples that will not be seen at test time. To combat this, SAGE (Ma et al., 2022) proposes a data augmentation technique that uses visual saliency to perform optimal image blending at each spatial location, and optimizes the relative image position such that the resulting visual saliency is maximized. Another strand of research tailors the augmentation strategy through a learning process to fit the training data (Ding et al., 2024; Cubuk et al., 2019b;a). Unlike the above methods, we propose to utilize pre-trained latent diffusion models to generate hard negatives (in contrast to generic augmentations) through a noise adaptation strategy discussed in Section 3. Data Augmentation with Generative Models. Using synthesized images from generative models to augment training data has been studied before in many domains (Frid-Adar et al., 2018; Sankaranarayanan et al., 2018), including domain adaptation (Huang et al., 2018), visual alignment (Peebles et al., 2022), and mitigation of dataset bias (Sharmanska et al., 2020; Hemmat et al., 2023; Prabhu et al., 2024). For example, (Prabhu et al., 2024) introduces a methodology aimed at enhancing test set evaluation through augmentation. While previous methods predominantly relied on GANs (Zhang et al., 2021c; Li et al., 2022b; Tritrong et al., 2021) as the generative model, more recent studies promote using diffusion models to augment the data (Rombach et al., 2021a; He et al., 2022b; Shipard et al., 2023; Trabucco et al., 2024; Azizi et al., 2023; Luo et al., 2023; Roy et al., 2023; Jain et al., 2022; Feng et al., 2023; Dunlap et al., 2023b; Chegini & Feizi, 2023). More specifically, (Trabucco et al., 2024; Roy et al., 2023; He et al., 2022b; Azizi et al., 2023) study the effectiveness of text-to-image diffusion models in data augmentation by diversification of each class with synthetic images. (Roy et al., 2023) also utilizes a text-to-image diffusion model, but with a BLIP (Li et al., 2022d) model to generate meaningful captions from the existing images. (Jain et al., 2022) utilizes diffusion models for augmentation to correct model mistakes. (Feng et al., 2023) uses CLIP (Radford et al., 2021) to filter generated images. Generative models for data augmentation may produce out-of-distribution samples if the downstream task s data distribution differs. Fine-tuning on a small downstream dataset can address this. For example, DAFusion (Trabucco et al., 2024) fine-tunes a diffusion model using textual inversion (Gal et al., 2022a), while Si STA (Thopalli et al., 2023) adapts a GAN for the task. (Graikos et al., 2023a) propose adapting generative models to downstream tasks by leveraging the internal representations of the denoiser network. Investigations by (Tian et al., 2023) explore the use of text-to-image synthetic images for generating positive samples in contrastive learning. (Dunlap et al., 2023b) utilizes text-based diffusion and a large language model (LLM) to diversify the training data. (Chegini & Feizi, 2023) uses an LLM to generate text descriptions of failure modes associated with spurious correlations, which are then used to generate synthetic data through generative models. The challenge here is that the LLM has little understanding of such failure scenarios and contexts. We take a completely different approach here, without replying on any extra source of information (e.g., through an LLM). Inspired by image editing approaches such as Boomerang (Luzi et al., 2022) and SDEdit (Meng et al., 2021), we propose to adaptively guide a latent diffusion model to generate hard negatives images (Mao et al., 2017; Xuan et al., 2021) on a per-sample basis per category. In a nutshell, the aforementioned studies focus on improving the diversity of each class with effective prompts and diffusion models, however, we focus on generating effective hard negative samples for each class by combining two sources of contradicting information (images from the source category and text prompt from the target category). Published in Transactions on Machine Learning Research (04/2025) Language Guided Recognition Models. Vision-Language foundation models (VLMs) (Alayrac et al., 2022; Radford et al., 2021; Rombach et al., 2021a; Saharia et al., 2022; Ramesh et al., 2022; 2021) utilize human language to guide the generation of images or to extract features from images that are aligned with human language. CLIP (Radford et al., 2021) excels in zero-shot tasks by aligning images with text, while recent works improve prompts (Dunlap et al., 2023a; Petryk et al., 2022) or use diffusion models as classifiers (Li et al., 2023). Similarly, we leverage Stable Diffusion 1.5 (Rombach et al., 2021a) to enhance downstream tasks by augmenting training data with hard negative samples based on category names. Few-Shot Learning. In Few-shot Learning (FSL), we pre-train a model with abundant data to learn a rich representation, then fine-tune it on new tasks with only a few available samples. In supervised FSL (Chen et al., 2019a; Afrasiyabi et al., 2019; Qiao et al., 2018; Ye et al., 2020; Dvornik et al., 2019; Li et al., 2020; Sung et al., 2018; Zhou et al., 2021; Singh & Jamali-Rad, 2023), pretraining is done on a labeled dataset, whereas in unsupervised FSL (Jang et al., 2022; Wang & Deng, 2022; Lu et al., 2022; Qin et al., 2020; Antoniou & Storkey, 2019; Khodadadeh et al., 2019; Hsu et al., 2018; Medina et al., 2020; Shirekar et al., 2023) the pretraining has to be conducted on an unlabeled dataset posing an extra challenge in the learning paradigm and neighboring these methods closer to the realm of self-supervised learning. 3 Proposed Method: Ge NIe Given a source image XS from category S = , we are interested in generating a target image Xr from category T = . In doing so, we intend to ensure the low-level visual features or background context of the source image are preserved, so that we generate samples that would serve as hard negatives for the source image. To this aim, we adopt a conditional latent diffusion model (such as Stable Diffusion, (Rombach et al., 2021a)) conditioned on a text prompt of the following format A photo of a T = . Key Idea. Ge NIe in its basic form is a simple yet effective augmentation sample generator for improving a classifier fθ(.) with the following two key aspects: (i) inspired by (Luzi et al., 2022; Meng et al., 2021) instead of adding the full amount of noise σmax and going through all Nmax (being typically 50) steps of denoising, we use less amount of noise (rσmax, with r (0, 1)) and consequently fewer number of denoising iterations ( r Nmax ); (ii) we prompt the diffusion model with a P mandating a target category T different than the source S. Hence, we denote the conditional diffusion process as Xr = STDiff(XS, P, r). In such a construct, the proximity of the final decoded image Xr to the source image XS or the target category defined through the text prompt P depends on r. Hence, by controlling the amount of noise, we can generate images that blend characteristics of both the text prompt P and the source image XS. If we do not provide much of visual details in the text prompt (e.g., desired background, etc.), we expect the decoded image Xr to follow the details of XS while reflecting the semantics of the text prompt P. We argue, and demonstrate later, that the newly generated samples can serve as hard negative examples for the source category S since they share the low-level features of XS while representing the semantics of the target category, T. Notably, the source category S can be randomly sampled or be carefully extracted from the confusion matrix of fθ(.) based on real training data. The latter might result in even harder negative samples being now cognizant of model confusions. Finally, we will append our initial dataset with the newly generated hard negative samples through Ge NIe and (re)train the classifier model. Enhancing Ge NIe: Ge NIe-Ada. One of the remarkable aspects of Ge NIe lies in its simple application, requiring only XS, P, and r. However, selecting the appropriate value for r poses a challenge as it profoundly influences the outcome. When r is small, the resulting Xr tends to closely resemble XS, and conversely, when r is large (closer to 1), it tends to resemble the semantics of the target category. This phenomenon arises because a smaller noise level restricts the capacity of the diffusion model to deviate from the semantics of the input XS. Thus, a critical question emerges: how can we select r for a particular source image to generate samples that preserve the low-level semantics of the source category S in XS while effectively representing the semantics of the target category T? We propose a method to determine an ideal value for r. Our intuition suggests that by varying the noise ratio r from 0 to 1, Xr will progressively resemble category S in the beginning and category T towards the end. However, somewhere between 0 and 1, Xr will undergo a rapid transition from category S to T. This phenomenon is empirically observed in our experiments with Published in Transactions on Machine Learning Research (04/2025) Figure 2: Effect of noise ratio, r, in Ge NIe: we employ Ge NIe to generate augmentations for the target classes (motorcycle and cat) with varying r. Smaller r yields images closely resembling the source semantics, creating an inconsistency with the intended target label. By tracing r from 0 to 1, augmentations gradually transition from source image characteristics to the target category. However, a distinct shift from the source to the target occurs at a specific r that may vary for different source images or target categories. For more examples, please refer to Fig. A9. Algorithm 1: Ge NIe-Ada Require: XS, XT , fθ(.), STDiff(.), M Extract ZS fθ(Xs), ZT fθ(XT ) for m [1, M] do M , Zr fθ( STDiff(X, P, r) ) dm (Zr ZS)T (ZT ZS) ||ZT ZS||2 m argmaxm |dm dm 1|, m [2, M] r m M Return: Xr = STDiff(XS, P, r ) Figure 3: Ge NIe-Ada: To choose r adaptively for each (source image, target category) pair, we propose tracing the semantic trajectory from ZS (source image embeddings) to ZT (target embeddings) through backbone feature extractor fθ( ) (Algorithm 1). We adaptively select the sample right after the largest semantic shift. varying r, as depicted in Fig. 2. Although the exact reason for this rapid change remains uncertain, one possible explanation is that the intermediate points between two categories reside far from the natural image manifold, thus, challenging the diffusion model s capability to generate them Sclocchi et al. (2024). Ideally, we should select r corresponding to just after this rapid semantic transition, as at this point, Xr exhibits the highest similarity to the source image while belonging to the target category. We propose to trace the semantic trajectory between XS and XT through backbone feature extractor fθ(.). As shown in Algorithm 1, assuming access to the classifier backbone fθ(.) and at least one example XT from the target category, we convert both XS and XT into their respective latent vectors ZS and ZT by passing them through fθ(.). Then, we sample M values for r uniformly distributed (0, 1), generating their corresponding Xr and their latent vectors Zr for all those r. Subsequently, we calculate dr = (Zr ZS)T (ZT ZS) ||ZT ZS||2 as the distance between Zr and ZS projected onto the vector connecting ZS and ZT . Our hypothesis posits that the rapid semantic transition corresponds to a sharp change in this projected distance. Therefore, we sample n values for r uniformly distributed between 0 and 1, and analyze the variations in dr. We identify the largest gap in dr and select the r value just after the gap when increasing r, as detailed in Algorithm 1 and illustrated in Fig. 3. 3.1 Psuedocode of Ge NIe: As illustrated in Algorithm 2, we provide a detailed pytorch-style pseudocode for Ge NIe. First, a SDv1.5 pipeline initialized by loading all the components such as the VQ-VAE encoder and decoder, the CLIP text Published in Transactions on Machine Learning Research (04/2025) encoder and the DPM scheduler for the forward and reverse diffusion process. Then, the source image is input to the encoder to encode the image into latent space for the diffusion model. Next, the encoded image is partially noised based on the noise ratio r using the scheduler. The diffusion model then de-noises the partially noised latent embedding for a total of NUM INFERENCE STEPS r steps, with an additional input of a text prompt from a contradictory target class. Finally, the decoder decodes the de-noised latent embedding into the generated hard-negative image, that contains the low-level features of the source image and the class/category of the contradictory text-prompt. Algorithm 2: Py Torch-style Pseudocode of Ge NIe. # Stable Diffusion Pipeline: Pre-trained diffusion model # DPMSolver Multistep Scheduler: Scheduler for forward and reverse diffusion # encode_latents: Encodes an image into latent space # decode_latents: Decodes latents back into an image def Augment Ge NIe(source_image, target_prompt, percent_noise): NUM_INFERENCE_STEPS = 50 # Number of steps for reverse diffusion NUM_TRAIN_STEPS = 1000 # Number of steps for forward diffusion # Initialize the stable diffusion pipeline and scheduler pipe = Stable Diffusion Pipeline.from_pretrained("stable-diffusion-v1-5") scheduler = DPMSolver Multistep Scheduler.from_config(pipe.scheduler.config) # Encode the source image into latent space latents = encode_latents(source_image) # Forward Diffusion noise = torch.randn(latents.shape) # Generate random noise timestep = torch.Tensor([int(NUM_TRAIN_STEPS * percent_noise)]) # Calculate timestep latents_noise = scheduler.add_noise(latents, noise, timestep) # Add noise to latents # Reverse Diffusion latents = pipe( prompt=target_prompt, percent_noise=percent_noise, latents=latents_noise, num_inference_steps=NUM_INFERENCE_STEPS ) # Decode latents back into an augmented image augmented_image = decode_latents(latents) return augmented_image 4 Experiments Since the impact of augmentation is more pronounced when the training data is limited, we evaluate the impact of Ge NIe on Few-Shot classification in Section 4.1, Long-Tailed classification in Section 4.3, and fine-grained classification in Section 4.2. For Ge NIe-Ada in all scenarios, we utilize Ge NIe to generate augmentations from the noise level set {0.5, 0.6, 0.7, 0.8, 0.9}. The selection of the appropriate noise level per source image and target is adaptive, achieved through Algorithm 1. Baselines. We use Stable Diffusion 1.5 (Rombach et al., 2021a) as our base diffusion model. In all settings, we use the same prompt format to generate images for the target class: i.e., A photo of a , where we replace the target category with the target category label. We generate 512 512 images for all methods. For fairness, we generate the same number of new images for each class. We use a single NVIDIA RTX 3090 for image generation. We consider 4 diffusion-based baselines and a suite of traditional data augmentation baselines. Img2Img (Luzi et al., 2022; Meng et al., 2021): We sample an image from a target class, add noise to its latent representation and then pass it along with a prompt for the target category through reverse diffusion. The focus here is on a target class for which we generate extra positive samples. Adding large amount of noise leads to generating an image less similar to the original image. We use two different noise magnitudes for this baseline: r = 0.3 and r = 0.7 and denote them by Img2Img L and Img2Img H, respectively. Txt2Img (Azizi et al., 2023; He et al., 2022b): For this baseline, we omit the forward diffusion process and only use the reverse process starting from a text prompt for the target class of interest. This is similar to the Published in Transactions on Machine Learning Research (04/2025) base text-to-image generation strategy adopted in (Rombach et al., 2021a; He et al., 2022b; Shipard et al., 2023; Azizi et al., 2023; Luo et al., 2023). Fig. 4 illustrates a set of generated augmentation examples for Txt2Img, Img2Img, and Ge NIe. DAFusion (Trabucco et al., 2024): In this method, an embedding is optimized with a set of images for each class to correspond to the classes in the dataset. This approach is introduced in Textual Inversion (Gal et al., 2022c). We optimize an embedding for 5000 iterations for each class in the dataset, followed by augmentation similar as the DAFusion method. Cap2Aug(Roy et al., 2023): It is a recent diffusion-based data augmentation strategy that uses image captions as text prompts for an image-to-image diffusion model. Traditional Data Augmentation: We consider both weak and strong traditional augmentations. More specifically, for weak augmentation we use random resize crop with scaling [0.2, 1.0] and horizontal flipping. For strong augmentation, we consider random color jitter, random grayscale, and Gaussian blur. For the sake of completeness, we also compare against data augmentations such as Cut Mix (Yun et al., 2019a) and Mix Up (Zhang et al., 2018) that naively combine two images together, and SAGE (Ma et al., 2022), that mixes image pairs based on their visual saliency to promote discriminative foreground objects in the mix. 4.1 Few-shot Classification We assess the impact of Ge NIe compared to other augmentations in a number of few-shot classification (FSL) scenarios, where the model has to learn only from the samples contained in the (N-way, K-shot) support set and infer on the query set. Note that this implies that the Ge NIe, Ge NIe-Ada augmentations are only applied during few-shot inference on the test-dataset, and not during the pre-training stage, where the backbone is unsupervised pre-trained on an abundant (few-shot training) dataset. The goal is to assess how well the model can benefit from the augmentations while keeping the original N K samples intact. Datasets. We conduct our few-shot experiments on two most commonly adopted few-shot classification datasets: mini-Imagenet (Ravi & Larochelle, 2017) and tiered-Imagenet (Ren et al., 2018). mini-Imagenet is a subset of Image Net (Deng et al., 2009) for few-shot classification. It contains 100 classes with 600 samples each. We follow the predominantly adopted settings of (Ravi & Larochelle, 2017; Chen et al., 2019a) where we split the entire dataset into 64 classes for training, 16 for validation and 20 for testing. tiered-Imagenet is a larger subset of Image Net with 608 classes and a total of 779, 165 images, which are grouped into 34 higher-level nodes in the Image Net human-curated hierarchy. This set of nodes is partitioned into 20, 6, and 8 disjoint sets of training, validation, and testing nodes, and the corresponding classes form the respective meta-sets. Evaluation. We evaluate the test-set accuracies of a state-of-the-art unsupervised few-shot learning method with Ge NIe and compare them against the accuracies obtained using other augmentation methods. Specifically, we use Uni Siam (Lu et al., 2022) pre-trained with Res Net-18, Res Net-34 and Res Net-50 backbones and follow its evaluation strategy of fine-tuning a logistic regressor to perform (N-way, K-shot) classification on the test sets of miniand tiered-Imagenet. Following (Ravi & Larochelle, 2017), an episode consists of a labeled support-set and an unlabelled query-set. The support-set contains N randomly sampled classes where each class contains K samples, whereas the query-set contains Q randomly sampled unlabeled images per class. We conduct our experiments on the two most commonly adopted settings: (5-way, 1-shot) and (5-way, 5-shot) classification settings. Following the literature, we sample 16-shots per class for the query set in both settings. We report the test accuracies along with the 95% confidence interval over 600 and 1000 episodes for mini-Image Net and tiered-Image Net, respectively. Implementation Details: Ge NIe generates augmented images for each class using images from all other classes as the source image. We use r = 0.8 in our experiments. We generate 4 samples per class as augmentations in the 5-way, 1-shot setting and 20 samples per class as augmentations in the 5-way, 5shot setting. For the sake of a fair comparison, we ensure that the total number of labelled samples in the support set after augmentation remains the same across all different traditional and generative augmentation methodologies. Due to the expensive training of embeddings for each class in each episode, we only evaluated the DA-Fusion baseline on the first 100 episodes. Published in Transactions on Machine Learning Research (04/2025) Figure 4: Visualization of Generative Samples: We compare Ge NIe with two baselines: Img2Img L augmentation: both image and text prompt are from the same category. Adding noise does not change the image much, so they are not hard examples. Txt2Img augmentation: We simply use the text prompt only to generate an image for the desired category (e.g., using a text2image method). Such images may be far from the domain of our task since the generation is not informed by any visual data from our task. Ge NIe augmentation: We use the target category name in the text prompt only along with the source image. Results: The results on mini-Imagenet and tiered-Imagenet for both (5-way, 1 and 5-shot) settings are summarized in Table 1 and Table 3, respectively. Regardless of the choice of backbone, we observe that Ge NIe helps consistently improve Uni Siam s performance and outperform other supervised and unsupervised few-shot classification methods as well as other diffusion-based (Trabucco et al., 2024; Luzi et al., 2022; Rombach et al., 2021b; He et al., 2022b) and classical (Yun et al., 2019a; Zhang et al., 2018; Ma et al., 2022) data augmentation techniques on both datasets, across both (5-way, 1 and 5-shot) settings. Our noise adaptive method of selecting optimal augmentations per source image (Ge NIe-Ada) further improves Ge NIe s performance across all three backbones, both few-shot settings, and both datasets (mini and tiered Imagenet). 4.2 Fine-grained Few-shot Classification To further investigate the impact of the proposed method, we compare Ge NIe with other text-based data augmentation techniques across four distinct fine-grained datasets in a 20-way, 1-shot classification setting. We employ the pre-trained DINOV2 Vi T-G (Oquab et al., 2023) backbone as a feature extractor to derive Published in Transactions on Machine Learning Research (04/2025) Table 1: mini-Image Net: We use our augmentations on (5-way, 1-shot) and (5-way, 5-shot) few-shot settings of mini Imagenet dataset with 3 different backbones (Res Net-18, 34, and 50). We compare with various baselines and show that our augmentations with Uni Siam outperform all the baselines including Txt2Img and DAFusion augmentation. The number of generated images per class is 4 for 1-shot and 20 for 5-shot settings. Augmentation Method Pre-training 1-shot 5-shot - i De Me-Net 2019b sup. 59.1 0.9 74.6 0.7 - Robust + dist 2019 sup. 63.7 0.6 81.2 0.4 - AFHN 2020 sup. 62.4 0.7 78.2 0.6 Weak Proto Net+SSL 2020 sup.+ssl - 76.6 Weak Neg-Cosine 2020 sup. 62.3 0.8 80.9 0.6 - Centroid Align2019 sup. 59.9 0.7 80.4 0.7 - Baseline 2019a sup. 59.6 0.8 77.3 0.6 - Baseline++ 2019a sup. 59.0 0.8 76.7 0.6 Weak PSST 2021 sup.+ssl 59.5 0.5 77.4 0.5 Weak UMTRA 2019 unsup. 43.1 0.4 53.4 0.3 Weak Proto CLR 2020 unsup. 50.9 0.4 71.6 0.3 Weak Sim CLR 2020 unsup. 62.6 0.4 79.7 0.3 Weak Sim Siam 2021 unsup. 62.8 0.4 79.9 0.3 Weak Uni Siam+dist 2022 unsup. 64.1 0.4 82.3 0.3 Weak Uni Siam 2022 unsup. 63.1 0.8 81.4 0.5 Strong Uni Siam 2022 unsup. 62.8 0.8 81.2 0.6 Cut Mix 2019a Uni Siam 2022 unsup. 62.7 0.8 80.6 0.6 Mix Up 2018 Uni Siam 2022 unsup. 62.1 0.8 80.7 0.6 Img2Img L2022 Uni Siam 2022 unsup. 63.9 0.8 82.1 0.5 Img2Img H2022 Uni Siam 2022 unsup. 69.1 0.7 84.0 0.5 Txt2Img2023; 2022b Uni Siam 2022 unsup. 74.1 0.6 84.6 0.5 DAFusion 2024 Uni Siam 2022 unsup. 64.3 1.8 82.0 1.4 Ge NIe (Ours) Uni Siam 2022 unsup. 75.5 0.6 85.4 0.4 Ge NIe-Ada (Ours) Uni Siam 2022 unsup. 76.8 0.6 85.9 0.4 Augmentation Method Pre-training 1-shot 5-shot Weak Baseline 2019a sup. 49.8 0.7 73.5 0.7 Weak Baseline++ 2019a sup. 52.7 0.8 76.2 0.6 Weak Sim CLR 2020 unsup. 64.0 0.4 79.8 0.3 Weak Sim Siam 2021 unsup. 63.8 0.4 80.4 0.3 Weak Uni Siam+dist 2022 unsup. 65.6 0.4 83.4 0.2 Weak Uni Siam 2022 unsup. 64.3 0.8 82.3 0.5 Strong Uni Siam 2022 unsup. 64.5 0.8 82.1 0.6 Cut Mix 2019a Uni Siam 2022 unsup. 64.0 0.8 81.7 0.6 Mix Up 2018 Uni Siam 2022 unsup. 63.7 0.8 80.1 0.8 Img2Img L2022 Uni Siam 2022 unsup. 65.5 0.8 82.9 0.5 Img2Img H2022 Uni Siam 2022 unsup. 70.5 0.8 84.8 0.5 Txt2Img2023; 2022b Uni Siam 2022 unsup. 75.4 0.6 85.5 0.5 DAFusion 2024 Uni Siam 2022 unsup. 64.7 1.9 83.2 1.4 Ge NIe (Ours) Uni Siam 2022 unsup. 77.1 0.6 86.3 0.4 Ge NIe-Ada (Ours) Uni Siam 2022 unsup. 78.5 0.6 86.6 0.4 Weak PDA+Net 2021 unsup. 63.8 0.9 83.1 0.6 Weak Meta-DM 2023 unsup. 66.7 0.4 85.3 0.2 Weak Uni Siam 2022 unsup. 64.6 0.8 83.4 0.5 Strong Uni Siam 2022 unsup. 64.8 0.8 83.2 0.5 Cut Mix 2019a Uni Siam 2022 unsup. 64.3 0.8 83.2 0.5 Mix Up 2018 Uni Siam 2022 unsup. 63.8 0.8 84.6 0.5 SAGE 2022 Uni Siam 2022 unsup. 75.7 0.8 84.9 0.4 Img2Img L2022 Uni Siam 2022 unsup. 66.0 0.8 84.0 0.5 Img2Img H2022 Uni Siam 2022 unsup. 71.1 0.7 85.7 0.5 Txt2Img2023; 2022b Uni Siam 2022 unsup. 76.4 0.6 86.5 0.4 DAFusion 2024 Uni Siam 2022 unsup. 65.7 1.8 83.9 1.2 Ge NIe (Ours) Uni Siam 2022 unsup. 77.3 0.6 87.2 0.4 Ge NIe-Ada (Ours) Uni Siam 2022 unsup. 78.6 0.6 87.9 0.4 features from training images. Subsequently, an SVM classifier is trained on these features, and we report the Top-1 accuracy of the model on the test set. Table 2: Few-shot Learning on Fine-grained dataset: We utilize an SVM classifier trained atop the DINOV2 Vi TG pretrained backbone, reporting Top-1 accuracy for the test set of each dataset. The baseline is an SVM trained on the same backbone using weak augmentation. Method Birds Cars Foods Aircraft CUB200 2011 Cars196 2013 Food101 2014 Aircraft 2013 Baseline 90.3 49.8 82.9 29.2 Img2Img L2022 90.7 50.4 87.4 31.0 Img2Img H2022 91.3 56.4 91.7 34.7 Txt2Img2022b 92.0 81.3 93.0 41.7 Ge NIe (r=0.5) 92.0 84.6 91.5 39.8 Ge NIe (r=0.6) 92.2 87.1 92.5 45.0 Ge NIe (r=0.7) 92.5 87.9 92.9 47.0 Ge NIe (r=0.8) 92.5 87.7 93.1 46.5 Ge NIe (r=0.9) 92.4 87.1 93.1 45.7 Ge NIe-Ada 92.6 87.9 93.1 46.9 Results: Table 2 summarizes the results. Additional details about this experiment can be found in Section A.9. Ge NIe outperforms all other baselines, including Txt2Img, by margins upto 0.5% on CUB200, 6.6% on Cars196, 0.1% on Food101 and 5.3% on FGVC-Aircraft. Ge NIe exhibits great effectiveness in more challenging datasets, outperforming the baseline with traditional augmentation by about 38% for the Cars dataset and by roughly 17% for the Aircraft dataset. It can be observed here that Ge NIe-Ada performs on-par with Ge NIe with a fixed noise level, eliminating the necessity for noise level search in Ge NIe. 4.3 Long-Tailed Classification We evaluate our method on long-tailed data, where the number of instances per class is unbalanced, with most categories having limited samples (tail). Our goal is to mitigate this bias by augmenting the tail of the distribution with generated samples. We evaluate Ge NIe using two backbones: Vi T with LVi T (Xu et al., 2023) and Res Net50 with VL-LTR (Tian et al., 2022). Following LVi T, we first train an MAE (He et al., 2021) and Vi T on the unbalanced dataset without any augmentation. Next, we train the Balanced Fine Tuning stage of LVi T by incorporating the augmentation data generated using Ge NIe or other baselines. For Res Net50, we use VL-LTR code to fine-tune the CLIP Res Net50 with generated augmentations by Ge NIe. Dataset: We perform experiments on Image Net-LT (Liu et al., 2019). It contains 115.8K images from 1, 000 categories. The number of images per class varies from 1280 to 5. Imagenet-LT classes can be divided into 3 groups: Few with less than 20 images, Med with 20 100 images, and Many with more than 100 images. Imagenet-LT uses the same validation set as Image Net. We augment Few categories only and limit the number of generated images to 50 samples per class. For Ge NIe, instead of randomly sampling Published in Transactions on Machine Learning Research (04/2025) Table 3: tiered-Image Net: Accuracies (% std) for 5way, 1-shot and 5-way, 5-shot classification settings on the test-set. We compare against various SOTA supervised and unsupervised few-shot classification baselines as well as other augmentation methods, with Uni Siam (Lu et al., 2022) pre-trained Res Net-18,50 backbones. Augmentation Method Pre-training 1-shot 5-shot Weak Sim CLR2020 unsup. 63.4 0.4 79.2 0.3 Weak Sim Siam 2021 unsup. 64.1 0.4 81.4 0.3 Weak Uni Siam 2022 unsup. 63.1 0.7 81.0 0.5 Strong Uni Siam 2022 unsup. 62.8 0.7 80.9 0.5 Cut Mix 2019a Uni Siam 2022 unsup. 62.1 0.7 78.9 0.6 Mix Up 2018 Uni Siam 2022 unsup. 62.1 0.7 78.4 0.6 Img2Img L2022 Uni Siam 2022 unsup. 63.9 0.7 81.8 0.5 Img2Img H2022 Uni Siam 2022 unsup. 68.7 0.7 83.5 0.5 Txt2Img2022b Uni Siam 2022 unsup. 72.9 0.6 84.2 0.5 DAFusion 2024 Uni Siam 2022 unsup. 62.6 2.1 81.0 1.5 Ge NIe(Ours) Uni Siam 2022 unsup. 73.6 0.6 85.0 0.4 Ge NIe-Ada(Ours) Uni Siam 2022 unsup. 75.1 0.6 85.5 0.5 Weak PDA+Net 2021 unsup. 69.0 0.9 84.2 0.7 Weak Meta-DM 2023 unsup. 69.6 0.4 86.5 0.3 Weak Uni Siam + dist 2022 unsup. 69.6 0.4 86.5 0.3 Weak Uni Siam 2022 unsup. 66.8 0.7 84.7 0.5 Strong Uni Siam 2022 unsup. 66.5 0.7 84.5 0.5 Cut Mix 2019a Uni Siam 2022 unsup. 66.0 0.7 83.3 0.5 Mix Up 2018 Uni Siam 2022 unsup. 66.1 0.5 84.1 0.8 Img2Img L2022 Uni Siam 2022 unsup. 67.8 0.7 85.3 0.5 Img2Img H2022 Uni Siam 2022 unsup. 72.4 0.7 86.7 0.4 Txt2Img2022b Uni Siam 2022 unsup. 77.1 0.6 87.3 0.4 DAFusion 2024 Uni Siam 2022 unsup. 66.5 2.2 84.8 1.4 Ge NIe (Ours) Uni Siam 2022 unsup. 78.0 0.6 88.0 0.4 Ge NIe-Ada (Ours) Uni Siam 2022 unsup. 78.8 0.6 88.6 0.6 Table 4: Long-Tailed Image Net-LT: We compare different augmentation methods on Image Net LT and report Top-1 accuracy for Few , Medium , and Many sets. On the Few set and Li VT method, our augmentations improve the accuracy by 11.7 points compared to Li VT original augmentation and 4.4 points compared to Txt2Img. Ge NIe-Ada outperforms Cap2Aug baseline in Few categories by 7.6%. Refer to Table A11 for a full comparison with prior Long-Tailed methods. Method Many Med. Few Overall Acc Res LT 2022 63.3 53.3 40.3 55.1 Pa Co 2021b 68.2 58.7 41.0 60.0 LWS 2019 62.2 48.6 31.8 51.5 Zero-shot CLIP 2021 60.8 59.3 58.6 59.8 DRO-LT 2021 64.0 49.8 33.1 53.5 VL-LTR 2022 77.8 67.0 50.8 70.1 Cap2Aug 2023 78.5 67.7 51.9 70.9 Ge NIe-Ada 79.2 64.6 59.5 71.5 Method Many Med. Few Overall Acc Vi T 2021 50.5 23.5 6.9 31.6 MAE 2022a 74.7 48.2 19.4 54.5 Dei T 2022 70.4 40.9 12.8 48.4 Li VT 2023 73.6 56.4 41.0 60.9 Li VT + Img2Img L 74.3 56.4 34.3 60.5 Li VT + Img2Img H 73.8 56.4 45.3 61.6 Li VT + Txt2Img 74.9 55.6 48.3 62.2 Li VT + Ge NIe-Ada 74.0 56.9 52.7 63.1 the source images from other classes, we use a confusion matrix on the training data to find the top-4 most confused classes and only consider those classes for random sampling of the source image. Results: Augmenting training data with Ge NIe-Ada improves accuracy on the Few set by 11.7% and 4.4% compared with LVi T only and LVi T with Txt2Img augmentation baselines respectively. In Res Net50, Ge NIe-Ada outperforms Cap2Aug baseline in Few categories by 7.6%. The results are summarized in Table 4. Please refer to Section A.11 for implementation details. 4.4 Ablation and Further Analysis Semantic Shift from Source to Target Class. The core motivation behind Ge NIe-Ada is that by varying the noise ratio r from 0 to 1, augmented sample Xr will progressively shift its semantic category from source (S) in the beginning to target category (T) towards the end. However, somewhere between 0 and 1, Xr will undergo a rapid transition from S to T. To demonstrate this hypothesis empirically, in Figs. 5 and A7, we visualize pairs of source images and target categories with their respective Ge NIe generated augmentations for different noise ratios r, along with their corresponding PCA-projected embedding scatter plots (on the far left). We extract embeddings for all the images using a DINOv2 Vi T-G pretrained backbone, which we assume as an oracle model in identifying the right category. We observe that as r increases from 0.3 to 0.8, the images transition to embody more of the target category s semantics while preserving the contextual features of the source image. This transition of semantics can also be observed in the embedding plots (on the left) where they consistently shift from the proximity of the source image (blue star) to the target class s centroid (red cross) as the noise ratio r increases. The sparse distribution of points within r = [0.4, 0.6] for the first and second images and r = [0.2, 0.4] for the third image aligns with our intuition of a rapid transition from category S to T, thus empirically affirming our motivation behind Ge NIe-Ada. This semantic shift also corroborates with the insight provided in (Sclocchi et al., 2025) - that at denoising timesteps beyond the transition, the class has completely changed/transitioned, but the generated sample may still be composed of low-level elements of the initial image. To further establish this, in Fig. 6, we demonstrate the efficacy of Ge NIe in generating hard negatives at the decision boundaries of an SVM classifier, which is trained on the labelled support set of the few-shot tasks of mini-Imagenet, without any augmentations. We then plot source and target class probabilities (P(YS|Xr) Published in Transactions on Machine Learning Research (04/2025) Figure 5: Embedding visualizations of generative augmentations: We pass all generative augmentations through DINOv2 Vi T-G (serving as an oracle) to extract their corresponding embeddings and visualize them with PCA. As shown, the extent of semantic shifts varies based on both the source image and the target class. Figure 6: Why Ge NIe augmentations are challenging? While deciding which class the generated augmentations (Xr) belong to is already difficult within r = [0.6, 0.7] (due to high overlap between P(YS|Xr) and P(YT |Xr)), Ge NIe-Ada selects the best noise threshold (r ) offering the hardest negative sample. and P(YT |Xr), respectively) of the generated augmentation samples Xr. For both r = 0.6 and 0.7, there is significant overlap between P(YS|Xr) and P(YT |Xr), making it difficult for the classifier to decide the correct class. On the right-hand-side, Ge NIe-Ada automatically selects the best r resulting in the most overlap between the two distributions, thus offering the hardest negative sample among the considered r values (for more details see A.1). Note that a large overlap between distributions is not sufficient to call the generated samples hard negatives because they should also belong to the target category. This is, however, confirmed by the high Oracle accuracy in Table 5 (elaborated in detail in the following paragraph) which verifies that majority of the generated augmentation samples do belong to the target category. Label consistency of the generated samples. The choice of noise ratio r is important in producing hard negative examples. In Table 5, we present the accuracy of the Ge NIe model across various noise ratios, alongside the oracle accuracy, which is an Image Net pre-trained Dei T-Base (Touvron et al., 2021b) classifier. We observe a decline in the label consistency of generated data (quantified by the performance of the oracle model) when decreasing the noise level. Reducing r also results in a degradation in the performance of the final few-shot model (87.2% 77.6%) corroborating that an appropriate choice of r plays a crucial role. We investigate this further in the following paragraph. Effect of Noise in Ge NIe. We examine the impact of noise on the performance of the few-shot model in Table 5. Noise levels r [0.7, 0.8] yield the best performance. Conversely, utilizing noise levels below 0.7 diminishes performance due to label inconsistency, as is demonstrated in Table 5 and Fig 5. As such, determining the appropriate noise level is pivotal for the performance of Ge NIe to be able to generate challenging hard negatives while maintaining label consistency. An alternative approach to finding the optimal noise level involves using Ge NIe-Ada to adaptively select the noise level for each source image and Published in Transactions on Machine Learning Research (04/2025) Table 5: Effect of Noise and Diffusion Models in Ge NIe: We use the same setting as in Table 1 to study the effect of the amount of noise. As expected (also shown in Fig 5), small noise results in worse accuracy since some generated images may be from the source category rather than the target one. For r = 0.5 only 73% of the generated data is from the target category. This behaviour is also shown in Fig. 2. Notably, reducing the noise level below 0.7 is associated with a decline in oracle accuracy and subsequent degradation in the performance of the final few-shot model. Note that the high oracle accuracy of Ge NIe-Ada demonstrates its capability to adaptively select the noise level per source and target, ensuring semantic consistency with the intended target. To further demonstrate Ge NIe s ability to generalize across different diffusion models, we replace the diffusion model with SD3 and SDXL-Turbo. The resulting accuracies follow a similar trend to those in Table 1, confirming Ge NIe s advantage over Txt2Img across various diffusion models. Method Generative Noise Res Net-18 Res Net-34 Res Net-50 Oracle Model r= 1-shot 5-shot 1-shot 5-shot 1-shot 5-shot Acc Txt2Img SD 1.5 - 74.1 0.6 84.6 0.5 75.4 0.6 85.5 0.5 76.4 0.6 86.5 0.4 - Ge NIe SD 1.5 0.5 60.4 0.8 74.1 0.6 62.0 0.8 75.8 0.6 63.7 0.9 77.6 0.6 73.4 0.5 Ge NIe SD 1.5 0.6 69.7 0.7 80.7 0.5 71.1 0.7 82.2 0.5 72.1 0.7 82.8 0.5 85.8 0.4 Ge NIe SD 1.5 0.7 74.5 0.6 83.3 0.5 76.4 0.6 84.4 0.5 77.1 0.6 85.0 0.4 94.5 0.2 Ge NIe SD 1.5 0.8 75.5 0.6 85.4 0.4 77.1 0.6 86.3 0.4 77.3 0.6 87.2 0.4 98.2 0.1 Ge NIe SD 1.5 0.9 75.0 0.6 85.3 0.4 77.6 0.6 86.2 0.4 77.7 0.6 87.0 0.4 99.3 0.1 Ge NIe-Ada SD 1.5 Adaptive 76.8 0.6 85.9 0.4 78.5 0.6 86.6 0.4 78.6 0.6 87.9 0.4 98.9 0.2 Txt2Img SDXL-Turbo - 72.5 0.3 82.1 0.6 76.2 0.2 84.4 0.3 76.7 0.6 85.9 0.5 - Ge NIe SDXL-Turbo 0.5 61.2 0.5 73.5 0.2 61.5 0.2 74.9 0.3 63.1 0.2 76.5 0.6 - Ge NIe SDXL-Turbo 0.6 70.2 0.2 79.3 0.4 71.2 0.7 81.4 0.6 73.2 0.2 82.4 0.5 - Ge NIe SDXL-Turbo 0.7 73.1 0.3 83.5 0.5 76.1 0.6 85.3 0.4 77.2 0.6 84.2 0.4 - Ge NIe SDXL-Turbo 0.8 74.2 0.3 85.1 0.3 76.9 0.4 85.5 0.5 78.7 0.6 87.7 0.4 - Ge NIe SDXL-Turbo 0.9 73.9 0.4 84.9 0.7 76.6 0.7 84.2 0.6 78.1 0.5 87.0 0.4 - Ge NIe-Ada SDXL-Turbo Adaptive 75.1 0.3 87.1 0.8 78.9 0.5 85.2 0.5 79.0 0.6 88.6 0.2 - Txt2Img SD 3 - 73.6 1.7 82.9 1.2 76.7 1.5 85.5 1.3 77.2 1.9 85.0 1.2 - Ge NIe SD 3 0.5 62.0 1.2 72.9 1.1 62.5 0.9 73.9 1.0 64.1 0.5 76.1 1.9 - Ge NIe SD 3 0.6 70.8 1.5 79.1 1.9 71.8 1.2 82.1 1.3 74.1 1.5 83.4 1.8 - Ge NIe SD 3 0.7 74.6 0.8 84.5 1.2 76.5 1.9 86.2 1.6 78.5 1.9 84.0 1.1 - Ge NIe SD 3 0.8 75.9 1.2 86.3 1.7 77.8 1.9 85.5 1.9 79.2 1.7 88.3 1.9 - Ge NIe SD 3 0.9 75.1 0.5 85.2 1.2 78.1 1.3 86.2 1.2 77.1 1.9 88.9 0.8 - Ge NIe-Ada SD 3 Adaptive 76.8 1.3 87.5 1.5 78.9 1.3 87.7 1.5 79.1 1.4 89.5 1.0 - target class. As demonstrated in Tables 5 and 2, Ge NIe-Ada matches or outperforms Ge NIe with fixed noise levels. Effect of Diffusion Models in Ge NIe. We have tried experimenting with both smaller as well as more recent diffusion models. More specifically, we have used Stable Diffusion XL-Turbo to generate hard-negatives through Ge NIe and Ge NIe-Ada. Few-shot classification results on mini Imagenet with these augmentations are shown in Table 5. The accuracies follow a similar trend to that of Table 1, where Stable Diffusion 1.5 was used to generate augmentations. Ge NIe-Ada improves Uni Siam s few-shot performance the most as compared to Ge NIe with different noise ratios r, and even when compared to Txt2Img. This empirically indicates the robustness of Ge NIe and Ge NIe-Ada to different diffusion engines. Note that, Stable Diffusion XL-Turbo by default uses 4 steps for the sake of optimization, and to ensure we can have the right granularity for the choice of r we have set the number of steps to 10. That is already 5 times faster than the standard Stable Diffusion v1.5 with 50 steps. Our experiments with Stable Diffusion v3 (which is a totally different model with a Transformers backbone) also in Table 5 also convey the same message. As such, we believe our approach is generalizable across different diffusion models. Ge NIe s applicability vs. VLMs and Large Scale Pre-Trained Backbones. Ge NIe enables downstream task performance improvements by fine-tuning the pre-trained backbone using our challenging hardnegative samples. This Ge NIe fine-tuned backbone could then be used for various different test samples at inference time. This allows the user to benefit from performance improvements, while using a computationally cheap pre-trained backbone (such as Res Net-18,34,50), as opposed to a compute and memory intensive pre-trained VLMs. Using VLMs is even more challenging in compute-constrained edge-device applications, where only a few gigabytes of memory is available for model inference upon deployment. That being said, we include comparisons against a LAION (Schuhmann et al., 2022) vision-language pre-trained CLIP (Radford et al., 2021) Res Net50 in Appendix A.5, Table A3 to evaluate if Ge NIe can still render performance improvements using hard negative samples. We observe that while this new CLIP backbone s performance Published in Transactions on Machine Learning Research (04/2025) appears saturated on this task ( 93% 1-shot accuracy), Ge NIe-Ada still achieves 1.3%, 0.5% improvements in the 1-shot, 5-shot settings, respectively, thus corroborating the efficacy of our hard-negative samples in data-deficient settings. 5 Concluding Remarks Ge NIe, for the first time to our knowledge, combines contradictory sources of information (a source image, and a different target category prompt) through a noise adjustment strategy into a conditional latent diffusion model to generate challenging augmentations, which can serve as hard negatives. Limitation. The required time to create augmentations through Ge NIe is on par with any typical diffusionbased competitors (Azizi et al., 2023; He et al., 2022b); however, this is naturally slower than traditional augmentation techniques (Yun et al., 2019a; Zhang et al., 2018). This is not a bottleneck in offline augmentation strategies, but can be considered a limiting factor in real-time scenarios. Recent studies are already mitigating this through advancements in diffusion model efficiency (Sauer et al., 2023; Meng et al., 2023; Liu et al., 2023). Another challenge present in any generative AI-based augmentation technique is the domain shift between the distribution of training data and the downstream context they might be used for augmentation. A possible remedy is to fine-tune the diffusion backbone on a rather small dataset from the downstream task. Broader Impact. Ge NIe can have a significant impact when it comes to generating challenging augmentations and thus enhancing downstream tasks beyond classification. Like any other generative model, Ge NIe can also introduce inherent biases stemming from the training data used to build its diffusion backbone, which can reflect and amplify societal prejudices or inaccuracies. Therefore, it is crucial to carefully mitigate potential biases in generative models such as Ge NIe to ensure a fair and ethical deployment of deep learning systems. Arman Afrasiyabi, Jean-François Lalonde, and Christian Gagné. Associative alignment for few-shot image classification. In ECCV, 2019. Jean-Baptiste Alayrac, Jeff Donahue, Pauline Luc, Antoine Miech, Iain Barr, Yana Hasson, Karel Lenc, Arthur Mensch, Katie Millican, Malcolm Reynolds, Roman Ring, Eliza Rutherford, Serkan Cabi, Tengda Han, Zhitao Gong, Sina Samangooei, Marianne Monteiro, Jacob Menick, Sebastian Borgeaud, Andrew Brock, Aida Nematzadeh, Sahand Sharifzadeh, Mikolaj Binkowski, Ricardo Barreira, Oriol Vinyals, Andrew Zisserman, and Karen Simonyan. Flamingo: a visual language model for few-shot learning, 2022. Antreas Antoniou and Amos Storkey. Assume, augment and learn: Unsupervised few-shot meta-learning via random labels and data augmentation. arxiv:1902.09884, 2019. Shekoofeh Azizi, Simon Kornblith, Chitwan Saharia, Mohammad Norouzi, and David J. Fleet. Synthetic data from diffusion models improves imagenet classification, 2023. Lukas Bossard, Matthieu Guillaumin, and Luc Van Gool. Food-101 mining discriminative components with random forests. In European Conference on Computer Vision, 2014. Jiarui Cai, Yizhou Wang, Jenq-Neng Hwang, et al. Ace: Ally complementary experts for solving long-tailed recognition in one-shot. In ICCV, pp. 112 121, 2021. Kaidi Cao, Colin Wei, Adrien Gaidon, Nikos Arechiga, and Tengyu Ma. Learning imbalanced datasets with label-distribution-aware margin loss. Neur IPS, 32, 2019. Atoosa Chegini and Soheil Feizi. Identifying and mitigating model failures through few-shot clip-aided diffusion generation. ar Xiv preprint ar Xiv:2312.05464, 2023. Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. A simple framework for contrastive learning of visual representations. In ICML, 2020. Published in Transactions on Machine Learning Research (04/2025) Wei-Yu Chen, Yen-Cheng Liu, Zsolt Kira, Yu-Chiang Frank Wang, and Jia-Bin Huang. A closer look at few-shot classification. In ICLR, 2019a. Wentao Chen, Chenyang Si, Wei Wang, Liang Wang, Zilei Wang, and Tieniu Tan. Few-shot learning with part discovery and augmentation from unlabeled images. ar Xiv preprint ar Xiv:2105.11874, 2021. Xinlei Chen and Kaiming He. Exploring simple siamese representation learning. In CVPR, 2021. Zhengyu Chen, Jixie Ge, Heshen Zhan, Siteng Huang, and Donglin Wang. Pareto self-supervised training for few-shot learning. In CVPR, 2021. Zitian Chen, Yanwei Fu, Yu-Xiong Wang, Lin Ma, Wei Liu, and Martial Hebert. Image deformation metanetworks for one-shot learning. In CVPR, 2019b. Ekin D. Cubuk, Barret Zoph, Dandelion Mane, Vijay Vasudevan, and Quoc V. Le. Autoaugment: Learning augmentation policies from data, 2019a. Ekin D. Cubuk, Barret Zoph, Jonathon Shlens, and Quoc V. Le. Randaugment: Practical automated data augmentation with a reduced search space, 2019b. Ekin Dogus Cubuk, Barret Zoph, Jon Shlens, and Quoc Le. Randaugment: Practical automated data augmentation with a reduced search space. In H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin (eds.), Advances in Neural Information Processing Systems, volume 33, pp. 18613 18624. Curran Associates, Inc., 2020. URL https://proceedings.neurips.cc/paper/2020/file/ d85b63ef0ccb114d0a3bb7b7d808028f-Paper.pdf. Jiequan Cui, Zhisheng Zhong, Shu Liu, Bei Yu, and Jiaya Jia. Parametric contrastive learning. In ICCV, pp. 715 724, 2021a. Jiequan Cui, Zhisheng Zhong, Shu Liu, Bei Yu, and Jiaya Jia. Parametric contrastive learning. In Proceedings of the IEEE/CVF international conference on computer vision, pp. 715 724, 2021b. Jiequan Cui, Shu Liu, Zhuotao Tian, Zhisheng Zhong, and Jiaya Jia. Reslt: Residual learning for long-tailed recognition. IEEE transactions on pattern analysis and machine intelligence, 45(3):3695 3706, 2022. Yin Cui, Menglin Jia, Tsung-Yi Lin, Yang Song, and Serge Belongie. Class-balanced loss based on effective number of samples. In CVPR, pp. 9268 9277, 2019. Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pp. 248 255. Ieee, 2009. Guneet S. Dhillon, Pratik Chaudhari, Avinash Ravichandran, and Stefano Soatto. A baseline for few-shot image classification. In ICLR, 2020. Mucong Ding, Bang An, Yuancheng Xu, Anirudh Satheesh, and Furong Huang. SAFLEX: Self-adaptive augmentation via feature label extrapolation. In The Twelfth International Conference on Learning Representations, 2024. URL https://openreview.net/forum?id=q L6brr BDk2. Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, Jakob Uszkoreit, and Neil Houlsby. An image is worth 16x16 words: Transformers for image recognition at scale. In ICLR, 2021. Lisa Dunlap, Clara Mohri, Han Zhang, Devin Guillory, Trevor Darrell, Joseph E. Gonzalez, Anna Rohrbach, and Aditi Raghunathan. Using language to extend to unseen domains. International Conference on Learning Representations (ICLR), 2023a. Lisa Dunlap, Alyssa Umino, Han Zhang, Jiezhi Yang, Joseph E. Gonzalez, and Trevor Darrell. Diversify your vision datasets with automatic diffusion-based augmentation, 2023b. Published in Transactions on Machine Learning Research (04/2025) Nikita Dvornik, Julien Mairal, and Cordelia Schmid. Diversity with cooperation: Ensemble methods for few-shot classification. In ICCV, 2019. Chun-Mei Feng, Kai Yu, Yong Liu, Salman Khan, and Wangmeng Zuo. Diverse data augmentation with diffusions for effective test-time prompt tuning, 2023. Chelsea Finn, Pieter Abbeel, and Sergey Levine. Model-agnostic meta-learning for fast adaptation of deep networks. In Proceedings of the 34th International Conference on Machine Learning, pp. 1126 1135, 2017. Maayan Frid-Adar, Idit Diamant, Eyal Klang, Michal Amitai, Jacob Goldberger, and Hayit Greenspan. Ganbased synthetic medical image augmentation for increased cnn performance in liver lesion classification. Neurocomputing, 2018. Rinon Gal, Yuval Alaluf, Yuval Atzmon, Or Patashnik, Amit H Bermano, Gal Chechik, and Daniel Cohen Or. An image is worth one word: Personalizing text-to-image generation using textual inversion. ar Xiv preprint ar Xiv:2208.01618, 2022a. Rinon Gal, Yuval Alaluf, Yuval Atzmon, Or Patashnik, Amit H. Bermano, Gal Chechik, and Daniel Cohen Or. An image is worth one word: Personalizing text-to-image generation using textual inversion, 2022b. URL https://arxiv.org/abs/2208.01618. Rinon Gal, Yuval Alaluf, Yuval Atzmon, Or Patashnik, Amit H. Bermano, Gal Chechik, and Daniel Cohen Or. An image is worth one word: Personalizing text-to-image generation using textual inversion, 2022c. URL https://arxiv.org/abs/2208.01618. Alexandros Graikos, Srikar Yellapragada, and Dimitris Samaras. Conditional generation from pre-trained diffusion models using denoiser representations. In 34th British Machine Vision Conference 2023, BMVC 2023, Aberdeen, UK, November 20-24, 2023. BMVA, 2023a. URL https://papers.bmvc2023.org/0478. pdf. Alexandros Graikos, Srikar Yellapragada, and Dimitris Samaras. Conditional generation from unconditional diffusion models using denoiser representations. ar Xiv preprint ar Xiv:2306.01900, 2023b. Kaiming He, Xinlei Chen, Saining Xie, Yanghao Li, Piotr Dollár, and Ross Girshick. Masked autoencoders are scalable vision learners, 2021. Kaiming He, Xinlei Chen, Saining Xie, Yanghao Li, Piotr Dollár, and Ross B. Girshick. Masked autoencoders are scalable vision learners. In CVPR, pp. 15979 15988. IEEE, 2022a. Ruifei He, Shuyang Sun, Xin Yu, Chuhui Xue, Wenqing Zhang, Philip Torr, Song Bai, and Xiaojuan Qi. Is synthetic data from generative models ready for image recognition? ar Xiv preprint ar Xiv:2210.07574, 2022b. Reyhane Askari Hemmat, Mohammad Pezeshki, Florian Bordes, Michal Drozdzal, and Adriana Romero Soriano. Feedback-guided data synthesis for imbalanced classification, 2023. Dan Hendrycks, Norman Mu, Ekin D. Cubuk, Barret Zoph, Justin Gilmer, and Balaji Lakshminarayanan. Aug Mix: A simple data processing method to improve robustness and uncertainty. Proceedings of the International Conference on Learning Representations (ICLR), 2020. Yan Hong, Jianfu Zhang, Zhongyi Sun, and Ke Yan. Safa: Sample-adaptive feature augmentation for long-tailed image classification. In ECCV, 2022. Kyle Hsu, Sergey Levine, and Chelsea Finn. Unsupervised learning via meta-learning. In ICLR, 2018. Wentao Hu, Xiurong Jiang, Jiarun Liu, Yuqi Yang, and Hui Tian. Meta-dm: Applications of diffusion models on few-shot learning, 2023. Sheng-Wei Huang, Che-Tsung Lin, Shu-Ping Chen, Yen-Yi Wu an Po-Hao Hsu, and Shang-Hong Lai. Auggan: Cross domain adaptation with gan-based data augmentation. European Conference on Computer Vision, 2018. Published in Transactions on Machine Learning Research (04/2025) Saachi Jain, Hannah Lawrence, Ankur Moitra, and Aleksander Madry. Distilling model failures as directions in latent space, 2022. Huiwon Jang, Hankook Lee, and Jinwoo Shin. Unsupervised meta-learning via few-shot pseudo-supervised contrastive learning. In The Eleventh International Conference on Learning Representations, 2022. Bingyi Kang, Saining Xie, Marcus Rohrbach, Zhicheng Yan, Albert Gordo, Jiashi Feng, and Yannis Kalantidis. Decoupling representation and classifier for long-tailed recognition. ar Xiv preprint ar Xiv:1910.09217, 2019. Bingyi Kang, Saining Xie, Marcus Rohrbach, Zhicheng Yan, Albert Gordo, Jiashi Feng, and Yannis Kalantidis. Decoupling representation and classifier for long-tailed recognition. In ICLR, 2020. Siavash Khodadadeh, Ladislau Boloni, and Mubarak Shah. Unsupervised meta-learning for few-shot image classification. In Neur IPS, 2019. Jang-Hyun Kim, Wonho Choo, and Hyun Oh Song. Puzzle mix: Exploiting saliency and local statistics for optimal mixup. In International Conference on Machine Learning, pp. 5275 5285. PMLR, 2020. Jonathan Krause, Michael Stark, Jia Deng, and Li Fei-Fei. 3D object representations for fine-grained categorization. In Workshop on 3D Representation and Recognition, Sydney, Australia, 2013. Alexander C. Li, Mihir Prabhudesai, Shivam Duggal, Ellis Brown, and Deepak Pathak. Your diffusion model is secretly a zero-shot classifier, 2023. Bolian Li, Zongbo Han, Haining Li, Huazhu Fu, and Changqing Zhang. Trustworthy long-tailed classification. In CVPR, pp. 6970 6979, 2022a. Daiqing Li, Huan Ling, Seung Wook Kim, Karsten Kreis, Adela Barriuso, Sanja Fidler, and Antonio Torralba. Bigdatasetgan: Synthesizing imagenet with pixel-wise annotations, 2022b. Jun Li, Zichang Tan, Jun Wan, Zhen Lei, and Guodong Guo. Nested collaborative learning for long-tailed visual recognition. In CVPR, pp. 6949 6958, 2022c. Junnan Li, Dongxu Li, Caiming Xiong, and Steven Hoi. Blip: Bootstrapping language-image pre-training for unified vision-language understanding and generation, 2022d. Kai Li, Yulun Zhang, Kunpeng Li, and Yun Fu. Adversarial feature hallucination networks for few-shot learning. In CVPR, 2020. Mengke Li, Yiu-ming Cheung, Yang Lu, et al. Long-tailed visual recognition via gaussian clouded logit adjustment. In CVPR, pp. 6929 6938, 2022e. Tianhong Li, Peng Cao, Yuan Yuan, Lijie Fan, Yuzhe Yang, Rogerio S Feris, Piotr Indyk, and Dina Katabi. Targeted supervised contrastive learning for long-tailed recognition. In CVPR, pp. 6918 6928, 2022f. Bin Liu, Yue Cao, Yutong Lin, Qi Li, Zheng Zhang, Mingsheng Long, and Han Hu. Negative margin matters: Understanding margin in few-shot classification. In ECCV, 2020. Xingchao Liu, Xiwen Zhang, Jianzhu Ma, Jian Peng, et al. Instaflow: One step is enough for high-quality diffusion-based text-to-image generation. In The Twelfth International Conference on Learning Representations, 2023. Zicheng Liu, Siyuan Li, Di Wu, Zihan Liu, Zhiyuan Chen, Lirong Wu, and Stan Z Li. Automix: Unveiling the power of mixup for stronger classifiers. In Computer Vision ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23 27, 2022, Proceedings, Part XXIV, pp. 441 458. Springer, 2022. Ziwei Liu, Zhongqi Miao, Xiaohang Zhan, Jiayun Wang, Boqing Gong, and Stella X. Yu. Large-scale longtailed recognition in an open world. In CVPR, 2019. Published in Transactions on Machine Learning Research (04/2025) Yuning Lu, Liangjian Wen, Jianzhuang Liu, Yajing Liu, and Xinmei Tian. Self-supervision can be a good few-shot learner. In European Conference on Computer Vision, pp. 740 758. Springer, 2022. Xue-Jing Luo, Shuo Wang, Zongwei Wu, Christos Sakaridis, Yun Cheng, Deng-Ping Fan, and Luc Van Gool. Camdiff: Camouflage image augmentation via diffusion model, 2023. Lorenzo Luzi, Ali Siahkoohi, Paul M Mayer, Josue Casco-Rodriguez, and Richard Baraniuk. Boomerang: Local sampling on image manifolds using diffusion models, 2022. Avery Ma, Nikita Dvornik, Ran Zhang, Leila Pishdad, Konstantinos G Derpanis, and Afsaneh Fazly. Sage: Saliency-guided mixup with optimal rearrangements. ar Xiv preprint ar Xiv:2211.00113, 2022. Subhransu Maji, Esa Rahtu, Juho Kannala, Matthew B. Blaschko, and Andrea Vedaldi. Fine-grained visual classification of aircraft. ar Xiv preprint ar Xiv:1306.5151, 2013. Jiayuan Mao, Tete Xiao, Yuning Jiang, and Zhimin Cao. What can help pedestrian detection?, 2017. Carlos Medina, Arnout Devos, and Matthias Grossglauser. Self-supervised prototypical transfer learning for few-shot classification. In ICMLW, 2020. Chenlin Meng, Yutong He, Yang Song, Jiaming Song, Jiajun Wu, Jun-Yan Zhu, and Stefano Ermon. Sdedit: Guided image synthesis and editing with stochastic differential equations. ar Xiv preprint ar Xiv:2108.01073, 2021. Chenlin Meng, Robin Rombach, Ruiqi Gao, Diederik Kingma, Stefano Ermon, Jonathan Ho, and Tim Salimans. On distillation of guided diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 14297 14306, 2023. Aditya Krishna Menon, Sadeep Jayasumana, Ankit Singh Rawat, Himanshu Jain, Andreas Veit, and Sanjiv Kumar. Long-tail learning via logit adjustment. In ICLR, 2021. Maxime Oquab, Timothée Darcet, Théo Moutakanni, Huy Vo, Marc Szafraniec, Vasil Khalidov, Pierre Fernandez, Daniel Haziza, Francisco Massa, Alaaeldin El-Nouby, Mahmoud Assran, Nicolas Ballas, Wojciech Galuba, Russell Howes, Po-Yao Huang, Shang-Wen Li, Ishan Misra, Michael Rabbat, Vasu Sharma, Gabriel Synnaeve, Hu Xu, Hervé Jegou, Julien Mairal, Patrick Labatut, Armand Joulin, and Piotr Bojanowski. Dinov2: Learning robust visual features without supervision, 2023. William Peebles, Jun-Yan Zhu, Richard Zhang, Antonio Torralba, Alexei Efros, and Eli Shechtman. Gansupervised dense visual alignment. In CVPR, 2022. Suzanne Petryk, Lisa Dunlap, Keyan Nasseri, Joseph Gonzalez, Trevor Darrell, and Anna Rohrbach. On guiding visual attention with language specification. In Conference on Computer Vision and Pattern Recognition (CVPR), 2022. doi: 10.48550/ARXIV.2202.08926. URL https://arxiv.org/abs/2202. 08926. Viraj Prabhu, Sriram Yenamandra, Prithvijit Chattopadhyay, and Judy Hoffman. Lance: Stress-testing visual models by generating language-guided counterfactual images. Advances in Neural Information Processing Systems, 36, 2024. Siyuan Qiao, Chenxi Liu, Wei Shen, and Alan Yuille. Few-shot image recognition by predicting parameters from activations. In CVPR, 2018. Tiexin Qin, Wenbin Li, Yinghuan Shi, and Gao Yang. Unsupervised few-shot learning via distribution shift-based augmentation. arxiv:2004.05805, 2020. Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language supervision. In International Conference on Machine Learning, pp. 8748 8763. PMLR, 2021. Published in Transactions on Machine Learning Research (04/2025) Aditya Ramesh, Mikhail Pavlov, Gabriel Goh, Scott Gray, Chelsea Voss, Alec Radford, Mark Chen, and Ilya Sutskever. Zero-shot text-to-image generation. In International Conference on Machine Learning, pp. 8821 8831. PMLR, 2021. Aditya Ramesh, Prafulla Dhariwal, Alex Nichol, Casey Chu, and Mark Chen. Hierarchical text-conditional image generation with clip latents. ar Xiv preprint ar Xiv:2204.06125, 2022. Sachin Ravi and H. Larochelle. Optimization as a model for few-shot learning. In ICLR, 2017. Mengye Ren, Sachin Ravi, Eleni Triantafillou, Jake Snell, Kevin Swersky, Josh B. Tenenbaum, Hugo Larochelle, and Richard S. Zemel. Meta-learning for semi-supervised few-shot classification. In International Conference on Learning Representations, 2018. URL https://openreview.net/forum?id= HJc Szz-CZ. Robin Rombach, Andreas Blattmann, Dominik Lorenz, Patrick Esser, and Björn Ommer. High-resolution image synthesis with latent diffusion models, 2021a. Robin Rombach, Andreas Blattmann, Dominik Lorenz, Patrick Esser, and Björn Ommer. High-resolution image synthesis with latent diffusion models, 2021b. Aniket Roy, Anshul Shah, Ketul Shah, Anirban Roy, and Rama Chellappa. Cap2aug: Caption guided image to image data augmentation, 2023. Chitwan Saharia, William Chan, Saurabh Saxena, Lala Li, Jay Whang, Emily L Denton, Kamyar Ghasemipour, Raphael Gontijo Lopes, Burcu Karagol Ayan, Tim Salimans, et al. Photorealistic textto-image diffusion models with deep language understanding. Advances in Neural Information Processing Systems, 35:36479 36494, 2022. Dvir Samuel and Gal Chechik. Distributional robustness loss for long-tail learning. In ICCV, 2021. Swami Sankaranarayanan, Yogesh Balaji, Carlos Domingo Castillo, and Rama Chellappa. Generate to adapt: Aligning domains using generative adversarial networks. Conference on Computer Vision and Pattern Recognition (CVPR), 2018. Axel Sauer, Dominik Lorenz, Andreas Blattmann, and Robin Rombach. Adversarial diffusion distillation. ar Xiv preprint ar Xiv:2311.17042, 2023. Christoph Schuhmann, Romain Beaumont, Richard Vencu, Cade Gordon, Ross Wightman, Mehdi Cherti, Theo Coombes, Aarush Katta, Clayton Mullis, Mitchell Wortsman, Patrick Schramowski, Srivatsa Kundurthy, Katherine Crowson, Ludwig Schmidt, Robert Kaczmarczyk, and Jenia Jitsev. Laion-5b: An open large-scale dataset for training next generation image-text models, 2022. Antonio Sclocchi, Alessandro Favero, and Matthieu Wyart. A phase transition in diffusion models reveals the hierarchical nature of data, 2024. Antonio Sclocchi, Alessandro Favero, and Matthieu Wyart. A phase transition in diffusion models reveals the hierarchical nature of data. Proceedings of the National Academy of Sciences, 122(1):e2408799121, 2025. Viktoriia Sharmanska, Lisa Anne Hendricks, Trevor Darrell, and Novi Quadrianto. Contrastive examples for addressing the tyranny of the majority. Co RR, abs/2004.06524, 2020. URL https://arxiv.org/abs/ 2004.06524. Jordan Shipard, Arnold Wiliem, Kien Nguyen Thanh, Wei Xiang, and Clinton Fookes. Boosting zero-shot classification with synthetic data diversity via stable diffusion. ar Xiv preprint ar Xiv:2302.03298, 2023. Ojas Kishorkumar Shirekar, Anuj Singh, and Hadi Jamali-Rad. Self-attention message passing for contrastive few-shot learning. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), pp. 5426 5436, January 2023. Published in Transactions on Machine Learning Research (04/2025) Connor Shorten and Taghi M Khoshgoftaar. A survey on image data augmentation for deep learning. Journal of big data, 6(1):1 48, 2019. Anuj Rajeeva Singh and Hadi Jamali-Rad. Transductive decoupled variational inference for few-shot classification. Transactions on Machine Learning Research, 2023. ISSN 2835-8856. URL https: //openreview.net/forum?id=bomd Tc9Hy L. Jake Snell, Kevin Swersky, and Richard Zemel. Prototypical networks for few-shot learning. In Advances in Neural Information Processing Systems, 2017. Jong-Chyi Su, Subhransu Maji, and Bharath Hariharan. When does self-supervision improve few-shot learning? In ECCV, 2020. Flood Sung, Yongxin Yang, Li Zhang, Tao Xiang, Philip H.S. Torr, and Timothy M. Hospedales. Learning to compare: Relation network for few-shot learning. In CVPR, 2018. Kaihua Tang, Jianqiang Huang, and Hanwang Zhang. Long-tailed classification by keeping the good and removing the bad momentum causal effect. Neur IPS, 33:1513 1524, 2020. Kowshik Thopalli, Rakshith Subramanyam, Pavan Turaga, and Jayaraman J. Thiagarajan. Target-aware generative augmentations for single-shot adaptation. In Proceedings of the 40th International Conference on Machine Learning, ICML 23. JMLR.org, 2023. Changyao Tian, Wenhai Wang, Xizhou Zhu, Jifeng Dai, and Yu Qiao. Vl-ltr: Learning class-wise visuallinguistic representation for long-tailed visual recognition. In ECCV 2022, 2022. Yonglong Tian, Lijie Fan, Phillip Isola, Huiwen Chang, and Dilip Krishnan. Stablerep: Synthetic images from text-to-image models make strong visual representation learners, 2023. URL https://arxiv.org/ abs/2306.00984. Hugo Touvron, Matthieu Cord, Matthijs Douze, Francisco Massa, Alexandre Sablayrolles, and Hervé Jégou. Training data-efficient image transformers & distillation through attention. In International Conference on Machine Learning (ICML), 2021a. Hugo Touvron, Matthieu Cord, Matthijs Douze, Francisco Massa, Alexandre Sablayrolles, and Hervé Jégou. Training data-efficient image transformers and distillation through attention, 2021b. Hugo Touvron, Matthieu Cord, and Hervé Jégou. Deit iii: Revenge of the vit. In ECCV, 2022. Brandon Trabucco, Kyle Doherty, Max A Gurinas, and Ruslan Salakhutdinov. Effective data augmentation with diffusion models. In The Twelfth International Conference on Learning Representations, 2024. URL https://openreview.net/forum?id=ZWz UA9ze Ag. Nontawat Tritrong, Pitchaporn Rewatbowornwong, and Supasorn Suwajanakorn. Repurposing gans for one-shot semantic part segmentation. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2021. Catherine Wah, Steve Branson, Peter Welinder, Pietro Perona, and Serge Belongie. The caltech-ucsd birds200-2011 dataset, 2011. Haoqing Wang and Zhi-Hong Deng. Contrastive prototypical network with wasserstein confidence penalty. In European Conference on Computer Vision, pp. 665 682. Springer, 2022. Hualiang Wang, Siming Fu, Xiaoxuan He, Hangxiang Fang, Zuozhu Liu, and Haoji Hu. Towards calibrated hyper-sphere representation via distribution overlap coefficient for long-tailed learning. In ECCV, 2022. Xudong Wang, Long Lian, Zhongqi Miao, Ziwei Liu, and Stella X. Yu. Long-tailed recognition by routing diverse distribution-aware experts. In ICLR. Open Review.net, 2021. Yue Xu, Yong-Lu Li, Jiefeng Li, and Cewu Lu. Constructing balance from imbalance for long-tailed image recognition. In ECCV, pp. 38 56. Springer, 2022. Published in Transactions on Machine Learning Research (04/2025) Zhengzhuo Xu, Ruikang Liu, Shuo Yang, Zenghao Chai, and Chun Yuan. Learning imbalanced data with vision transformers, 2023. Hong Xuan, Abby Stylianou, Xiaotong Liu, and Robert Pless. Hard negative examples are hard, but useful, 2021. Han-Jia Ye, Hexiang Hu, De-Chuan Zhan, and Fei Sha. Few-shot learning via embedding adaptation with set-to-set functions. In CVPR, 2020. Sangdoo Yun, Dongyoon Han, Seong Joon Oh, Sanghyuk Chun, Junsuk Choe, and Youngjoon Yoo. Cutmix: Regularization strategy to train strong classifiers with localizable features. In ICCV, pp. 6023 6032, 2019a. Sangdoo Yun, Dongyoon Han, Seong Joon Oh, Sanghyuk Chun, Junsuk Choe, and Youngjoon Yoo. Cutmix: Regularization strategy to train strong classifiers with localizable features. In International Conference on Computer Vision (ICCV), 2019b. Hongyi Zhang, Moustapha Cisse, Yann N. Dauphin, and David Lopez-Paz. mixup: Beyond empirical risk minimization. In ICLR, 2018. Songyang Zhang, Zeming Li, Shipeng Yan, Xuming He, and Jian Sun. Distribution alignment: A unified framework for long-tail visual recognition. In CVPR, pp. 2361 2370, 2021a. Yifan Zhang, Bryan Hooi, Lanqing Hong, and Jiashi Feng. Test-agnostic long-tailed recognition by test-time aggregating diverse experts with self-supervision. ar Xiv preprint ar Xiv:2107.09249, 2021b. Yuxuan Zhang, Huan Ling, Jun Gao, Kangxue Yin, Jean-Francois Lafleche, Adela Barriuso, Antonio Torralba, and Sanja Fidler. Datasetgan: Efficient labeled data factory with minimal human effort. In CVPR, 2021c. Zhisheng Zhong, Jiequan Cui, Shu Liu, and Jiaya Jia. Improving calibration for long-tailed recognition. In CVPR, pp. 16489 16498. Computer Vision Foundation / IEEE, 2021. Ziqi Zhou, Xi Qiu, Jiangtao Xie, Jianan Wu, and Chi Zhang. Binocular mutual learning for improving few-shot classification. In ICCV, 2021. Jianggang Zhu, Zheng Wang, Jingjing Chen, Yi-Ping Phoebe Chen, and Yu-Gang Jiang. Balanced contrastive learning for long-tailed visual recognition. In CVPR, pp. 6908 6917, 2022. Published in Transactions on Machine Learning Research (04/2025) A.1 Analyzing Ge NIe, Ge NIe-Ada s Class-Probabilities The core aim of Ge NIe and Ge NIe-Ada is to address the failure modes of a classifier by generating challenging samples located near the decision boundary of each class pair, which facilitates the learning process in effectively enhancing the decision boundary between classes. As summarized in Table 5 and illustrated in Fig. 5, we have empirically corroborated that Ge NIe and Ge NIe-Ada can respectively produce samples Xr, Xr that are negative with respect to the source image XS, while semantically belonging to the class T. To further analyze the effectiveness of Ge NIe and Ge NIe-Ada, we compare the Figure A1: P(YS|Xr) and P(YT |Xr) for r {0.5, 0.6, 0.7, 0.8, 0.9}. On average, the classifier confidently predicts the source class more than the target class for Xr for r = 0.5, and vice-versa for r = 0.8, 0.9. However, for r = 0.6, 0.7, the classifier struggles to classify Xr, indicating that the augmented samples are located closer to the decision boundary. source class-probabilities P(YS|Xr) and target-class probabilities P(YS|Xr) of augmented samples Xr. Figure A2: Significant overlap between P(YS|Xr ) and P(YT |Xr ) indicates high classconfusion for augmented samples generated by Ge NIe-Ada. To compute these class probabilities, we first fit an SVM classifier (as followed in Uni Siam (Lu et al., 2022)) only on the labelled support set embeddings of each episode in the mini Imagenet test dataset. Then, we perform inference using each episode s SVM classifier on its respective Xr s and extract its class probabilities of belonging to its source class S and target class T. These per augmentation-sample source and target class probabilities are then averaged for each episode for each r {0.5, 0.6, 0.7, 0.8, 0.9} in the case of Ge NIe and for the optimal r = r per sample in the case of Ge NIe-Ada, plotted as density plots in Fig. A1, Fig. A2, respectively. Fig. A1 illustrates that P(YS|Xr) and P(YT |Xr) have significant overlap in the case of r {0.6, 0.7} indicating class-confusion for Xr. Furthermore, Fig. A2 illustrates that when using the optimal r = r found by Ge NIe-Ada per sample, P(YS|Xr) and P(YT |Xr) significantly overlap around probability scores of 0.2 0.45, indicating class confusion for Ge NIe-Ada augmentations. This corroborates with our analysis in Section 4.4, Table 5 and additionally empirically proves that the augmented samples generated by Ge NIe for r {0.6, 0.7} and Ge NIe-Ada for r = r are actually located near the decision boundary of each class pair. A.2 Independence of Generated Augmentations from Downstream Test Sets Here we analyzed whether the augmented samples generated by Ge NIe using the diffusion model overlap with the test set of the downstream task. To set the stage, we extracted the latent embeddings corresponding to the train set (i.e., support), test set (i.e., query), and augmentations generated by Ge NIe. Fig A3 illustrates the distribution of distances between train-test and augmentation-test pairs across 600 episodes. Notably, the mean distance of augmentation-test pairs is higher than that of train-test pairs, indicating that the Published in Transactions on Machine Learning Research (04/2025) Figure A3: Comparison of embedding distributions and UMAP visualization for train, test, and Ge NIe-augmented samples. augmented samples are distinct from the test set. This observation aligns with the fundamental assumption of train and test sets being mutually exclusive. Additionally, Fig A3 provides further evidence through a UMAP embedding plot of a randomly selected episode, where the embeddings of train, test, and augmented samples are visualized. The plot reveals clear separations between the test set and augmented samples, further confirming that the augmented samples do not overlap with or resemble the test set in embedding space. These findings validate that the diffusion-generated augmentations are independent of the downstream task s test set, ensuring the integrity of the evaluation process. A.3 Computational Complexity of Ge NIe and Ge NIe-Ada Table A1: Computational Complexity Augmentation Inf. Steps Runtime [sec/img] Txt2Img T 4.12 Ge NIe(r=0.5) 0.5 T 2.17 Ge NIe(r=0.6) 0.6 T 2.59 Ge NIe(r=0.7) 0.7 T 2.98 Ge NIe(r=0.8) 0.8 T 3.46 Ge NIe-Ada 2.1 T 9.22 Here we provide details on the computational complexity of Ge NIe across multiple noising ratios r and Ge NIe-Ada when operating on a search space of r [0.6, 0.8]. Computational complexity has been reported in terms of the total number of inference/denoising-diffusion steps and the runtime in seconds per generated image. The runtime has been averaged over 10 different image-generations on an NVIDIA Tesla-V100 GPU with 16GB VRAM with 50 steps of denoising using a DPM scheduler with Stable Diffusion v1.5. As can be seen in Tab. A1, Ge NIe is approximately 1/r times faster than the base diffusion model (referred to as the Txt2Img augmentation baseline). This empirically corroborates with the total number of denoising steps using in Ge NIe vs. Txt2Img. Since, Ge NIe-Ada scans for the best hard-negative in r [0.6, 0.8], it incurs a computational cost of 2.2 the Txt2Img. Note that the runtime for Ge NIe-Ada reported in Tab. A1 also includes the runtime of performing a batched forward pass through a Res Net-50 feature extraction backbone. A.4 Additional Augmentation Comparisons We compute few-shot classification scores on mini-Image Net with additional combinations of traditional augmentations. We introduce a Mixed augmentation scheme where we use a combination of Weak and Strong augmentations together. We also experiment the scenario where Cut Mix and Mix Up are used alongside the Mixed augmentation strategy as indicated by Mixed+Cut Mix and Mixed+Mix Up. Finally, we experiment with a combination of Ge NIe along with Mix Up, similar to (Graikos et al., 2023b). As can be seen in Tab. A2, we notice marginal improvements of upto 0.6% by using the Mixed augmentations either with or without the Cut Mix, Mix Up counterparts. We also notice a drop in performance of upto 0.9% when Mix Up is used along with Ge NIe. This follows the general trend of drop in performance when using Cut Mix or Mix Up, as reported in Tab. 1. A.5 Comparison against large-scale pre-trained CLIP backbone We evaluate Ge NIe and Ge NIe-Ada with few-shot pre-trained Res Net-18,34,50 backbones to demonstrate that our hard-negative augmentations can boost downstream-task performance of any data-deficient pre- Published in Transactions on Machine Learning Research (04/2025) Table A2: mini-Image Net: We use our augmentations on (5-way, 1-shot) and (5-way, 5-shot) few-shot settings of mini-Imagenet dataset with 2 different backbones (Res Net-18 and 50). We compare with additional combinations of traditional augmentations, with and without Ge NIe. The number of generated images per class is 4 for 1-shot and 20 for 5-shot settings. Augmentation Method Pre-training 1-shot 5-shot Weak Uni Siam 2022 unsup. 63.1 0.8 81.4 0.5 Strong Uni Siam 2022 unsup. 62.8 0.8 81.2 0.6 Mixed Uni Siam 2022 unsup. 63.2 0.5 81.9 0.4 Cut Mix 2019a Uni Siam 2022 unsup. 62.7 0.8 80.6 0.6 Mix Up 2018 Uni Siam 2022 unsup. 62.1 0.8 80.7 0.6 Mixed+Mix Up 2018 Uni Siam 2022 unsup. 65.7 0.9 82.1 0.2 Mixed+Cut Mix 2018 Uni Siam 2022 unsup. 64.9 0.8 81.6 0.5 DAFusion 2024 Uni Siam 2022 unsup. 64.3 1.8 82.0 1.4 Ge NIe+Mix Up Uni Siam 2022 unsup. 74.8 0.5 84.5 0.3 Ge NIe (Ours) Uni Siam 2022 unsup. 75.5 0.6 85.4 0.4 Ge NIe-Ada (Ours) Uni Siam 2022 unsup. 76.8 0.6 85.9 0.4 Augmentation Method Pre-training 1-shot 5-shot Weak Uni Siam 2022 unsup. 64.6 0.8 83.4 0.5 Strong Uni Siam 2022 unsup. 64.8 0.8 83.2 0.5 Mixed Uni Siam 2022 unsup. 64.5 0.5 83.8 0.5 Cut Mix 2019a Uni Siam 2022 unsup. 64.3 0.8 83.2 0.5 Mix Up 2018 Uni Siam 2022 unsup. 63.8 0.8 84.6 0.5 Mixed+Mix Up 2018 Uni Siam 2022 unsup. 64.9 0.7 84.5 0.7 Mixed+Cut Mix 2018 Uni Siam 2022 unsup. 63.5 0.5 83.0 0.8 DAFusion 2024 Uni Siam 2022 unsup. 65.7 1.8 83.9 1.2 Ge NIe+Mix Up Uni Siam 2022 unsup. 76.4 0.5 85.9 0.7 Ge NIe (Ours) Uni Siam 2022 unsup. 77.3 0.6 87.2 0.4 Ge NIe-Ada (Ours) Uni Siam 2022 unsup. 78.6 0.6 87.9 0.4 trained backbones. In this subsection, we conduct an experiment where we use a CLIP (Radford et al., 2021) pre-trained Res Net50 backbone rather than a Uni Siam pre-trained Res Net50. We then evaluate if Ge NIe, Ge NIe-Ada can improve CLIP s few-shot accuracies to clarify whether the benefits of our approach come primarily from the generated negative samples or from the additional label information obtained by training on LAION. As can be seen in Table A3, we observe that while the CLIP backbone s performance appears saturated on this task ( 93% 1-shot accuracy), Ge NIe-Ada still achieves 1.3%, 0.5% improvements in the 1-shot, 5-shot settings, respectively. Table A3: Performance comparison against CLIP pre-trained Res Net-50. Augmentation Method Pre-training 1-shot 5-shot Mixed CLIP Res Net-50 unsup. 93.1 0.3 95.6 0.2 Ge NIe CLIP Res Net-50 unsup. 93.9 0.3 95.8 0.3 Ge NIe-Ada CLIP Res Net-50 unsup. 94.4 0.2 96.1 0.2 A.6 Effect of Linear Classification Head vs. SVM Following (Dhillon et al., 2020), we use an SVM classification head for few-shot and fine-grained datasets. However, our augmentation method is not restricted to this choice of classification head. In Tables A4, A5, we use a single layer linear classification (LL) head on top of the embedding extraction backbone, following (Dhillon et al., 2020). We notice that Ge NIe and Ge NIe-Ada offer similar on-par performances with SVM and linear-layer (LL) classification heads for 1-shot and 5-shot mini Imagenet classification tasks as well as on the fine-grained few-shot classification datasets. This corroborates Ge NIe s robustness to both - SVM and linear-layer fine-tuning based classification methods. Table A4: We use our augmentations on (5-way, 1-shot) and (5-way, 5-shot) few-shot settings of mini-Imagenet dataset with 2 different backbones (Res Net-18 and 50) with a linear-layer (LL) and SVM classification head. The number of generated images per class is 4 for 1-shot and 20 for 5-shot settings. Augmentation Method Pre-training 1-shot 5-shot Ge NIe + LL Uni Siam (2024) unsup. 75.9 0.3 85.1 0.7 Ge NIe-Ada + LL Uni Siam (2024) unsup. 76.9 0.5 85.7 0.6 Ge NIe + SVM Uni Siam (2024) unsup. 75.5 0.6 85.4 0.4 Ge NIe-Ada + SVM (Ours) Uni Siam (2024) unsup. 76.8 0.6 85.9 0.4 Augmentation Method Pre-training 1-shot 5-shot Ge NIe + LL Uni Siam (2024) unsup. 77.8 0.8 87.5 0.5 Ge NIe-Ada + LL Uni Siam (2024) unsup. 78.1 0.5 87.8 0.2 Ge NIe + SVM Uni Siam (2024) unsup. 77.3 0.6 87.2 0.4 Ge NIe-Ada + SVM Uni Siam (2024) unsup. 78.6 0.6 87.9 0.4 A.7 Effect of Backbone for Noise Ratio Selector in Ge NIe-Ada To analyze the effect of the backbone feature extractor fθ on selecting the optimal hard-negative using Ge NIe-Ada, we use a pre-trained Dei T-B (Touvron et al., 2021a) instead of the Uni Siam pretrained Res Net backbone. However, we still utilize the same Res Net backbone for few-shot classification. As shown in Tab. A6, we notice a marginal improvement of upto 0.7% when using Ge NIe-Ada+Dei T-B as compared to Ge NIe-Ada which uses the Uni Siam pre-trained Res Net backbone. This suggests that there is still potential Published in Transactions on Machine Learning Research (04/2025) Table A5: Few-shot accuracies with linear-layer (LL) and SVM classification heads on fine-grained datasets. Method Birds (CUB200) Cars (Cars196) Foods (Food101) Aircraft Baseline 90.3 49.8 82.9 29.2 Img2Img L 90.7 50.4 87.4 31.0 Img2Img H 91.3 56.4 91.7 34.7 Txt2Img 92.0 81.3 93.0 41.7 Ge NIe (r=0.5)-SVM 92.0 84.6 91.5 39.8 Ge NIe (r=0.6)-SVM 92.2 87.1 92.5 45.0 Ge NIe (r=0.7)-SVM 92.5 87.9 92.9 47.0 Ge NIe (r=0.8)-SVM 92.5 87.7 93.1 46.5 Ge NIe (r=0.9)-SVM 92.4 87.1 93.1 45.7 Ge NIe-Ada-SVM 92.6 87.9 93.1 46.9 Ge NIe (r=0.5)-LL 92.4 84.1 91.4 40.1 Ge NIe (r=0.6)-LL 91.8 87.5 92.1 44.7 Ge NIe (r=0.7)-LL 92.9 88.2 92.4 47.2 Ge NIe (r=0.8)-LL 92.3 87.9 92.7 46.4 Ge NIe (r=0.9)-LL 92.8 87.6 93.5 46.2 Ge NIe-Ada-LL 92.1 87.4 93.7 46.3 to develop more effective strategies for selecting noise ratios to further enhance Ge NIe. However, in this paper, we limit our exploration to Ge NIe-Ada and leave these improvements for future work. Table A6: Effect of Backbone for Noise Ratio Selector in Ge NIe-Ada: We evaluate the impact of the noise ratio selector used in Ge NIe-Ada (fθ(.)). Note that in all experiments presented in this paper, we use the same backbone for fθ(.) that is subsequently fine-tuned for few-shot classification tasks. However, to analyze the effect of fθ(.) on sampled augmentations, we replace it with a more powerful backbone, specifically Dei T-B pretrained on Image Net1K. It is important to note that this is not a practical assumption; if Dei T-B were available for noise selection, it could also be used as the classifier in few-shot experiments, outperforming the weaker backbones employed in our study. Nevertheless, this experiment demonstrates that using a stronger backbone can result in more accurate selection of augmentations in Ge NIe, thereby enhancing the final accuracy. To clarify, Dei T-B is utilized solely as fθ(.) for sampling augmentations and not as the classifier. Therefore, the observed improvement is attributed exclusively to better augmentation sampling. Augmentation Noise Ratio Selector Method 1-shot 5-shot Backbone fθ(.) [Classifier Backbone] Ge NIe (Ours) - Uni Siam[Res Net18] 75.5 0.6 85.4 0.4 Ge NIe-Ada Uni Siam[Res Net18] Uni Siam[Res Net18] 76.8 0.6 85.9 0.4 Ge NIe-Ada IN-1K[Dei T-B] Uni Siam[Res Net18] 77.5 0.5 86.3 0.2 Augmentation Noise Selector Method 1-shot 5-shot Backbone fθ(.) [Classifier Backbone] Ge NIe - Uni Siam[Res Net50] 77.3 0.6 87.2 0.4 Ge NIe-Ada Uni Siam[Res Net50] Uni Siam[Res Net50] 78.6 0.6 87.9 0.4 Ge NIe-Ada IN-1K[Dei T-B] Uni Siam[Res Net50] 79.2 0.4 88.3 0.5 A.8 Impact of Ge NIe with Fine-Tuning: For all our experiments regarding Ge NIe and Ge NIe-Ada, we assume that the base diffusion model is aware/has been trained on some samples of the target class. This facilitates the addition of the target class (input as text-prompt) into the generated augmentation, while retaining the low-level features of the source image through partial noising. However, there can be a scenario where the base diffusion model does not understand the contradictory text prompt and thus fails to incorporate it into the generated image. As a solution, we can use textual inversion (Gal et al., 2022b) to fine-tune the diffusion model on few images belonging the unknown target class to learn the corresponding embeddings for the target categories. This fine-tuning allows us to learn embeddings specific to the target class, enabling the generation of the desired hard-negative examples. To empirically demonstrate the robustness of Ge NIe on these scenarios, we present few-shot classification results on mini-Imagenet using Ge NIe hard-negative augmentations in Tab. A7, generated by textual-inversion fine-tuning the diffusion model on images of the target class. Note that once the diffusion model is fine-tuned, the procedure to generate hard-negatives using partial noising and a contradictory text-prompt remains the same. As can be seen in Tab. A7, Ge NIe+Txt Inv performs significantly better than DAFusion baseline. It is important to note that, in this case, we do not utilize any information about the target category labels. DAFusion also employs textual-inversion-based fine-tuning; however, it does so without generating hard-negative samples. This indicates that Ge NIe is effective even in scenario where the diffusion model is unaware of the target-class. Published in Transactions on Machine Learning Research (04/2025) Table A7: mini-Image Net: We use our augmentations on (5-way, 1-shot) and (5-way, 5-shot) few-shot settings of mini-Imagenet dataset with 2 different backbones (Res Net-18 and 50), by using Textual-Inversion (Gal et al., 2022b) on the target-classes. The number of generated images per class is 4 for 1-shot and 20 for 5-shot settings. Augmentation Method Pre-training 1-shot 5-shot DAFusion 2024 Uni Siam 2022 unsup. 64.3 1.8 82.0 1.4 Ge NIe+Txt Inv Uni Siam 2022 unsup. 73.9 0.8 84.6 0.9 Augmentation Method Pre-training 1-shot 5-shot DAFusion 2024 Uni Siam 2022 unsup. 65.7 1.8 83.9 1.2 Ge NIe+Txt Inv Uni Siam 2022 unsup. 76.2 1.2 86.2 0.9 A.9 Details of Fine-grained Few-shot Classification Here we provide details of Fine-grained Few-shot Classification experiments. Datasets: We assess our method on several datasets: Food101 (Bossard et al., 2014) with 101 classes of foods, CUB200 (Wah et al., 2011) with 200 bird species classes, Cars196 (Krause et al., 2013) with 196 car model classes, and FGVC-Aircraft (Maji et al., 2013) with 41 aircraft manufacturer classes. We provide detailed information around fine-grained datasets in Table A8. The reported metric is the average Top-1 accuracy over 100 episodes. Each episode involves sampling 20 classes and 1-shot from the training set, with the final model evaluated on the respective test set. Implementation Details: We enhance the basic prompt by incorporating the superclass name for the finegrained dataset: A photo of a , a type of ". For instance, in the food dataset and the burger class, our prompt reads: A photo of a burger, a type of food." No additional augmentation is used for generative methods in this context. We generate 19 samples for both cases of our method and also the baseline with weak augmentation. Table A8: Train and test split details of the fine-grained datasets. We use the provided train set for few-shot task generation, and the provided test sets for our evaluation. Aircraft dataset uses the manufacturer hierarchy. Dataset Classes Train Test samples samples CUB200 (Wah et al., 2011) 200 5994 5794 Food101 (Bossard et al., 2014) 101 75750 25250 Cars (Krause et al., 2013) 196 8144 8041 Aircraft (Maji et al., 2013) 41 6,667 3333 A.10 Few-shot Classification with Res Net-34 on tiered Imagenet Table A9: tiered-Image Net: Accuracies (% std) for 5-way, 1-shot and 5-way, 5-shot classification settings on the test-set. We compare against various SOTA supervised and unsupervised few-shot classification baselines as well as other augmentation methods, with Uni Siam (Lu et al., 2022) pre-trained Res Net-34 backbone. Augmentation Method Pre-training 1-shot 5-shot Weak MAML + dist (Finn et al., 2017) sup. 51.7 1.8 70.3 1.7 Weak Proto Net (Snell et al., 2017) sup. 52.0 1.2 72.1 1.5 Weak Uni Siam + dist (Lu et al., 2022) unsup. 68.7 0.4 85.7 0.3 Weak Uni Siam (Lu et al., 2022) unsup. 65.0 0.7 82.5 0.5 Strong Uni Siam (Lu et al., 2022) unsup. 64.8 0.7 82.4 0.5 Cut Mix (Yun et al., 2019a) Uni Siam (Lu et al., 2022) unsup. 63.8 0.7 80.3 0.6 Mix Up (Zhang et al., 2018) Uni Siam (Lu et al., 2022) unsup. 64.1 0.7 80.0 0.6 Img2Img L(Luzi et al., 2022) Uni Siam (Lu et al., 2022) unsup. 66.1 0.7 83.1 0.5 Img2Img H(Luzi et al., 2022) Uni Siam (Lu et al., 2022) unsup. 70.4 0.7 84.7 0.5 Txt2Img(He et al., 2022b) Uni Siam (Lu et al., 2022) unsup. 75.0 0.6 85.4 0.4 DAFusion (Trabucco et al., 2024) Uni Siam (Lu et al., 2022) unsup. 64.1 2.1 82.8 1.4 Ge NIe (Ours) Uni Siam (Lu et al., 2022) unsup. 75.7 0.6 86.0 0.4 Ge NIe-Ada (Ours) Uni Siam (Lu et al., 2022) unsup. 76.9 0.6 86.3 0.2 We follow the same evaluation protocol here as mentioned in section 4.1. As summarized in Table A9, Ge NIe and Ge NIe-Ada outperform all other data augmentation techniques. Published in Transactions on Machine Learning Research (04/2025) Table A10: Few-shot classification comparison of Ge NIe-Ada with Txt2Img on mini Imagenet. Method Res Net-18 Res Net-34 Res Net-50 1-shot 5-shot 1-shot 5-shot 1-shot 5-shot Txt2Img 76.9 1.0 86.5 0.9 77.1 0.8 86.7 1.0 77.2 1.3 86.8 0.9 Ge NIe-Ada 77.7 0.8 87.4 1.0 78.3 0.9 87.8 0.9 79.1 1.1 88.4 1.2 Figure A4: Key components of Ge NIe: (i) careful choice of r and (ii) contradictory prompt are two key idea behind Ge NIe Figure A5: Analyzing the semantic trajectory of Ge NIe augmentations across 10 different images of class Mushroom (source image) to class Volcano (target class). A.11 Additional details of Long-Tail experiments We present a comprehensive version of Table 4 to benchmark the performance with different backbone architectures (e.g., Res Net50) and to compare against previous long-tail baselines; this is detailed in Table A11. Implementation Details of LVi T: We download the pre-trained Vi T-B of LVi T (Xu et al., 2023) and finetune it with Bal-BCE loss proposed therein on the augmented dataset. Training takes 2 hours on four NVIDIA RTX 3090 GPUs. We use the same hyperparameters as in (Xu et al., 2023) for finetuning: 100 epochs, lr = 0.008, batch size of 1024, Cut Mix and Mix Up for the data augmentation. Implementation Details of VL-LTR: We use the official code of VL-LTR (Tian et al., 2022) for our experiments. We use a pre-trained CLIP Res Net-50 backbone. We followed the hyperparameters reported in VL-LTR (Tian et al., 2022). We augment only Few category and train the backbone with the VL-LTR (Tian et al., 2022) method. Training takes 4 hours on 8 NVIDIA RTX 3090 GPUs. Published in Transactions on Machine Learning Research (04/2025) A.12 Extra Computation of Ge NIe-Ada Given that Ge NIe-Ada searches for the best hard-negative between multiple noise-ratios r s, it naturally requires a higher compute budget than txt2Img that only uses r = 1. For this experiment, we use Ge NIe Ada with r {0.6, 0.7, 0.8} to compare with Txt2Img. Based on this, we only have 3 paths, with steps of 0.1), and for each of which we go through partial reverse diffusion process. E.g. for r = 0.6 we do 30 steps instead of standard 50 steps of Stable Diffusion. This practically breaks down the total run-time of Ge NIe-Ada to approximately 2 times that of the standard reverse diffusion (Ge NIe-Ada: total r = 0.6 + 0.7 + 0.8 = 2.1 vs Txt2Img total r = 1). Thus, to be fair, we generate twice as many Txt2Img augmentations as compared to Ge NIe-Ada to keep a constant compute budget across the methods, following your suggestion. The results are shown in Table A10. As can be seen, even in this new setting, Ge NIe-Ada offers a performance improvement of 0.8% to 1.9% across different backbones. Table A11: Long-Tailed Image Net-LT: We compare different augmentation methods on Image Net-LT and report Top-1 accuracy for Few , Medium , and Many sets. indicates results with Res Ne Xt50. : indicates training with 384 resolution so is not directly comparable with other methods with 224 resolution. On the Few set and Li VT method, our augmentations improve the accuracy by 11.7 points compared to Li VT original augmentation and 4.4 points compared to Txt2Img. Method Many Med. Few Overall Acc CE (Cui et al., 2019) 64.0 33.8 5.8 41.6 LDAM (Cao et al., 2019) 60.4 46.9 30.7 49.8 c-RT (Kang et al., 2020) 61.8 46.2 27.3 49.6 τ-Norm (Kang et al., 2020) 59.1 46.9 30.7 49.4 Causal (Tang et al., 2020) 62.7 48.8 31.6 51.8 Logit Adj. (Menon et al., 2021) 61.1 47.5 27.6 50.1 RIDE(4E) (Wang et al., 2021) 68.3 53.5 35.9 56.8 Mi SLAS (Zhong et al., 2021) 62.9 50.7 34.3 52.7 Dis Align (Zhang et al., 2021a) 61.3 52.2 31.4 52.9 ACE (Cai et al., 2021) 71.7 54.6 23.5 56.6 Pa Co (Cui et al., 2021a) 68.0 56.4 37.2 58.2 TADE (Zhang et al., 2021b) 66.5 57.0 43.5 58.8 TSC (Li et al., 2022f) 63.5 49.7 30.4 52.4 GCL (Li et al., 2022e) 63.0 52.7 37.1 54.5 TLC (Li et al., 2022a) 68.9 55.7 40.8 55.1 BCL (Zhu et al., 2022) 67.6 54.6 36.6 57.2 NCL (Li et al., 2022c) 67.3 55.4 39.0 57.7 SAFA (Hong et al., 2022) 63.8 49.9 33.4 53.1 DOC (Wang et al., 2022) 65.1 52.8 34.2 55.0 DLSA (Xu et al., 2022) 67.8 54.5 38.8 57.5 Res LT (Cui et al., 2022) 63.3 53.3 40.3 55.1 Pa Co (Cui et al., 2021b) 68.2 58.7 41.0 60.0 LWS (Kang et al., 2019) 62.2 48.6 31.8 51.5 Zero-shot CLIP (Radford et al., 2021) 60.8 59.3 58.6 59.8 DRO-LT (Samuel & Chechik, 2021) 64.0 49.8 33.1 53.5 VL-LTR (Tian et al., 2022) 77.8 67.0 50.8 70.1 Cap2Aug (Roy et al., 2023) 78.5 67.7 51.9 70.9 Ge NIe-Ada 79.2 64.6 59.5 71.5 Li VT* (Xu et al., 2023) 76.4 59.7 42.7 63.8 Vi T (Dosovitskiy et al., 2021) 50.5 23.5 6.9 31.6 MAE (He et al., 2022a) 74.7 48.2 19.4 54.5 Dei T (Touvron et al., 2022) 70.4 40.9 12.8 48.4 Li VT (Xu et al., 2023) 73.6 56.4 41.0 60.9 Li VT + Img2Img L 74.3 56.4 34.3 60.5 Li VT + Img2Img H 73.8 56.4 45.3 61.6 Li VT + Txt2Img 74.9 55.6 48.3 62.2 Li VT + Ge NIe (r=0.8) 74.5 56.7 50.9 62.8 Li VT + Ge NIe-Ada 74.0 56.9 52.7 63.1 Published in Transactions on Machine Learning Research (04/2025) A.13 Further analysis of semantic shifts using Ge NIe In Fig. 5, we empirically demonstrate that by increasing the noise ratio from 0 to 1, the semantic category of the source image transitions gradually from the source class to the text-prompt s target class. To establish this further, we now choose 10 samples of a source class of Mushroom and generate Ge NIe augmentations with the target class of a Volcano. The generated images corresponding to each r [0, 1] are passed through a DINOv2 encoder and their embeddings are projected onto their 2 principle eigen vectors using PCA. The trajectories extracted from each of these 10 source images is depicted collectively and individually in Fig. A5. It can be noticed that each of the trajectories demonstrate a gradual transition of semantic category from the source to the target class, with a sparse distribution of points usually observed within [0.4, 0.6]. This is also observed in the plot on the bottom-right side of the figure where all trajectories are collectively plotted. Here, however, there is no clear range of r where a sparse distribution of points can be observed, thus indicating that each source image has its own optimal r value. This can be attributed to the inter-sample variances of images belonging to the same class. Since Ge NIe-Ada operates on each individual source image and target class text-prompt, it facilitates the selection of the best hard-negative per sample. A.14 How does Ge NIe control which features are retained or changed? We instruct the diffusion model to generate an image by combining the latent noise of the source image with the textual prompt of the target category. This combination is controlled by the amount of added noise and the number of reverse diffusion iterations. This approach aims to produce an image that aligns closely with the semantics of the target category while preserving the background and features from the source image that are unrelated to the target. To demonstrate this, in Figure A4, We are progessivley moving towards the two key components of Ge NIe: (i) careful choice of r and (ii) contradictory prompt. The input image is a bird in a cage. The top row shows a Stable Diffusion model, unprompted. As can be seen, such a model can generate anything (irrespective of the input image) with a large r. Now prompting the same model with a photo of a bird allows the model to preserve low-level and contextual features of the input image (up to r = 0.7 and 0.8), until for a large r 0.9 it returns a bird but the context has nothing to do with the source input. This illustrates how a careful choice of r can help preserve such low-level features, and is a key idea behind Ge NIe. However, we also need a semantic switch to a different target class as shown in the last row where a hardly seen image of a dog in a cage is generated by a combination of a careful choice of r and the contradictory prompt - leading to the full mechanics of Ge NIe. This sample now serves as hard negative for the source image (bird class). A.15 Analyzing Noise Effects in Bi-Directional Transformations with Ge NIe To further explore the effect of noise ratio r in Ge NIe, we conducted an experiment where Ge NIe was applied twice to transform between a source image and a target category. For this experiment, images from the mushroom category were used as the source, and volcano served as the target category. In the first step, we applied Ge NIe using a mushroom image as the source and a volcano prompt as the target. In the second step, we reversed the process: the Ge NIe-generated volcano image from the first step was used as the source, with the target prompt set to mushroom. Importantly, using a smaller noise ratio, r during the generation of the volcano image helps preserve more low-level visual features from the original mushroom source image. Consequently, when the roles of source and target are flipped in the second step, the final image retains a stronger resemblance to the original mushroom source image for lower noise ratios. This phenomenon is visualized in Fig. A6. As shown, a lower noise ratio during the first step results in the preservation of more visual features, leading to a final image that more closely resembles the original mushroom source. A.16 More Visualizations Additional qualitative results resembling the style presented in Fig. 4 are presented in Fig. A8, and more visuals akin to Fig. 2 can be found in Fig. A9. Moreover, we also present more visualization similar to the style in Fig. 5 in Fig. A7. Published in Transactions on Machine Learning Research (04/2025) Figure A6: Trajectory of Ge NIe augmentations: To further analyze the effect of noise ratio r in Ge NIe, we conducted an experiment using a set of augmentations generated from 10 different source images in the "mushroom" category, with a target label of "Volcano," across varying noise ratios. Similar to Fig. 5, all generated augmentations were processed through the Dino V2 Vi T-G model, which serves as our oracle, to extract their embeddings. For visualization, we applied PCA to these embeddings. Next, we selected one augmentation with a specific noise ratio, (r), and used it as the source image in for the "volcano" category in Ge NIe, with the target prompt set to "mushroom." As observed, using a lower noise ratio samples as the source for "volcano" preserves more low-level visual features from the original mushroom source image. Consequently, after a second round of applying Ge NIe, the resulting augmentations (even rows) tend to more closely resemble the original source image (first image in the corresponding odd rows above). The left plot presents the embeddings of all 10 samples, while the right plot provides a detailed visualization of one sample, showcasing the impact of varying noise ratios used in the second step of applying Ge NIe. Published in Transactions on Machine Learning Research (04/2025) Figure A7: Effect of noise in Ge NIe: Similar to Fig. 5, we pass all the generated augmentations through the Dino V2 Vi T-G model, which acts as our oracle model, to obtain their associated embeddings. Subsequently, we employ PCA for visualization purposes. The visualization reveals that the magnitude of semantic transformations is contingent upon both the source image and the specified target category. Published in Transactions on Machine Learning Research (04/2025) Figure A8: Visualization of Generative Samples: More visualization akin to Fig. 4. We compare Ge NIe with two baselines: Img2Img L augmentation uses both image and text prompt from the same category, resulting in less challenging examples. Txt2Img augmentation generates images based solely on a text prompt, potentially deviating from the task s visual domain. Ge NIe augmentation incorporates the target category name in the text prompt along with the source image, producing desired images with an optimal amount of noise, and balancing the impact of the source image and text prompt. 31 Published in Transactions on Machine Learning Research (04/2025) Figure A9: Effect of noise in Ge NIe: Akin to Fig. 2, we use Ge NIe to create augmentations with varying noise levels. As is illustrated in the examples above, a reduced amount of noise leads to images closely mirroring the semantics of the source images, causing a misalignment with the intended target label.