# contrastive_visual_data_augmentation__a24fbc65.pdf Contrastive Visual Data Augmentation Yu Zhou 1 * Bingxuan Li 1 * Mohan Tang 1 * Xiaomeng Jin 2 Te-Lin Wu 1 Kuan-Hao Huang 2 3 Heng Ji 2 Kai-Wei Chang 1 Nanyun Peng 1 Large multimodal models (LMMs) often struggle to recognize novel concepts, as they rely on pre-trained knowledge and have limited ability to capture subtle visual details. Domain-specific knowledge gaps in training also make them prone to confusing visually similar, commonly misrepresented, or low-resource concepts. To help LMMs better align nuanced visual features with language, improving their ability to recognize and reason about novel or rare concepts, we propose a Contrastive visual Data Augmentation (Co DA) strategy. Co DA extracts key contrastive textual and visual features of target concepts against the known concepts they are misrecognized as, and then uses multimodal generative models to produce targeted synthetic data. Automatic filtering of extracted features and augmented images is implemented to guarantee their quality, as verified by human annotators. We show the effectiveness of Co DA on low-resource concept and diverse scene recognition datasets including INaturalist and SUN. We additionally collect Novel Species, a benchmark dataset consisting of newly discovered animal species that are guaranteed to be unseen by LMMs. LLa VA-1.6 1-shot updating results on these three datasets show Co DA significantly improves SOTA visual data augmentation strategies by 12.3% (Novel Species), 5.1% (SUN), and 6.0% (i Nat) absolute gains in accuracy. Code and data at contrastive-visual-data-augmentation.github.io 1. Introduction Recent advancements in multimodal pre-training (Open AI, 2023; Google, 2023; Hurst et al., 2024) and visual instruction tuning (Liu et al., 2023b;a; 2024c) have enabled impressive LMM abilities. However, as shown in Figure 1, it still * Equal contribution, interchangeable ordering. 1UCLA 2UIUC 3TAMU. Correspondence to: Yu Zhou . Proceedings of the 42 nd International Conference on Machine Learning, Vancouver, Canada. PMLR 267, 2025. Copyright 2025 by the author(s). (b) Confusing Concept Resupply Base (a) Novel Concept Clouded Tiger Cat African Leopard Wholesale Store Clouded Tiger Cat Resupply Base Plain décor, Cardboard boxes, People in uniform, Bushy tail, Elongated slim body, Round ears, Rosette Patterns, Carts, Shelves, Figure 1. Co DA uses diffusion-generated synthetic data to help LMMs recognize novel and confusing concepts in the wild. The Clouded Tiger Cat (L. pardinoides) is a new animal species first described in April 2024, while Resupply Base is an example of a confusing concept for LMMs. Based on model failures (collected from GPT4o-2024-08-06 and LLa VA-Ne XT 34B), Co DA extracts contrastive visual and textual features to generate synthetic image data for model updating. remains a challenge for current state-of-the-art proprietary and open-source models to robustly recognize novel visual concepts (e.g. Clouded Tiger Cat Figure 1a) and confusing / low-resource / commonly misrepresented visual concepts (e.g. Resupply Base Figure 1b). In order to help models better acquire new visual concepts and distinguish confusable concepts, existing approaches straightforwardly 1). Fine-tune text decoder on new textual corpora to expand the concept base; and 2). Fine-tune both vision and text components on new web image-text pairs Contrastive Visual Data Augmentation for visual concept acquisition. These approaches are ineffective due to data scarcity for certain concepts and data inefficiency caused by not knowing what precisely confused the models (3.1). As depicted in Figure 1, it can be difficult to obtain ample high-quality real images for novel concepts such as new animal species. While for confusing concepts, the problem usually lies with biased concept representation in web image-text data. For example: online images of Resupply Base mostly only consist of exterior views of the architecture without the interior details, which may cause the models to confuse it with a Wholesale Store that shares some interior features. To help LMMs recognize and reason about novel and confusing concepts more robustly and efficiently, we propose Co DA, a Contrastive Visual Data Augmentation technique. For each target concept, Co DA first identifies a confusable concept that the LMMs finds most similar to the target. Then, it extracts contrastive textual and visual features of the target concept with respect to the confusable concept. The extracted features go through a filtering process based on discriminability and generability to make sure that: 1). The features are possessed by the target concept but not the confusable concept; and 2). The feature can be reliably generated by the text-to-image generative model and recognized by the LMMs. Afterwards, the features are passed to the text-to-image generative model to produce augmented visual instances of the target concept. To make sure that the features are indeed generated and recognizable by the LMMs, Co DA again uses the LMMs zero-shot inference to rank and filter the augmented images. Finally, the resulting augmented images can be used to update the LMMs via low-rank adaptation, basic fine-tuning, in-context learning, or any other method of choice. In addition to evaluating on existing datasets INaturalist and SUN, we create Novel Species, an annotated image dataset of newly discovered animal species in recent years. Novel Species allows the simple selection of species discovered after any model s latest knowledge cutoff date, ensuring the selected species were never seen by the model. Therefore, Novel Species is the perfect testbed for methods aimed at improving LMMs novel concept recognition ability. Comprehensive experiments with LLa VA-Ne XT on the 3 datasets show Co DA performs surprisingly well in teaching LMMs novel and confusing concepts, significantly improving data efficiency compared to existing methods. In additional experiments, we show that Co DA is also able to improve novel concept recognition for traditional classifiers like Vi T and proprietary LMMs such as GPT4o-mini. Finally, ablation experiments show that Co DA can be significantly improved by simply replacing its off-the-shelf components such as the text-to-image generation model with superior versions of similar models. Our key contributions include: Co DA, a simple plug-and-play contrastive visual data augmentation method that can be used to effectively and efficiently improve LMMs ability to recognize novel and confusing concepts. Co DA is also the first widely successful method using text-to-image generation for visual data augmentation. Novel Species, a new benchmark dataset of novel animal species discovered in recent years, providing an ideal benchmark for novel concept recognition. Novel Species currently consists of 2240 annotated images and will continue to be updated with future discoveries. 2. Related Works Few-shot image recognition is a long-standing problem in the vision community. Early works in this area focused on improving traditional image classifiers on classifying existing concepts (Vinyals et al., 2016; Finn et al., 2017; Nichol, 2018; Dhillon et al., 2019; Tian et al., 2020; Bhagat et al., 2023; Afrasiyabi et al., 2022). On the other hand, while recent advancements in the training of vision language models (VLMs) and large multimodal models (LMMs) (Open AI, 2023; Google, 2023; Hurst et al., 2024; Liu et al., 2023b;a; 2024c) have shown great promise and extensibility, they still severely lag behind traditional models in image classification, especially for low-resource, novel, and confusing concepts (Zhang et al., 2024; Cooper et al., 2024; Wu et al., 2023b; Yang et al., 2024; Ha et al., 2025). While commonly used text-side VLM data augmentation strategies (Yuksekgonul et al., 2022; Yang et al., 2023; Liu et al., 2024e; Sharifzadeh et al., 2024) have little effect on this issue, a more promising technique to solve this is through visual data augmentation. This includes basic visual manipulations such as cropping, flipping, and rotation (Yang et al., 2022; Kumar et al., 2024); and more advanced modelbased augmentation such as style transfer (Zheng et al., 2019; Chun & Park, 2021) and image mixing (Uddin et al., 2020; Xie et al., 2021; Hao et al., 2023). More recently, with the rise of controllable and promptable visual generative models, knowledge and feature editing-based augmentation methods (Liu et al., 2022; Wu et al., 2023a; Jin et al., 2024) have gained in popularity. Such methods generally focus on using multimodal data and general knowledge bases to guide image-editing models in creating augmented visual data based on existing images. One main issue with current methods is that the augmented images they produce must be closely based on existing real images, which makes them unhelpful for novel concepts where real images are extremely rare, and mis-represented concepts where existing real images do not accurately depict the concept. Additionally, due to their close connection Contrastive Visual Data Augmentation to existing images, such augmented images usually lack visual frame structure and view variation. In contrast, our method Co DA can extract accurate and meaningful features from extremely limited multimodal data, and use text-toimage generative models to produce diverse high-quality augmented data for LMM updating. As shown in Figure 2, Co DA consists of 4 major steps including contrastive textual and visual feature extraction, feature filtering, feature-controlled image generation, and augmented image filtering. Together these steps ensure Co DA reliably generates informative and high-quality augmented images that help LMMs recognize novel and confusing concepts. 3.1. Feature Extraction Textual Feature Extraction In our exploratory experiments, we find that significant mis-recognition errors occur on low-resource or commonly mis-represented concepts in vision-language instruction fine-tuning and multimodal pretraining datasets, which the LMMs are trained on. For example, the LLa VA 1.6 (34B) model (Liu et al., 2024c), mainly tuned on LAION-GPT-4V(LAION, 2024) and Share GPT4V (Chen et al., 2023) datasets, has a strong tendency to misrecognize interior images of Resupply Base as Wholesale Store (Figure 1). Unsurprisingly, we find that all related references of Resupply Base across the 3 instruction-tuning datasets only depict exterior views of the concept rather than interior views. While the concept itself is not a low-resource concept in existing text corpora, it is severely low-resource and also commonly mis-represented in vision-language instruction fine-tuning datasets. To address this issue, we prompt LLMs to directly generate feature attributes of the target concept based on their existing knowledge, focusing on visual appearance, and avoiding hallucination for unfamiliar concepts. For this task, we use the cost-efficient GPT4o-mini model with chain of thought reasoning. Generally, textual feature extraction is most applicable for concepts that are high-resource in existing textual corpora, yet low-resource and/or commonly misrepresented in vision-language instruction-tuning and pretraining datasets. Here we do not try to classify which concepts fall under this criteria, but rather apply this step for all concepts. To ensure extracted feature quality and filter out hallucinated and/or non-visually-recognizable features, we pass all extracted features through an automatic filtering step, as described in 3.2. We also considered other methods for textual feature extraction, including using knowledge bases (Jin et al., 2024), retrieval augmented generation, and LLMs with internet search. However, we believe currently the advantages brought by these methods do not out-weigh their complexity overhead, thus we opt for simplicity. Visual Feature Extraction While textual feature extraction generally works well for pre-existing and non-hyperdomain-specific concepts that are prevalent in textual data sources, it tends to fail when either of the conditions are not met. For example, a large language model with a knowledge cutoff prior to June 2023 would not be able to provide meaningful features regarding the Apple Vision Pro device announced in July, or the new animal species Clouded Tiger Cat (L. pardinoides) first described by scientists in April 2024 (Figure 1). In addition to this weakness, LLM-based textual feature extraction is also unreliable when asked to provide detailed information regarding hyper-domainspecific concepts like the Mazda MX-5 Miata RF or the Lear s Macaw (Anodorhynchus Leari) . In practice, we observe that for novel and hyper-domain-specific concepts, most of the LLM extracted textual features end up being filtered out by our automatic feature filtering module. To address this weakness, we implement an additional visual feature extraction module based on VLMs. Given a single image of the target concept, the VLM is asked to extract its key visual features. When there is more than one image containing the target concept available, we use a LM to de-duplicate and summarize the combined extracted visual features from all images. For simplicity and costefficiency, we use the GPT4o-mini model for both visual feature extraction and feature de-duplication. In contrast to textual feature extraction, visual feature extraction is most effective for hyper-domain-specific and novel concepts that are very rare or non-existent in textual corpora but have a limited number of visual examples. Thus, it well-complements textual feature extraction. Similarly, we do not attempt to classify which concepts fall under this criterion; instead, we apply this step to all concepts and rely on automatic filtering (3.2) to remove low-quality features. Contrastive Feature Extraction While basic textual and visual feature extraction both aim to exhaustively list identifying features of the target concept, this is essentially an intractable task for complex concepts as it usually requires a huge number of features to fully describe them. For novel or low-resource concepts the LMM has likely never seen before, it is extremely difficult to teach the LLM the new concept using an incomplete description. There are two potential solutions to this problem: (1). Leveraging hierarchical information to narrow down concept category and reduce descriptional features. (2). Illustrating the new concept based on contrastive differences from a similar existing concept the LMM already understands. Previous works in language and visual data augmentation (Jin et al., 2024) tend to use solution (1). However, its feasibility Contrastive Visual Data Augmentation Anodorhynchus Leari (Lear's Macaw) Target Concept Cyanopsitta Spixii (Spix's Macaw) Feature-controlled Augmentation Augmented Image Filtering Model Updating Lo RA, Fine-tuning, ICL, VL-Adaptor, Anodorhynchus Leari (Lear's Macaw) Cyanopsitta Spixii (Spix's Macaw) Anodorhynchus Leari (Lear's Macaw) Cyanopsitta Spixii (Spix's Macaw) Yellow eye ring Yellow cheek skin Deep cobalt blue Yellow eye ring Yellow cheek skin Deep cobalt blue Zero-Shot Inference Feature Filtering Coal black curved beak Discriminability: 0.45 Generability: 0.93 Bright Yellow eye ring Discriminability: 0.98 Generability: 0.86 Native to Brazil Discriminability: 0.48 Generability: 0.14 Feature Extraction Shorter life, Cliff-nesting, Faster flight, Light blue facial skin, Bright yellow eye rings, Coal black curved beak Figure 2. The Co DA method. Including Feature Extraction, Feature Filtering, Feature-controlled Augmentation, and Augmented Image Filtering. The target concept and misidentified concept are highlighted respectively. Specific feature filtering scores are for illustration only. Here the example concepts Anodorhynchus Leari (Lear s Macaw) and Cyanopsitta Spixii (Spix s Macaw) are from the i Naturalist (Van Horn et al., 2018) dataset, and augmented images are produced by the Recraft V3 model (Recraft.AI, 2024). is contingent on the existence of a comprehensive textual knowledge base or tree-like structure that already includes the target concept. As discussed in Section 3.1, this is often not the case for novel concepts such as new electronic products (e.g. Apple Vision Pro) or new animal species (eg. Clouded Tiger Cat). To enable the handling of novel concepts and remove the need for external databases, we adopt solution (2) and perform contrastive multimodal feature extraction for all target concepts. First, we use the LMM s zero-shot inference on the target concept CT to obtain the misidentified concept CM. Then, we perform contrastive textual and visual feature extraction by querying LLMs and VLMs for visually identifying features that belong to CT but not CM. 3.2. Feature Filtering Automatic Feature Filtering After obtaining visually identifying features from contrastive textual and visual feature extraction, we filter them based on two key criteria: 1. Discriminability (D(f, CT , CM)) : measures whether a feature f indeed differentiates the target class CT from the misidentified concept CM (check whether f is a valid feature of CT but not CM) . 2. Generability (G(f, CT , CM)) : measures whether a feature f can be properly generated by the text-toimage generative model. To calculate the Discriminability of a feature f given the target concept CT and misidentified concept CM, we compute the likelihood that CLIP (Radford et al., 2021) associates this feature with real images of the target concept compared to the likelihood that it is associated with real images of the misidentified class: D(f, CT , CM) = X CLIP(f, ireal CT ) CLIP(f, ireal CT ) + CLIP(f, ireal CM ) Here we use an equal number of images of the target and misidentified concepts. A score below 0.5 indicates that the feature is more likely to be associated with the misidentified class rather than the target class. To ensure that selected features are more strongly associated with the target class, we filter out all features with Discriminability below 0.6. This method avoids the CLIP score bias against smaller features by only comparing feature association with the two classes and not relying on the absolute CLIP score. Generability is calculated for all features that pass the Discriminability threshold. We prompt the T2I generative model g to generate synthetic images of the target concept that contains the feature f, and then compare the average CLIP similarity between f and those generated images against the average CLIP similarity between f and ireal CM : G(f, CT , CM, g) = X CLIP(f, isynthetic CT ) CLIP(f, isynthetic CT ) + CLIP(f, ireal CM ) Contrastive Visual Data Augmentation Here we rank all remaining features by their Generability score and select the top 5 features to be passed to the text-toimage generative model (as current diffusion models usually have limited text encoder attention span). This step identifies features that not only help distinguish the target concept, but also can be effectively rendered by the text-to-image generative model in synthetic images, which is critical to the success of synthetic data augmentation. Our automatic feature filtering module based on Discriminability and Generability ensures feature quality and limits the information loss between features and the generated augmented images. The remaining features are used for image generation and improving in-context recognition ability in inference prompts. We further verify the quality of remaining features with human evaluation in Sec.3.4. 3.3. Image Generation and Verification Image Generation After feature extraction and filtering based on Discriminability and Generability, we pass the selected features to a text-to-image generative model to generate augmented visual data. We experiment with both SOTA open-weights (Esser et al., 2024; Stability AI, 2024) and proprietary (Recraft.AI, 2024) models. Verification To ensure final images for augmentation contain our extracted and filtered target concept features, we propose a simple automatic verification metric that checks whether desired features are recognized in the augmented images by the LMM we want to update: Given the vanilla LMM M, a set of features F, and an augmented image isynthetic, the feature satisfaction rate S(isynthetic, F, M) for each augmented image: S(isynthetic, F, M) = f F 1{M(f, isynthetic)} Here M(f, isynthetic)} returns true if the feature f is recognized in the image isynthetic. Afterwards, we filter out all images with S(isynthetic, F, M) <1.0, keeping only augmented images that fully match all target concept features. 3.4. Human Evaluation Image Target Misidentified Inter-Annotator Type Concept (%) Concept (%) Agreement (κ) Real 92.51 14.32 0.87 Synthetic 83.97 - 0.82 Table 1. Human eval of extracted features and augmented images. 3 external annotators are asked to answer (yes/no) to whether the extracted and filtered features are present in the corresponding real and synthetic images. IAA based on Fleiss Kappa. To verify the reliability of our feature filtering and augmented image verification modules, we conduct human evaluation on a subset of i Naturalist and the novel animal species dataset. For target concepts, we select 100 image-feature pairs for both real and augmented synthetic images. We also select 100 image-feature pairs for real images of misidentified concepts. 3 external human annotators are asked to label whether they believe the given feature belongs to the concept in the corresponding image. Results in Table 1 show human annotators overwhelmingly agree that the final extracted features belong to the target concept (92%) but not the misidentified concept (14%). The augmented synthetic images of the target concept also likely contains the desired features (83%), though as expected, there is some information loss between the text-to-image generation step. In addition, the three independent annotators generally agreed in their response (>0.8 IAA). 3.5. In-Context Inference for Enhanced Recognition In addition to updating the LMM with augmented data, we can further boost performance by integrating the extracted features into the inference prompt. For each query, we can append a concise list of the most discriminative and generable features of the target and confusable classes. These features serve as an in-context guide, focusing the LMM s attention on critical distinguishing attributes. By explicitly highlighting what to look for (and what not to mistake it for), the model more reliably identifies the correct concept. 4. Novel Species Dataset Proprietary LMMs like GPT4o (Hurst et al., 2024) and Gemini (Google, 2023) are trained on vast online text-image data and proprietary data, both non-public and impossible to inspect. Some open-source and open-data LMMs such as LLa VA (Liu et al., 2024b;d) are trained on publicly available image-text datasets. However, the text encoders used by such models are often not open-data, for example LLa VA1.6 34B uses the closed-data Yi-34B model as its language backbone. Even in the rare cases where both image-text training data and text encoder training data are publicly available, it is still difficult to ascertain whether concepts in your benchmark were seen by your LMM through indirect data leakage (i.e. partial / paraphrased mentions). Due to the above issues, it is difficult to evaluate true novel concept recognition ability with existing datasets. One way to bypass this problem with 100% guaranteed success is to use a dataset that only contains concepts created / discovered after the LMM s knowledge cutoff, i.e. the latest knowledge cutoff date among all of its textual / visual subcomponents. Based on this idea, we curate Novel Species, a dataset of novel animal species discovered in each recent Contrastive Visual Data Augmentation year, starting with 2023 and 2024. We provide detailed information for each species, including time of discovery, latin name, common name, family category, textual description, and more. Data will be released upon publication. To create Novel Species, we start by collecting the list of species first described in each year by Wikidata (Wikidata, 2024). Then, to make sure we can curate a visual benchmark of novel species, we manually annotate and filter out extinct species and species with too few publicly available images. After filtering, we end up with a dataset of 64 new species, each consisting of 35 human-verified image instances, thus a total of 2240 images. The images are split into training, validation, and test sets. For each species, there are 5 training images, 15 validation images, and 15 test images. This data split is consistent with our goal of creating a benchmark dataset for novel concept recognition, where the maximum number of training instances for a completely unseen concept can range from 1 to 5. 5. Experiments 5.1. Datasets and Baselines To evaluate Co DA s ability to improve novel and confusing concept recognition in LMMs, we experimented with Co DA and other relevant baselines on three different datasets: 1. The i Naturalist Dataset (Van Horn et al., 2018) is a challenging natural world concept recognition benchmark for LMMs due to its extensive highly domainspecific and fine-grained species categories and inclusion of rare and low-resource species classes. 2. The SUN Dataset (Xiao et al., 2010) is a widely used large-scale scene recognition dataset that contains rich and confusing visual scenes. Correctly recognizing the scenes requires fine-grained visual reasoning and understanding of the scenes. 3. Novel Species Dataset (Sec.4) is our new dataset consisting only of novel animal species concepts that LMMs are guaranteed to have never encountered in their training or fine-tuning. For each dataset, we use an automatic data selection strategy A.2 to find a subset of challenging concepts that the model fails to recognize. Then, we apply Co DA along with 3 other visual data augmentation baselines: 1. All Real uses an all real augmented image set. In the Fixed Real Data setting, this means using the 5 real images provided. In the Fixed Compute setting, this means using unlimited real images to match the total number of real + synthetic images in other settings. 2. Cropping and Flipping are widely used traditional visual data augmentation strategies. We include them here for direct comparison with Co DA and other existing feature-based augmentation methods. 3. ARMADA (Jin et al., 2024) is the current state-of-theart feature-based visual data augmentation strategy for concept recognition and image classification. In addition to these 3 baselines, we also include ablations of Co DA with non-contrastive textual and visual features, i.e. w/o contrastive guidance from confusable concepts (3.1) nor discriminability-based feature filtering (3.2). 5.2. Main Experiment For our main experiment, we consider two different resource settings that correspond to common real-world scenarios: Fixed Real Data Under the fixed real data setting, we only have access to 5 real images for each concept. Each data augmentation strategy may generate 1-5 synthetic images. Then, the model is Lo RA-adapted on the combined real and synthetic images. This setting simulates real-world scenarios, where there isn t sufficient real training data for certain concepts. This is common for novel concepts, hyperdomain-specific concepts, and long-tail distributed datasets. In these scenarios, the quality and effectiveness of synthetic augmented data is especially instrumental to the updated model s performance. Experiment results across the 3 datasets show that Co DA consistently outperforms existing traditional and featurebased data augmentation methods in the Fixed Real Data setting. When augmenting the training set with just a single synthetic image, Co DA is able to achieve 11.8% (Novel Species), 10.0% (SUN), and 17.8% (i Nat) absolute gains in accuracy compared to using all real images. It further outperforms the best existing baseline augmentation methods by 5-12% absolute gains. We also observe that the ablated performance of Co DA (w/o contrastive) is still significantly above traditional and image-editing-based augmentation baselines while being almost consistently below Co DA s performance. This shows the benefits of text-to-image generative augmentation methods compared to existing methods, as well as the benefits of fine-grained textual features during inference. This also highlights the need for contrastive feature selection and discriminability-based feature filtering. We find that increasing the number of augmented synthetic images does not necessarily improve updated model performance; this may be attributed to the fact that all generated images are ranked and selected from the same pool, with the first image being of the highest quality. Finally, the largest improvement over existing baselines can be seen in Novel Species, where Co DA methods involving visual features achieve the highest performance. This makes sense as the visual feature extraction method is designed to be robust to novel concepts with little textual documentation. Contrastive Visual Data Augmentation Dataset Augmentation Method Feature Type Fixed Real Data (Real:Syn) Fixed Compute (Real:Syn) 5:0 5:1 5:3 5:5 20:0 10:10 0:20 Novel Species (Sec.4) All Real 61.2 - - - - - - Cropping - 60.4 60.4 59.5 - - - Flipping - 60.7 62.9 60.1 - - - ARMADA - 60.7 60.2 61.2 - - - Co DA (w/o contrastive) Textual - 69.1 68.6 70.5 - - - Visual - 71.8 72.6 71.7 - - - T+V - 70.3 65.1 70.1 - - - Co DA Textual - 72.0 69.2 70.3 - - - Visual - 73.0 72.8 71.8 - - - T+V - 70.1 72.6 73.0 - - - SUN (Xiao et al., 2010) All Real 73.4 - - - 74.3 - - Cropping - 78.3 75.8 76.3 - 77.3 76.4 Flipping - 75.7 78.4 74.8 - 75.2 76.1 ARMADA - 75.9 78.3 77.6 - 76.2 76.8 Co DA (w/o contrastive) Textual - 80.6 79.7 79.4 - 81.3 80.8 Visual - 81.3 81.6 79.3 - 80.0 80.8 T+V - 82.7 80.7 80.4 - 82.8 82.1 Co DA Textual - 79.2 83.2 82.3 - 82.8 82.1 Visual - 82.3 81.7 82.2 - 81.8 83.1 T+V - 83.4 81.7 82.6 - 83.3 82.1 i Naturalist (Van Horn et al., 2018) All Real 49.2 - - - 64.3 - - Cropping - 59.7 58.8 62.2 - 61.4 63.9 Flipping - 61.0 61.1 62.3 - 62.1 62.7 ARMADA - 60.1 60.7 61.1 - 61.6 58.5 Co DA (w/o contrastive) Textual - 63.9 64.6 66.5 - 65.6 63.2 Visual - 65.0 64.7 64.3 - 65.6 63.2 T+V - 62.8 64.4 62.3 - 64.4 63.4 Co DA Textual - 63.9 67.8 62.6 - 65.0 64.9 Visual - 67.0 66.0 65.1 - 62.5 60.9 T+V - 63.5 65.0 64.6 - 67.0 64.1 Table 2. Main experiment results on INaturalist, SUN, and Novel Species under Fixed Real Data and Fixed Compute settings: Experiments are defined by the number of Real:Synthetic images used. For example, 5:1 means the model uses 5 real images and 1 synthetic image for each concept class at training time. All results are in terms of LLa VA-1.6 34B concept recognition accuracy (%). Best performance scores for each setting and scores using all real data are highlighted in Red and Green, respectively. Fixed Compute Under the fixed compute setting, we assume access to unlimited real and synthetic images. However, the fine-tuning budget can only support a total of 20 images, allowing different percentages of real and synthetic images, from 0% synthetic (20:0) to 100% synthetic (0:20). This setting simulates real-world scenarios, where there is abundant real data. In such cases, the question is whether to just use all real data to update the model, or to include a nontrivial amount of augmented synthetic data. Traditionally, real data is always preferred due to perceived higher-quality. However, Co DA s effectiveness in the Fixed Real Data setting prompts us to test the possibility of it being beneficial to include synthetic data even when real data is abundant. This hypothesis is tested by whether any of the models fine-tuned with mixed real/synthetic data can outperform the model fine-tuned with all real data. Experiments on i Naturalist show diverging results between Co DA and other baseline augmentation methods: While including synthetic images generated by baseline methods generally led to lower performance, using Co DA augmented images can actually lead to improvements over using all real data. Furthermore, a 50-50 real-synthetic data mix generally outperforms all real or all synthetic data. We attribute the success of mixing synthetic and real data to the fact that Co DA generated synthetic data is aimed to highlight Contrastive Visual Data Augmentation Augmentation Method Feature Type LLa VA-Ne XT GPT4o-mini Vi T 5:0 5:1 5:3 5:5 5:0 5:1 5:3 5:5 5:0 5:1 5:3 5:5 All Real 61.2 - - - 84.3 - - - 75.4 - - - Cropping - 60.4 60.4 59.5 - 84.8 86.3 85.9 - 78.3 77.6 79.6 Flipping - 60.7 62.9 60.1 - 83.2 83.5 84.3 - 76.9 77.9 78.2 ARMADA - 60.7 60.2 61.2 - 84.1 84.3 83.9 - 76.3 76.4 78.6 Co DA (w/o contrastive) Textual - 74.8 75.1 74.7 - 87.6 87.2 87.0 - 82.5 84.5 84.7 Visual - 76.5 77.9 76.2 - 88.3 89.6 88.2 - 82.5 83.0 82.6 T+V - 77.6 78.9 78.8 - 89.5 91.2 87.9 - 84.3 84.9 82.5 Co DA Textual - 76.4 75.9 76.8 - 87.1 87.9 87.4 - 84.6 85.0 84.5 Visual - 77.5 78.1 77.9 - 91.3 90.8 92.6 - 85.5 84.6 85.7 T+V - 78.8 78.7 79.2 - 91.6 90.8 91.4 - 85.3 85.8 86.3 Table 3. Experiments on Novel Species with open-weight VLM (LLa VA-Ne XT), proprietary LMM (GPT4o-mini), and traditional classifier (Vi T) under the Fixed Real Data setting. Results are in terms of accuracy (%). Synthetic image data generated by Recraft V3. Best performance scores for each setting and scores using all real data are highlighted in Red and Green, respectively. discriminable features of the confusing / novel concepts, making them more prominent and visible compared to real images. On the other hand, real images provide valuable style information and is a more accurate reflection of the testtime distribution, helping to ground the updated model. 5.3. Additional Experiments For additional experiments, we focus on Novel Species as it most closely resembles real-world scenarios, where over time, models are required to learn novel concepts without access to sufficient real training data. Advanced T2I Model As explained in Sec.3, off-the-shelf model components used in Co DA can be easily swapped for superior versions of similar models to improve performance. To demonstrate this, we replace the open-weight Stable Diffusion 3.5 Large Turbo model (Stability AI, 2024) with the SOTA proprietary Recraft V3 Model (Recraft.AI, 2024) and run the same LLa VA-updating experiments as in Table 2. Here we note that Recraft V3 has better instructionfollowing ability as well as better image generation quality compared to Stable Diffusion 3.5 Large Turbo. More details on these differences can be found in Sec.6. Our experiment results in Table 3 show a significant performance boost when Lo RA fine-tuning LLa VA with Recraft V3 produced synthetic images compared to fine-tuning on all-real data (28.7%) and also compared to fine-tuning on Stable Diffusion 3.5 Large Turbo produced synthetic data (7.9%). This demonstrates the potential increase of Co DA s effectiveness along with improvements in Text-to-Image generative models. We believe it is also possible to achieve similar improvements by replacing the LLM/VLM components of Co DA with superior models in the future. Proprietary LMM While proprietary LMMs like GPT4omini (Hurst et al., 2024) tend to have relatively strong 0shot performance on existing datasets such as SUN and i Naturalist, their performance significantly degrades on Novel Species due to having never encountered the novel concepts. To test whether Co DA can effectively improve novel concept recognition performance for such proprietary LMMs, we fine-tune the gpt-4o-mini-2024-07-18 model using Co DA and relevant augmentation baselines. Results in Table 3 demonstrate a significant performance gain (9.5%) for GPT4o-mini after being fine-tuned on Co DA augmented synthetic images. While this improvement is not as significant compared to the LLa VA-1.6 model (20.3%), it is due to GPT4o-mini s better base performance. Traditional Classifier In addition to evaluating Co DA on LMMs which take image-text input and produce text output, we also test whether it can help traditional image classifiers recognize novel concepts. We run the widelyused Vi T classifier (Alexey, 2020) on Novel Species with Co DA and other augmentation baselines. Results in Table 3 show that Co DA is able to achieve a consistent performance gain over existing baselines for Vi T-base (9.1% for singleshot augmentation). The Vi T classifier provides stronger base performance compared to general VLMs, thus offering less room for improvement. However, we note here that our main focus on improving LMMs instead of such traditional classifiers stems from LMMs superior extensibility and generalizability to other related tasks such as recognitionbased reasoning and explanation. 6. Discussions In Figure 3, we compare example synthetic images generated by Co DA and baseline visual data augmentation methods including Cropping and ARMADA (Jin et al., 2024). While providing localized feature emphasis, Cropping often results in the loss of crucial visual details necessary for concept identification. For instance, for Phyllobates Samperi, cropping occludes the black spots on the frog s skin, Contrastive Visual Data Augmentation Real Images Co DA (SD-3.5) Co DA (Recraft V3) Phyllobates Co DA Features: Prominent orange stripes, Dark shiny coloration, Black spots on legs, Co DA Features: Vibrant horizontal stripes along body, Greenish-yellow cyan accents, Figure 3. Qualitative Comparison of Co DA and baseline visual data augmentation methods. Phyllobates Samperi and Tail-Spot Wrasse are example concepts from the Novel Species dataset. All Co DA images are generated using contrastive textual + visual features. an essential distinguishing feature. Without such essential distinguishing features, the cropped images provide less helpful learning signals compared to other methods. Unlike Cropping, ARMADA successfully retains some structural features of the target concept, using Wiki Data text features to guide its image-editing backbone (Brooks et al., 2023). However, this setup also induces two significant issues: (1) Leveraging only existing textual features present in Wiki Data leads to an incomplete feature set, especially for novel concepts. This is apparent in images generated for Phyllobates Samperi, where generated images contained black spots but failed to specify their location on the legs of the frog instead of the body. In addition, the model completely failed to generate Phyllobates Samperi s iconic orange stripes due to the feature not being recorded in Wiki Data. (2) Image-editing models are not as strong in depicting precise details compared to text-to-image generative models. In the case of the Tail-Spot Wrasse, the ARMADA generated image fails to accurately depict vibrant horizontal stripes along the body, leading to a visually inconsistent and less biologically accurate representation. In contrast to existing baselines, Figure 3 shows that Co DA is much better at generating high quality synthetic images of the target novel concept that depict accurate and realistic details. Both versions of Co DA using different backbone models (Esser et al., 2024; Recraft.AI, 2024) are able to produce significantly more realistic images compared to the two baselines, maintaining general biological consistency. However, we should note that Co DA s performance is inherently bounded by the instruction-following ability of its image generation backbone model, more specifically the ability to accurately generate multiple feature details in a single image. For example, while Co DA-Recraft-V3 is able to accurately generate all three extracted features including prominent orange stripes , dark shiny coloration , and black spots on legs ; Co DA-SD-3.5 is only able to generate the first two features while failing to capture black spots on legs . With such limitations in mind, we give Co DA an extremely modularized design. This allows each pre-trained model component in Co DA to be easily replaced for newer and stronger versions of similar models, including more perceptive VLMs and T2V generative models with stronger instruction-following ability and higher generation quality. 7. Conclusion In this work, we propose Co DA, a contrastive visual data augmentation approach that helps LMMs recognize novel, confusing, and low-resource concepts through efficient and effective model updating. Co DA is a plug-and-play method which utilizes off-the-shelf models for contrastive feature extraction, feature filtering, text-to-image generation, and image filtering. We evaluate Co DA against four existing baselines and self-ablations on three datasets: INaturalist, SUN, and Novel Species, which we created in this work. Consisting only of animal species discovered in recent years, Novel Species offers an ideal testbed for LMMs novel concept recognition. We provide comprehensive additional experiments demonstrating Co DA s effectiveness for traditional classifiers and proprietary LMMs. Finally, we show that Co DA can be easily improved by replacing off-the-shelf components, such as text-to-image generation model with superior versions of similar models in the future. Contrastive Visual Data Augmentation Acknowledgment This material is based on research supported by the ECOLE program under Cooperative Agreement HR00112390060 with the US Defense Advanced Research Projects Agency (DARPA), an award from Office of Naval Research with #N00014-23-1-2780, Apple Research Award, and Amazon AGI Research Award. The views and conclusions contained herein are those of the authors and should not be interpreted as necessarily representing DARPA, or the U.S. Government. Impact Statement This paper presents work whose goal is to advance the field of Machine Learning. There are many potential societal consequences of our work, none which we feel must be specifically highlighted here. Afrasiyabi, A., Larochelle, H., Lalonde, J.-F., and Gagn e, C. Matching feature sets for few-shot image classification. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 9014 9024, 2022. Alexey, D. An image is worth 16x16 words: Transformers for image recognition at scale. ar Xiv preprint ar Xiv: 2010.11929, 2020. Bhagat, S., Stepputtis, S., Campbell, J., and Sycara, K. Sample-efficient learning of novel visual concepts. In Conference on Lifelong Learning Agents, pp. 637 657. PMLR, 2023. Brooks, T., Holynski, A., and Efros, A. A. Instructpix2pix: Learning to follow image editing instructions. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 18392 18402, 2023. Chen, L., Li, J., Dong, X., Zhang, P., He, C., Wang, J., Zhao, F., and Lin, D. Sharegpt4v: Improving large multi-modal models with better captions. ar Xiv preprint ar Xiv:2311.12793, 2023. Chun, S. and Park, S. Styleaugment: Learning texture de-biased representations by style augmentation without pre-defined textures. ar Xiv preprint ar Xiv:2108.10549, 2021. Cooper, A., Kato, K., Shih, C.-H., Yamane, H., Vinken, K., Takemoto, K., Sunagawa, T., Yeh, H.-W., Yamanaka, J., Mason, I., et al. Rethinking vlms and llms for image classification. ar Xiv preprint ar Xiv:2410.14690, 2024. Dhillon, G. S., Chaudhari, P., Ravichandran, A., and Soatto, S. A baseline for few-shot image classification. ar Xiv preprint ar Xiv:1909.02729, 2019. Dubey, A., Jauhri, A., Pandey, A., Kadian, A., Al-Dahle, A., Letman, A., Mathur, A., Schelten, A., Yang, A., Fan, A., et al. The llama 3 herd of models. ar Xiv preprint ar Xiv:2407.21783, 2024. Esser, P., Kulal, S., Blattmann, A., Entezari, R., M uller, J., Saini, H., Levi, Y., Lorenz, D., Sauer, A., Boesel, F., et al. Scaling rectified flow transformers for high-resolution image synthesis. In Forty-first International Conference on Machine Learning, 2024. Finn, C., Abbeel, P., and Levine, S. Model-agnostic metalearning for fast adaptation of deep networks. In International conference on machine learning, pp. 1126 1135. PMLR, 2017. Gidaris, S. and Komodakis, N. Dynamic few-shot visual learning without forgetting. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 4367 4375, 2018. Google. Gemini: a family of highly capable multimodal models. ar Xiv preprint ar Xiv:2312.11805, 2023. Ha, H., Jin, X., Kim, J., Liu, J., Wang, Z., Nguyen, K. D., Blume, A., Peng, N., Chang, K.-W., and Ji, H. Synthia: Novel concept design with affordance composition. ar Xiv preprint ar Xiv:2502.17793, 2025. Hao, X., Zhu, Y., Appalaraju, S., Zhang, A., Zhang, W., Li, B., and Li, M. Mixgen: A new multi-modal data augmentation. In Proceedings of the IEEE/CVF winter conference on applications of computer vision, pp. 379 389, 2023. Hurst, A., Lerer, A., Goucher, A. P., Perelman, A., Ramesh, A., Clark, A., Ostrow, A., Welihinda, A., Hayes, A., Radford, A., et al. Gpt-4o system card. ar Xiv preprint ar Xiv:2410.21276, 2024. Jin, X., Kim, J., Zhou, Y., Huang, K.-H., Wu, T.-L., Peng, N., and Ji, H. Armada: Attribute-based multimodal data augmentation. ar Xiv preprint ar Xiv:2408.10086, 2024. Kumar, T., Brennan, R., Mileo, A., and Bendechache, M. Image data augmentation approaches: A comprehensive survey and future directions. IEEE Access, 2024. LAION. Laion-gpt-4v, 2024. URL https: //huggingface.co/datasets/laion/ gpt4v-dataset. Liu, A., Feng, B., Xue, B., Wang, B., Wu, B., Lu, C., Zhao, C., Deng, C., Zhang, C., Ruan, C., et al. Deepseekv3 technical report. ar Xiv preprint ar Xiv:2412.19437, 2024a. Liu, H., Li, C., Li, Y., and Lee, Y. J. Improved baselines with visual instruction tuning, 2023a. Contrastive Visual Data Augmentation Liu, H., Li, C., Wu, Q., and Lee, Y. J. Visual instruction tuning, 2023b. Liu, H., Li, C., Li, Y., and Lee, Y. J. Improved baselines with visual instruction tuning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 26296 26306, 2024b. Liu, H., Li, C., Li, Y., Li, B., Zhang, Y., Shen, S., and Lee, Y. J. Llava-next: Improved reasoning, ocr, and world knowledge, January 2024c. URL https://llava-vl.github.io/blog/ 2024-01-30-llava-next/. Liu, H., Li, C., Wu, Q., and Lee, Y. J. Visual instruction tuning. Advances in neural information processing systems, 36, 2024d. Liu, Z., Tang, Z., Shi, X., Zhang, A., Li, M., Shrivastava, A., and Wilson, A. G. Learning multimodal data augmentation in feature space. ar Xiv preprint ar Xiv:2212.14453, 2022. Liu, Z., Liang, H., Huang, X., Xiong, W., Yu, Q., Sun, L., Chen, C., He, C., Cui, B., and Zhang, W. Synthvlm: High-efficiency and high-quality synthetic data for vision language models. ar Xiv preprint ar Xiv:2407.20756, 2024e. Nichol, A. On first-order meta-learning algorithms. ar Xiv preprint ar Xiv:1803.02999, 2018. Open AI. Gpt-4v(ision) system card, 2023. URL https: //api.semanticscholar.org/Corpus ID: 263218031. Radford, A., Kim, J. W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., Sastry, G., Askell, A., Mishkin, P., Clark, J., et al. Learning transferable visual models from natural language supervision. In International conference on machine learning, pp. 8748 8763. PMLR, 2021. Recraft.AI. Recraft v3 model, 2024. URL https://www. recraft.ai/. Sharifzadeh, S., Kaplanis, C., Pathak, S., Kumaran, D., Ilic, A., Mitrovic, J., Blundell, C., and Banino, A. Synth2 : Boosting visual-language models with synthetic captions and image embeddings. ar Xiv preprint ar Xiv:2403.07750, 2024. Stability AI. Introducing stable diffusion 3.5, 2024. URL https://stability.ai/news/ introducing-stable-diffusion-3-5. Sung, Y.-L., Cho, J., and Bansal, M. Vl-adapter: Parameterefficient transfer learning for vision-and-language tasks. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 5227 5237, 2022. Tian, Y., Wang, Y., Krishnan, D., Tenenbaum, J. B., and Isola, P. Rethinking few-shot image classification: a good embedding is all you need? In Computer Vision ECCV 2020: 16th European Conference, Glasgow, UK, August 23 28, 2020, Proceedings, Part XIV 16, pp. 266 282. Springer, 2020. Uddin, A., Monira, M., Shin, W., Chung, T., Bae, S.-H., et al. Saliencymix: A saliency guided data augmentation strategy for better regularization. ar Xiv preprint ar Xiv:2006.01791, 2020. Van Horn, G., Mac Aodha, O., Song, Y., Cui, Y., Sun, C., Shepard, A., Adam, H., Perona, P., and Belongie, S. The inaturalist species classification and detection dataset. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 8769 8778, 2018. Vinyals, O., Blundell, C., Lillicrap, T., Wierstra, D., et al. Matching networks for one shot learning. Advances in neural information processing systems, 29, 2016. Wikidata. Wikidata: A free, collaborative, multilingual, secondary database, 2024. URL https://www. wikidata.org. Wu, D., Wan, Y., and Chang, K.-W. Visualized text-toimage retrieval. ar Xiv preprint ar Xiv:2505.20291, 2025. Wu, Q., Zhao, M., He, Y., Huang, L., Ono, J., Wakaki, H., and Mitsufuji, Y. Towards reporting bias in visual-language datasets: bimodal augmentation by decoupling object-attribute association. ar Xiv preprint ar Xiv:2310.01330, 2023a. Wu, T.-L., Zhou, Y., and Peng, N. Localizing active objects from egocentric vision with symbolic world knowledge. ar Xiv preprint ar Xiv:2310.15066, 2023b. Wu, Z., Chen, X., Pan, Z., Liu, X., Liu, W., Dai, D., Gao, H., Ma, Y., Wu, C., Wang, B., Xie, Z., Wu, Y., Hu, K., Wang, J., Sun, Y., Li, Y., Piao, Y., Guan, K., Liu, A., Xie, X., You, Y., Dong, K., Yu, X., Zhang, H., Zhao, L., Wang, Y., and Ruan, C. Deepseek-vl2: Mixture-ofexperts vision-language models for advanced multimodal understanding, 2024. URL https://arxiv.org/ abs/2412.10302. Xiao, J., Hays, J., Ehinger, K. A., Oliva, A., and Torralba, A. Sun database: Large-scale scene recognition from abbey to zoo. In 2010 IEEE computer society conference on computer vision and pattern recognition, pp. 3485 3492. IEEE, 2010. Xie, T., Cheng, X., Wang, X., Liu, M., Deng, J., Zhou, T., and Liu, M. Cut-thumbnail: A novel data augmentation for convolutional neural network. In Proceedings of the 29th ACM International Conference on Multimedia, pp. 1627 1635, 2021. Contrastive Visual Data Augmentation Yang, C.-F., Yin, D., Hu, W., Peng, N., Zhou, B., and Chang, K.-W. Verbalized representation learning for interpretable few-shot generalization. ar Xiv preprint ar Xiv:2411.18651, 2024. Yang, K., Deng, J., An, X., Li, J., Feng, Z., Guo, J., Yang, J., and Liu, T. Alip: Adaptive language-image pre-training with synthetic caption. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 2922 2931, 2023. Yang, S., Xiao, W., Zhang, M., Guo, S., Zhao, J., and Shen, F. Image data augmentation for deep learning: A survey. ar Xiv preprint ar Xiv:2204.08610, 2022. Yuksekgonul, M., Bianchi, F., Kalluri, P., Jurafsky, D., and Zou, J. When and why vision-language models behave like bags-of-words, and what to do about it? ar Xiv preprint ar Xiv:2210.01936, 2022. Zhang, Y., Unell, A., Wang, X., Ghosh, D., Su, Y., Schmidt, L., and Yeung-Levy, S. Why are visually-grounded language models bad at image classification? ar Xiv preprint ar Xiv:2405.18415, 2024. Zheng, X., Chalasani, T., Ghosal, K., Lutz, S., and Smolic, A. Stada: Style transfer as data augmentation. ar Xiv preprint ar Xiv:1909.01056, 2019. Contrastive Visual Data Augmentation A. Appendix A.1. Limitations and Future Work Our work is not without limitations. First, in our experiments, we focus on the fine-tuning use case as it is the most general and intuitive way to utilize our augmented visual data. In the future, we plan to investigate other conceivable use cases for our augmented data, include model adaptation (Sung et al., 2022), test-time augmentation (Gidaris & Komodakis, 2018), visual information retrieval (Wu et al., 2025), and more. Third, the modularity of our method also invites other researchers to replace components of Co DA with superior models to achieve better performance. The Novel Species dataset, which can be updated with new species in future years, may also be used to evaluate future VLMs novel concept recognition abilities. Finally, we also expect improved versions of T2I generation-based visual data augmentation techniques to eventually surpass Co DA in effectiveness and efficiency. Potential improvements may include more robust image / feature filtering and more controllable text-conditioned image generation like multi-view synthesis. We hope our work can pave the way for future downstream advancements by demonstrating effective uses of our augmented visual data for enhancing model capabilities. A.2. Data Selection Strategy For each dataset, we focus on a randomly selected subset of concepts that the model is unable to recognize. The data selection strategy is as follows: In each iteration, we select a random subset of 15 species across different supercategories, including Birds, Mammals, and Reptiles. This strategy allows us to identify confusing pairs without overloading the system, progressively building a collection of challenging cases from each subset. For each species within a subset, we create prompts in a multiple-choice format, incorporating the image and a randomized list of options from all species in the subset. Based on the response from the LMM, we are able to highlight specific species that are commonly mistaken for each other, guiding us in selecting pairs for further analysis. In particular, misclassification happens when an image of one species is identified by the LLM to be an image of another species. A pair (A, B) is considered as a confusing pair if rate of misclassification on either direction is above the threshold 0.2. The process is repeated across new subsets, incrementally building an ample dataset of concepts the model has difficulty recognizing. A.3. Experiment Details A.3.1. FEATURE EXTRACTION For textual feature extraction, we use GPT-4o-mini with chain-of-thought reasoning, running with Open AI API calls. Each API call processes up to 2048 tokens, costing approximately 0.0025 per 1K input tokens and 0.005 per 1K output tokens. Given an average of 500 tokens per query and 10 queries per concept, the estimated cost per concept is around $0.0375. For visual feature extraction, we utilize GPT-4o-mini running with Open AI API calls. Images are preprocessed to a resolution of 336x336 pixels and normalized before feature embedding extraction. Each image query incurs a cost similar to textual feature extraction. With an estimated 5 images processed per concept, the cost per concept amounts to approximately 0.1875. With the rapid advancement of open-weights large language models and vision language models including Deep Seek V3 (Liu et al., 2024a), Deep Seek VL2 (Wu et al., 2024), Llama 3.2 (Dubey et al., 2024), and more; we expect that feature extraction LLMs and VLMs can be replaced with these models with none or minimal impact to performance. We plan to perform experiments on some of these models and provide comparison results in the next updated version of our work. A.3.2. FEATURE FILTERING We employ CLIP for automatic feature filtering, evaluating Discriminability and Generability scores. Discriminability is computed using cosine similarity between feature embeddings of target and misidentified concepts, with a threshold of 0.6. Generability is assessed by comparing feature presence in synthetic images using an ensemble of Stable Diffusion 3.5 Large and Recraft V3 models. The feature selection step is executed on an NVIDIA A100 GPU, processing features in approximately 2 hours. Top 5 ranked features are selected per concept. A.3.3. IMAGE GENERATION AND VERIFICATION For synthetic image generation, we employ Stable Diffusion 3.5 Large, running on a single A100 GPU. Additionally, we also integrate the Recraft V3 model through an API call. Image generation is performed at a resolution of 512x512 pixels Contrastive Visual Data Augmentation with a guidance scale of 7.5. The pipeline generates 50 images per concept in approximately 1.2 seconds per image. Post-generation, we perform automated verification using LLa VA V1.6-34b, running on an A6000 GPU. Each image would take approximately 1 minutes to run for feature presence using a feature-matching confidence threshold of 0.85. Images with a satisfaction rate S(isynthetic, F, M) < 1.0 are discarded. A.3.4. MODEL UPDATING We train V1.6-34b with supervised fine-tuning (SFT) using Lo RA with rank 128 and alpha 256, optimizing memory efficiency while maintaining model expressiveness. The training runs on two NVIDIA A6000 GPUs, leveraging Deep Speed Zero-3 for distributed optimization and mixed precision (bf16) for efficiency. The vision encoder is CLIP-Vi T-Large Patch14-336, with an MLP projector aligning visual and text features. We use a cosine learning rate scheduler with a 3% warmup ratio, training for 30 epochs with a batch size of 5 and a learning rate of 2e-4. Images are padded for aspect ratio consistency, and gradient checkpointing is enabled to reduce memory usage. Checkpoints are saved every 50,000 steps, retaining only the most recent one. A.3.5. EVALUATION Automatic evaluation measures zero-shot classification accuracy on a held-out test set. Inference runs on a single A6000 GPU with a batch size of 20, taking approximately 1 hour to complete. The prompt templates for evaluation are attached to Appendix A.4 A.4. Prompt Construction ## Prompt for Visual/Text Feature Extractions: # Contrastive Visual GPT_Contrastive_Visual_Prompt = "You are an experienced and meticulously observant biological scientist who is asked to carefully assess the provided image. As labelled in the image, the left half of the image contains a picture of the animal {main_class} and the right half contains a picture of the animal {confusing_class}. Now, your task is summarize the key distinctive visual attributes possessed by {main_class} (on the left of the image) that makes uniquely discernible from the {confusing_class} (on the right half of the image). Reason step by step to produce an answer. Finally, output the key visual attributes of a {main_class} (that make it distinct from a { confusing_class}) in a python list format containing short phrases of less than 8 words each. Do not output any features of the {confusing_class} in your python list. Make sure not to name the {main_class} or the {confusing_class} in any of the attributes in your list. Also, please try not to use negation in the visual attributes you generate: for example, change features like lack of facial markings to plain brown face . Additionally, do not use comparative form in any of the features you output, for example, change features like thinner body than the other class to thin # Visual GPT_Visual_Prompt = "You are an experienced and meticulously observant biological scientist who is asked to carefully assess the provided image. The image contains a picture of the animal {main_class}. Now, your task is summarize the key distinctive visual attributes possessed by {main_class}. Reason step by step to produce an answer. Finally, output the key visual attributes of a {main_class} in a python list format containing short phrases of less than 8 words each. Make sure not to name the { main_class} in any of the attributes in your list. Also, please try not to use negation in the visual attributes you generate: for example, change features like lack of facial markings to plain brown face . Additionally, do not use comparative form in any of the features you output, for example, change features like thinner body than the other class to thin body ." # Contrastive Text GPT_Contrastive_Text_Prompt = "You are an experienced and knowledgeable scene classification specialist who is tasked to summarize the key distinctive visual attributes possessed by {main_class} that makes uniquely discernible from the { confusing_class} (just based on a visual image). First retrieve your knowledge about the two different types of scenes, then reason step by step to produce an answer. Contrastive Visual Data Augmentation Finally, output the key visual attributes of a {main_class} (distinct from a { confusing_class}) in a python list format containing short phrases of less than 8 words each. Do not output any features of the {confusing_class} in your python list. Make sure not to name the {main_class} or the {confusing_class} in any of the attributes in your list. Also, please try not to use negation in the visual attributes you generate: for example, instead of saying no bright lights, use dark environment. Additionally, do not use comparative forms in any of the features you provide. For instance, instead of saying smaller windows than the other place, use small windows. " # Text GPT_Text_Prompt = "You are an experienced and knowledgeable scene classification specialist who is tasked to summarize the key distinctive visual attributes possessed by {main_class}. First retrieve your knowledge about the {main_class}, then reason step by step to produce an answer. Finally, output the key visual attributes of a { main_class} in a python list format containing short strings of less than 8 words each . Make sure not to name the {main_class} in any of the attributes in your list. Do not output any features of the {confusing_class} in your python list. Also, please try not to use negation in the visual attributes you generate: for example, instead of saying no bright lights, use dark environment. Additionally, do not use comparative forms in any of the features you provide. For instance, instead of saying smaller windows than the other place, use small windows. " ## Text to Image Generation Prompt Stable_Diffusion_Text_to_Image_Generation_Prompt = "Generate a 4K realistic image of { main_class} that contains the following attributes: " + , .join(attributes) ## Feature Verification Prompt llava_Verification_Prompt = "You are an image verification specialist. Your task is to meticulously assess the image for specific attributes and confirm their presence. For each attribute in the list, carefully check the image, examine visual elements such as color, shape, texture, position, and context clues that might indicate whether the attribute is present. Provide a binary python output list, where each element is either 1 (attribute is present) or 0 (attribute is absent), corresponding exactly to the order of attributes provided.\\n\\n Attributes to Verify:{attributes}\\n\\n Expected Output: A list of 0s and 1s indicating the presence or absence of each attribute, in the same order as listed. Here is an example output: [0, 1, 1]." ## Finetune and Evaluation Prompt llava_Finetune_and_Evaluation_Prompt = "You are an image classification specialist with expertise in categorizing images into specific groups. Given an image, identify its category from the following options: " + ", ".join(provided_options_capitalized[:-1]) + ", or " + provided_options_capitalized[-1] + ". Provide your answer as only one category name for precise classification. Please response with the category name only. " ## Deduplication Prompt GPT_Deduplication_Prompt = "You are an experienced and knowledgeable biological scientist who is tasked to summarize the key distinctive visual attributes possessed by { main_class} into a coherent list. Given the following list of attributes describing the animal species {main_class}: {attributes_list}. You task is to combine the duplicate features (which have the same or very similar meanings) into one. Then, you will order the remaining features in order of visual importance, the most visually significant / observable features will be at the front of the list while the least visually observable features will be at the back. Finally, output the key visual attributes of a {main_class} in a python list format containing short phrases of less than 8 words each. Make sure not to name the {main_class} in any of the attributes in your list. Also, please try not to use negation in the visual attributes you generate: for example, change features like lack of facial markings to plain brown face . Additionally, do not use comparative form in any of the features you output, for example, change features like thinner body than the other class to thin body ." ## System Prompt GPT_System_Prompt = "You are a helpful assistant."