# diffclip_differential_attention_meets_clip__a63548eb.pdf Published in Transactions on Machine Learning Research (08/2025) Diff CLIP: Differential Attention Meets CLIP Hasan Abed Al Kader Hammoud hasanabedalkader.hammoud@kaust.edu.sa King Abdullah University of Science and Technology (KAUST) Bernard Ghanem bernard.ghanem@kaust.edu.sa King Abdullah University of Science and Technology (KAUST) Reviewed on Open Review: https: // openreview. net/ forum? id= 2I2f Tehry2 We propose Diff CLIP, a novel vision-language model that extends the differential attention mechanism to CLIP architectures. Differential attention was originally developed for large language models to amplify relevant context while canceling out noisy information. In this work, we integrate this mechanism into CLIP s dual encoder (image and text) framework. With minimal additional parameters, Diff CLIP achieves superior performance on imagetext understanding tasks. Across zero-shot classification, retrieval, and robustness benchmarks, Diff CLIP consistently outperforms baseline CLIP models. Notably, these gains come with negligible computational overhead, demonstrating that differential attention can significantly enhance multi-modal representations without sacrificing efficiency. Code and models can be found at https://github.com/hammoudhasan/Diff CLIP. 1 Introduction Vision-language models (VLMs) have made remarkable progress in bridging the gap between textual and visual modalities, enabling powerful capabilities such as zero-shot image classification, image-text retrieval, and descriptive captioning (Radford et al., 2021; Jia et al., 2021). By aligning images and text in a joint embedding space, these models capture broad semantic relationships across modalities and often excel at outof-distribution generalization. Among VLMs, Contrastive Language-Image Pre-training (CLIP) (Radford et al., 2021) stands out as a foundational approach, demonstrating strong zero-shot performance on numerous benchmarks with minimal fine-tuning. While CLIP s contrastive training regime has been widely adopted, its attention mechanism can sometimes focus on irrelevant or spurious features in both the image and text encoders. This attention noise can hamper fine-grained understanding, particularly when precise localization or explicit contextual knowledge is required. Interestingly, recent language modeling research has proposed a differential attention mechanism (Ye et al., 2024), which subtracts complementary attention distributions to suppress noise and highlight salient tokens. However, whether a similar strategy would be effective for multimodal tasks has remained an open question. Can differential attention be adapted to vision-language models in a way that meaningfully improves their ability to focus on relevant features across modalities? Motivated by this question, we introduce Diff CLIP, an extension of CLIP that integrates differential attention into both the vision and text encoders. By learning two attention maps and subtracting one from the other, Diff CLIP effectively cancels out misaligned or noisy signals, enabling a more precise alignment of images and text. Crucially, this enhancement introduces only a negligible overhead in model parameters and computational cost. Our results show that Diff CLIP consistently outperforms standard CLIP in a wide range of tasks, including linear probing, few-shot classification, image text retrieval, outside-domain robustness, and fine-grained visual understanding, highlighting the efficacy of differential attention in a multimodal Published in Transactions on Machine Learning Research (08/2025) CLIP Diff CLIP 0 Accuracy (%) Linear Probing CLIP Diff CLIP 0 CLIP Diff CLIP 0 Image Retrieval CLIP Diff CLIP 0 Accuracy (%) Text Retrieval CLIP Diff CLIP 0 Image Net Zero-Shot CLIP Diff CLIP 0.0 Zero Shot OOD CC3M Pretraining CLIP vs. Diff CLIP Figure 1: CC3M Pretraining: CLIP vs. Diff CLIP Across Six Tasks. We compare standard CLIP (blue) and our Diff CLIP variant (pink) on linear probing, few-shot classification, image/text retrieval, zeroshot Image Net, and zero-shot OOD. In each case, Diff CLIP consistently outperforms CLIP, highlighting the effectiveness of differential attention with only 0.003% extra parameters. setting. As shown in Figure 1, Diff CLIP is capable of improving performance across various benchmarks with only 0.003% extra parameters. Figure 2 also shows how Diff CLIP is capable of suppressing attention noise compared to CLIP models with vanilla non-differential attention. Our contributions are threefold: We propose Diff CLIP, the first integration of differential attention into CLIP-based VLMs, yielding a simple yet effective approach to reducing attention noise in both vision and text streams. Through extensive experiments on Conceptual Captions 3M/12M pretraining, we demonstrate consistent gains over baseline CLIP across a diverse suite of tasks, with a minimal parameter overhead of roughly 0.003%. We perform detailed ablations, showing that (i) dynamic initialization can boost zero-shot performance, and (ii) applying differential attention solely in the vision encoder already captures most of the benefits, suggesting a flexible and cost-effective path to improved multimodal learning. The remainder of this paper is organized as follows. Section 2 surveys previous work on training-centric, model-centric, and data-centric strategies for enhancing CLIP. Section 3 provides an overview of the standard Transformer attention mechanism, the differential attention concept, and the CLIP framework. Section 4 details our experimental setup, empirical results, and ablation studies, while Sections 5 and 7 concludes with a discussion of future research directions and wrapping up of the paper. 2 Related Work Vision-language pre-training (VLP) has advanced our ability to learn joint representations of images and text, leading to improvements in tasks such as image retrieval, visual question answering, and zero-shot classification (Gan et al., 2022; Zhang et al., 2024). CLIP (Radford et al., 2021) has been central to this progress by using a contrastive loss to align image and text embeddings from large-scale image-caption data. Published in Transactions on Machine Learning Research (08/2025) Figure 2: Comparing CLIP vs. Diff CLIP Attention Maps. For two images (rows), we visualize where CLIP and Diff CLIP attend when matching each image against two different textual queries. While CLIP allocates attention to irrelevant background regions, Diff CLIP more effectively centers on query-relevant objects, highlighting how differential attention can reduce noise and improve focus. Queries: First Row: Mug , Lamp ; Second Row: Flower , Dog . Despite CLIP s strong zero-shot performance, researchers continue to explore improvements in its training, architecture, and data collection strategies. These efforts generally fall into three categories: training-centric, model-centric, and data-centric approaches. Training-Centric Approaches A common strategy is to enrich CLIP s contrastive framework with additional objectives. For example, SLIP (Mu et al., 2022) adds masked image modeling to boost downstream results, while De CLIP (Li et al., 2021) uses nearest-neighbor supervision to enhance data efficiency. Sig LIP (Zhai et al., 2023) replaces the standard softmax temperature with a sigmoid loss, allowing larger batch training and improving generalization and robustness to noisy labels. Retrieval-Enhanced CLIP (Iscen et al., 2023) leverages external memory of image-text pairs at inference, achieving significant gains on fine-grained zero-shot tasks. Further, novel training objectives, such as those proposed by Yang et al. (Yang et al., 2022) and Pyramid CLIP (Gao et al., 2022), aggregate information across multiple semantic levels, highlighting the benefit of diversified training signals for improved CLIP performance. Model-Centric Approaches Another line of work modifies CLIP s architecture for greater efficiency or accuracy. The original CLIP (Radford et al., 2021) employs a Transformer (Vaswani et al., 2017) for text and either a Res Net (He et al., 2016) or Vision Transformer (Vi T) (Dosovitskiy et al., 2020) for images. Subsequent studies incorporate ideas from object detection and segmentation to capture finer visual details, such as region-level representations (Xu et al., 2022; Zhong et al., 2022). Recently, Vi Tamin (Chen et al., 2024) proposed a specialized vision transformer architecture tailored specifically for multimodal models, demonstrating improved zero-shot results compared to standard Vi Ts under similar training setups. Other researchers attempt to unify image and text encoders into a single Transformer (Tschannen et al., 2022), although this approach is less common. Notably, few methods have altered the core attention mechanism within CLIP. Our work addresses this gap by adapting Differential Attention (Ye et al., 2024), originally proposed for language models, to CLIP s multimodal setting. This adaptation aims to reduce attention noise and enhance representation quality. Data-Centric Approaches Data-centric methods emphasize improving the size, diversity, and quality of pre-training datasets. Initial efforts focused on scaling datasets (Jia et al., 2021; Radford et al., 2021), while more recent approaches prioritize richer and cleaner supervision. Ve CLIP (Lai et al., 2024) uses large language models (LLMs) to generate detailed and enriched captions, enhancing textual supervision. Published in Transactions on Machine Learning Research (08/2025) Similarly, CLIPS (Liu et al., 2024b) utilizes truncated synthetic captions to improve visual grounding and retrieval performance, showing that carefully controlled synthetic textual inputs can surpass standard image-caption pairs. Synth CLIP (Hammoud et al., 2024) explores training entirely on synthetic image-text pairs. Further methods employ filtering techniques to eliminate noisy or irrelevant samples (Gadre et al., 2023; Abbas et al., 2023), while Cluster Masking (Wei et al., 2024) proposes masking clusters of similar image patches, leading to faster training and improved representation quality. These efforts underline the potential of data curation and augmentation strategies in bolstering the efficacy of CLIP-based models. Beyond performance, fairness and compositionality have also received increased attention. Fair CLIP (Luo et al., 2024) addresses demographic biases found in models like CLIP by using optimal-transport-based feature alignment across demographic groups. Meanwhile, iterated learning approaches (Zheng et al., 2024) tackle the compositional limitations of large vision-language models, promoting representations that generalize more reliably to complex and compositional visual-linguistic scenarios. In this paper, we contribute to the model-centric direction by adapting Differential Attention (Ye et al., 2024) to CLIP s dual-encoder architecture. Through this adaptation, we aim to reduce attention noise and enhance performance across various image-text understanding tasks. 3 Preliminaries In this section, we outline the fundamental concepts that are essential for our approach. We begin by reviewing the Transformer self-attention mechanism (Vaswani et al., 2017), which is widely used in modern sequence modeling. Next, we introduce differential attention (Ye et al., 2024), a technique designed to reduce attention noise by leveraging complementary attention distributions. Finally, we summarize the Contrastive Language-Image Pre-training (CLIP) framework (Radford et al., 2021), which learns to align images and text in a shared representation space. These components form the basis for our model and experiments. 3.1 Transformer Attention Transformer networks (Vaswani et al., 2017) capture relationships among elements in a sequence through a self-attention operation. Let X RN d be an input sequence of N tokens (or image patches), each embedded in a d-dimensional space. The Transformer maps X to queries (Q), keys (K), and values (V) using learned weight matrices: Q = X W Q, K = X W K, V = X W V , where W Q, W K, W V Rd d. Self-attention scores are then computed via scaled dot-products: A = softmax QK and these scores are used to weight V: Attn(X) = A V. To capture different types of relationships, Transformers use multi-head attention (MHA). An MHA module with h heads splits each projection into lower-dimensional parts of size dh = d/h. In each head i, Attni(X) = softmax Qi K i dh where Qi, Ki, Vi RN dh. The head outputs are concatenated and projected back: MHA(X) = Attn1(X) . . . Attnh(X) W O, with W O R(h dh) d. Despite remarkable success in many areas, standard attention can assign nonnegligible weights to irrelevant tokens (often called attention noise) (Kamradt, 2023; Liu et al., 2024a), which can degrade performance in settings requiring precise focus. Published in Transactions on Machine Learning Research (08/2025) 3.2 Differential Attention Differential attention (Ye et al., 2024) addresses attention noise by learning two separate attention distributions and subtracting one from the other, effectively canceling out spurious alignments. Single-Head Differential Attention. Let X RN d be the input to a single attention head. We split Q and K into two halves, denoted by subscripts 1 and 2: [Q1; Q2] = X W Q, [K1; K2] = X W K, V = X W V , where Q1, Q2, K1, K2 RN d 2 . Each half computes its own attention distribution: A1 = softmax Q1K 1 , A2 = softmax Q2K 2 The output is formed by subtracting the second distribution (scaled by a learnable parameter λ) from the first: Diff Attn(X) = A1 λ A2 V. The parameter λ is trained to control how strongly the second distribution is subtracted: λ = exp λq1 λk1 exp λq2 λk2 + λinit, where λq1, λk1, λq2, λk2 are learnable weights and λinit is a hyperparameter. This subtraction often yields a sparser, more focused attention map, which can improve results in scenarios sensitive to background or redundant signals (Ye et al., 2024). Multi-Head Extension. Like standard attention, differential attention can be extended to multiple heads. In Differential Multi-Head Attention (Diff MHA), each head i applies the differential step independently: Diff Attni(X) = softmax Q1,i K 1,i p λ softmax Q2,i K 2,i p where Q1,i, Q2,i, K1,i, K2,i RN (dh/2). The final output is then Diff MHA(X) = Diff Attn1(X) . . . Diff Attnh(X) W O. By learning complementary attention maps in each head and subtracting them, Diff MHA aims to amplify relevant patterns while reducing noise. 3.3 CLIP Training Contrastive Language-Image Pre-training (CLIP) (Radford et al., 2021) learns image and text embeddings in a shared space using a large collection of paired image-text examples {(Ik, Tk)}M k=1. It consists of two encoders: one for images fθ and one for text gϕ . Their outputs are normalized to unit length: ui = fθ(Ii) fθ(Ii) 2 , vi = gϕ(Ti) gϕ(Ti) 2 . For a batch of N pairs, CLIP forms a similarity matrix Sij = u i vj where τ is a (learned or fixed) temperature parameter. The text-to-image contrastive loss is i=1 log exp(Sii) PN j=1 exp(Sij) , Published in Transactions on Machine Learning Research (08/2025) Table 1: Classification Performance (Linear Probing and Few-Shot). We compare CLIP and Diff CLIP on nine classification tasks with two pretraining sets (CC3M and CC12M). The top block reports linear probing accuracy, while the bottom block shows few-shot results. Numbers in parentheses indicate absolute gains or drops for Diff CLIP relative to CLIP. Pretraining Model Caltech-101 DTD Pets Flowers SUN397 Aircraft CIFAR10 CIFAR100 Food-101 Avg. Linear Probing CC3M CLIP 72.5 58.7 61.0 85.8 54.1 35.7 83.5 63.4 59.1 63.8 CC3M Diff CLIP 76.2 (+3.7) 60.2 (+1.5) 62.2 (+1.2) 86.6 (+0.8) 56.2 (+2.1) 34.6 (-1.1) 83.9 (+0.4) 63.7 (+0.3) 59.4 (+0.3) 64.8 (+1.0) CC12M CLIP 88.3 71.2 79.5 92.6 68.3 48.8 92.0 74.7 77.5 77.0 CC12M Diff CLIP 89.5 (+1.2) 71.8 (+0.6) 83.0 (+3.5) 93.5 (+0.9) 69.4 (+1.1) 46.4 (-2.4) 90.7 (-1.3) 73.3 (-1.4) 77.7 (+0.2) 77.3 (+0.3) CC3M CLIP 90.4 72.9 69.6 92.5 91.8 44.6 63.4 72.8 67.0 73.9 CC3M Diff CLIP 91.6 (+1.2) 73.2 (+0.3) 71.6 (+2.0) 92.9 (+0.4) 92.8 (+1.0) 45.4 (+0.8) 62.4 (-1.0) 73.5 (+0.7) 68.3 (+1.3) 74.6 (+0.7) CC12M CLIP 97.4 81.9 86.3 96.9 96.5 56.1 81.3 85.1 86.0 85.3 CC12M Diff CLIP 97.6 (+0.2) 82.2 (+0.3) 88.2 (+1.9) 97.3 (+0.4) 96.8 (+0.3) 55.2 (-0.9) 80.3 (-1.0) 83.3 (-1.8) 87.5 (+1.5) 85.4 (+0.1) Table 2: Zero-Shot Retrieval and Image Net Zero-shot Accuracy. We report image and text retrieval (Recall@5, %) and zero-shot Image Net accuracy (%) for CLIP vs. Diff CLIP, using CC3M or CC12M as pretraining data. Values in parentheses reflect absolute gains or drops for Diff CLIP relative to CLIP. Pretraining Model Image Retrieval (R@5) Text Retrieval (R@5) Zero-Shot Flickr30k Flickr8k MSCOCO Avg. Flickr30k Flickr8k MSCOCO Avg. Image Net CC3M CLIP 31.8 35.4 19.4 28.9 43.4 46.2 25.4 38.3 13.6 CC3M Diff CLIP 32.9 (+1.1) 36.5 (+1.1) 20.9 (+1.5) 30.1 (+1.2) 44.7 (+1.3) 47.8 (+1.6) 27.6 (+2.2) 40.1 (+1.8) 14.4 (+0.8) CC12M CLIP 62.5 62.1 41.3 55.3 76.8 77.7 53.8 69.4 31.8 CC12M Diff CLIP 62.2 (-0.3) 61.5 (-0.6) 42.3 (+1.0) 55.3 (+0.0) 77.4 (+0.6) 77.4 (-0.3) 55.5 (+1.7) 70.1 (+0.7) 33.8 (+2.0) and the image-to-text counterpart is i=1 log exp(Sii) PN j=1 exp(Sji). The overall objective is 2 Lti + Lit . By encouraging matching image-text pairs to have high similarity (and non-matching pairs to have low similarity), CLIP learns robust features that often transfer well to downstream tasks like zero-shot classification and retrieval. 4 Experiments We present an extensive empirical study to investigate whether differential attention can benefit CLIP-style vision-language models. We first describe our dataset sources and training configurations, then evaluate both standard CLIP and our Diff CLIP variant under linear probing, few-shot classification, and image-text retrieval. We also test robustness to distribution shifts (via OOD Image Net) and fine-grained features (via MMVP), and conclude with ablation studies on the initialization of the differential attention parameter λinit and on applying differential attention to only the vision encoder. 4.1 Experimental Setup Datasets. We pretrain on Conceptual Captions 3M (CC3M) (Sharma et al., 2018) and Conceptual Captions 12M (CC12M) (Changpinyo et al., 2021). After downloading using img2dataset (Beaumont, Published in Transactions on Machine Learning Research (08/2025) Image Net IN-v2 IN-A IN-R IN-Sketch Avg. 0 Accuracy (%) CC3M Pretraining Image Net IN-v2 IN-A IN-R IN-Sketch Avg. 0 CC12M Pretraining CLIP Diff CLIP Figure 3: OOD Zero-Shot Image Net Performance. Comparison of zero-shot accuracy (%) on Image Net, Image Net-V2, Image Net-A, Image Net-R, and Image Net-Sketch, plus the average. Bars show performance of CLIP (blue) versus Diff CLIP (pink), trained on CC3M (left) or CC12M (right). Numerical deltas above the bars indicate the absolute improvement or drop for Diff CLIP relative to CLIP. Diff CLIP improves on average the zero-shot performance on OOD Image Net datasets as compared to CLIP. 2021) (with shorter edge resized to 224), we end up with about 2.3M image-text pairs for CC3M and 7.9M for CC12M. For CC3M, we train on four A100 GPUs, while CC12M uses eight A100 GPUs to reduce training time. Text data is minimally processed, limited to basic tokenization. Training Parameters. All models train for 40 epochs, using one epoch of linear warmup, a global batch size of 4096, and Adam W optimizer (Loshchilov & Hutter, 2017). We set the base learning rate to 5 10 4 with weight decay of 0.5. For Diff CLIP, every attention layer in both the vision and text encoders is replaced with differential attention. We initialize each layer s λ at 0.8 unless stated otherwise. This setup introduces only a minor parameter overhead: roughly 0.003% additional parameters relative to a standard CLIP-B/16. Training parameters are chosen similar to Synth CLIP (Hammoud et al., 2024) and training code is adopted from SLIP (Mu et al., 2022). Evaluation Protocol. We follow established practices for linear probing and few-shot evaluation (El Banani et al., 2023) on nine image-classification datasets: DTD (Cimpoi et al., 2014), Flowers (Nilsback & Zisserman, 2008), Pets, Caltech-101 (Li et al., 2022), Aircraft (Maji et al., 2013), CIFAR-10 (Krizhevsky et al., 2009), SUN397 (Xiao et al., 2010), CIFAR-100 (Krizhevsky et al., 2009), and Food-101 (Bossard et al., 2014). For retrieval (image-to-text and text-to-image) on Flickr8k (Rashtchian et al., 2010), Flickr30k (Young et al., 2014), and MSCOCO (Lin et al., 2014), we use the LAION CLIP Benchmark framework (Schuhmann et al., 2022). We measure zero-shot robustness on Image Net (Russakovsky et al., 2015) and its variants (Image Net V2 (Recht et al., 2019), Image Net-A (Hendrycks et al., 2021b), Image Net-R (Hendrycks et al., 2021a), and Image Net-Sketch (Wang et al., 2019)). Finally, we use the MMVP-VLM benchmark (Tong et al., 2024) to check how well each model focuses on fine-grained visual details. 4.2 Do CLIP Models Benefit from Differential Attention? Motivation. To evaluate the effectiveness of our proposed Diff CLIP, we test its performance across tasks involving image classification, image-text retrieval, and zero-shot generalization, following common benchmarks established in prior literature (Hammoud et al., 2024). Results. We compare a baseline CLIP-B/16 to our Diff CLIP-B/16 (with differential attention in both vision and text encoders). Table 1 shows linear probing and few-shot classification results for models pretrained on both CC3M and CC12M. Diff CLIP outperforms standard CLIP on almost every dataset. For example, with CC3M pretraining, Diff CLIP achieves about +1% gain in linear probing and +0.7% in fewshot accuracy. Published in Transactions on Machine Learning Research (08/2025) Table 2 presents retrieval metrics and zero-shot Image Net. Diff CLIP again surpasses CLIP on image and text retrieval: for CC3M, we see an average improvement of about 1.2% (image retrieval) and 1.8% (text retrieval). On zero-shot Image Net, Diff CLIP-CC3M increases accuracy by 0.8%, with even larger gains of +2.0% when using CC12M. Conclusion. Even though Diff CLIP only adds a tiny fraction of extra parameters, it consistently outperforms standard CLIP on classification and retrieval benchmarks. This suggests that differential attention is a lightweight yet effective way to enhance vision-language representation. 4.3 Does Differential Attention Improve Out-of-Domain Robustness? Motivation. Having observed improvements on in-distribution Image Net, we ask if these gains carry over to more challenging out-of-domain variants. Real-world applications often involve domain shifts, and CLIP s zero-shot adaptability has been tested on Image Net-V2, Image Net-A, Image Net-R, and Image Net Sketch benchmarks known to stress model robustness beyond standard Image Net. Understanding how differential attention influences robustness in such scenarios is crucial for assessing its practical utility in deployment settings. We aim to see if differential attention helps maintain or improve performance under such shifts. Results. Figure 3 summarizes zero-shot performance across Image Net-V2, Image Net-A, Image Net-R, and Image Net-Sketch. Models with differential attention outperform standard CLIP by an average of 2.1%, suggesting that subtracting noisy attention patterns yields features that generalize more robustly, even under significant distribution shifts. Conclusion. Diff CLIP not only enhances in-distribution performance but also strengthens zeroshot robustness against substantial domain shifts, further demonstrating the benefits of differential attention. 4.4 Does Diff CLIP Improve Fine-Grained Visual Understanding? MMVP-VLM Benchmark. To test fine-grained visual understanding, we employ the MMVP-VLM benchmark (Tong et al., 2024). This benchmark measures how well vision-language models capture nuanced visual properties, such as object orientation, presence, and relational context, beyond straightforward recognition. Both CLIP and Diff CLIP are pretrained on CC12M under identical settings. Results. On average, Diff CLIP improves MMVP-VLM accuracy by 5.7% relative to baseline CLIP. A radar plot (Figure 4) shows Diff CLIP surpassing or matching CLIP on nearly all categories except one (state and condition). This suggests that subtracting noisy attention patterns (via differential attention) helps the model attend to more subtle details in images. Interpretation. The MMVP benchmark is specifically designed to require precise visual focus on relevant object details in order to answer fine-grained questions correctly. Therefore, the observed +5.7% absolute improvement over baseline CLIP provides strong quantitative evidence that Diff CLIP s reduced attentional noise translates directly into more accurate and discriminative visual focus. This aligns with our qualitative visualizations, where Diff CLIP consistently attends to the most semantically relevant regions, supporting our central claim that differential attention enhances the quality of attention maps in CLIP-style models. Conclusion. By mitigating extraneous context through differential attention, Diff CLIP achieves stronger fine-grained visual understanding. These gains highlight the effectiveness of explicitly canceling irrelevant attention weights in multimodal settings. Published in Transactions on Machine Learning Research (08/2025) Orientation and Presence of Specific Features State and Condition Positional and Color and Appearance Structural Characteristics Viewpoint and Perspective Model Performance CLIP (avg: 21.9%) Diff CLIP (avg: 27.6%) Figure 4: MMVP-VLM Benchmarking. Radar plot illustrating performance on different fine-grained visual categories. Both models (CLIP in blue, Diff CLIP in pink) are evaluated on properties like orientation, positional context, and color appearance. Diff CLIP (average 27.6%) consistently outperforms CLIP (average 21.9%), demonstrating more focused attention on subtle visual details. 4.5 Dynamic or Static λinit? Motivation. All previous experiments used a fixed initialization λinit = 0.8 for differential attention. However, (Ye et al., 2024) proposes a dynamic schedule: λinit(l) = 0.8 0.6 exp( 0.3 l), where l is the layer index. We denote the model using this schedule as Diff CLIP . Results. Figure 5 summarizes six tasks: linear probing, few-shot classification, image retrieval, text retrieval, zero-shot Image Net, and zero-shot OOD. Compared to the baseline CC12M CLIP, Diff CLIP improves zero-shot Image Net by +2.8% and text retrieval by +1.5%. It also raises zero-shot OOD accuracy by +1.3%. However, relative to standard Diff CLIP (with fixed λinit = 0.8), Diff CLIP is +0.8% better on zero-shot Image Net and +0.8% on text retrieval, but it underperforms or only slightly improves on other tasks. For instance, in zero-shot OOD, Diff CLIP is -0.8% behind standard Diff CLIP. Conclusion. A dynamic λ schedule yields notable gains on zero-shot Image Net and text retrieval, though it lags behind the simpler constant initialization on several other benchmarks. Future work might explore how best to tune or combine these schedules to achieve consistent improvements. 4.6 Does Applying Differential Attention to Vision Only Suffice? Motivation. Because the vision encoder often plays a dominant role in CLIP models, one might ask whether differential attention is necessary in both encoders. We define Diff CLIP as a variant that integrates differential attention only in the vision encoder, leaving the text encoder with regular attention. Results. Figure 5 compares CLIP, Diff CLIP, and Diff CLIP across six tasks: linear probing, few-shot classification, image retrieval, text retrieval, zero-shot Image Net, and zero-shot OOD. Diff CLIP improves upon baseline CLIP by +0.1% in linear probing, +0.3% in few-shot, +0.4% in image retrieval, +1.2% in text Published in Transactions on Machine Learning Research (08/2025) Linear Probing Image Retrieval Text Retrieval Image Net Zero-Shot Zero Shot OOD Accuracy (%) CC12M Pretraining CLIP Diff CLIP Diff CLIP* Diff CLIP Figure 5: Comparing Different Diff CLIP Variants. We evaluate four models on six tasks (linear probing, few-shot, image retrieval, text retrieval, Image Net zero-shot, and zero-shot OOD), all pretrained on CC12M. CLIP (blue) is the baseline, Diff CLIP (pink) uses a fixed differential attention parameter, Diff CLIP (purple) employs a dynamic schedule for differential attention, and Diff CLIP (yellow) applies differential attention only to the vision encoder. retrieval, +1.9% on zero-shot Image Net, and +2.3% on zero-shot OOD. Compared to Diff CLIP, Diff CLIP surpasses or matches performance on few-shot, image retrieval, text retrieval, and zero-shot OOD, but is slightly behind on linear probing and standard zero-shot Image Net. Conclusion. Applying differential attention solely to the vision encoder already brings sizable gains. Interestingly, Diff CLIP can even match or exceed full Diff CLIP on certain tasks, suggesting that most of the performance boost may come from more robust visual feature extraction. 5 Future Directions & Limitations 5.1 Beyond CLIP An intriguing question for future research is how a vision encoder trained with differential attention within the CLIP framework would perform when integrated into larger, more sophisticated vision-language models such as LLa VA (Liu et al., 2023) or Tiny LLa VA (Zhou et al., 2024). To provide initial insights into this possibility, we conducted preliminary experiments by combining our Diff CLIP-CC12M vision encoder with the Qwen-2.5-Instruct-0.5B (Yang et al., 2024) language encoder. We followed a typical two-stage training procedure: first, a linear projector was trained to align visual tokens with the language embedding space, freezing all other components; second, both the projector and the language encoder underwent instruction fine-tuning. For the projection training, we utilized the LAION-CC-SBU dataset (558K image-text pairs) used in the LLa VA training setup. For instruction fine-tuning, we adopted the COCO (Lin et al., 2014) subset (approximately 350K pairs) also used by LLa VA. All experiments were conducted using the Tiny LLa VA repository Table 3: CLIP vs. Diff CLIP on POPE Hallucination Benchmark. We compare models across three POPE categories, showing accuracy, precision, and recall. Absolute improvements of Diff CLIP over CLIP are highlighted in parentheses. CLIP Diff CLIP POPE Accuracy Precision Recall Accuracy Precision Recall Random 50.14 50.07 98.21 50.41 (+0.27) 50.21 (+0.14) 98.56 (+0.35) Popular 50.27 50.13 99.33 50.27 (+0.00) 50.13 (+0.00) 99.47 (+0.14) Adversarial 50.17 50.08 99.33 50.20 (+0.03) 50.10 (+0.02) 99.47 (+0.14) Published in Transactions on Machine Learning Research (08/2025) on 4 A100-80GB GPUs. The hyperparameters for fine-tuning included a batch size of 48 samples per GPU, a learning rate of 2 10 5, zero weight decay, a warm-up ratio of 0.03, and cosine decay scheduling. The projection pretraining similarly employed 48 samples per GPU, a learning rate of 1 10 3, no weight decay, a warm-up ratio of 0.03, and cosine decay scheduling. We evaluated the resulting models on the POPE (Li et al., 2023) hallucination dataset, which assesses models susceptibility to visual hallucinations. Despite the modest size of observed improvements, Diff CLIP-CC12M consistently outperformed the CLIP-CC12M baseline across all metrics (Table 3). These initial findings suggest that differential attention-trained vision encoders could enhance performance when integrated into broader vision-language architectures, making it a promising direction for further exploration. 5.2 Scaling Data and Architecture Training CLIP models with a Vi T-B/16 backbone on the CC12M dataset (7.9M samples) currently requires approximately 10 GPU-days on A100 GPUs, translating to roughly $600 using Google Cloud Platform (GCP). A natural future direction would involve exploring how differential attention performs when scaling to larger architectures (e.g., Vi T-L or Vi T-H) and substantially bigger datasets (e.g., LAION-400M). Investigating such scaling could reveal whether the performance gains observed with Diff CLIP persist or even amplify as model size and dataset scale increase, offering insights into the broader applicability and benefits of differential attention in vision-language pretraining. 6 Limitations and Future Work Although Diff CLIP demonstrates consistent gains in a wide range of tasks, several limitations remain. First, due to computational constraints inherent to academic settings, we are unable to perform large-scale multiseed pre-training runs (e.g., on LAION-400M). Instead, we focus on smaller scale datasets and we provide statistical measures only for nondeterministic evaluations (linear probing, few-shot), and we rely on the breadth and consistency of results as supporting evidence. Second, while preliminary results on integrating Diff CLIP-trained encoders into larger vision-language models (e.g., LLa VA-style) are promising, exploring this direction at scale remains a future work. We believe that these extensions, particularly scaling to larger datasets and architectures, will further reveal the potential of differential attention in multimodal learning. 7 Conclusion We introduced Diff CLIP, which integrates differential attention into CLIP-based vision-language models to better filter out noisy alignments. Through extensive experiments on classification, retrieval, robustness, and fine-grained benchmarks, Diff CLIP consistently improves over standard CLIP with minimal overhead. Further ablations highlight the flexibility of dynamic attention schedules and vision-only setups. We hope these findings inspire future research on more efficient, robust attention mechanisms in large multimodal learning. 8 Acknowledgements The research reported in this publication was supported by funding from King Abdullah University of Science and Technology (KAUST) - Center of Excellence for Generative AI, under award number 5940. Amro Abbas, Kushal Tirumala, Dániel Simig, Surya Ganguli, and Ari S. Morcos. Semdedup: Data-efficient learning at web-scale through semantic deduplication, 2023. 4 Romain Beaumont. img2dataset: Easily turn large sets of image urls to an image dataset. https://github. com/rom1504/img2dataset, 2021. 6 Published in Transactions on Machine Learning Research (08/2025) Lukas Bossard, Matthieu Guillaumin, and Luc Van Gool. Food-101 mining discriminative components with random forests. In ECCV, pp. 446 461. Springer, 2014. 7 Soravit Changpinyo, Piyush Sharma, Nan Ding, and Radu Soricut. Conceptual 12M: Pushing web-scale image-text pre-training to recognize long-tail visual concepts. In CVPR, 2021. 6 Jieneng Chen, Qihang Yu, Xiaohui Shen, Alan Yuille, and Liang-Chieh Chen. Vitamin: Designing scalable vision models in the vision-language era. In CVPR, pp. 12954 12966, 2024. 3 M. Cimpoi, S. Maji, I. Kokkinos, S. Mohamed, and A. Vedaldi. Describing textures in the wild. In CVPR, 2014. 7 Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. ar Xiv preprint ar Xiv:2010.11929, 2020. 3 Mohamed El Banani, Karan Desai, and Justin Johnson. Learning Visual Representations via Language Guided Sampling. In CVPR, 2023. 7 Samir Yitzhak Gadre, Gabriel Ilharco, Alex Fang, Jonathan Hayase, Georgios Smyrnis, Thao Nguyen, Ryan Marten, Mitchell Wortsman, Dhruba Ghosh, Jieyu Zhang, Eyal Orgad, Rahim Entezari, Giannis Daras, Sarah Pratt, Vivek Ramanujan, Yonatan Bitton, Kalyani Marathe, Stephen Mussmann, Richard Vencu, Mehdi Cherti, Ranjay Krishna, Pang Wei Koh, Olga Saukh, Alexander Ratner, Shuran Song, Hannaneh Hajishirzi, Ali Farhadi, Romain Beaumont, Sewoong Oh, Alex Dimakis, Jenia Jitsev, Yair Carmon, Vaishaal Shankar, and Ludwig Schmidt. Datacomp: In search of the next generation of multimodal datasets, 2023. 4 Zhe Gan, Linjie Li, Chunyuan Li, Lijuan Wang, Zicheng Liu, Jianfeng Gao, et al. Vision-language pretraining: Basics, recent advances, and future trends. Foundations and Trends in Computer Graphics and Vision, 14(3 4):163 352, 2022. 2 Yuting Gao, Jinfeng Liu, Zihan Xu, Jun Zhang, Ke Li, Rongrong Ji, and Chunhua Shen. Pyramidclip: Hierarchical feature alignment for vision-language model pretraining. Neur IPS, 35:35959 35970, 2022. 3 Hasan Abed Al Kader Hammoud, Hani Itani, Fabio Pizzati, Philip Torr, Adel Bibi, and Bernard Ghanem. Synthclip: Are we ready for a fully synthetic clip training? ar Xiv preprint ar Xiv:2402.01832, 2024. 4, 7 Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In CVPR, pp. 770 778, 2016. 3 Dan Hendrycks, Steven Basart, Norman Mu, Saurav Kadavath, Frank Wang, Evan Dorundo, Rahul Desai, Tyler Zhu, Samyak Parajuli, Mike Guo, et al. The many faces of robustness: A critical analysis of out-of-distribution generalization. In ICCV, pp. 8340 8349, 2021a. 7 Dan Hendrycks, Kevin Zhao, Steven Basart, Jacob Steinhardt, and Dawn Song. Natural adversarial examples. In CVPR, pp. 15262 15271, 2021b. 7 Ahmet Iscen, Mathilde Caron, Alireza Fathi, and Cordelia Schmid. Retrieval-enhanced contrastive visiontext models. ar Xiv preprint ar Xiv:2306.07196, 2023. 3 Chao Jia, Yinfei Yang, Ye Xia, Yi-Ting Chen, Zarana Parekh, Hieu Pham, Quoc Le, Yun-Hsuan Sung, Zhen Li, and Tom Duerig. Scaling up visual and vision-language representation learning with noisy text supervision. In ICML, pp. 4904 4916. PMLR, 2021. 1, 3 Greg Kamradt. Needle in a Haystack - pressure testing LLMs. https://github.com/gkamradt/LLMTest_ Needle In AHaystack/tree/main, 2023. 4 Alex Krizhevsky, Geoffrey Hinton, et al. Learning multiple layers of features from tiny images. 2009. 7 Published in Transactions on Machine Learning Research (08/2025) Zhengfeng Lai, Haotian Zhang, Bowen Zhang, Wentao Wu, Haoping Bai, Aleksei Timofeev, Xianzhi Du, Zhe Gan, Jiulong Shan, Chen-Nee Chuah, et al. Veclip: Improving clip training via visual-enriched captions. In ECCV, pp. 111 127. Springer, 2024. 3 Fei-Fei Li, Marco Andreeto, Marc Aurelio Ranzato, and Pietro Perona. Caltech 101, Apr 2022. 7 Yangguang Li, Feng Liang, Lichen Zhao, Yufeng Cui, Wanli Ouyang, Jing Shao, Fengwei Yu, and Junjie Yan. Supervision exists everywhere: A data efficient contrastive language-image pre-training paradigm. ar Xiv preprint ar Xiv:2110.05208, 2021. 3 Yifan Li, Yifan Du, Kun Zhou, Jinpeng Wang, Xin Zhao, and Ji-Rong Wen. Evaluating object hallucination in large vision-language models. In Houda Bouamor, Juan Pino, and Kalika Bali (eds.), ACL, pp. 292 305, Singapore, December 2023. Association for Computational Linguistics. doi: 10.18653/v1/2023. emnlp-main.20. URL https://aclanthology.org/2023.emnlp-main.20/. 11 Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C Lawrence Zitnick. Microsoft coco: Common objects in context. In ECCV, pp. 740 755. Springer, 2014. 7, 10 Haotian Liu, Chunyuan Li, Qingyang Wu, and Yong Jae Lee. Visual instruction tuning. Neur IPS, 36: 34892 34916, 2023. 10 Nelson F Liu, Kevin Lin, John Hewitt, Ashwin Paranjape, Michele Bevilacqua, Fabio Petroni, and Percy Liang. Lost in the middle: How language models use long contexts. Transactions of the Association for Computational Linguistics, 12:157 173, 2024a. 4 Yanqing Liu, Xianhang Li, Zeyu Wang, Bingchen Zhao, and Cihang Xie. Clips: An enhanced clip framework for learning with synthetic captions. ar Xiv preprint ar Xiv:2411.16828, 2024b. 4 Ilya Loshchilov and Frank Hutter. Decoupled weight decay regularization. ar Xiv preprint ar Xiv:1711.05101, 2017. 7 Yan Luo, Min Shi, Muhammad Osama Khan, Muhammad Muneeb Afzal, Hao Huang, Shuaihang Yuan, Yu Tian, Luo Song, Ava Kouhana, Tobias Elze, et al. Fairclip: Harnessing fairness in vision-language learning. In CVPR, pp. 12289 12301, 2024. 4 S. Maji, J. Kannala, E. Rahtu, M. Blaschko, and A. Vedaldi. Fine-grained visual classification of aircraft. Technical report, 2013. 7 Norman Mu, Alexander Kirillov, David Wagner, and Saining Xie. Slip: Self-supervision meets languageimage pre-training. In ECCV, pp. 529 544. Springer, 2022. 3, 7 Maria-Elena Nilsback and Andrew Zisserman. Automated flower classification over a large number of classes. In 2008 Sixth Indian conference on computer vision, graphics & image processing, pp. 722 729. IEEE, 2008. 7 Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language supervision. In ICML, pp. 8748 8763. Pm LR, 2021. 1, 2, 3, 4, 5 Cyrus Rashtchian, Peter Young, Micah Hodosh, and Julia Hockenmaier. Collecting image annotations using amazon s mechanical turk. In Proceedings of the NAACL HLT 2010 workshop on creating speech and language data with Amazon s Mechanical Turk, pp. 139 147, 2010. 7 Benjamin Recht, Rebecca Roelofs, Ludwig Schmidt, and Vaishaal Shankar. Do imagenet classifiers generalize to imagenet? In ICML, pp. 5389 5400. PMLR, 2019. 7 Olga Russakovsky, Jia Deng, Hao Su, Jonathan Krause, Sanjeev Satheesh, Sean Ma, Zhiheng Huang, Andrej Karpathy, Aditya Khosla, Michael Bernstein, et al. Imagenet large scale visual recognition challenge. IJCV, 115:211 252, 2015. 7 Published in Transactions on Machine Learning Research (08/2025) Christoph Schuhmann, Romain Beaumont, Richard Vencu, Cade Gordon, Ross Wightman, Mehdi Cherti, Theo Coombes, Aarush Katta, Clayton Mullis, Mitchell Wortsman, et al. Laion-5b: An open large-scale dataset for training next generation image-text models. Neur IPS, 35:25278 25294, 2022. 7 Piyush Sharma, Nan Ding, Sebastian Goodman, and Radu Soricut. Conceptual captions: A cleaned, hypernymed, image alt-text dataset for automatic image captioning. In ACL, 2018. 6 Shengbang Tong, Zhuang Liu, Yuexiang Zhai, Yi Ma, Yann Le Cun, and Saining Xie. Eyes wide shut? exploring the visual shortcomings of multimodal llms. In CVPR, pp. 9568 9578, 2024. 7, 8 Michael Tschannen, Basil Mustafa, and Neil Houlsby. Image-and-language understanding from pixels only. ar Xiv preprint ar Xiv:2212.08045, 2022. 3 Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. Neur IPS, 30, 2017. 3, 4 Haohan Wang, Songwei Ge, Zachary Lipton, and Eric P Xing. Learning robust global representations by penalizing local predictive power. Neur IPS, 32, 2019. 7 Zihao Wei, Zixuan Pan, and Andrew Owens. Efficient vision-language pre-training by cluster masking. In CVPR, pp. 26815 26825, 2024. 4 Jianxiong Xiao, James Hays, Krista A Ehinger, Aude Oliva, and Antonio Torralba. Sun database: Large-scale scene recognition from abbey to zoo. In CVPR, pp. 3485 3492. IEEE, 2010. 7 Jiarui Xu, Shalini De Mello, Sifei Liu, Wonmin Byeon, Thomas Breuel, Jan Kautz, and Xiaolong Wang. Groupvit: Semantic segmentation emerges from text supervision. ar Xiv preprint ar Xiv:2202.11094, 2022. 3 An Yang, Baosong Yang, Beichen Zhang, Binyuan Hui, Bo Zheng, Bowen Yu, Chengyuan Li, Dayiheng Liu, Fei Huang, Haoran Wei, et al. Qwen2. 5 technical report. ar Xiv preprint ar Xiv:2412.15115, 2024. 10 Jianwei Yang, Chunyuan Li, Pengchuan Zhang, Bin Xiao, Ce Liu, Lu Yuan, and Jianfeng Gao. Unified contrastive learning in image-text-label space. In CVPR, pp. 19163 19173, 2022. 3 Tianzhu Ye, Li Dong, Yuqing Xia, Yutao Sun, Yi Zhu, Gao Huang, and Furu Wei. Differential transformer. ar Xiv preprint ar Xiv:2410.05258, 2024. 1, 3, 4, 5, 9 Peter Young, Alice Lai, Micah Hodosh, and Julia Hockenmaier. From image descriptions to visual denotations: New similarity metrics for semantic inference over event descriptions. Transactions of the Association for Computational Linguistics, 2:67 78, 2014. 7 Xiaohua Zhai, Basil Mustafa, Alexander Kolesnikov, and Lucas Beyer. Sigmoid loss for language image pre-training. In ICCV, pp. 11975 11986, 2023. 3 Jingyi Zhang, Jiaxing Huang, Sheng Jin, and Shijian Lu. Vision-language models for vision tasks: A survey. IEEE TPAMI, 2024. 2 Chenhao Zheng, Jieyu Zhang, Aniruddha Kembhavi, and Ranjay Krishna. Iterated learning improves compositionality in large vision-language models. In CVPR, pp. 13785 13795, 2024. 4 Yiwu Zhong, Jianwei Yang, Pengchuan Zhang, Chunyuan Li, Noel Codella, Liunian Harold Li, Luowei Zhou, Xiyang Dai, Lu Yuan, Yin Li, et al. Regionclip: Region-based language-image pretraining. In CVPR, pp. 16793 16803, 2022. 3 Baichuan Zhou, Ying Hu, Xi Weng, Junlong Jia, Jie Luo, Xien Liu, Ji Wu, and Lei Huang. Tinyllava: A framework of small-scale large multimodal models. ar Xiv preprint ar Xiv:2402.14289, 2024. 10 Published in Transactions on Machine Learning Research (08/2025) A Extra Visualizations In this appendix, we provide more visualizations of the attention maps similar to Figure 2. Figure 6: Additional Visualizations of Diff CLIP Attention vs CLIP Attention On the Difficulty of Quantitatively Evaluating Attention Maps. While qualitative comparisons (as in Figure 6) clearly show that Diff CLIP produces more focused and object-centric attention patterns than CLIP, turning these visual differences into a robust quantitative metric is challenging. A common approach would be to compare attention maps to ground-truth object masks (e.g., from segmentation datasets) using Published in Transactions on Machine Learning Research (08/2025) Io U scores. However, this requires converting soft attention maps into binary masks, which introduces several issues: (i) selecting a single threshold is problematic because optimal values vary widely across images, especially between small and large objects; (ii) attention scores may be diffuse by design for some heads, making them poorly suited to binary evaluation; and (iii) per-image threshold tuning risks introducing bias and undermining reproducibility. Given these constraints, we rely instead on downstream task performance (e.g., the +5.7% gain on the MMVP benchmark) as an indirect but more meaningful quantitative indicator that the attention mechanism improves focus on semantically relevant regions. B Statistical Significance In this appendix, we run the non-deterministic linear probing and few-shot evaluation from Table 1 on three seeds and report the mean and standard deviation performance. The following tables (4 and 5) present the aggregated results, enabling a clearer comparison that accounts for non-determinism. Table 4: Three-seed Linear Probing (mean sd, %). Each cell reports the average accuracy and standard deviation over three seeds for non-deterministic evaluations. Pretraining Model Caltech-101 DTD Pets Flowers SUN397 Aircraft CIFAR10 CIFAR100 Food-101 Avg. CC3M CLIP 72.5 0.1 58.7 0.0 61.0 0.0 85.8 0.0 54.1 0.0 35.6 0.0 83.5 0.0 63.4 0.0 59.1 0.0 63.7 0.0 CC3M Diff CLIP 76.2 0.0 60.2 0.0 62.2 0.0 86.6 0.0 56.2 0.0 34.6 0.0 83.9 0.0 63.7 0.0 59.4 0.0 64.8 0.0 CC12M CLIP 88.4 0.2 71.2 0.0 79.9 0.8 92.6 0.1 68.3 0.0 48.8 0.0 92.0 0.0 74.7 0.0 77.5 0.0 77.1 0.1 CC12M Diff CLIP 89.5 0.0 71.8 0.0 83.0 0.0 93.5 0.0 69.4 0.0 46.8 0.0 90.7 0.0 73.3 0.0 77.7 0.0 77.3 0.0 Table 5: Three-seed Few-Shot (mean sd, %). Average accuracy and standard deviation over three seeds. Pretraining Model Caltech-101 DTD Pets Flowers SUN397 Aircraft CIFAR10 CIFAR100 Food-101 Avg. CC3M CLIP 90.3 0.4 72.5 0.5 70.6 0.2 92.6 0.4 91.7 0.1 44.6 0.3 63.6 0.2 72.6 0.3 66.9 0.5 73.9 0.1 CC3M Diff CLIP 91.4 0.3 72.9 0.5 72.6 0.2 93.0 0.3 92.7 0.2 45.3 0.3 62.6 0.2 73.2 0.4 68.3 0.3 74.7 0.1 CC12M CLIP 97.3 0.1 81.9 0.4 86.8 0.2 96.5 0.3 96.5 0.2 56.5 0.3 80.7 0.2 84.8 0.2 86.1 0.1 85.2 0.0 CC12M Diff CLIP 97.6 0.1 82.2 0.4 88.6 0.3 97.1 0.2 96.6 0.1 55.1 0.2 79.8 0.1 83.2 0.2 87.4 0.1 85.3 0.0 C Intuitive Explanation of Differential Attention While Section 3.2 presents the formal definition of differential attention, here we offer a more intuitive explanation to complement the mathematical formulation. Differential attention learns two separate attention maps, A1 and A2, for the same input: A1 = softmax A2 = softmax The final attention output is computed as: Diff Attn(X) = (A1 λ A2) V. (3) Interpretation. A1 is encouraged by the contrastive objective to highlight the most salient, task-relevant features (e.g., the dog in an image). Published in Transactions on Machine Learning Research (08/2025) A2 is learned through separate projections and, via the subtraction, is implicitly encouraged to capture more diffuse, high-entropy patterns often corresponding to noise such as background textures ( grass, sky ) or features spuriously correlated with the main object. Subtracting a scaled version of A2 from A1 reduces these irrelevant activations, producing a sparser and more discriminative attention distribution. This mechanism helps the model cancel out noisy or distracting visual patterns while amplifying the signals that are most relevant to the text query. In practice, we find that this results in attention maps that more tightly focus on semantically important regions, which contributes to the improved performance and robustness observed in Section 4.