# mobileclip2_improving_multimodal_reinforced_training__37ea13ba.pdf Published in Transactions on Machine Learning Research (08/2025) Mobile CLIP2: Improving Multi-Modal Reinforced Training Fartash Faghri fartash@apple.com Apple Pavan Kumar Anasosalu Vasu panasosaluvasu@apple.com Apple Cem Koc cem_koc@apple.com Apple Vaishaal Shankar Work done while at Apple Alexander Toshev toshev@apple.com Apple Oncel Tuzel otuzel@apple.com Apple Hadi Pouransari mpouransari@apple.com Apple Reviewed on Open Review: https: // openreview. net/ forum? id= We F9zolng8 Foundation image-text models such as CLIP with zero-shot capabilities enable a wide array of applications. Mobile CLIP is a recent family of image-text models at 3 15ms latency and 50 150M parameters with state-of-the-art zero-shot accuracy. The main ingredients in Mobile CLIP were its low-latency and light architectures and a novel multi-modal reinforced training that made knowledge distillation from multiple caption-generators and CLIP teachers efficient, scalable, and reproducible. In this paper, we improve the multi-modal reinforced training of Mobile CLIP through: 1) better CLIP teacher ensembles trained on the DFN dataset, 2) improved captioner teachers trained on the DFN dataset and fine-tuned on a diverse selection of high-quality image-caption datasets. We discover new insights through ablations such as the importance of temperature tuning in contrastive knowledge distillation, the effectiveness of caption-generator fine-tuning for caption diversity, and the additive improvement from combining synthetic captions generated by multiple models. We train a new family of models called Mobile CLIP2 and achieve state-of-the-art Image Net-1k zero-shot accuracies at low latencies. In particular, we observe 2.2% improvement in Image Net-1k accuracy for Mobile CLIP2-B compared with Mobile CLIP-B architecture. Notably, Mobile CLIP2-S4 matches the zero-shot accuracy of Sig LIP-SO400M/14 on Image Net-1k while being 2 smaller and improves on DFN Vi T-L/14 at 2.5 lower latency. We release our pretrained models 1 and the data generation code 2. The data generation code makes it easy to create new reinforced datasets with arbitrary teachers using distributed scalable processing. Equal contribution. 1https://github.com/apple/ml-mobileclip 2https://github.com/apple/ml-mobileclip-dr Published in Transactions on Machine Learning Research (08/2025) 1 Introduction CLIP (Radford et al., 2021) is an image-text model that maps images and text inputs to a shared embedding space, where a text describing an image, also called caption, is mapped close to an image matching its description but far from dissimilar images. Building on a vast literature (Frome et al., 2013; Socher et al., 2014; Karpathy & Fei-Fei, 2015; Kiros et al., 2014; Faghri et al., 2018), CLIP substantially increased the scale of training data and models. Consequentially, along with improved image-text retrieval performance, new zero-shot classification capabilities emerged with non-trivial accuracy on classification tasks without any explicit supervised training with classification labels through linear probing. The image-encoder can be further specialized to a new task by either linear probing (fixed encoder), or full fine-tuning to achieve state-of-the-art performance on a diverse set of tasks (Wortsman et al., 2022). CLIP is one of the first to be called a foundation model given the diversity of its capabilities and applications (Bommasani et al., 2021). 5 10 15 20 25 30 Latency Sum of Image/Text Encoders (ms) Image Net-1k Zero-shot Accuracy (%) Open AI Data Comp Sig LIP LAION Tiny CLIP ACED Sig LIP2 Mobile CLIP Mobile CLIP2 (ours) 20M 100M 200M Figure 1: Mobile CLIP2 models trained on DFNDR2B achieve state-of-the-art accuracy at low latencies. Mobile CLIP2-S4 matches the accuracy of Sig LIP-SO400M/14 with 2 fewer parameters and surpasses DFN Vi T-L/14 at 2.5 lower latency measured on i Phone12 Pro Max. Mobile CLIPS3/S4 are our new architectures trained on Mobile CLIP s training dataset, Data Comp DR-1B (dashed lines). The success of CLIP resulted in an increase in the sizes of models and datasets, leading to a gradual increase in performance (Fang et al., 2024b; Zhai et al., 2023; Gadre et al., 2023; Fang et al., 2024a). Recently, this trend has been reversed to models with small size and low latency for applications on mobile devices. Notably, Tiny CLIP (Wu et al., 2023) and Mobile CLIP (Vasu et al., 2024c) proposed models with as few as 50M total parameters (sum of image and text encoder parameters). For example, Mobile CLIP-S0, with total latency of 3ms (sum of image and text encoder latencies), achieves similar average performance to the original Open AI Vi T-B/16 CLIP while being 3x smaller and 5x faster. It also demonstrates improved performance compared to prior state-of-the-art larger models, such as Sig LIP (Zhai et al., 2023). In this paper, we present ablations of multi-modal reinforced training and present an improved training recipe. We train a new family of models, Mobile CLIP2, that establishes new state-of-the-art Image Net-1k accuracy at a range of latencies matching the performance of larger Sig LIP (Zhai et al., 2023) and DFN (Fang et al., 2024a) models while up to 4 smaller (our Mobile CLIP2-S2 compared with Sig LIP2-B/32) and up to 2.5 faster (our Mobile CLIP2-S4 compared with DFN Vi T-L/14). Moreover, we release efficient distributed code for generating reinforced datasets using arbitrary teacher models. 2 Improved Training Mobile CLIP introduced a family of low-latency image-text models consisting of S0, S1, S2, B, and B-LT variants with aggregate image-text latencies spanning 3.8-13.7ms. These low latencies were achieved with specialized architectures based on Fast Vi T (Vasu et al., 2023b) and an improved training method called multi-modal reinforced training. We seek to further explore and improve each step of multi-modal reinforced training. We additionally consider a more diverse family of architectures that cover a wider range of latencies. Reinforced training is a method for achieving better performance from a base dataset through improvements from additional sources such as pretrained models (Faghri et al., 2023). Multi-modal reinforced training introduced in Vasu et al. (2024c) adds information to an image-text dataset from pretrained image-text models Published in Transactions on Machine Learning Research (08/2025) Table 1: Summary of Mobile CLIP2 training improvements. Co Ca models are pretrained on a large dataset for 13B seen samples then fine-tuned for 12M seen samples (denoted by ). The architecture for all CLIP teachers in this table is Vi T-L/14. We report mean and standard deviations of 5 runs when available. Name Dataset CLIP Teacher Datasets Co Ca Dataset IN-val Flickr30k Avg. 38 Mobile CLIP (Vasu et al., 2024c) Data Comp-1B12M Open AI + Data Comp-XL LAION-2B MSCOCO-123k 61.6 72.8 53.5 Table 2 DFN-5B12M Open AI + Data Comp-XL LAION-2B MSCOCO-123k 63.10.2 73.30.6 54.10.4 Table 4 DFN-5B12M DFN-2B + DFN-2B-s39B LAION-2B MSCOCO-123k 65.40.4 75.80.3 56.20.6 Mobile CLIP2 (Tab. 6) DFN-5B12M DFN-2B + DFN-2B-s39B DFN-2B MSCOCO-38k 65.90.3 75.40.2 56.50.3 Table 6 DFN-5B12M DFN-2B + DFN-2B-s39B DFN-2B Syn. 10 66.00.1 75.10.6 56.50.3 as well as a pretrained synthetic caption generator. Specifically, they add the following additional information to Data Comp-1B dataset: 1) image embeddings from two CLIP teachers on 10 random augmentations of each image 2) text embeddings from two CLIP teachers on the original text as well as 5 synthetic captions generated from a Co Ca caption generator. Given a reinforced dataset, they modify the training loss to include a knowledge distillation loss given the embeddings from teachers on each sample (Hinton et al., 2015). To ensure consistency between the teacher and student, the same image augmentation is reproduced via stored augmentation parameters (Beyer et al., 2022; Faghri et al., 2023). They perform ablations to find the set of CLIP teachers, caption generator, and image augmentations that provide the largest performance gain on Image Net as well as the average accuracy on 38 evaluations from Data Comp (Gadre et al., 2023). We follow a similar multi-modal reinforced training to Mobile CLIP while improving all aspects and call the resulting model family Mobile CLIP2. Table 1 summarizes the gains from each major improvement. In short, we use better training data, better CLIP teacher models, and better and diverse synthetic caption generators compared to Mobile CLIP. In all ablations, we train Mobile CLIP-B for 30k iterations ( 20 epochs) on datasets with 12.8M images. We provide a summary of datasets in this paper in Tab. 15. Figure 2 demonstrates the efficiency gains compared with DFN (Fang et al., 2024a), Data Comp (Gadre et al., 2023) and Data Comp DR (Vasu et al., 2024c) datasets during training. Training on DFNDR-2B12M for 30M seen samples is 5x more efficient than training on Data Comp-1B12M, i.e., we reach the Image Net-1k zero-shot accuracy of training on Data Comp-1B12M for 30M after seeing only 6M samples of DFNDR-2B12M. Similarly, training on Data Comp DR-12M is 3.3x more efficient compared to DFN-2B12M and 1.3x more efficient compared with Data Comp DR-12M. We also observe 1.6x speedup when training on DFNDR-2B compared with training on Data Comp DR-1B for 13B seen samples. Similar to Data Comp DR, training on DFNDR datasets do not have any wall-clock time overhead, i.e., each training step of training on Data Comp, DFN, Data Comp DR, and DFNDR takes the same amount of time. That means any efficiency gains in terms of the number of samples and training iterations directly translate to wall-clock time efficiency gains. 2.1 Multi-Modal Reinforced Training Dataset Reinforcement (DR) (Faghri et al., 2023) is a method for improving a dataset to achieve higher accuracy with minimal changes to the training code and minimal computational overhead. DR was first introduced for training image classifiers where Faghri et al. (2023) improved the Image Net dataset by storing classification probabilities efficiently from a strong ensemble of classifiers. Given stored probabilities, the training was essentially Knowledge Distillation (Hinton et al., 2015) with no overhead for computing the teacher predictions. The cost efficiency makes it feasible to train longer for larger gains as observed in Beyer et al. (2022). Vasu et al. (2024c) adopted DR for training image-text CLIP models by storing knowledge from a strong ensemble of CLIP models and additionally synthetic captions from an image caption generator. They demonstrated up to 1000 improved learning efficiency compared with non-reinforced CLIP training. Given a batch of b image-text pairs, we denote the embeddings of a target student model by Φimg, Φtxt Rb d, where d is the dimensionality of the shared embedding space. We utilize two types of teachers, an image-text teacher ensemble that maps images and texts to a shared space similar to CLIP (Radford et al., 2021) and Co Ca-based captioners that generate a caption given an image using an encoder-decoder architecture (Yu et al., 2022). Let Ψ(k) img, Ψ(k) txt Rb dk denote the image-text embeddings from the k-th CLIP-based teacher Published in Transactions on Machine Learning Research (08/2025) 5000 10000 15000 20000 25000 30000 Training step Image Net-1k Zero-shot Accuracy (%) DFNDR-12M DFN-12M Data Comp-12M Data Comp DR-12M 20000 40000 60000 80000 100000 120000 140000 160000 180000 200000 Training step Image Net-1k Zero-shot Accuracy (%) DFNDR-2B Data Comp DR-1B Figure 2: Left: Training on DFNDR-12M is up to 5x more efficient compared with Data Comp-1B12M, 3.3x compared with DFN-12M, and 1.3x compared to Data Comp DR-12M. All models are trained for 30k iterations and global batch size 8192 (246M seen samples). DFN-12M consists of 12M uniformly sampled image-text pairs from DFN-2B and DFNDR-12M consists of additional reinforcements per sample in DFN12M. Right: Training on DFNDR-2B is up to 1.7x more efficient compared with Data Comp DR-1B. Models are trained for 200k iterations and gloabl batch size 65536 (13B seen samples). where dk is the dimensionality of the shared space. The distillation loss is defined as LDistill = 1 2b K k=1 KL(Sτk(Ψ(k) img, Ψ(k) txt) Sbτ(Φimg, Φtxt)) | {z } Image to Text + KL(Sτk(Ψ(k) txt, Ψ(k) img) Sbτ(Φtxt, Φimg)) | {z } Text to Image where KL denotes Kullback-Leibler divergence, and Sτ(U, V ) is the row-wise Softmax operation applied to UV /τ with temperature τ. The total loss is LTotal = (1 λ)LCLIP + λLDistill , that is the sum of a standard CLIP loss and the distillation loss with coefficients 1 λ and λ, respectively. 2.2 Better Base Dataset: DFN Multi-modal reinforced training starts from a base dataset containing real image-text pairs commonly collected from the web. Data Comp (Gadre et al., 2023) demonstrated that the quality of large-scale image-text datasets can be significantly improved through filtering based on scores such as compatibility of image and text. Their proposed Best Pool filtering applied on a pool of 12B samples resulted in the Data Comp-1B dataset that was used as the base dataset in Mobile CLIP. Data Comp also released the original 12B samples as a benchmark for dataset curation and filtering methods. DFN (Fang et al., 2024a) proposed to filter data using a filtering network trained on high-quality data. Applying their model on Data Comp-12B pool resulted the DFN-2B dataset. They additionally collected a larger set of images from the web disjoint from Data Comp-12B and after filtering resulted in another 3B samples and collectively created the DFN-5B dataset. We study the impact of replacing the base dataset in Mobile CLIP with DFN-5B. We ablate using 12M uniformly sampled subset of Data Comp-1B referred to as Data Comp-1B12M that was introduced in (Vasu et al., 2024c) for rapid experimentation. We similarly sample a 12M subset from DFN-5B referred to as DFN-5B12M. Table 2 compares the performance of training with and without distillation/synthetic captions. We observe that DFN-5B12M results in up to 1.4% gain together with distillation and synthetic captions. Although this gain is smaller compared to the up to 6% gain without distillation/synthetic captions, it is still more than standard deviation. 2.3 DFN CLIP Teachers One source of reinforcement in multi-modal reinforced training is the embeddings from CLIP teachers that are used as targets in CLIP distillation. (Vasu et al., 2024c) performed a comprehensive study of existing strong CLIP teachers at the time of publication and found the ensemble of Vi T-L-14-openai Published in Transactions on Machine Learning Research (08/2025) Table 2: Training on DFN is better than Data Comp with and without distillation/synthetic captions. CLIP teachers and synthetic caption generators are the same as Mobile CLIP (Open AI+Data Comp XL CLIP-Vi T-L/14, and Co Ca-Vi T-L/14). For distillation, the coefficient λ is set to 1.0 (no CLIP loss) and use strong image augmentations. Dataset Distill. Syn. Caps. IN-val Flickr30k Avg. 38 Data Comp-1B12M 44.6 42.4 40.1 DFN-5B12M 49.9 48.5 43.5 Data Comp-1B12M 51.9 71.8 47.8 DFN-5B12M 54.9 70.7 49.6 Data Comp-1B12M 56.3 57.8 48.7 DFN-5B12M 59.5 60.4 50.0 Data Comp-1B12M 61.6 72.8 53.7 DFN-5B12M 63.0 74.1 54.6 and Vi T-L-14-datacomp_xl_s13b_b90k to result the best student performance. Here we investigate the effectiveness of DFN-pretrained models as teachers. DFN-pretrained CLIP models with Vi T-L-14 and Vi T-H-14 achieve the state-of-the-art performance on Avg. 38 evaluations of Data Comp (Fang et al., 2024a) better than other popular models such as Sig LIP (Zhai et al., 2023). As the choice of the caption generator and the CLIP teachers may depend on each other, we reduce the complexity of our analysis by analyzing the effect of the CLIP teachers on synthetic captions from a Co Ca model without fine-tuning (See Sec. 2.4). We explore the diversity of synthetic captions through fine-tuning in Sec. 2.5. Table 3: Optimal logit scale for distillation varies across teachers. The dataset is DFN-5B12M with synthetic captions generated from Co Ca-DFN-2B in Sec. 2.4. The loss coefficient λ is set to 1.0 and trained using strong image augmentations. Teacher Logit Scale IN-val Flickr30k Avg. 38 datacomp_xl_s13b_b90k-CLIP-Vi T-L-14 50 62.6 65.6 53.3 DFN2B-CLIP-Vi T-L-14 70 65.5 68.0 56.5 DFN5B-CLIP-Vi T-H-14 90 64.0 65.9 54.7 DFN5B-CLIP-Vi T-H-14-384 55 64.6 67.6 54.4 DFN2B-CLIP-Vi T-L-14-s39b 60 65.2 67.5 54.8 Table 4: Ensemble of DFN CLIP teachers improve Image Net-1k validation accuracy by 2.8%. The dataset is DFN-5B12M with synthetic captions generated from Co Ca-DFN-2B in Sec. 2.4. The loss coefficient λ is set to 1.0 and trained using strong image augmentations. The optimal logit scales for each model is set independently based on Tab. 3. Teacher 1 Teacher 2 IN-val Flickr30k Avg. 38 Vi T-L-14-openai Vi T-L-14-datacomp_xl_s13b_b90k 63.1 64.7 55.2 Vi T-L-14-datacomp_xl_s13b_b90k DFN5B-CLIP-Vi T-H-14-384 64.5 67.8 54.5 Vi T-L-14-datacomp_xl_s13b_b90k DFN5B-CLIP-Vi T-H-14 64.4 67.3 55.3 Vi T-L-14-datacomp_xl_s13b_b90k DFN2B-CLIP-Vi T-L-14 65.3 68.1 56.2 DFN5B-CLIP-Vi T-H-14-384 DFN5B-CLIP-Vi T-H-14 64.7 66.9 54.9 DFN5B-CLIP-Vi T-H-14-384 DFN2B-CLIP-Vi T-L-14 65.8 68.6 56.2 DFN5B-CLIP-Vi T-H-14 DFN2B-CLIP-Vi T-L-14 65.2 68.0 55.8 DFN2B-CLIP-Vi T-L-14-s39b datacomp_xl_s13b_b90k 65.1 67.6 55.7 DFN2B-CLIP-Vi T-L-14-s39b DFN5B-CLIP-Vi T-H-14-384 65.7 67.3 55.1 DFN2B-CLIP-Vi T-L-14-s39b DFN5B-CLIP-Vi T-H-14 65.7 68.2 55.7 DFN2B-CLIP-Vi T-L-14-s39b DFN2B-CLIP-Vi T-L-14 65.9 68.7 55.9 Logit scaling. CLIP models are trained with a logit scale that is tuned during the training in the range 0-100. Mobile CLIP used the same logit scalar as the temperature scaling in the KD loss. We observe that Published in Transactions on Machine Learning Research (08/2025) the logit scalar in DFN and Data Comp models is not optimal for KD and tune that further. Table 3 shows the optimal logit scale used for each teacher to train a Mobile CLIP-B model. We observe that the logit scale is not a sensitive hyperparameter where values within a range of 5 points achieve similar performance. Ensemble teachers. We construct ensembles of size two using Data Comp and DFN teachers. Table 4 shows the performance of training a Mobile CLIP-B model using embeddings from various ensembles. We observe significant improvements compared with teachers used in Mobile CLIP. Specifically, IN-val and Flickr30k improve by up to 3%. We choose the ensemble of DFN2B-CLIP-Vi T-L-14-s39b and DFN2B-CLIP-Vi T-L-14 for Mobile CLIP2 based on its performance and cost efficiency compared to other larger or higher resolution ensembles. We utilize the optimal logit scales for each member of the ensemble that is found independently. It is possible that the optimal logit scales for ensemble would vary when used together but we do not further optimize logit scales jointly. 2.4 DFN Caption Generators Another source of reinforcements for training Mobile CLIP2 is synthetic captions generated from an image caption generator. Mobile CLIP used a single Co Ca captioner which has a two-tower image-text architecture coupled with a text decoder (Yu et al., 2022). Compared with most recent VLMs, the text-decoder is fairly light-weight that results in an overall relatively faster caption generator compared with more recent VLMs (Liu et al., 2024b; Vasu et al., 2024a). As Mobile CLIP generated multiple synthetic captions on billions of images, the cost of running Co Ca was an important decision factor. They did not provide analysis on the choice of captioner but observed significant gains from training on synthetic captions compared with not using synthetic captions (7.4% for 30k training iterations). Mobile CLIP generated 5 synthetic captions per image although they observed the majority of the gain comes from the first 1-2 synthetic captions. We explore training a new Co Ca model using the DFN dataset to improve the quality of synthetic captions. We adopt the same architecture as the Co Ca model utilized in Mobile CLIP based on the Vi T-L/14 image encoder. They utilized the model trained on LAION-2B dataset and fine-tuned on MSCOCO-128k dataset. We pretrain the same architecture on DFN-2B for 13B seen samples using Open CLIP (Ilharco et al., 2021). Table 5: Pretraining Co Ca on DFN-2B without fine-tuning results in similar IN-1k performance but worse robustness and retrieval. The dataset is DFN-5B12M, CLIP teachers are the same as Mobile CLIP (Open AI+Data Comp-XL CLIP-Vi T-L/14) and the architecture of the Co Ca model is the same as Co Ca-Vi T-L/14. For distillation, the coefficient λ is set to 1.0 (no CLIP loss) and use strong image augmentations. Values within one standard deviation of the best of each group are highlighted. Distill. High Aug. Co Ca IN-val Flickr30k Avg. 38 LAION-2B MSCOCO-128k DFN-2B 49.9 48.5 43.5 54.9 70.6 49.6 51.1 65.7 45.3 54.6 55.1 46.2 56.8 67.2 48.4 59.5 60.3 50.0 63.0 74.1 54.6 63.1 64.7 55.2 63.4 72.0 55.1 Table 5 demonstrates the impact of DFN-Co Ca synthetic captions on the performance with and without distillation. We observe that utilizing DFN-Co Ca synthetic captions results in improved IN-val and Avg. 38 performance but negatively impacts retrieval. As we observe in Sec. 2.5, the retrieval performance recovers with fine-tuning on high-quality datasets such as MSCOCO. We further observe the synthetic captions from the original Co Ca model can be used together with DFN-Co Ca captions to provide additional gains but these gains are small with distillation. Published in Transactions on Machine Learning Research (08/2025) 2.5 Fine-tuning Caption Generators In Sec. 2.4, we showed that pretraining a Co Ca model on DFN-2B results in improved IN-val and Avg. 38 performance when utilized for multi-modal reinforced training. However, the retrieval performance falls behind which is due to the lack of fine-tuning on a high-quality dataset. Mobile CLIP used a Co Ca model fine-tuned on MSCOCO (Chen et al., 2015). MSCOCO-2017 contains 123k images with captions that have higher quality compared to average image-text pairs in Data Comp and DFN datasets. In this section, we study the impact of fine-tuning on various high-quality datasets. In addition to 123k samples from MSCOCO which we refer to as MSCOCO-123k, we also consider a subset of 38k samples with permissive licenses (CC Attribution 2.0, CC Attribution-Share Alike 2.0, and CC Attribution-No Derivs 2.0) which we refer to as MSCOCO-38k. We also consider GBC-1M/10M (Hsieh et al., 2024), DOCCI-9kshort/extended/complete (Onoe et al., 2025), DCI-8k (Urbanek et al., 2024), and Re Cap-COCO-30k (Li et al., 2024). We fine-tune DFN-Co Ca on each dataset for 12M seen samples using the same loss as Co Ca pretraining. Table 6: The dataset is DFN-5B12M, CLIP teachers are our selected DFN models (DFN2B-CLIP-Vi T-L-14s39b and DFN2B-CLIP-Vi T-L-14) and the architecture of the Co Ca model is the same as Co Ca-Vi T-L/14. For distillation, the coefficient is set to 1.0 (no CLIP loss) and use strong image augmentations. Base Dataset FT Dataset Context len. IN-val Flickr30k Avg. 38 LAION-2B MSCOCO-123k 77 65.40.4 75.80.3 56.20.6 DFN-2B - 77 65.9 68.7 55.9 DFN-2B MSCOCO-123k 77 65.9 76.0 56.2 DFN-2B MSCOCO-38k 77 65.90.3 75.40.2 56.50.3 DFN-2B GBC1M-short 77 65.8 75.0 56.6 DFN-2B DOCCI 77 66.3 72.6 57.3 DFN-2B DCI-short 77 65.9 74.0 56.3 DFN-2B DCI-extended 77 65.7 73.5 56.1 DFN-2B DCI-complete 77 65.8 73.8 56.2 DFN-2B Recap-COCO-30K 77 65.1 73.5 55.5 DFN-2B GBC-1M-long 255 64.7 72.4 55.1 DFN-2B GBC-10M-short-relation 255 65.2 73.8 55.4 DFN-2B GBC-10M-long 255 64.6 71.9 54.6 DFN-2B DOCCI 255 66.1 74.0 57.2 DFN-2B DCI-extended 255 65.7 75.1 55.9 DFN-2B DCI-complete 255 65.6 74.0 56.8 DFN-2B 5 2 77 65.90.2 74.70.4 56.30.2 DFN-2B 10 1 77 66.00.1 75.10.6 56.50.3 Fine-tuning on MSCOCO38k and MSCOCO128k. We observe that restricting fine-tuning to MSCOCO samples with permissive licenses does not have a negative impact on performance. Ablation on number of synthetic captions and beam search. (Vasu et al., 2024c) observed that even though one can generate multiple synthetic captions from a Co Ca model, their effectiveness saturates at 2 per sample for classification tasks. We observe similar results using a single Co Ca model with various sampling strategies. We explore varying the generation method and hyperparameters. Specifically, we used top-p, top-k, and beam-search and observed that beam-search results in qualitatively more diverse captions, however, we did not observe any improvement in downstream performance when utilized for reinforced training. Fine-tuning on GBC1M, GBC12M, DOCCI, DCI, Re Cap-COCO30k. We observe that most fine-tuning datasets underperform MSCOCO fine-tuning or perform on-par within one standard deviation. An exception is fine-tuning on DOCCI results in 0.8% improvement in average of 38 evaluations which is more than one standard deviation from the MSCOCO-38k results. Effect of context length. The context length for training CLIP and Co Ca models is typically set to 77. We explore training Co Ca models to generate longer captions by setting the context length for training and generation to 255. Most results stay within one standard deviation. Recent works have improved the support for long captions in CLIP models with improved loss functions and training strategies (Zhang et al., 2024; Zheng et al., 2024; Najdenkoska et al., 2024). We leave extending these modifications to Co Ca models for future work. Published in Transactions on Machine Learning Research (08/2025) Patch Embed. Stride 2 Patch Embed. Stride 2 Patch Embed. Stride 2 Patch Embed. Stride 2 Convolutional Stem Rep Mixer Stage Self Attention Stage Patch Embed. Stride 2 Patch Embed. Stride 2 Patch Embed. Stride 2 To Projection To Projection (a) MCi architectures for new larger variants. For smaller variants, the image encoder has four distinct stages of compute and for larger variants, we use five stages. The projection layer for MCi models include a Global Avg. Pooling layer followed by a linear layer. 256 512 768 1024 Image Resolution (px) Latency (ms) MCi2 Scaled MCi3 (b) 5-Stage achieves lower latency at higher resolutions. Both models MCi2Scaled and MCi3 are of the same size, with MCi3 using a 5-stage design. Figure 3: Mobile CLIP2 architecture and latency. Effect of synthetic caption diversity. We further explore training with a diverse collection of captions generated from an ensemble of Co Ca models fine-tuned on different datasets. The motivation is the diversity in fine-tuning datasets would increase the divresity in synthetic captions and hence an increase in the effectiveness of additional synthetic captions. We observe that utilizing up to 10 different Co Ca models results in a performance that is still within one standard deviation of the best performance. Reinforced DFN datasets. Our final datasets small DFNDR-5B12M and DFNDR-2B12M consist of 5 synthetic captions with MSCOCO-38k fine-tuning, and embeddings from the ensemble of two DFN2B-Vi TL/14 teachers discussed in Sec. 2.3 for 30 image augmentations as well as ground-truth and synthetic captions. We explored training on only the 2B subset of DFN versus the full 5B set that was expanded with 3B samples outside of the 12B pool of Data Comp. Tab. 7 shows that the average performance on 38 evaluations is within standard deviation for both datasets while Image Net-1k validation accuracy is better with the 12M samples from 5B. However, we did not observe the improvement to hold when training at larger scales and restricted our recipe to the 2B dataset. Table 7: DFNDR-5B12M and DFNDR-2B12M perform similarly on average 38 evaluations. Dataset IN-val Flickr30k Avg. 38 DFNDR-5B12M 65.90.3 75.40.2 56.50.3 DFNDR-2B12M 65.5 74.8 56.4 3 Architecture Our Mobile CLIP2 consists of similar architectures to Mobile CLIP as well as two new variants. Specifically, we train Mobile CLIP2-S0, Mobile CLIP2-S2, and Mobile CLIP2-B where we utilize the standard Base text encoder for Mobile CLIP2-S0 and drop the S1 variant. In addition to architectures introduced in Mobile CLIP, we introduce two new variants in Mobile CLIP2 family, i.e., Mobile CLIP2-S3 and Mobile CLIP2-S4. The text encoders for these variants are pure transformer-based architectures and the image encoders are based on Fast Vi T (Vasu et al., 2023b), which uses train-time overparameterization blocks introduced in (Vasu et al., 2023a). The smaller variants, MCi0, MCi1, and MCi2 are hybrid vision transformers with four distinct stages of compute. We introduce an additional transformer stage for MCi3 and MCi4 preceded by 4 down-sampling of input tensor as shown in Fig. 3a. The 5-stage design has two advantages when scaled up; First, the Published in Transactions on Machine Learning Research (08/2025) parameters can be distributed across five stages with the largest layers operating on four times fewer tokens. Second, the design scales more effectively to higher resolutions. We empirically validate our design choices across various image resolutions. In Fig. 3b, we scale MCi2 to match the size of MCi3 (125 million parameters) and benchmark its performance across four input resolutions. Our results show that MCi3, with its five-stage design, offers a significantly better trade-off compared to a scaled MCi2. At low image resolution, i.e., 256 256, MCi3 is 1.9 faster than similar sized MCi2 and for larger input resolutions, i.e., as 1024 1024, MCi3 is 7.1 faster than a similar sized MCi2. Responsiveness at higher resolutions is particularly important when the image encoder is fine-tuned for dense prediction tasks such as image segmentation, where the input image resolution is 512 512. 4 Experiments In this section, we train a new family of efficient CLIP models, Mobile CLIP2, and evaluate on a diverse set of tasks. Following our findings in Sec. 2, we create the reinforced dataset, DFNDR-2B, which contains five synthetic captions generated from our Co Ca-Vi T-L/14 model pretrained on DFN-2B and fine-tuned on MSCOCO-38K. DFNDR-2B also contains image-text embeddings from an ensemble of CLIP models, DFN2B-CLIP-Vi T-L-14-s39b and DFN2B-CLIP-Vi T-L-14, for all images, ground-truth captions, and synthetic captions. We train a more diverse family of architectures compared with Mobile CLIP and evaluate their performance on 38 zero-shot classification tasks (Gadre et al., 2023). Particularly, we introduce Mobile CLIP2S3 and Mobile CLIP2-S4 architectures trained on DFNDR-2B as well as variants trained on Data Comp DR-1B which we refer to as Mobile CLIP-S3 and Mobile CLIP-S4. Table 8 shows our results compared with other models with similar latencies. Details of training and hyperparameters are described in Appx. A. We compare Mobile CLIP2 to prior small CLIP architectures Tiny CLIP (Wu et al., 2023) trained on LAION (Schuhmann et al., 2022; 2021) and ACED (Udandarao et al., 2024). We also compare with larger models from Open AI s CLIP (Radford et al., 2021), Data Comp (Gadre et al., 2023), Ve CLIP (Lai et al., 2023), EVA (Sun et al., 2023), DFN (Fang et al., 2024a), Sig LIP (Zhai et al., 2023), and Sig LIP2 (Tschannen et al., 2025). We evaluate all models using Open CLIP (Ilharco et al., 2021) and Data Comp (Gadre et al., 2023). In some cases such as Sig LIP2, we observe positive/negative gaps with reported results in their paper. Mobile CLIP2 achieves state-of-the-art Image Net-1k validation zero-shot accuracies at various latencies. Notably, Mobile CLIP2-S4 matches the zero-shot accuracy of Sig LIP-SO400M/14 on Image Net validation set while being 2 smaller and improves on DFN Vi T-L/14 at 2.5 lower latency. We also improve on Image Net-1k performance of ACED models considering their latencies. As ACED optimized their models for low inference flops, the latency of both ACED-F1 and ACED-F2 are comparable to our Mobile CLIP2-S2 architecture while still have higher latency and more parameters. Sig LIP-B/16 and Sig LIP2-B/16 models are more comparable in size and latency to our new larger architectures. Particularly, Sig LIP2 models have substantially larger text-encoders compared to Sig LIP models. We note that our models pretrained on DFNDR-2B do not always achieve state-of-the-art retrieval performance. We attribute this to the bias of DFNDR-2B dataset towards zero-shot classification tasks and particularly Image Net-1k. We observe that models trained on Data Comp, Web LI, and their derivatives may achieve higher retrieval performance compared to DFN datasets and derivatives while lower on Avg. 38 performance. As such, we also train our new architectures on Data Comp DR-1B referred to as Mobile CLIP-S3 and Mobile CLIP-S4. The combination of these two families of architectures will provide flexibility for broader applications. 4.1 VLM evaluations We report vision-language evaluations using Mobile CLIP2 pretrained models in the LLa VA-1.5 setup (Liu et al., 2024a). We keep the vision backbone frozen for all the runs and use Qwen2-7B instead of Vicuna-7B. All other training details are the same as the original LLa VA-1.5 setup, more details are provided in appendix. We evaluate Vi T-B/16 models pretrained on Data Comp, DFN, Data Comp DR, and DFNDR for 13B seen samples. In Tab. 9 we observe that on average training on DFNDR achieves 3.5% higher accuracy compared with DFN pretrained model, 1.6% better than Data Comp pretrained model, and 0.6% better than Data Comp DR pretrained model. Published in Transactions on Machine Learning Research (08/2025) Table 8: Mobile CLIP2 family of models has the best average performance at various latencies. Retrieval performances are reported @1. Last column shows average performance on 38 datasets as in Open CLIP (Ilharco et al., 2021). Models are grouped by their total latency in increasing order and by performance within each group. Base refers to standard CLIP Transformer-based (Vaswani et al., 2017) text encoder with 12 layers, and Custom stands for customized text encoder used in the respective method. Models with substantially higher latencies and/or larger model sizes are grayed out. Name Dataset Seen Samples Resolution Image Encoder Text Encoder Params (M) (img+txt) Latency (ms) (img+txt) Zero-shot CLS Flickr30k Ret. COCO Ret. Avg. Perf. on 38 IN-val IN-shift T I I T T I I T Tiny CLIP-RN19M LAION-400M 15.2B 224 Res Net-19M Custom 18.6 + 44.8 1.9 + 1.9 56.3 43.6 58.0 75.4 30.9 47.8 48.3 Tiny CLIP-RN30M LAION-400M 15.2B 224 Res Net-30M Custom 29.6 + 54.2 2.6 + 2.6 59.1 45.7 61.5 80.1 33.8 51.6 50.2 Tiny CLIP-40M/32 LAION-400M 15.2B 224 Vi T-40M/32 Custom 39.7 + 44.5 3.0 + 1.9 59.8 46.5 59.1 76.1 33.5 48.7 51.2 Mobile CLIP-S0 Data Comp DR-1B 13B 256 MCi0 MCt 11.4 + 42.4 1.5 + 1.6 67.8 55.1 67.7 85.9 40.4 58.7 58.1 ACED-F0 Data Comp-1B 13B 256 Vi T-S/32 Small 22.7 + 28.8 2.1 + 1.8 68.5 (-) 71.4 87.6 41.2 60.8 (-) Mobile CLIP2-S0 DFNDR-2B 13B 256 MCi0 Base 11.4 + 63.4 1.5 + 3.3 71.5 57.6 69.2 86.6 43.7 62.7 59.7 Open AI-RN50 Open AI-400M 13B 224 Res Net-50 Base 38.3 + 63.4 3.3 + 3.3 59.8 45.1 57.4 80.0 28.5 48.8 48.1 Tiny CLIP-61M/32 LAION-400M 15.2B 224 Vi T-61M/32 Custom 61.4 + 54.0 4.3 + 2.6 62.4 48.7 62.6 78.7 36.5 52.8 53.0 Tiny CLIP-63M/32 LAION-400M YFCC-15M 15.8B 224 Vi T-63M/32 Custom (-) (-) 64.5 (-) 66.0 84.9 38.5 56.9 (-) Mobile CLIP-S1 Data Comp DR-1B 13B 256 MCi1 Base 21.5 + 63.4 2.5 + 3.3 72.6 60.7 71.0 89.2 44.0 62.2 61.3 Open AI-RN101 Open AI-400M 13B 224 Res Net-101 Base 56.3 + 63.4 4.3 + 3.3 62.3 48.5 58.0 79.0 30.7 49.8 50.3 Open AI-B/32 Open AI-400M 13B 224 Vi T-B/32 Base 86.2 + 63.4 5.9 + 3.3 63.3 48.5 58.8 78.9 30.4 50.1 52.5 LAION-B/32 LAION-2B 32B 224 65.7 51.9 66.4 84.4 39.1 56.2 54.8 Data Comp-B/32 Data Comp-1B 13B 224 69.2 55.2 61.1 79.0 37.1 53.5 58.0 Data Comp-B/32-256 Data Comp-1B 34B 256 Vi T-B/32 Base 86.2 + 63.4 6.2 + 3.3 72.8 58.7 64.9 84.8 39.9 57.9 60.9 Sig LIP2-B/32 Web LI-10B 40B 256 Vi T-B/32 Custom 94.6 + 282.3 6.3 + 6.3 73.8 57.8 73.2 88.0 47.9 64.9 61.9 Mobile CLIP-S2 Data Comp DR-1B 13B 256 MCi2 Base 35.7 + 63.4 3.6 + 3.3 74.4 63.1 73.4 90.3 45.4 63.4 63.7 ACED-F1 Data Comp-1B 13B 256 Vi T-B/32 Small 86.2 + 28.8 6.2 + 1.8 74.9 (-) 77.9 90.3 47.3 74.9 (-) ACED-F2 Data Comp-1B 13B 256 Vi T-B/24 Small 86.2 + 28.8 6.5 + 1.8 76.9 (-) 79.5 91.1 49.7 66.9 (-) Mobile CLIP2-S2 DFNDR-2B 13B 256 MCi2 Base 35.7 + 63.4 3.6 + 3.3 77.2 64.7 74.8 90.4 48.8 66.7 64.1 Ve CLIP-B/16 WIT-200M 6.4B 224 Base 86.2 + 63.4 11.5 + 3.3 64.6 (-) 76.3 91.1 48.4 67.2 (-) Open AI-B/16 WIT-400M 13B 224 Base 86.2 + 63.4 11.5 + 3.3 68.3 55.9 67.7 85.9 40.4 58.7 58.1 LAION-B/16 LAION-2B 34B 224 Base 86.2 + 63.4 11.5 + 3.3 70.2 56.6 69.8 86.3 42.3 59.4 58.7 EVA02-B/16 Merged-2B 8B 224 Base 86.2 + 63.4 (-) 74.7 59.6 71.5 86.0 42.2 58.7 58.9 DFN-B/16 DFN-2B 13B 224 Base 86.2 + 63.4 11.5 + 3.3 76.2 62.3 69.1 85.4 43.4 60.4 60.9 Data Comp-B/16 Data Comp-1B 13B 224 Base 86.2 + 63.4 11.5 + 3.3 73.5 60.8 69.8 86.3 42.3 59.4 61.5 Mobile CLIP-B Data Comp DR-1B 13B 224 Base 86.3 + 63.4 10.4 + 3.3 76.8 65.6 77.3 91.4 50.6 68.8 65.2 Mobile CLIP-B (LT) Data Comp DR-1B 39B 224 Base 86.3 + 63.4 10.4 + 3.3 77.2 66.1 76.9 92.3 50.0 68.7 65.8 Mobile CLIP2-B DFNDR-2B 13B 224 Base 86.3 + 63.4 10.4 + 3.3 79.4 66.4 76.5 89.7 49.9 67.5 65.8 Sig LIP-B/16 Web LI 40B 224 Vi T-B/16 Custom 92.9 + 110.3 9.9 + 5.8 76.0 61.0 74.7 89.1 47.8 65.7 62.3 Sig LIP-B/16-256 Web LI 40B 256 Vi T-B/16 Custom 92.9 + 110.3 11.4 + 5.8 76.5 62.0 75.0 90.4 48.4 66.1 62.3 Sig LIP2-B/16 Web LI-10B 40B 224 Vi T-B/16 Custom 92.9 + 282.3 9.9 + 6.3 78.5 63.9 79.3 93.1 53.2 69.4 64.6 Sig LIP2-B/16-256 Web LI-10B 40B 256 Vi T-B/16 Custom 92.9 + 282.3 11.4 + 6.3 79.3 65.3 80.2 93.2 54.1 70.8 64.6 Mobile CLIP-S3 Data Comp DR-1B 13B 256 MCi3 Large 125.1 + 123.6 8.0 + 6.6 78.3 68.2 77.9 93.1 51.3 68.8 66.3 Mobile CLIP2-S3 DFNDR-2B 13B 256 MCi3 Large 125.1 + 123.6 8.0 + 6.6 80.7 68.9 77.3 91.6 50.9 68.4 66.8 Sig LIP-L/16 Web LI 40B 256 Vi T-L/16 Custom 316.0 + 336.2 38.2 + 19.1 80.4 66.6 79.0 91.8 52.3 70.8 65.6 DFN-L/14-quickgelu DFN-2B 13B 224 Vi T-L/14 Large 304.3 + 123.6 57.9 + 6.6 81.4 68.8 78.5 89.0 53.7 66.8 66.9 Mobile CLIP-L/14 Data Comp DR-1B 13B 224 Vi T-L/14 Large 304.3 + 123.6 57.9 + 6.6 79.5 69.9 75.3 91.3 47.6 66.5 66.9 Mobile CLIP2-S4 DFNDR-2B 13B 256 MCi4 Large 321.6 + 123.6 19.6 + 6.6 81.9 70.3 78.0 92.4 51.5 69.3 67.5 Mobile CLIP2-L/14 DFNDR-2B 13B 224 Vi T-L/14 Large 304.3 + 123.6 57.9 + 6.6 81.9 70.2 77.2 92.0 51.6 69.0 67.8 Mobile CLIP-S4 Data Comp DR-1B 13B 256 MCi4 Large 321.6 + 123.6 19.6 + 6.6 79.4 69.7 79.5 94.9 52.1 70.3 68.1 Sig LIP-SO400M/14 Web LI 40B 224 So-400M Custom 427.7 + 449.7 (-) 82.0 69.5 75.2 91.0 51.8 69.7 68.1 Sig LIP2-L/16 Web LI-10B 40B 256 Vi T-L/16 Custom 316.0 + 565.6 38.2 + 19.8 82.3 70.5 81.8 94.6 54.7 72.0 68.3 Sig LIP2-SO400M/14 Web LI-10B 40B 224 So-400M Custom 427.7 + 707.8 (-) 83.2 72.0 82.8 93.9 55.5 71.9 69.1 Table 9: VLM evaluations in LLa VA-1.5 setup. Vi T-B/16 pretrained models reach 3.5% higher accuracy compared with DFN pretrained model, 1.6% better than Data Comp pretrained model, and 0.6% better than Data Comp DR pretrained model. Dataset GQA SQA Text VQA POPE MMMU MMB Viz Wiz VQAv2 Avg. Data Comp-1B 59.6 71.5 50.5 81.8 42.6 59.1 51.8 70.7 61.0 DFN-2B 56.9 71.3 46.0 81.4 41.9 52.2 56.1 66.9 59.1 Data Comp DR-1B 60.3 73.1 50.4 81.7 43.6 60.2 54.9 72.1 62.0 DFNDR-2B 60.4 72.9 49.9 83.3 45.2 61.9 54.5 72.4 62.6 4.2 Dense Prediction tasks We evaluate the quality of the visual representations learned by finetuning the image encoder on dense prediction tasks like object detection, semantic segmentation and depth estimation. In Table 10, we report performance of Vi T-B/16 model with Mask RCNN He et al. (2017) head for instance segmentation on MS-COCO Chen et al. (2015) dataset. All models were trained using MMDetection library Chen et al. (2019) using 1 schedule with single scale testing as described in Wei et al. (2023). We follow finetuning setup described in Wei et al. (2023), more details in appendix. In Table Table 11, we report performance of Vi T-B/16 model with Uper Net Xiao et al. (2018) head, trained using the same setup described in Liu et al. (2024c) on ADE20k (Zhou et al., 2017) dataset. In Table 12, we report Root Mean Square Error (RMSE) on NYUv2 dataset Nathan Silberman & Fergus (2012). We use the same settings as described in Vasu et al. (2024b), more details are provided in appendix. Published in Transactions on Machine Learning Research (08/2025) Table 12: Results on NYUv2 for depth estimation following the same settings as Wei et al. (2023). All results are for Vi T-B/16 models. Method Dataset RMSE( ) Cat LIP Mehta et al. (2024) Data Comp 0.394 MAE He et al. (2022) IN-1K 0.383 MAWS Singh et al. (2023) IG-3B 0.371 FD-CLIP Wei et al. (2023) Open AI-WIT + IN-1K 0.352 MAE Singh et al. (2023) IG-3B 0.348 CLIP Radford et al. (2021) Open AI-WIT 0.416 Mobile CLIP2 DFNDR-2B 0.356 Table 13: Comparison pretraining methods for semantic segmentation on ADE-20k. For reference, we have included recent state-of-the-art semantic segmentation models (in gray). Encoder Decoder Pre-Training Resolution # Params(M) m Io U Intern Image-B Wang et al. (2023) Uper Net Xiao et al. (2018) Sup. IN-1K 512 512 128.0 50.8 Vi T-Adapter-B Chen et al. (2023) Semantic FPN Kirillov et al. (2019) Sup. IN-22K 512 512 104.6 50.7 Vi T-Adapter-B Chen et al. (2023) Uper Net Xiao et al. (2018) Sup. IN-22K 512 512 133.9 51.9 Swin-L Liu et al. (2021) Uper Net Xiao et al. (2018) Sup. IN-22K 640 640 234.1 52.1 MCi0 Semantic FPN Kirillov et al. (2019) Sup. IN-1K 512 512 14.5 44.8 MCi2 Semantic FPN Kirillov et al. (2019) Sup. IN-1K 512 512 38.5 48.9 MCi0 Semantic FPN Kirillov et al. (2019) Mobile CLIP2 512 512 14.5 47.0 (+2.2) MCi2 Semantic FPN Kirillov et al. (2019) Mobile CLIP2 512 512 38.5 51.6 (+2.7) Additionally, we assess the performance of smaller Mobile CLIP2 variants on dense prediction tasks. Popular pretraining methods like MAE (He et al., 2022), are not directly applicable to hierarchical convolutional and hybrid architectures such as our MCi models, hence we compare Mobile CLIP2 pretraining with supervised pretraining for the same architectures. In Tabs. 13 and 14, we see that Mobile CLIP2 pretraining is significantly better than supervised pretraining and can serve as a good pretraining choice for hierarchical architectures. Table 10: Object detection and instance segmentation results on MS-COCO with Mask-RCNN head trained for 1 schedule. All models are Vi T-B/16. Method Dataset m APbox m APmask Cat LIP Mehta et al. (2024) Data Comp 45.7 40.6 MAE He et al. (2022) IN-1K 46.5 40.9 MAE Singh et al. (2023) IG-3B 46.4 42.1 MAWS Singh et al. (2023) IG-3B 48.0 43.4 FD-CLIP Wei et al. (2023) Open AI-WIT + IN-1K 48.2 42.5 CLIP Radford et al. (2021) Open AI-WIT 45.0 39.8 Mobile CLIP2 DFNDR-2B 47.0 41.8 Table 11: Semantic segmentation results on ADE20k using Uper Net decoder. All models are Vi T-B/16. Method Dataset m Io U m Acc MAE He et al. (2022) IN-1K 48.1 58.9 d BOT Liu et al. (2024c) IN-1K 49.5 60.7 MAWS Singh et al. (2023) IG-3B 50.4 61.5 Cat LIP Mehta et al. (2024) Data Comp 50.6 61.8 FD-CLIP Wei et al. (2023) Open AI-WIT + IN-1K 51.7 - CLIP Radford et al. (2021) Open AI-WIT 49.5 - Mobile CLIP2 DFNDR-2B 52.8 64.0 5 Related Work Improving the training of multi-modal models focus on three aspects: data, objective function and architecture. Our Mobile CLIP2 builds on Mobile CLIP and provides improvements in all three aspects. Data approaches either filter a dataset or augment it with additional information. Basic filtering methods begin by selecting or crawling a large dataset of candidate image-text pairs and filter using ad-hoc rules based on the URLs or statics of the images and captions (Radford et al., 2021; Schuhmann et al., 2021; 2022; Xu et al., 2024). More advanced filtering methods involve filtering models trained on high-quality data utilized to remove low-quality image-text pairs. These methods may utilize a pretrained CLIP model (Gadre et al., 2023) or more specialized filtering models (Fang et al., 2024a). The challenge with data methods is that the biases introduced by ad-hoc rules or pretrained models. For example, most publicly available datasets such as Data Comp are filtered for English-only data which limits the capabilities of models on non-English tasks (Carlsson et al., 2022; Nguyen et al., 2024; Pouget et al., 2024). Alternatively, pretrained models may be used for active data selection based on the sample difficulty (Evans et al., 2024a;b). It has also been observed that repeating high-quality data achieves higher utilization (Goyal et al., 2024). Published in Transactions on Machine Learning Research (08/2025) Table 14: Comparison pretraining methods for object detection task on MS-COCO using Mask RCNN He et al. (2017) detection head. All models are trained for 1 schedule. For reference we have included recent state-of-the-art object detection models (in gray). Model Pre-Training # Params(M) m APbox m APmask Vi T-Adapter-B Chen et al. (2023) Sup. IN-1K 284 47.0 41.8 Intern Image-B Wang et al. (2023) Sup. IN-1K 115 48.8 44.0 Vi T-Adapter-L Chen et al. (2023) Sup. IN-22K 347.9 48.7 43.3 MCi0 Sup. IN-1K 31.0 41.8 38.0 MCi2 Sup. IN-1K 55.0 46.6 41.7 MCi0 Mobile CLIP2 31.0 44.4 (+2.6) 39.6 (+1.6) MCi2 Mobile CLIP2 55.0 49.1 (+2.5) 43.2 (+1.5) More broadly, the output of pretrained models can be stored as part of a new augmented dataset. For example, various works utilize image-captioning models to generate synthetic captions for images in a dataset (Yang et al., 2023a; Nguyen et al., 2023; Lai et al., 2023; Liu et al., 2024d; Li et al., 2024). Large language models can also be used to rewrite ground-truth captions (Fan et al., 2023) as well as together with text-to-image models to generate fully synthetic datasets (Hammoud et al., 2024). Mobile CLIP introduced the multi-modal dataset reinforcement where they utilized an image-caption model to generate synthetic captions as well as an ensemble of large CLIP models to store CLIP embeddings for multiple image augmentations and synthetic captions and store them efficiently (Vasu et al., 2024c). We follow a similar approach while improving both the caption generator and CLIP embedding generators through better DFN models (Fang et al., 2024a). Another approach is to improve the objective function of multi-modal training. The original CLIP paper utilized a contrastive loss that encourages the representations of images and texts paired in the dataset to be kept close to each other while staying farther away from other images and texts in a mini-batch (Radford et al., 2021). Sig LIP introduced a variant based on Sigmoid instead of Softmax that achieves higher training efficiency at larger batch sizes (Zhai et al., 2023; Tschannen et al., 2025). Other methods utilize objectives based on image masking (Yang et al., 2023b; Fang et al., 2023; Sun et al., 2023; Li et al., 2023b) and unimodal self-supervision (Mu et al., 2022; Li et al., 2021) as well as multi-resolution training (Li et al., 2023a) for cost-effective training. Multi-modal distillation achieves more significant improvements, particularly for smaller architecture variants (Wang et al., 2022b; Kuang et al., 2023; Wang et al., 2022a; Wu et al., 2023). Notably, Mobile CLIP (Vasu et al., 2024c) achieved high training efficiency by utilizing an offline knowledge distillation method (Shen & Xing, 2022; Yun et al., 2021; Faghri et al., 2023). We utilize a similar objective function as Mobile CLIP that includes embedding distillation on image-text pairs and synthetic captions. Lastly, architectural improvements seek improved inference efficiency and higher performance given a parameter, flops, or latency budget. CLIP architectures are often borrowed from uni-modal image and text models. Particularly, the original CLIP and various followup works utilized standard Vi T architectures together with a modified BERT text encoder (Dosovitskiy et al., 2020; Devlin et al., 2019; Radford et al., 2021). Efficient architectures for CLIP include Tiny CLIP that prunes Vi T (Wu et al., 2023), Cao et al. (2023) that reduce tokens, and Evans et al. (2024b) that reduce the parameters for lower flops. Mobile CLIP introduced efficient architectures specifically design for CLIP where they introduced a low latency convolution-transformer hybrid architectures for both their image and text encoders. We further improve on their architectures by introducing two new variants that fill the large latency gap between common B and L architectures. 6 Conclusion We introduce Mobile CLIP2, a new family of low latency image-text models, achieving state-of-the-art Image Net-1k zero-shot validation accuracy. Our methodology improves multi-modal reinforced training by utilizing stronger CLIP teachers as well as our newly trained image-captioning models. We particularly perform a comprehensive study of tuning and ensembling CLIP teachers as well as training and fine-tuning efficient image-captioning models. Notably, Mobile CLIP2-S4 matches the zero-shot accuracy of Sig LIPSO400M/14 on Image Net-1k while being 2 smaller and improves on DFN Vi T-L/14 at 2.5 lower latency. We release our model checkpoints and data generation code that facilitates dataset generation at scale. Published in Transactions on Machine Learning Research (08/2025) Broader Impact Statement Our work introduces a family of foundation models particularly optimized for deployment on mobile and edge devices. As such, it facilitates broader use of foundation models and development of applications for wider user bases. Mobile CLIP2 may be used for various applications such as image classification where its output is impacted by the existing biases of the training datasets and teacher models. Acknowledgments We would like to thank Albin Madappally Jose, Barry Theobald, Chen Huang, Rick Chang, and Apple Machine Learning Research team for their help and discussions throughout this project. Lucas Beyer, Xiaohua Zhai, Amélie Royer, Larisa Markeeva, Rohan Anil, and Alexander Kolesnikov. Knowledge distillation: A good teacher is patient and consistent. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 10925 10934, 2022. Rishi Bommasani, Drew A Hudson, Ehsan Adeli, Russ Altman, Simran Arora, Sydney von Arx, Michael S Bernstein, Jeannette Bohg, Antoine Bosselut, Emma Brunskill, et al. On the opportunities and risks of foundation models. ar Xiv preprint ar Xiv:2108.07258, 2021. Qingqing Cao, Bhargavi Paranjape, and Hannaneh Hajishirzi. Pu Mer: Pruning and merging tokens for efficient vision language models. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2023. Fredrik Carlsson, Philipp Eisen, Faton Rekathati, and Magnus Sahlgren. Cross-lingual and multilingual clip. In Proceedings of the thirteenth language resources and evaluation conference, pp. 6848 6854, 2022. Kai Chen, Jiaqi Wang, Jiangmiao Pang, Yuhang Cao, Yu Xiong, Xiaoxiao Li, Shuyang Sun, Wansen Feng, Ziwei Liu, Jiarui Xu, Zheng Zhang, Dazhi Cheng, Chenchen Zhu, Tianheng Cheng, Qijie Zhao, Buyu Li, Xin Lu, Rui Zhu, Yue Wu, Jifeng Dai, Jingdong Wang, Jianping Shi, Wanli Ouyang, Chen Change Loy, and Dahua Lin. MMDetection: Open mmlab detection toolbox and benchmark. ar Xiv preprint ar Xiv:1906.07155, 2019. Xinlei Chen, Hao Fang, Tsung-Yi Lin, Ramakrishna Vedantam, Saurabh Gupta, Piotr Dollár, and C Lawrence Zitnick. Microsoft coco captions: Data collection and evaluation server. ar Xiv preprint ar Xiv:1504.00325, 2015. Zhe Chen, Yuchen Duan, Wenhai Wang, Junjun He, Tong Lu, Jifeng Dai, and Yu Qiao. Vision transformer adapter for dense predictions. In ICLR, 2023. MMSegmentation Contributors. MMSegmentation: Openmmlab semantic segmentation toolbox and benchmark. https://github.com/open-mmlab/mmsegmentation, 2020. Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: human language technologies, volume 1 (long and short papers), pp. 4171 4186, 2019. Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. ar Xiv preprint ar Xiv:2010.11929, 2020. Talfan Evans, Nikhil Parthasarathy, Hamza Merzic, and Olivier J. Hénaff. Data curation via joint example selection further accelerates multimodal learning. In Amir Globersons, Lester Mackey, Danielle Belgrave, Angela Fan, Ulrich Paquet, Jakub M. Tomczak, and Cheng Zhang (eds.), Advances in Neural Information Processing Systems 38: Annual Conference on Neural Information Processing Systems 2024, Neur IPS 2024, Vancouver, Published in Transactions on Machine Learning Research (08/2025) BC, Canada, December 10 - 15, 2024, 2024a. URL http://papers.nips.cc/paper_files/paper/2024/ hash/ff8d608f6dcebec401df78ca76617e95-Abstract-Datasets_and_Benchmarks_Track.html. Talfan Evans, Shreya Pathak, Hamza Merzic, Jonathan Schwarz, Ryutaro Tanno, and Olivier J Henaff. Bad students make great teachers: Active learning accelerates large-scale visual understanding. In European Conference on Computer Vision, pp. 264 280. Springer, 2024b. Fartash Faghri, David J Fleet, Jamie Ryan Kiros, and Sanja Fidler. Vse++: Improving visual-semantic embeddings with hard negatives. 2018. URL https://github.com/fartashf/vsepp. Fartash Faghri, Hadi Pouransari, Sachin Mehta, Mehrdad Farajtabar, Ali Farhadi, Mohammad Rastegari, and Oncel Tuzel. Reinforce data, multiply impact: Improved model accuracy and robustness with dataset reinforcement. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), October 2023. Lijie Fan, Dilip Krishnan, Phillip Isola, Dina Katabi, and Yonglong Tian. Improving clip training with language rewrites. Advances in Neural Information Processing Systems, 36:35544 35575, 2023. Alex Fang, Albin Madappally Jose, Amit Jain, Ludwig Schmidt, Alexander T. Toshev, and Vaishaal Shankar. Data filtering networks. 2024a. URL https://openreview.net/forum?id=KAk6ng Z09F. Yuxin Fang, Wen Wang, Binhui Xie, Quan Sun, Ledell Wu, Xinggang Wang, Tiejun Huang, Xinlong Wang, and Yue Cao. Eva: Exploring the limits of masked visual representation learning at scale. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 19358 19369, 2023. Yuxin Fang, Quan Sun, Xinggang Wang, Tiejun Huang, Xinlong Wang, and Yue Cao. Eva-02: A visual representation for neon genesis. Image and Vision Computing, pp. 105171, 2024b. Andrea Frome, Greg S Corrado, Jon Shlens, Samy Bengio, Jeff Dean, Tomas Mikolov, et al. Devise: A deep visual-semantic embedding model. In Neur IPS, pp. 2121 2129, 2013. Samir Yitzhak Gadre, Gabriel Ilharco, Alex Fang, Jonathan Hayase, Georgios Smyrnis, Thao Nguyen, Ryan Marten, Mitchell Wortsman, Dhruba Ghosh, Jieyu Zhang, et al. Datacomp: In search of the next generation of multimodal datasets. ar Xiv preprint ar Xiv:2304.14108, 2023. Sachin Goyal, Pratyush Maini, Zachary C Lipton, Aditi Raghunathan, and J Zico Kolter. Scaling laws for data filtering data curation cannot be compute agnostic. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 22702 22711, 2024. Hasan Abed Al Kader Hammoud, Hani Itani, Fabio Pizzati, Philip Torr, Adel Bibi, and Bernard Ghanem. Synthclip: Are we ready for a fully synthetic clip training? ar Xiv preprint ar Xiv:2402.01832, 2024. Kaiming He, Georgia Gkioxari, Piotr Dollár, and Ross Girshick. Mask r-cnn. In ICCV, 2017. Kaiming He, Xinlei Chen, Saining Xie, Yanghao Li, Piotr Dollár, and Ross Girshick. Masked autoencoders are scalable vision learners. In CVPR, 2022. Geoffrey Hinton, Oriol Vinyals, and Jeff Dean. Distilling the knowledge in a neural network. ar Xiv preprint ar Xiv:1503.02531, 2015. Yu-Guan Hsieh, Cheng-Yu Hsieh, Shih-Ying Yeh, Louis Béthune, Hadi Pour Ansari, Pavan Kumar Anasosalu Vasu, Chun-Liang Li, Ranjay Krishna, Oncel Tuzel, and Marco Cuturi. Graph-based captioning: Enhancing visual descriptions by interconnecting region captions. ar Xiv preprint ar Xiv:2407.06723, 2024. Gabriel Ilharco, Mitchell Wortsman, Ross Wightman, Cade Gordon, Nicholas Carlini, Rohan Taori, Achal Dave, Vaishaal Shankar, Hongseok Namkoong, John Miller, Hannaneh Hajishirzi, Ali Farhadi, and Ludwig Schmidt. Openclip, July 2021. URL https://doi.org/10.5281/zenodo.5143773. If you use this software, please cite it as below. Published in Transactions on Machine Learning Research (08/2025) Andrej Karpathy and Li Fei-Fei. Deep visual-semantic alignments for generating image descriptions. In CVPR, pp. 3128 3137, 2015. Alexander Kirillov, Ross Girshick, Kaiming He, and Piotr Dollár. Panoptic feature pyramid networks. In CVPR, 2019. Ryan Kiros, Ruslan Salakhutdinov, and Richard S Zemel. Unifying visual-semantic embeddings with multimodal neural language models. 2014. Huafeng Kuang, Jie Wu, Xiawu Zheng, Ming Li, Xuefeng Xiao, Rui Wang, Min Zheng, and Rongrong Ji. Dlip: Distilling language-image pre-training. ar Xiv preprint ar Xiv:2308.12956, 2023. Zhengfeng Lai, Haotian Zhang, Wentao Wu, Haoping Bai, Aleksei Timofeev, Xianzhi Du, Zhe Gan, Jiulong Shan, Chen-Nee Chuah, Yinfei Yang, et al. From scarcity to efficiency: Improving clip training via visual-enriched captions. ar Xiv preprint ar Xiv:2310.07699, 2023. Xianhang Li, Zeyu Wang, and Cihang Xie. An inverse scaling law for clip training. ar Xiv preprint ar Xiv:2305.07017, 2023a. Xianhang Li, Haoqin Tu, Mude Hui, Zeyu Wang, Bingchen Zhao, Junfei Xiao, Sucheng Ren, Jieru Mei, Qing Liu, Huangjie Zheng, et al. What if we recaption billions of web images with llama-3? ar Xiv preprint ar Xiv:2406.08478, 2024. Yangguang Li, Feng Liang, Lichen Zhao, Yufeng Cui, Wanli Ouyang, Jing Shao, Fengwei Yu, and Junjie Yan. Supervision exists everywhere: A data efficient contrastive language-image pre-training paradigm. ar Xiv preprint ar Xiv:2110.05208, 2021. Yanghao Li, Haoqi Fan, Ronghang Hu, Christoph Feichtenhofer, and Kaiming He. Scaling language-image pre-training via masking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 23390 23400, 2023b. Haotian Liu, Chunyuan Li, Yuheng Li, and Yong Jae Lee. Improved baselines with visual instruction tuning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 26296 26306, 2024a. Haotian Liu, Chunyuan Li, Qingyang Wu, and Yong Jae Lee. Visual instruction tuning. Advances in neural information processing systems, 36, 2024b. Xingbin Liu, Jinghao Zhou, Tao Kong, Xianming Lin, and Rongrong Ji. Exploring target representations for masked autoencoders. In ICLR, 2024c. Yanqing Liu, Xianhang Li, Zeyu Wang, Bingchen Zhao, and Cihang Xie. Clips: An enhanced clip framework for learning with synthetic captions. ar Xiv preprint ar Xiv:2411.16828, 2024d. Ze Liu, Yutong Lin, Yue Cao, Han Hu, Yixuan Wei, Zheng Zhang, Stephen Lin, and Baining Guo. Swin transformer: Hierarchical vision transformer using shifted windows. ar Xiv preprint ar Xiv:2103.14030, 2021. Sachin Mehta, Maxwell Horton, Fartash Faghri, Mohammad Hossein Sekhavat, Mahyar Najibi, Mehrdad Farajtabar, Oncel Tuzel, and Mohammad Rastegari. Catlip: Clip-level visual recognition accuracy with 2.7x faster pre-training on web-scale image-text data, 2024. Norman Mu, Alexander Kirillov, David Wagner, and Saining Xie. Slip: Self-supervision meets language-image pre-training. In European Conference on Computer Vision, pp. 529 544. Springer, 2022. Ivona Najdenkoska, Mohammad Mahdi Derakhshani, Yuki M Asano, Nanne van Noord, Marcel Worring, and Cees GM Snoek. Tulip: Token-length upgraded clip. ar Xiv preprint ar Xiv:2410.10034, 2024. Pushmeet Kohli Nathan Silberman, Derek Hoiem and Rob Fergus. Indoor segmentation and support inference from rgbd images. In ECCV, 2012. Published in Transactions on Machine Learning Research (08/2025) Thao Nguyen, Samir Yitzhak Gadre, Gabriel Ilharco, Sewoong Oh, and Ludwig Schmidt. Improving multimodal datasets with image captioning. ar Xiv preprint ar Xiv:2307.10350, 2023. Thao Nguyen, Matthew Wallingford, Sebastin Santy, Wei-Chiu Ma, Sewoong Oh, Ludwig Schmidt, Pang Wei W Koh, and Ranjay Krishna. Multilingual diversity improves vision-language representations. Advances in Neural Information Processing Systems, 37:91430 91459, 2024. Yasumasa Onoe, Sunayana Rane, Zachary Berger, Yonatan Bitton, Jaemin Cho, Roopal Garg, Alexander Ku, Zarana Parekh, Jordi Pont-Tuset, Garrett Tanzer, et al. Docci: Descriptions of connected and contrasting images. In European Conference on Computer Vision, pp. 291 309. Springer, 2025. Angéline Pouget, Lucas Beyer, Emanuele Bugliarello, Xiao Wang, Andreas Steiner, Xiaohua Zhai, and Ibrahim M Alabdulmohsin. No filter: Cultural and socioeconomic diversity in contrastive vision-language models. Advances in Neural Information Processing Systems, 37:106474 106496, 2024. Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language supervision. In International conference on machine learning, pp. 8748 8763. PMLR, 2021. Christoph Schuhmann, Richard Vencu, Romain Beaumont, Robert Kaczmarczyk, Clayton Mullis, Aarush Katta, Theo Coombes, Jenia Jitsev, and Aran Komatsuzaki. Laion-400m: Open dataset of clip-filtered 400 million image-text pairs. ar Xiv preprint ar Xiv:2111.02114, 2021. Christoph Schuhmann, Romain Beaumont, Richard Vencu, Cade Gordon, Ross Wightman, Mehdi Cherti, Theo Coombes, Aarush Katta, Clayton Mullis, Mitchell Wortsman, et al. Laion-5b: An open large-scale dataset for training next generation image-text models. Advances in Neural Information Processing Systems, 35:25278 25294, 2022. Zhiqiang Shen and Eric Xing. A fast knowledge distillation framework for visual recognition. In European Conference on Computer Vision, pp. 673 690. Springer, 2022. Mannat Singh, Quentin Duval, Kalyan Vasudev Alwala, Haoqi Fan, Vaibhav Aggarwal, Aaron Adcock, Armand Joulin, Piotr Dollár, Christoph Feichtenhofer, Ross Girshick, Rohit Girdhar, and Ishan Misra. The effectiveness of mae pre-pretraining for billion-scale pretraining. In ICCV, 2023. Richard Socher, Andrej Karpathy, Quoc V Le, Christopher D Manning, and Andrew Y Ng. Grounded compositional semantics for finding and describing images with sentences. 2:207 218, 2014. Quan Sun, Yuxin Fang, Ledell Wu, Xinlong Wang, and Yue Cao. Eva-clip: Improved training techniques for clip at scale. ar Xiv preprint ar Xiv:2303.15389, 2023. Michael Tschannen, Alexey Gritsenko, Xiao Wang, Muhammad Ferjad Naeem, Ibrahim Alabdulmohsin, Nikhil Parthasarathy, Talfan Evans, Lucas Beyer, Ye Xia, Basil Mustafa, et al. Siglip 2: Multilingual vision-language encoders with improved semantic understanding, localization, and dense features. ar Xiv preprint ar Xiv:2502.14786, 2025. Vishaal Udandarao, Nikhil Parthasarathy, Muhammad Ferjad Naeem, Talfan Evans, Samuel Albanie, Federico Tombari, Yongqin Xian, Alessio Tonioni, and Olivier J Hénaff. Active data curation effectively distills large-scale multimodal models. ar Xiv preprint ar Xiv:2411.18674, 2024. Jack Urbanek, Florian Bordes, Pietro Astolfi, Mary Williamson, Vasu Sharma, and Adriana Romero-Soriano. A picture is worth more than 77 text tokens: Evaluating clip-style models on dense captions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 26700 26709, 2024. Pavan Kumar Anasosalu Vasu, James Gabriel, Jeff Zhu, Oncel Tuzel, and Anurag Ranjan. Mobileone: An improved one millisecond mobile backbone. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 7907 7917, 2023a. Published in Transactions on Machine Learning Research (08/2025) Pavan Kumar Anasosalu Vasu, James Gabriel, Jeff Zhu, Oncel Tuzel, and Anurag Ranjan. Fastvit: A fast hybrid vision transformer using structural reparameterization. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), October 2023b. Pavan Kumar Anasosalu Vasu, Fartash Faghri, Chun-Liang Li, Cem Koc, Nate True, Albert Antony, Gokul Santhanam, James Gabriel, Peter Grasch, Oncel Tuzel, et al. Fastvlm: Efficient vision encoding for vision language models. ar Xiv preprint ar Xiv:2412.13303, 2024a. Pavan Kumar Anasosalu Vasu, Hadi Pouransari, Fartash Faghri, and Oncel Tuzel. Clip with quality captions: A strong pretraining for vision tasks. ar Xiv preprint ar Xiv:2405.08911, 2024b. Pavan Kumar Anasosalu Vasu, Hadi Pouransari, Fartash Faghri, Raviteja Vemulapalli, and Oncel Tuzel. Mobileclip: Fast image-text models through multi-modal reinforced training. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 15963 15974, 2024c. Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. Advances in neural information processing systems, 30, 2017. Peng Wang, Shuai Bai, Sinan Tan, Shijie Wang, Zhihao Fan, Jinze Bai, Keqin Chen, Xuejing Liu, Jialin Wang, Wenbin Ge, Yang Fan, Kai Dang, Mengfei Du, Xuancheng Ren, Rui Men, Dayiheng Liu, Chang Zhou, Jingren Zhou, and Junyang Lin. Qwen2-vl: Enhancing vision-language model s perception of the world at any resolution. ar Xiv preprint ar Xiv:2409.12191, 2024. Wenhai Wang, Jifeng Dai, Zhe Chen, Zhenhang Huang, Zhiqi Li, Xizhou Zhu, Xiaowei Hu, Tong Lu, Lewei Lu, Hongsheng Li, et al. Internimage: Exploring large-scale vision foundation models with deformable convolutions. In CVPR, 2023. Zhecan Wang, Noel Codella, Yen-Chun Chen, Luowei Zhou, Xiyang Dai, Bin Xiao, Jianwei Yang, Haoxuan You, Kai-Wei Chang, Shih-fu Chang, et al. Multimodal adaptive distillation for leveraging unimodal encoders for vision-language tasks. ar Xiv preprint ar Xiv:2204.10496, 2022a. Zhecan Wang, Noel Codella, Yen-Chun Chen, Luowei Zhou, Jianwei Yang, Xiyang Dai, Bin Xiao, Haoxuan You, Shih-Fu Chang, and Lu Yuan. Clip-td: Clip targeted distillation for vision-language tasks. ar Xiv preprint ar Xiv:2201.05729, 2022b. Yixuan Wei, Han Hu, Zhenda Xie, Ze Liu, Zheng Zhang, Yue Cao, Jianmin Bao, Dong Chen, and Baining Guo. Improving clip fine-tuning performance. In ICCV, 2023. Mitchell Wortsman, Gabriel Ilharco, Jong Wook Kim, Mike Li, Simon Kornblith, Rebecca Roelofs, Raphael Gontijo Lopes, Hannaneh Hajishirzi, Ali Farhadi, Hongseok Namkoong, et al. Robust finetuning of zero-shot models. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 7959 7971, 2022. Kan Wu, Houwen Peng, Zhenghong Zhou, Bin Xiao, Mengchen Liu, Lu Yuan, Hong Xuan, Michael Valenzuela, Xi Stephen Chen, Xinggang Wang, et al. Tinyclip: Clip distillation via affinity mimicking and weight inheritance. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 21970 21980, 2023. Tete Xiao, Yingcheng Liu, Bolei Zhou, Yuning Jiang, and Jian Sun. Unified perceptual parsing for scene understanding. In ECCV, 2018. Hu Xu, Saining Xie, Xiaoqing Ellen Tan, Po-Yao Huang, Russell Howes, Vasu Sharma, Shang-Wen Li, Gargi Ghosh, Luke Zettlemoyer, and Christoph Feichtenhofer. Demystifying CLIP data. 2024. URL https://openreview.net/forum?id=5BCFlnf E1g. Kaicheng Yang, Jiankang Deng, Xiang An, Jiawei Li, Ziyong Feng, Jia Guo, Jing Yang, and Tongliang Liu. Alip: Adaptive language-image pre-training with synthetic caption. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 2922 2931, 2023a. Published in Transactions on Machine Learning Research (08/2025) Shusheng Yang, Yixiao Ge, Kun Yi, Dian Li, Ying Shan, Xiaohu Qie, and Xinggang Wang. Rils: Masked visual reconstruction in language semantic space. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 23304 23314, 2023b. Jiahui Yu, Zirui Wang, Vijay Vasudevan, Legg Yeung, Mojtaba Seyedhosseini, and Yonghui Wu. Coca: Contrastive captioners are image-text foundation models. Trans. Mach. Learn. Res., 2022, 2022. URL https://openreview.net/forum?id=Ee277P3AYC. Sangdoo Yun, Seong Joon Oh, Byeongho Heo, Dongyoon Han, Junsuk Choe, and Sanghyuk Chun. Re-labeling imagenet: from single to multi-labels, from global to localized labels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 2340 2350, 2021. Xiaohua Zhai, Basil Mustafa, Alexander Kolesnikov, and Lucas Beyer. Sigmoid loss for language image pre-training. ar Xiv preprint ar Xiv:2303.15343, 2023. Beichen Zhang, Pan Zhang, Xiaoyi Dong, Yuhang Zang, and Jiaqi Wang. Long-clip: Unlocking the long-text capability of clip. In European Conference on Computer Vision, pp. 310 325. Springer, 2024. Kecheng Zheng, Yifei Zhang, Wei Wu, Fan Lu, Shuailei Ma, Xin Jin, Wei Chen, and Yujun Shen. Dreamlip: Language-image pre-training with long captions. In European Conference on Computer Vision, pp. 73 90. Springer, 2024. Bolei Zhou, Hang Zhao, Xavier Puig, Sanja Fidler, Adela Barriuso, and Antonio Torralba. Scene parsing through ade20k dataset. In CVPR, 2017. Published in Transactions on Machine Learning Research (08/2025) A Experimental Setup Table 15 provides a summary of datasets used in our ablations and experiments. Table 15: Summary of pretraining datasets. Dataset Num. Samples CLIP Teachers Syn. Captioner Num. Image Augs. Num. Syn. Caps. BFloat16 Size (TBs) Data Comp-1B12M 12.8M 0.9 DFN-2B12M 12.8M 0.2 DFN-5B12M 12.8M 1.5 Data Comp DR-12M 12.8M Open AI + Data Comp-XL LAION-2B MSCOCO-123k 30 5 2.1 DFNDR-2B12M 12.8M DFN-2B + DFN-2B-s39B DFN-2B MSCOCO-38k 30 5 1.3 DFNDR-5B12M 12.8M DFN-2B + DFN-2B-s39B DFN-2B MSCOCO-38k 30 5 2.6 Data Comp-1B 1.3B 90 DFN-2B 1.9B 65 Data Comp DR-1B 1.3B Open AI + Data Comp-XL LAION-2B MSCOCO-123k 10 5 140 DFNDR-2B 1.9B DFN-2B + DFN-2B-s39B DFN-2B MSCOCO-38k 2 5 162 Table 16 summarizes the hyperparameters we used to train Mobile CLIP2. For training on 13B seen samples, we use either a setup with 32x8x A100-40GB GPUs or a setup with 16x8x H100-80GB GPUs. For our ablations we train for 30k seen samples using 4x8x H100-80GB GPUs and global batch size 8192. Table 16: Training hyperparameters for our CLIP experiments on DFNDR-2B. Hyperparameter S0 S2 B S3 S4 Input resolution 2562 2562 2242 2562 2562 Context length 77 Data augmentation Rand Augment Random resize crop scale [0.08, 1.0] Random resized crop ratio [0.75, 1.33] Range Augment target value (40, 20) Train iterations 200k Warmup iterations 10k 10k 2k 2k 2k Global batch size 65536 65536 65536 114688 114688 Optimizer Adam W Adam W beta1 0.9 Adam W beta2 0.95 Max learning rate 1e-3 Min learning rate 1e-6 1e-6 1e-6 0 0 LR. decay schedule cosine Weight decay rate 0.2 Gradient clipping 1.0 Mixed precision BFloat16 EMA decay rate 0.9995 No EMA No EMA No EMA No EMA CLIP loss weight 0.0 0.0 0.0 0.0 0.0 KD loss weight 1.0 1.0 1.0 1.0 1.0 GT caption weight 1.0 Synth. caption weight 1.0 Synth. teacher Co Ca-Vi T-L/14 - DFN-2B MSCOCO-38k Teacher 1 DFN2B-CLIP-Vi T-L-14-s39b Teacher 2 DFN2B-CLIP-Vi T-L-14 Teacher 1 logit scale 70.0 Teacher 2 logit scale 60.0 Teacher resolution 224 224 A.1 Training details for Co Ca caption generators We use Open CLIP to train Co Ca-Vi T-L/14 architecture (coca_Vi T-L-14). We pretrain models on DFN-2B and fine-tune on various datasets. Table 17 summarizes the hyperparameters for our Co Ca pretraining and fine-tuning. Published in Transactions on Machine Learning Research (08/2025) Table 17: Training hyperparameters for our Co Ca models trained on DFN-2B. Hyperparameter DFN-2B Pretrain Fine-tune Input resolution 2242 2242 Context length 77 77, 255 Seen samples 12.8B 12M Train iterations 195k 3k, 6k Early stop iterations 143k N/A Warmup iterations 10k 1k Global batch size 65536 4092, 2048 Optimizer Adam W Adam W beta1 0.9 Adam W beta2 0.95 Max learning rate 1e-3 1e-5 Min learning rate 0 LR. decay schedule cosine Weight decay rate 0.2 0.1 Gradient clipping 1.0 Mixed precision amp Co Ca caption loss weight 2.0 Co Ca contrastive loss weight 1.0 GPU Setup 32x8x A100-40GBs 1x8x H100-80GBs A.2 Training details for VLM To assess the quality of the vision encoders, we adopt the LLa VA-1.5 (Liu et al., 2024a) training framework. This framework consists of two stages: (1) projector training, and (2) joint fine-tuning of the projector and the language model on 665K instruction-tuning samples. The hyperparameters used in our experiments are summarized in Table 18. For the language model, we use Qwen2-7B-Instruct Wang et al. (2024) as opposed to Vicuna-7B. In both the stages the vision encoder remains frozen. Stage-1 Stage-2 Data LLa VA-1.5 558K LLa VA-1.5 665k Learning Rate 1e-3 2e-5 Global Batch Size 256 128 Epochs 1 1 LR. schedule cosine decay cosine decay LR. warmup ratio 0.03 0.03 Optimizer Adam W Adam W Trainable Projector Projector + modules Language Model Table 18: LLa VA-1.5 training setup used in ablations for Table 9. A.3 Training details for dense prediction tasks A.3.1 Object detection We train object detection models with Mask RCNN detection heads. Along with detection, these models also perform instance segmentation. We follow the settings prescribed in recent works like Liu et al. (2024c); Wei et al. (2023); Singh et al. (2023); Vasu et al. (2024b). All evaluations reported in the main paper are from single-scale evaluations on MS COCO validation set following prior works. We sweep through stochastic depth rate in steps of 0.05 and peak learning rate for all the results reported in the paper and the ranges are listed in Table 19. For Vi T-B/16 models, we use Vi TDet style feature pyramid network. For MCi architectures, we follow the setup described in Vasu et al. (2023b). All models were trained using MMDetection library Chen et al. (2019) on a single node with 8 A100 NVIDIA GPUs. Published in Transactions on Machine Learning Research (08/2025) A.3.2 Semantic Segmentation We train segmentation models with Uper Net and Semantic FPN heads. These models are trained on ADE20k Zhou et al. (2017) dataset following the settings prescribed in Liu et al. (2024c); Wei et al. (2023); Singh et al. (2023); Vasu et al. (2024b). All evaluations reported in the main paper are from single-scale evaluations on validation set following prior works. For Vi T-B/16 models, we use Vi TDet style feature pyramid network with Uper Net head. For MCi architectures, we follow the setup described in Vasu et al. (2023b) and train models with only Semantic FPN head. We sweep through stochastic depth rate in steps of 0.05 and peak learning rate for all the results reported in the paper and the ranges are listed in Table 20. All models were trained using MMSegmentation library Contributors (2020) on a single node with 8 A100 NVIDIA GPUs. A.3.3 Depth Estimation We follow the experimental setup and architecture as described in Wei et al. (2023); Vasu et al. (2024b). The models are trained and evaluated on NYUv2 dataset Nathan Silberman & Fergus (2012). We sweep through stochastic depth rate in steps of 0.05 and peak learning rate for all the results reported in the paper and the ranges are listed in Table 21. Table 19: Training hyperparameters for object detection and instance segmentation experiments on MS COCO. RRC is Random Resized Crop. We sweep through stochastic depth rate in steps of 0.05. Hyperparameters Mask RCNN Stochastic depth rate [0.0, ..., 0.3] Data augmentation Multi scale RRC Train epochs 12 Batch size 16 Optimizer Adam W Peak learning rate [5e-4, 2e-4, 1e-4] LR. decay schedule type Step-wise LR. decay schedule [8, 11] Weight decay rate 0.1 Table 20: Training hyperparameters for semantic segmentation experiments on ADE20k. RRC is Random Resized Crop. We sweep through stochastic depth rate in steps of 0.05. Hyperparameters Uper Net Semantic FPN Stochastic depth rate [0.0, ..., 0.2] Data augmentation RRC Crop Size 512 512 Train iterations 160k 40k Batch size 16 64 Optimizer Adam W Peak learning rate [5e-4, 2e-4, 1e-4] LR. decay schedule type Polynomial Warmup iterations 1500 - Weight decay rate 0.01 5e-4 Table 21: Training hyperparameters for depth estimation experiments on NYUv2 dataset. RRC is Random Resized Crop. We sweep through stochastic depth rate in steps of 0.05. Hyperparameters Value Stochastic depth rate [0.0, ..., 0.2] Data augmentation RRC Crop Size 480 480 Train epochs 25 Batch size 24 Optimizer Adam W Peak learning rate [7e-4, 5e-4, 2e-4, 1e-4] Layer decay rate 0.8 Weight decay rate 0.05