# probing_visual_language_priors_in_vlms__128c494c.pdf Probing Visual Language Priors in VLMs Tiange Luo 1 * Ang Cao 1 * Gunhee Lee 2 Justin Johnson 1 Honglak Lee 1 2 Abstract Vision-Language Models (VLMs) may over-rely on visual language priors from their training data rather than true visual reasoning. To investigate this, we introduce Vi LP, a benchmark featuring deliberately out-of-distribution images synthesized via image generation models and out-ofdistribution Q&A pairs. Each question in Vi LP is coupled with three potential answers and three corresponding images: one that can be resolved by text priors alone and two that demand visual reasoning. Although humans achieve near-perfect accuracy, modern VLMs falter; for instance, GPT4o achieves only 66.17% on Vi LP. To alleviate this, we propose a self-improving framework in which models generate new VQA data and then apply pixel-level and semantic corruptions to form good-bad" image pairs for self-training. Our proposed training objective, Image-DPO, compels VLMs to focus more on the actual visual inputs, and we demonstrate its effectiveness in LLa VAv1.5 and Cambrian. Project Page: Vi LP. 1. Introduction Vision-Language Models (VLMs) have advanced text-image interaction, bridging the gap between visual and textual data (Achiam et al., 2023; Team et al., 2023). However, a persistent challenge for learning-based models, such as Res Nets and CLIPs, lies in their reliance on learned priors from the training data, sometimes overlooking visual cues when answering questions (Agrawal et al., 2016; Prabhu et al., 2023). For example, when shown a torus-shaped soccer ball (Figure 1), a model might incorrectly identify it as a sphere due to strong language priors. Simultaneously, these models may adhere to learned visual priors (Thrush et al., 2022; Sterz et al., 2024), which hampers their ability to comprehend out-of-distribution visual cues, such as a zebra *Joint fist authorship, Equal advising 1University of Michigan 2LGAI Research. Correspondence to: Tiange Luo , Ang Cao . Proceedings of the 42 nd International Conference on Machine Learning, Vancouver, Canada. PMLR 267, 2025. Copyright 2025 by the author(s). with atypical spot patterns (Figure 1), which humans would easily discern. This raises an important question: do today s VLMs still over-rely on learned visual language priors, especially given that they rely on far fewer image-text pairs than the internet-scale text corpora used for pretraining? To investigate this, we probe the Visual Language Priors of VLMs by constructing Question-Image-Answer (QIA) triplets that deliberately deviate from the training data distribution. Unlike existing benchmarks that typically rely on internet-sourced images (Goyal et al., 2017; Tong et al., 2024), which inadvertently favor the visual language priors embedded in VLM training data, we utilize modern image generation models, including DALL E-3 (Ramesh et al., 2021) and Flux, to synthesize out-of-distribution imageanswer pairs that also feature notable variation in texture, shape, conceptual combinations, hallucinated elements, and proverb-based contexts. Our benchmark, Vi LP, contains 300 carefully designed questions, each paired with three distinct answers: a Prior Answer and two Test Answers, resulting in a total of 900 QIA triplets. To further challenge the priors of VLMs, we amplify language priors in questions by introducing distractor facts: each question is structured to present a distractor fact followed by a question. The distractor fact directly leads to the Prior Answer. In contrast, the two Test Answers are crafted to challenge the priors by requiring both textual and visual cues for accurate reasoning. While human participants achieved 98% accuracy easily, current VLMs exhibit considerable difficulty, as evidenced by a significant performance drop on our benchmarks, with GPT-4o (Open AI, 2024) scoring only 66.17%. Motivated by the results of Vi LP, we propose Image-DPO, a self-improving method for enhancing VLM visual reasoning by increasing reliance on visual inputs. Our method employs self-generated VQAs using image generation and editing models (Podell et al., 2023; Ren et al., 2024; Brooks et al., 2023a) and applies controlled corruptions to create goodbad" image pairs for DPO-like training (Rafailov et al., 2024). Experiments with open-source VLMs, including LLa VA-v1.5 (Liu et al., 2024a) and Cambrian (Tong et al., 2024), demonstrate its effectiveness. Moreover, we theoretically prove our objective optimizes an upper bound on the RL fine-tuning objective (Rafailov et al., 2024). Probing Visual Language Priors in VLMs Sphere Cube Torus A soccer ball is typically a sphere. What is the geometric shape of the large soccer ball in the image? Flamingo Tiger Penguin Flamingos often stand on one leg. In the image, what pink animal is standing on one leg in the water? One Two Five A common phrase is "kill two birds with one stone". In the image, how many stones are being used to kill two birds? Stripes Spots Checkerboard Zebras usually have stripes. What pattern is on the zebra's body in the image? Brick Pitaya Fork Houses are usually built from bricks. What object is the house in the image built from? Umbrella Plunger Blender An umbrella keeps you dry in the rain. What does the center character in the picture use on his head to stay dry? Pisa Eiffel Bigben The Leaning Tower of Pisa is famous for it's tilt. What tower is leaning in the image? Sheep Pig Chicken Sheep usually produce wool. What farm animal in the middle of the image is producing wool? Giraffe Elephant Camel Giraffes are known for their long necks. What is the tallest long-necked animal shown in the image? Newton Einstein Tesla Einstein created the famous equation E=mc². Which scientist is shown drawing this equation in the image? Figure 1: Sample data from Vi LP. For the same question, Vi LP provides three distinct images and corresponding answers (upper-left corner). All questions follow a consistent structure, combining a distractor fact with a question. The Prior Answer (first column) can be directly inferred from the question, while Test Answers (second & third column) rely on visual cues. Our answers are designed to be single words, and both the model and human evaluators are tasked with open-domain answering, rather than selecting from predefined options. To support this, we have developed a robust synonym and plural detection pipeline, ensuring that open-ended responses do not hinder the evaluation process. This approach also enables evaluation without relying on LLMs. Please refer to Appendix B.1 for more data samples from Vi LP. We investigate the impact of image styles in Appendix B.3, where we generate more realistic images using 4o image generation. Furthermore, we include both qualitative and quantitative comparison results with Winoground (Thrush et al., 2022), Whoops!(Bitton-Guetta et al., 2023a), and Hallusion Bench(Guan et al., 2023) in Appendix B.4. Probing Visual Language Priors in VLMs 2. Related Work VQA Dataset: Significant efforts have produced VQA datasets from various angles, including general VQA (Agrawal et al., 2015; Gurari et al., 2018; Fu et al., 2023; Liu et al., 2023d; Li et al., 2023a; Yu et al., 2023b; Liu et al., 2024a), reading text or charts (Singh et al., 2019a; Mathew et al., 2020; 2021; Masry et al., 2022), complex reasoning (Lu et al., 2022; 2023), composition probing (Hudson & Manning, 2019; Ma et al., 2022; Thrush et al., 2022; Hsieh et al., 2023; Li et al., 2024a), hallucinations (Rohrbach et al., 2018; Li et al., 2023c; Guan et al., 2023), common-sense reasoning (Bitton-Guetta et al., 2023b;a), and more (Majumdar et al., 2024; Sterz et al., 2024). We propose a benchmark that tests VLMs visual reasoning when questions, answers, and images defy common patterns. Following the balanced dataset design of (Goyal et al., 2017), each question is accompanied by three answers: one aligns with language priors, and two deviate, prompting reliance on visual cues. By leveraging state-of-the-art image generation models, our benchmark challenges these priors more effectively than previous datasets built from internet images (Goyal et al., 2017; Tong et al., 2024). Furthermore, unlike the trick category in (Sterz et al., 2024), we first generate question-answer pairs before synthesizing images under specified constraints, creating more difficult visual reasoning tasks. Comprehensive comparisons with existing datasets appear in Appendix B.4. Vision Language Models and Language Priors: Multimodal reasoning is crucial for machine intelligence, with VLMs integrating visual perception, text reasoning, instruction following, and generation for complex tasks (Tan & Bansal, 2019; Li et al., 2019; Kim et al., 2021; Wang et al., 2021b;a; Alayrac et al., 2022; Li et al., 2023b; Chen et al., 2022; Jia et al., 2021; Shen et al., 2021; Singh et al., 2021; Liu et al., 2023c;a; Zhao et al., 2023a; Chen et al., 2023; Zhu et al., 2024b; Li et al., 2024c; Dai et al., 2023; Li et al., 2024c; Yu et al., 2024; Dai et al., 2024; Deitke et al., 2024). Inspired by the success of large language models (Brown et al., 2020; Open AI, 2023a; Touvron et al., 2023a;b; Chiang et al., 2023) and pre-trained visual encoders (Radford et al., 2021; Desai & Johnson, 2020; Caron et al., 2021; Chen et al., 2024), many recent methods leverage relatively small vision-language paired datasets (Liu et al., 2024a; Tong et al., 2024) to fine-tune connectors between LLMs and visual backbones (Liu et al., 2024a). However, these datasets are far smaller than the vast text corpora for LLM pre-training (Open AI, 2023b; Soldaini et al., 2024), and freezing the visual encoder and LLM parameters often preserves language biases, causing visual inputs to be overshadowed (Thrush et al., 2022; Sterz et al., 2024). This challenge is amplified by deliberately generated images that expose such biases, as shown in our study. Previous works and datasets (Goyal et al., 2016; Agrawal et al., 2017; Dancette et al., 2021; Wu et al., 2022; Ramakrishnan et al., 2018; Gouthaman & Mittal, 2020) addressed these issues with curated simulators (Johnson et al., 2016) or internet imagery (Zhang et al., 2016). In this paper, we present a novel VQA benchmark featuring carefully designed questions, fact-based distractors, rare-distribution answers, and image generation techniques to produce realistic visuals that challenge learned visual language priors (Figure 1). Self-Rewarding VLM: Self-rewarding LLM (Yuan et al., 2024) has shown that LLMs can generate and improving themselves in the process via Directed Peference Optimization (DPO) (Rafailov et al., 2024). This approach extends to VLMs by generating new answers for DPO training (Zhou et al., 2024a; Deng et al., 2024; Zhou et al., 2024b; Wang et al., 2024c;a; Yue et al., 2024; Liu et al., 2024b; Xiao et al., 2024). Our work aligns with these self-rewarding VLMs but differs in two key ways: (1) our proposed Image-DPO generates multiple images for a single question-image pair (rather than multiple answers); (2) rather than relying solely on existing images (Zhu et al., 2024a), Image-DPO creates diverse new images using pre-trained generative models (SDXL (Podell et al., 2023), Grounded SAM (Ren et al., 2024), Instruct Pix2Pix (Brooks et al., 2023a)). Furthermore, Image-DPO deliberately corrupts images to produce multiple degraded versions that serve as rejected data in DPO training. Concurrent works (Wang et al., 2024a; Xie et al., 2024) explore similar methods but lack a benchmark to verify enhanced visual focus, fail to establish theoretical connections between their proposed objective and DPO, and utilize limited image transformations (e.g., only randomly cropping). In contrast, we introduce Vi LP (Section 3) to assess visual reasoning and provide theoretical foundations (Appendix A), alongside multi-category image corruptions (semantic editing, Gaussian blurring, pixelation). 3. Vi LP Benchmark 3.1. Design Principles What s the tall animal with the longest neck shown? Humans readily guess giraffe based on learned priors, , yet as shown in the bottom-right of Figure 1, it could be an elephant or camel where visual reasoning corrects the answer. This highlights a potential shortfall in Vision-Language Models (VLMs), which may over-rely on learned visual language priors instead of true visual reasoning, particularly since VLMs are typically fine-tuned on limited image-text data, which is several orders of magnitude smaller than the trained text corpus (Liu et al., 2024a; Tong et al., 2024). Specifically for the scope of visual language priors in this paper, we target (1) strong language priors that lead VLMs to derive answers solely from text, and (2) potential visual priors causing models to overlook critical uncommon visual cues (e.g., unusual zebra spots in Figure 1). Probing Visual Language Priors in VLMs To evaluate how VLMs handle learned visual language priors, we introduce Vi LP, a specialized benchmark of out-ofdistribution Question-Image-Answer (QIA) triplets guided by two core principles. First, text-only inference ensures that each question can be answered with high confidence using textual clues alone. Second, visual inference requires that the correct answer sometimes contradicting common sense only emerges once an out-of-distribution image is considered. By forcing models to integrate both textual and visual information, Vi LP reveals whether they truly engage in visual reasoning or merely rely on memorized patterns. Mathematically, let Q be a question, I an image, and A = {aprior, . . . , atest, . . . } the set of possible answers. We define P(a | Q) as the probability of answer a given Q alone and P(a | Q, I) as the probability given both Q and I. We consider a prior model p, which may represent either human cognition (Phuman) or a VLM/LLM s learned visuallanguage prior (Pθ). For constructing our benchmark, we used the following guidances: Criterion One: The question Q alone should strongly favor aprior, where δ1 is a high-confidence threshold. aprior usually satisfies common knowledge, such as soccer ball is a sphere" and Einstein created E = mc2" (Figure 1). P(aprior | Q) δ1 (1) Criterion Two: With the image I, the correct answer shifts to atest, where δ2 is another high-confidence threshold. The image also significantly impacts the answer distribution, where we let p = P( | Q) and q = P( | Q, I) denote the answer distributions before and after seeing the image, respectively. Here, D is a divergence measure, and δ3 is a threshold indicating significant difference. For instance, the image in the 1st row and the 3rd column of Figure 1 turns the answer to torus. P(atest | Q, I) δ2, DJS p, q δ3 (2) Criterion Three: The answer atest should be rare and unlikely from Q alone, while aprior becomes clearly incorrect when considering I. This is enforced by a low-confidence threshold δ4 (e.g., Newton as atest inferred from the image, contradicting Einstein, shown by the image from the 3rd row and 2nd column of Figure 1). P(atest | Q) δ4, P(aprior | Q, I) δ4 (3) In designing Vi LP, we leverage the human cognition prior Phuman as our guiding principle, ensuring each QIA configuration aligns with typical human expectations while requiring visual evidence to override strong textual assumptions. We then compare the learned priors of VLMs and LLMs, denoted Pθ, Phuman to evaluate whether these models genuinely engage in visual reasoning rather than relying on memorized patterns. 3.2. Question-Image-Answer Generation Following Criterion Three, atest should be highly improbable based on Q alone yet the correct choice when paired with I. Since such images do not exist naturally, we use generative models like DALL E-3 (Ramesh et al., 2021) and Flux to blend unusual elements that override typical language priors. We incorporate substantial human input and leverage advanced LLMs such as Open AI-o1 and Claude3.5-Sonnet to ensure alignment with all the criteria. More details, including text prompts and average cost, are provided in Appendix B.2. Note that as more advanced image generative models such as the recently introduced 4o image generation become available, we anticipate generating increasingly abundant and high-quality data for our benchmark, yet future updates will remain consistent with the dataset construction criteria outlined in Section 3.1. For each question, we design three answers: one aprior inferred solely from Q, and two atest that defy language priors, requiring visual cues for correctness. We rely on GPT-4 to generate text prompts, produce large-scale images, and then conduct human filtering and refinement. This process faces two main challenges: (1) producing diverse out-ofdistribution QA pairs, and (2) synthesizing images that defy specific priors, sometimes necessitating hundreds of samples to find one that accurately matches Q and atest. Table 1: Category Statics. Type Frequency Texture 16 Shape 20 Conceptual combinations 276 Hallucinated Components 151 Proverbs 17 Ultimately, we curated 300 questions, each paired with three distinct image-answer sets, totaling 900 QIA triplets. These cover a broad range from low-level recognition (texture, shape) to high-level reasoning (conceptual combinations, hallucinated components, proverbs). Table 1 summarizes their categorical distribution, with each question spanning an average of 1.6 categories. To reinforce text priors, we present a distractor fact before the question. Rigorous human review ensures that all final QIA triplets are clear and interpretable, as reflected by our human evaluation results in Table 2. Besides, in Appendix B.3, we investigate the impact of image styles by generating more realistic images via 4o image generation and comparing them to those produced by DALL E-3 (Ramesh et al., 2021) and Flux. We find that realistic images can increase the difficulty of the task, highlighting their importance for future studies. 3.3. Dataset Evaluation All of our questions are designed to elicit single-word answers, an approach that is more efficient and more reliable than sentence-based evaluations that rely on LLM judgment. Probing Visual Language Priors in VLMs Image-DPO Text-DPO 1 3 2 (a) Image-DPO vs. Text-DPO: In the green dashed box, we illustrate Image-DPO, which uses a single Q&A pair paired with multiple corrupted images. In contrast, the purple dashed box presents Text-DPO, involving a single input image paired with multiple, distinct Q&A pairs. Q:What is the color scheme of the A: The train has a blue and yellow color scheme, with a blue and yellow passenger train engine. Good Image 𝐼! Corrupted Image 𝐼" Semantic Corrupted Gaussian Blurred Pixelated VLM Chosen Rejected (b) Illustration of Image DPO. We construct chosen and rejected pairs by corrupting the image with a set of perturbations while keeping the Q&A unchanged. Perturbations include semantic editing, Gaussian blurring, and pixelation. The mathematical formulations and implementation details are provided in Appendix A and Appendix C, respectively. By avoiding using LLM, we reduce API fees, computational overhead, and the risk of occasional inaccuracies due to incorrect model reasoning. We explicitly instruct the model to provide a single-word answer, and we evaluate the correctness of each response using a binary system. To ensure a fair evaluation, we devote significant efforts to building a comprehensive set of synonyms and plural for each answer to detect other valid alternative answers. This ensures that the model is only penalized for actual errors, not for providing synonymous or alternative correct responses. 4. Image DPO Inspired by our benchmark, we propose Image DPO, a selfimprovement method for enhancing VLMs visual reasoning, featuring a new objective and a data generation pipeline using VLMs themselves and pre-trained image models. 4.1. Objective Existing approaches for VLM self-improvement follow the way used in DPO paper (Rafailov et al., 2024), where the model is trained to distinguish between good and bad answers for a fixed image and question (Figure 2a right). However, this straightforward adaptation may not the best for vision models, as the model sometimes distinguish good and bad answers from the text alone without needing to analyze the image. In contrast, we propose Image DPO, a vision-focused objective that creates good and bad questionimage-answer pairs by corrupting the image while keeping the question and answer unchanged (Figure 2a left). An example of our synthetic data is illustrated in Figure 2b. Formally, given an image Iw, a question Q, and its corresponding answer A, we generate a corrupted image Il via image-editing operations, including Gaussian blur, pixelation, or semantic modifications. The triplet (Q, Il, A) forms a degraded question-image-answer pair compared to (Q, Iw, A). We train the model to distinguish between good and bad triplets using the objective 4, where πθ is the target VLM, πref is the reference VLM (typically an earlier version of πθ), S is the dataset of good and bad triplets, σ is the sigmoid function, and α is a scaling factor. In Proposition 1, we show that objective 4 optimizes an upper bound on the RL fine-tuning objective (Eq. (3) in (Rafailov et al., 2024)). L(πθ, πref) = EQ,Iw,Il,A S h log σ α πθ(A | Q, Iw) πref(A | Q, Iw) α πθ(A | Q, Il) πref(A | Q, Il) Intuitively, since the textual inputs and outputs are identical in both good and bad cases, the gradients of this objective push the model to rely more on the vision branch, driving a shift in gradient direction when processing normal images Iw compared to corrupted images Il (Figure 6). This behavior encourages the model to focus more on image inputs rather than relying solely on text-based reasoning, thereby enhancing its performance on visual-related tasks. Our experiments demonstrate Image-DPO objective (Eq. 4) outperforms varies self-improve VLM baselines on Vi LP. Proposition 1. Let LRLFT(πθ, πref; S) be the KLconstrained reward maximization objective used in the RL fine-tuning phase (Appendix Eq. 7), where the dataset S = {(Q, A, Iw, Il)} contains good images Iw and corrupted images Il. Let LImage DPO(πθ, πref; S) be the objective from Eq. 4, which compares Q, Iw, A against Q, Il, A . Then for any policy πθ and reference model πref, we have LRLFT(πθ, πref; S) LImage DPO(πθ, πref; S). Proof Sketch. Following (Rafailov et al., 2024), we express the optimal KL-constrained policy in terms of a latent reward function. Applying a Bradley Terry preference model to question-image-answer triplets Q, Iw/Il, A and using Jensen s inequality yields an upper bound whose minimization is equivalent to LImage DPO(πθ, πref; S). A full derivation appears in Appendix A. Probing Visual Language Priors in VLMs Flamingo Tiger Penguin Flamingos often stand on one leg. In the image, what pink animal is standing on one leg in the water? Moscow London Paris The Red Square is in Moscow. From the image, in which city is the Red Square located? Please answer with a single word or abbreviation, like NYC. Round Hexagon Dodecagon The moon is typically depicted as round. What shape does the moon appear to take in this image of the night sky? Camel Horse Tiger Camels have humps. Which animal in the image stores fat in its humps? GPT4o: Flamingo Claude: Flamingo Gemini: Flamingo Llama V: Flamingo GPT4o: Flamingo Claude: Flamingo Gemini: Flamingo Llama V: Flamingo GPT4o: Penguin Claude: Flamingo Gemini: Penguin Llama V: Penguin GPT4o: Hexagon Claude: Hexagon Gemini: Round Llama V: Hexagon GPT4o: Hexagon Claude: Hexagon Gemini: Hexagon Llama V: Hexagon GPT4o: Polygonal Claude: Octagonal Gemini: Octagonal Llama V: Octagonal GPT4o: Camel Claude: Camel Gemini: Camel Llama V: Camel GPT4o: Camel Claude: Camel Gemini: Camel Llama V: Camel GPT4o: Camel Claude: Camel Gemini: Camel Llama V: Camel GPT4o: Moscow Claude: Moscow Gemini: Moscow Llama V: Moscow GPT4o: Moscow Claude: Moscow Gemini: Moscow Llama V: Moscow GPT4o: Moscow Claude: Paris Gemini: Paris Llama V: Moscow Figure 3: Qualitative examples. We show the results from GPT-4o, Claude-3.5-Sonnet, Gemini-1.5-Pro, and Llama-3.2Vision-90B for some challenging cases. Please refer to Appendix B.6 for categorized failure case analysis. 4.2. Data Generation Training VLMs demands large-scale question-image-answer (QIA) triplets, which are often scarce. To address this, we introduce a scalable data generation pipeline (Appendix Figure 16) that repurposes existing image datasets via VLM themselves and image generative models. Given a seed image from COCO (Lin et al., 2014), Text2VQA (Singh et al., 2019b), or Visual Genome (Krishna et al., 2017), VLMs are tasked with simultaneously selecting appropriate functions (e.g., image generation or editing models) and generating corresponding instructions. These instructions are then used to produce new images, in addition to the seed image, as illustrated in Figure 17. The same VLMs are then employed to generate QA pairs for these newly created images. Following, we apply the mentioned three types of image corruptions to the generated images, constructing good bad pars (Q, Iw, A) and (Q, Il, A). Specifically, we employ Stable Diffusion XL (Podell et al., 2023; Rombach et al., 2022) for image generation, and use Instruct-Pix2Pix (Brooks et al., 2023a), and Grounded-SAM (Rombach et al., 2022; Ren et al., 2024) for image editing. Example generated data, prompts, and more details are included in Appendix C. 5. Experiments We introduce Vi LP, a new benchmark comprising 300 questions. Each question is paired with three unique images and their corresponding answers one QIAprior and two QIAtest for a total of 900 QIAs. The QIAprior examples (300 in total) align with common language priors (i.e., they can usually be answered correctly by relying on textual cues alone). In contrast, the QIAtest examples (600 in total) challenge these priors by requiring visual reasoning. Vi LP features two evaluation settings: Vi LPF, where both distractor facts and the questions are provided; Vi LPP, where only the questions themselves are given (i.e., no distractor facts). We report two metrics in Table 2: average accuracy on QIAtest (noted as Score) and average accuracy on QIAprior (noted as Prior). Our benchmark emphasizes the performance in Score. Is the QIA easy for humans? We begin by evaluating our benchmark through a human study. Participants achieved nearly perfect accuracy on Vi LPF-Prior and over 98% on Vi LPF-Score and Vi LPP-Score, confirming that our questionimage-answer combinations are unambiguous for human interpretation. Notably, despite QIAtest being designed as out-of-distribution examples, humans were still able to correctly distinguish them. Probing Visual Language Priors in VLMs Figure 4: Comparison of benchmark scores under different image transformations. Solid line and dotted line refer to Vi LPF-Score and Vi LPF-Prior, respectively. Table 2: Benchmarking on Vi LP. Please refer to the left text for symbol definitions. indicates the model often fails to follow the instructions. Model Vi LPF Vi LPP Score Prior Score Prior Baseline Human 98.33 99.67 98.67 96.67 GPT-4o (text only) 0.0 92.33 0.17 71.33 API call only GPT-4o 66.17 91.00 56.00 87.67 GPT-4V 57.67 88.33 38.33 85.33 GPT-4o-Mini 57.67 89.00 46.67 84.67 Claude-3.5-Sonnet 70.00 84.33 59.33 86.67 Claude-3-Opus 59.17 74.00 43.00 82.67 Claude-3-Sonnet 48.83 83.67 40.33 81.33 Claude-3-Haiku 43.67 82.67 34.83 82.33 Gemini-1.5-Pro 60.50 79.33 48.00 83.00 Gemini-1.5-Flash 54.50 83.33 69.17 79.67 Open weights Llama-3.2-Vision-11B 67.33 76.67 61.17 79.33 Llama-3.2-Vision-90B 64.00 91.67 63.17 83.33 Molmo E-1B 48.67 57.33 47.83 69.00 Molmo-7B-O 57.83 60.67 47.33 76.33 Molmo-7B-D 54.5 69.00 46.17 72.33 Molmo-72B 60.33 85.00 47.17 82.33 Qwen2-VL-7B 50.50 83.00 48.67 80.33 Qwen2-VL-72B 56.50 92.33 53.83 83.00 Intern VL2-8B 47.00 66.67 43.00 75.00 Intern VL2-76B 42.67 47.67 50.84 74.33 LLa VA-1.5-7B 29.67 71.33 37.67 65.67 LLa VA-1.5-13B 35.33 81.00 41.50 73.67 Cambrian-1-8B 8.67 43.67 32.50 63.67 LLa VA-One Vision-7B 54.17 82.33 49.67 75.00 LLa VA-One Vision-72B 1.67 3.00 5.22 11.67 Humans performed slightly better on Vi LPF-Prior when distractor facts were provided, as they could easily identify that these facts aligned with the correct answers. Moreover, Vi LPF-Score was marginally lower when facts were introduced, as the distractor facts added some noise and caused minor confusion, although the impact of this noise is relatively small. These findings are consistent with the design principles of our benchmark. Narwhal Dolphin Narwhal has long tusk. Which marine mammal with long tusk is depicted in the image? GPT4o: Narwhal LLa VA: Narwhal GPT4o: Narwhal LLa VA: Narwhal GPT4o: Dolphin LLa VA: Narwhal GPT4o: Narwhal LLa VA: Dolphin Octagon Round Stop signs are usually octagons. What is the shape of the stop sign in the image? GPT4o: Octagon LLa VA: Octagon GPT4o: Octagon LLa VA: Octagon GPT4o: Sphere LLa VA: Octagon GPT4o: Octagon LLa VA: Round Figure 5: Qualitative results before and after removing distactor facts. GPT-4o and LLa VA-1.5-13B models yield completely opposite behaviors. Are our QIAs aligned with the learned priors of VLMs? We tested GPT-4o (text only) on our questions (removing all image references). Despite no visual content, it correctly answered 92.33% on Vi LPF-Prior. The accuracy drops to 71.33% once distractor facts are removed, showing that these facts significantly guide the answer. For QIAtest, GPT4o (text only) accuracy nearly falls to 0% (Vi LPF-Score & Vi LPP-Score), indicating the QIAtest cannot be answered using text alone. How do VLMs perform on our benchmark? Although our benchmark questions are distinguishable for humans, they are challenging for VLMs. Even the advanced VLM models like GPT-4o, have a clear performance gap (66.17% v.s. 98.33%) compared to humans performance on Vi LPFScore, indicating the difficulty of these questions for VLMs. Claude-3.5-Sonnet achieved the best score 70%, while most of the commercial VLMs are below 60%. Figure 3 highlights sample outputs from top commercial and open-source models, including GPT-4o, Claude-3.5-Sonnet, Gemini-1.5Pro, and Llama-3.2-Vision-90B. They face significant challenges when addressing these cases in our Vi LP, whereas humans can arrive at correct answers after consideration. Notably, it is encouraging to see that some open-source models achieved over 60% accuracy on Vi LPF-Score, with performance nearing that of their commercial counterparts, including Llama-3.2-Vison and Molmo-72B. Additionally, we provide more detailed failure-case analysis in Appendix B.6, where we group the errors into six recurring modes (i) shape-recognition slips, (ii) counting mistakes, (iii) prior- Probing Visual Language Priors in VLMs Table 3: Effectiveness of Image-DPO on General VQA benchmarks. VLMs Vi LPFScore Vi LPPScore NBQ NBI NBG NBB MM-Vet CHAIRS CHAIRI LLa VA-1.5-7B 29.67 37.67 37.7 43.8 12.7 67.3 31.1 49.1 14.8 LLa VA-1.5-7B + Image-DPO 34.17 4.5 39.33 1.66 39.79 2.09 45.47 1.67 14.16 1.46 68.45 1.15 32.3 1.2 45 -4.1 12.3 -2.5 LLa VA-1.5-13B 35.33 41.5 39.6 44.6 14.8 68.9 36.1 48.3 14.1 LLa VA-1.5-13B + Image-DPO 38.17 2.84 42.5 1 42.68 3.08 47.37 2.77 17.16 2.36 70.36 1.46 37.5 1.4 42.6 -5.7 11.6 -2.5 Cambrian-8B 8.67 32.5 44.6 47.9 19.4 71.5 51.4 14.5 4.7 Cambrian-8B + Image-DPO 20.83 12.16 39.3 6.83 46.5 1.9 50.2 2.3 20 0.6 72 0.5 51.7 0.3 11.4 -3.1 4.4 -0.3 Table 4: Benchmark Performance with More Fine-Grained Categories. We extended our evaluation system to include three categories Correct, Wrong, and Invalid. Invalid responses (not single-word) are further split into Invalid-Correct (noted as Invalid- ) and Invalid-Wrong (noted as Invalid- ) based on GPT assessment. Model Vi LPF Vi LPP Score Prior Score Prior Correct Wrong Invalid Invalid Correct Wrong Invalid Invalid Correct Wrong Invalid Invalid Correct Wrong Invalid Invalid- Cambrian-8B 8.67% 28.83% 29.17% 33.33% 43.67% 2.33% 54.00% 6.00% 32.50% 37.50% 11.67% 18.33% 63.67% 8.33% 22.00% 6.00% LLa VA-One Vision-7B 54.17% 42.83% 1.33% 1.67% 82.33% 15.33% 2.33% 0.00% 49.67% 47.00% 1.33% 2.00% 75.00% 20.33% 3.67% 1.00% LLa VA-One Vision-72B 1.67% 1.00% 43.33% 54.00% 3.00% 0.33% 90.33% 6.33% 5.22% 5.78% 39.00% 50.00% 11.67% 1.33% 78.67% 8.33% GPT-4o 66.17% 33.77% 0.03% 0.03% 91.00% 8.33% 0.33% 0.33% 56.00% 43.17% 0.50% 0.03% 87.67% 11.33% 1.00% 0.00% biased rejection of counterfactual scenes, (iv) proverbdriven hallucinations, (v) misclassification of creative concepts, and (vi) text-over-vision errors on blended images. Do distractor facts really distract? In Vi LPF setting, we add a distractor fact before the question. Since these facts implicitly suggest incorrect answers for QIAtest, we expected this change to make the questions more suggestive and lower the Vi LPF-Score, as the distractors would mislead the VLMs. Surprisingly, GPT-4o benefits from distractor facts, improving accuracy on QIAtest. We hypothesize these facts highlight question focus, narrowing the search space. However, weaker models like LLa VA-1.5-13B (Liu et al., 2023b) often get misled by the distractors, hurting their Score but boosting Prior. For instance, as shown in Figure 5, with including distractor facts, LLa VA-1.5-13B consistently predicts the distractor fact as the answer. However, once the distractors are removed, it can then predict correctly. For bad instruction-following models like Cambrian8B (Tong et al., 2024), distractor facts significantly hinder adherence to explicit instructions, such as providing single-word answers. With facts, Cambrian-8B fails to follow instructions in 62% of cases, compared to 30% without (a nearly 2x increase). Manual review shows 59% of these failures are contextually correct, yielding an adjusted accuracy of 47.92%. Similarly, LLa VA-One Vision72B (Li et al., 2024b) often generates detailed analyses despite explicit single-word prompts. This trend highlights a concerning trend: focusing on improving performance on well-established benchmarks may come at the cost of basic instruction-following abilities, ultimately limiting the practical utility of these models in real-world applications. How image transformations affect the results? We also investigate how image transformations, including resizing, Gaussian blur, and pixelation, affect Vi LP performance. The results, shown in Figure 4, reveal that the Vi LPF-Score rapidly decreases as the severity of the transformations (xaxis) increases, while the Vi LPF-Prior score remains around 50%. Interestingly, GPT-4o, when using degraded images, performs worse in Vi LPF-Prior than when no images are used, i.e., GPT-4o (text only) in Table 2. Comparison of Vi LP with other VQA datasets: To highlight the distinctions between Vi LP and existing benchmarks, including Winoground (Thrush et al., 2022), Whoops! (Bitton-Guetta et al., 2023a), and Hallusion Bench (Guan et al., 2023), we conduct a comparative analysis of both their high-level design principles and low-level data formats. This comparison incorporates qualitative and quantitative insights as detailed in Appendix B.4. Finer-Grained Evaluation Currently, we explicitly ask VLMs to produce a single-word response for evaluation, and penalize responses that include long sentences, as we believe instruction-following is a foundation capability of VLMs thus, failure to comply warrants a penalty. Although this evaluation protocol is cost-efficient without calling LLM, fast, and reliable, it may conflate semantic errors with instruction-following failures. To address this, we introduce an optional finer-grained evaluation protocol with three categories: Correct, Wrong, and Invalid. Correct and Wrong denote semantically accurate or inaccurate answers that comply with the single-word constraint. Invalid captures responses that violate the format. For these, we further offer an optional GPT-based classification into Invalid-Correct or Invalid-Wrong, providing greater granularity at the cost of increased evaluation overhead. We show these evaluation results in Table 4. Probing Visual Language Priors in VLMs Commercial models such as GPT-4o follow instructions reliably. In contrast, LLa VA-One Vision-72B matches other models in reasoning ability (Correct + Invalid-Correct) yet ignores the single-word instruction in most instances. Cambrian-8B similarly yields a significant number of Invalid cases. Counterintuitively, LLa VA-One Vision-72B performs markedly worse at instruction following than its smaller counterpart, LLa VA-One Vision-7B, and shows a stronger bias toward learned priors reflected in its lower Vi LP Score and higher Vi LP Prior. We hypothesize that instruction following offers an additional lens for probing prior reliance: models dominated by priors tend to generate verbose, prior-driven answers (i.e., long responses) that disregard explicit single-word directives. 5.1. Image DPO In this section, we evaluate Image-DPO (Section 4.1) on both Vi LP and general VQA benchmarks. As an ablation baseline, we introduce Text-DPO, which uses the same Question-Image-Answe (QIA) generation process as we used in Image-DPO but applies LLM self-rewarding objective (Yuan et al., 2024) (the standard DPO objective). In Text-DPO, good and bad pairs stem from VLM-generated positive and negative answers, while the question and image remain fixed. As shown in the right of Figure 2a, the green box depicts Image-DPO, generating corrupted images via semantic edits, Gaussian blur, and pixelation while keeping the question and answer constant; the purple box illustrates Text-DPO, which fixes the image and varies the answers with associated ratings. This setup parallels other VLM self-rewarding work (Zhou et al., 2024a; Deng et al., 2024; Zhou et al., 2024b; Wang et al., 2024c;a). For baselines, we compare with VLM self-improvement methods, including SIMA (Wang et al., 2024c), HADPO (Zhao et al., 2023a), and EOS (Yue et al., 2024), by using their publicly available checkpoints. Additionally, we train models using the dataset and code provided in RLHF-V (Yu et al., 2023a) and V-DPO (Xie et al., 2024). All models use LLa VA-7B for a comprehensive comparisons as many paper only release 7B checkpoints. Table 5 shows that Image-DPO achieves the highest across all the metrics. Besides, we evaluate the proposed Image-DPO algorithm across three VLM models Cambrian-8B, LLa VA-1.5-7B, and LLa VA-1.5-13B using several popular VLM benchmarks that focus on different aspects, including compositionality & biases (Natural Bench (Li et al., 2024a)), general visual reasoning (MM-Vet (Yu et al., 2023c)), and hallucinations (CHAIR (Rohrbach et al., 2018)). The results, presented in Table 3, show consistent performance improvements across both datasets and models, further demonstrating the effectiveness of our Image DPO method. Table 5: Comparisons of Image-DPO on Vi LP. Model Vi LP F Vi LP P Score Prior Score Prior LLa VA-1.5-7B 29.67 71.33 37.67 65.67 +HADPO (Zhao et al., 2023b) 33.00 74.33 38.50 65.00 +RLHF-V (Yu et al., 2023a) 29.50 75.00 36.33 65.33 +EOS (Yue et al., 2024) 31.33 67.00 38.67 65.67 +SIMA (Wang et al., 2024b) 27.83 68.67 36.17 66.00 +V-DPO (Xie et al., 2024) 29.50 72.67 37.83 67.67 +Text-DPO 31.34 71.67 37.83 65.67 +Image-DPO 34.17 75.00 39.33 68.00 Text DPO on Corrupted Images. Does Image-DPO s improvement stem from the objective itself, or is it merely due to training on more perturbed data? To investigate this, we conduct an ablation where we train Text DPO with corrupted images. Specifically, for each pair of (Q, I, Aw) and (Q, I, Al) used in Text-DPO, we apply the same corruptions as Image-DPO to I to form (Q, I , Aw) and (Q, I , Al). Results show that LLa VA-v1.5-7B trained with this Text DPO variant achieves a Vi LPF-Score of 31, a Vi LPP-Score of 37.5, and 30.3 on MMVET. In contrast, Image DPO outperforms it across all metrics, achieving a Vi LPF-Score of 34.17, a Vi LPP-Score of 39.33, and 32.3 on MMVET. Impact of Synthetic Image Augmentation. When we remove the image-generation stage and fine-tune Image DPO only on QIA pairs drawn from the seed datasets (COCO, VG, Text2VQA), performance slips across every metric: Vi LPFScore declines from 34.17 to 31.17 and Vi LPPScore from 39.33 to 38.00; the four novel-bias scores (NBQ, NBI, NBG, NBB) each lose roughly one point; MM-Vet drops from 32.3 to 31.4; and hallucination rates rise (CHAIRS : 45 47, CHAIRI : 12.3 13.6). These uniform declines confirm that synthetic images by image generative models broaden the visual distribution, inject the priors of powerful generative models, and are crucial for achieving the full gains of Image DPO. 6. Conclusion In conclusion, we present the Vi LP benchmark to probe the challenge of visual language bias in Vision-Language Models (VLMs). By utilizing advanced image generation models and designing questions that demand visual cues for accurate responses, our benchmark includes images that defy language priors, revealing the limitations of current VLMs. Our method, Image-DPO, which incorporates selfgenerated VQA pairs and image corruption for training, has demonstrated promising improvements in enhancing visual reliance, as evidenced by performance gains on open-source VLMs, such as LLa VA-v1.5 and Cambrian. Probing Visual Language Priors in VLMs Acknowledgment This work was supported by the LG AI Research grant. We also extend our gratitude to the Open AI Researcher Access Program for providing credits used to access Open AI s APIs. Additionally, we appreciate Chris Rockwell and Jiaming Yang for their contributions to our human evaluation studies, and thank Yongyi Yang for discussions regarding the proposition proof. Impact Statement This work aims to advance the field of Vision-Language Models (VLMs) by diagnosing and mitigating their reliance on learned visual language priors. By introducing the Vi LP benchmark and Image-DPO training framework, our contributions promote more robust and interpretable visual reasoning in AI systems. While Vi LP dataset involves synthetic image generation, we have rigorous human review to avoid generating harmful or deceptive content. Our findings may improve AI alignment with human perception, benefiting applications in accessibility, education, and AI-assisted decision-making. Achiam, J., Adler, S., Agarwal, S., Ahmad, L., Akkaya, I., Aleman, F. L., Almeida, D., Altenschmidt, J., Altman, S., Anadkat, S., et al. Gpt-4 technical report. ar Xiv preprint ar Xiv:2303.08774, 2023. Agrawal, A., Lu, J., Antol, S., Mitchell, M., Zitnick, C. L., Parikh, D., and Batra, D. Vqa: Visual question answering. International Journal of Computer Vision, 123:4 31, 2015. URL https://api.semanticscholar.org/ Corpus ID:3180429. Agrawal, A., Batra, D., and Parikh, D. Analyzing the behavior of visual question answering models. ar Xiv preprint ar Xiv:1606.07356, 2016. Agrawal, A., Batra, D., Parikh, D., and Kembhavi, A. Don t just assume; look and answer: Overcoming priors for visual question answering. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4971 4980, 2017. URL https://api.semanticscholar. org/Corpus ID:19298149. Alayrac, J.-B., Donahue, J., Luc, P., Miech, A., Barr, I., Hasson, Y., Lenc, K., Mensch, A., Millican, K., Reynolds, M., Ring, R., Rutherford, E., Cabi, S., Han, T., Gong, Z., Samangooei, S., Monteiro, M., Menick, J., Borgeaud, S., Brock, A., Nematzadeh, A., Sharifzadeh, S., Binkowski, M., Barreira, R., Vinyals, O., Zisserman, A., and Simonyan, K. Flamingo: a visual language model for few- shot learning. Ar Xiv, abs/2204.14198, 2022. URL https: //api.semanticscholar.org/Corpus ID:248476411. Bitton-Guetta, N., Bitton, Y., Hessel, J., Schmidt, L., Elovici, Y., Stanovsky, G., and Schwartz, R. Breaking common sense: Whoops! a vision-and-language benchmark of synthetic and compositional images. 2023 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 2616 2627, 2023a. URL https://api. semanticscholar.org/Corpus ID:257496749. Bitton-Guetta, N., Bitton, Y., Hessel, J., Schmidt, L., Elovici, Y., Stanovsky, G., and Schwartz, R. Breaking common sense: Whoops! a vision-and-language benchmark of synthetic and compositional images. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 2616 2627, 2023b. Brooks, T., Holynski, A., and Efros, A. A. Instructpix2pix: Learning to follow image editing instructions. In CVPR, 2023a. Brooks, T., Holynski, A., and Efros, A. A. Instructpix2pix: Learning to follow image editing instructions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 18392 18402, 2023b. Brown, T. B., Mann, B., Ryder, N., Subbiah, M., Kaplan, J., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., Agarwal, S., Herbert-Voss, A., Krueger, G., Henighan, T., Child, R., Ramesh, A., Ziegler, D. M., Wu, J., Winter, C., Hesse, C., Chen, M., Sigler, E., Litwin, M., Gray, S., Chess, B., Clark, J., Berner, C., Mc Candlish, S., Radford, A., Sutskever, I., and Amodei, D. Language models are few-shot learners. Ar Xiv, abs/2005.14165, 2020. URL https: //api.semanticscholar.org/Corpus ID:218971783. Caron, M., Touvron, H., Misra, I., J egou, H., Mairal, J., Bojanowski, P., and Joulin, A. Emerging properties in selfsupervised vision transformers. 2021 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 9630 9640, 2021. URL https://api.semanticscholar. org/Corpus ID:233444273. Chen, L., Li, J., Dong, X., Zhang, P., He, C., Wang, J., Zhao, F., and Lin, D. Sharegpt4v: Improving large multi-modal models with better captions. ar Xiv preprint ar Xiv:2311.12793, 2023. Chen, X., Wang, X., Changpinyo, S., Piergiovanni, A. J., Padlewski, P., Salz, D. M., Goodman, S., Grycner, A., Mustafa, B., Beyer, L., Kolesnikov, A., Puigcerver, J., Ding, N., Rong, K., Akbari, H., Mishra, G., Xue, L., Thapliyal, A. V., Bradbury, J., Kuo, W., Seyedhosseini, M., Jia, C., Ayan, B. K., Riquelme, C., Steiner, A., Angelova, A., Zhai, X., Houlsby, N., and Soricut, Probing Visual Language Priors in VLMs R. Pali: A jointly-scaled multilingual language-image model. Ar Xiv, abs/2209.06794, 2022. URL https: //api.semanticscholar.org/Corpus ID:252222320. Chen, Z., Wu, J., Wang, W., Su, W., Chen, G., Xing, S., Zhong, M., Zhang, Q., Zhu, X., Lu, L., et al. Internvl: Scaling up vision foundation models and aligning for generic visual-linguistic tasks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 24185 24198, 2024. Chiang, W.-L., Li, Z., Lin, Z., Sheng, Y., Wu, Z., Zhang, H., Zheng, L., Zhuang, S., Zhuang, Y., Gonzalez, J. E., Stoica, I., and Xing, E. P. Vicuna: An opensource chatbot impressing gpt-4 with 90%* chatgpt quality, March 2023. URL https://lmsys.org/blog/ 2023-03-30-vicuna/. Dai, W., Li, J., Li, D., Tiong, A. M. H., Zhao, J., Wang, W., Li, B. A., Fung, P., and Hoi, S. C. H. Instructblip: Towards general-purpose vision-language models with instruction tuning. Ar Xiv, abs/2305.06500, 2023. URL https://api.semanticscholar.org/Corpus ID: 258615266. Dai, W., Lee, N., Wang, B., Yang, Z., Liu, Z., Barker, J., Rintamaki, T., Shoeybi, M., Catanzaro, B., and Ping, W. Nvlm: Open frontier-class multimodal llms. ar Xiv preprint ar Xiv:2409.11402, 2024. Dancette, C., Cadène, R., Teney, D., and Cord, M. Beyond question-based biases: Assessing multimodal shortcut learning in visual question answering. 2021 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 1554 1563, 2021. URL https://api. semanticscholar.org/Corpus ID:233168594. Deitke, M., Clark, C., Lee, S., Tripathi, R., Yang, Y., Park, J. S., Salehi, M., Muennighoff, N., Lo, K., Soldaini, L., et al. Molmo and pixmo: Open weights and open data for state-of-the-art multimodal models. ar Xiv preprint ar Xiv:2409.17146, 2024. Deng, Y., Lu, P., Yin, F., Hu, Z., Shen, S., Zou, J., Chang, K.-W., and Wang, W. Enhancing large vision language models with self-training on image comprehension. ar Xiv preprint ar Xiv:2405.19716, 2024. Desai, K. and Johnson, J. Virtex: Learning visual representations from textual annotations. 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 11157 11168, 2020. URL https: //api.semanticscholar.org/Corpus ID:219573658. Fu, C., Chen, P., Shen, Y., Qin, Y., Zhang, M., Lin, X., Qiu, Z., Lin, W., Yang, J., Zheng, X., Li, K., Sun, X., and Ji, R. Mme: A comprehensive evaluation benchmark for multimodal large language models. Ar Xiv, abs/2306.13394, 2023. URL https://api. semanticscholar.org/Corpus ID:259243928. Gouthaman, K. and Mittal, A. Reducing language biases in visual question answering with visually-grounded question encoder. Ar Xiv, abs/2007.06198, 2020. URL https: //api.semanticscholar.org/Corpus ID:220496567. Goyal, Y., Khot, T., Summers-Stay, D., Batra, D., and Parikh, D. Making the v in vqa matter: Elevating the role of image understanding in visual question answering. International Journal of Computer Vision, 127:398 414, 2016. URL https://api.semanticscholar.org/ Corpus ID:8081284. Goyal, Y., Khot, T., Summers-Stay, D., Batra, D., and Parikh, D. Making the v in vqa matter: Elevating the role of image understanding in visual question answering. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 6904 6913, 2017. Guan, T., Liu, F., Wu, X., Xian, R., Li, Z., Liu, X., Wang, X., Chen, L., Huang, F., Yacoob, Y., Manocha, D., and Zhou, T. Hallusionbench: An advanced diagnostic suite for entangled language hallucination and visual illusion in large vision-language models. 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 14375 14385, 2023. URL https: //api.semanticscholar.org/Corpus ID:265499116. Gurari, D., Li, Q., Stangl, A., Guo, A., Lin, C., Grauman, K., Luo, J., and Bigham, J. P. Vizwiz grand challenge: Answering visual questions from blind people. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 3608 3617, 2018. URL https://api. semanticscholar.org/Corpus ID:3831582. Hsieh, C.-Y., Zhang, J., Ma, Z., Kembhavi, A., and Krishna, R. Sugarcrepe: Fixing hackable benchmarks for visionlanguage compositionality. Ar Xiv, abs/2306.14610, 2023. URL https://api.semanticscholar.org/Corpus ID: 259251493. Hudson, D. A. and Manning, C. D. Gqa: A new dataset for real-world visual reasoning and compositional question answering. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6693 6702, 2019. URL https://api.semanticscholar. org/Corpus ID:152282269. Jaques, N., Gu, S., Bahdanau, D., Hernández-Lobato, J. M., Turner, R. E., and Eck, D. Sequence tutor: Conservative fine-tuning of sequence generation models with klcontrol. In International Conference on Machine Learning, pp. 1645 1654. PMLR, 2017. Probing Visual Language Priors in VLMs Jaques, N., Shen, J. H., Ghandeharioun, A., Ferguson, C., Lapedriza, A., Jones, N., Gu, S. S., and Picard, R. Humancentric dialog training via offline reinforcement learning. ar Xiv preprint ar Xiv:2010.05848, 2020. Jia, C., Yang, Y., Xia, Y., Chen, Y.-T., Parekh, Z., Pham, H., Le, Q. V., Sung, Y.-H., Li, Z., and Duerig, T. Scaling up visual and vision-language representation learning with noisy text supervision. Ar Xiv, abs/2102.05918, 2021. URL https://api.semanticscholar.org/Corpus ID: 231879586. Johnson, J., Hariharan, B., van der Maaten, L., Fei-Fei, L., Zitnick, C. L., and Girshick, R. B. Clevr: A diagnostic dataset for compositional language and elementary visual reasoning. 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1988 1997, 2016. URL https://api.semanticscholar.org/Corpus ID: 15458100. Kim, W., Son, B., and Kim, I. Vilt: Vision-and-language transformer without convolution or region supervision. In International Conference on Machine Learning, 2021. URL https://api.semanticscholar.org/Corpus ID: 231839613. Krishna, R., Zhu, Y., Groth, O., Johnson, J., Hata, K., Kravitz, J., Chen, S., Kalantidis, Y., Li, L.-J., Shamma, D. A., et al. Visual genome: Connecting language and vision using crowdsourced dense image annotations. International journal of computer vision, 123:32 73, 2017. Li, B., Wang, R., Wang, G., Ge, Y., Ge, Y., and Shan, Y. Seed-bench: Benchmarking multimodal llms with generative comprehension. Ar Xiv, abs/2307.16125, 2023a. URL https://api.semanticscholar.org/Corpus ID: 260334888. Li, B., Lin, Z., Peng, W., Nyandwi, J. d. D., Jiang, D., Ma, Z., Khanuja, S., Krishna, R., Neubig, G., and Ramanan, D. Naturalbench: Evaluating vision-language models on natural adversarial samples. ar Xiv preprint ar Xiv:2410.14669, 2024a. Li, B., Zhang, Y., Guo, D., Zhang, R., Li, F., Zhang, H., Zhang, K., Li, Y., Liu, Z., and Li, C. Llavaonevision: Easy visual task transfer. ar Xiv preprint ar Xiv:2408.03326, 2024b. Li, J., Li, D., Savarese, S., and Hoi, S. C. H. Blip-2: Bootstrapping language-image pre-training with frozen image encoders and large language models. In International Conference on Machine Learning, 2023b. URL https: //api.semanticscholar.org/Corpus ID:256390509. Li, L. H., Yatskar, M., Yin, D., Hsieh, C.-J., and Chang, K.-W. Visualbert: A simple and performant baseline for vision and language. Ar Xiv, abs/1908.03557, 2019. URL https://api.semanticscholar.org/Corpus ID: 199528533. Li, Y., Du, Y., Zhou, K., Wang, J., Zhao, W. X., and rong Wen, J. Evaluating object hallucination in large visionlanguage models. In Conference on Empirical Methods in Natural Language Processing, 2023c. URL https: //api.semanticscholar.org/Corpus ID:258740697. Li, Y., Zhang, Y., Wang, C., Zhong, Z., Chen, Y., Chu, R., Liu, S., and Jia, J. Mini-gemini: Mining the potential of multi-modality vision language models. Ar Xiv, abs/2403.18814, 2024c. URL https://api. semanticscholar.org/Corpus ID:268724012. Lin, T.-Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Dollár, P., and Zitnick, C. L. Microsoft coco: Common objects in context. In Computer Vision ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part V 13, pp. 740 755. Springer, 2014. Liu, H., Li, C., Li, Y., and Lee, Y. J. Improved baselines with visual instruction tuning. Ar Xiv, abs/2310.03744, 2023a. URL https://api.semanticscholar.org/Corpus ID: 263672058. Liu, H., Li, C., Li, Y., and Lee, Y. J. Improved baselines with visual instruction tuning. ar Xiv preprint ar Xiv:2310.03744, 2023b. Liu, H., Li, C., Wu, Q., and Lee, Y. J. Visual instruction tuning. Ar Xiv, abs/2304.08485, 2023c. URL https: //api.semanticscholar.org/Corpus ID:258179774. Liu, H., Li, C., Wu, Q., and Lee, Y. J. Visual instruction tuning. Advances in neural information processing systems, 36, 2024a. Liu, Y., Duan, H., Zhang, Y., Li, B., Zhang, S., Zhao, W., Yuan, Y., Wang, J., He, C., Liu, Z., Chen, K., and Lin, D. Mmbench: Is your multi-modal model an all-around player? Ar Xiv, abs/2307.06281, 2023d. URL https: //api.semanticscholar.org/Corpus ID:259837088. Liu, Z., Zang, Y., Dong, X., Zhang, P., Cao, Y., Duan, H., He, C., Xiong, Y., Lin, D., and Wang, J. Miadpo: Multi-image augmented direct preference optimization for large vision-language models. ar Xiv preprint ar Xiv:2410.17637, 2024b. Lu, P., Mishra, S., Xia, T., Qiu, L., Chang, K.-W., Zhu, S.-C., Tafjord, O., Clark, P., and Kalyan, A. Learn to explain: Multimodal reasoning via thought chains for science question answering. Ar Xiv, abs/2209.09513, 2022. URL https://api.semanticscholar.org/Corpus ID: 252383606. Probing Visual Language Priors in VLMs Lu, P., Bansal, H., Xia, T., Liu, J., yue Li, C., Hajishirzi, H., Cheng, H., Chang, K.-W., Galley, M., and Gao, J. Mathvista: Evaluating math reasoning in visual contexts with gpt-4v, bard, and other large multimodal models. Ar Xiv, abs/2310.02255, 2023. URL https: //api.semanticscholar.org/Corpus ID:267069069. Ma, Z., Hong, J., Gul, M. O., Gandhi, M., Gao, I., and Krishna, R. @ crepe: Can vision-language foundation models reason compositionally? 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 10910 10921, 2022. URL https: //api.semanticscholar.org/Corpus ID:254685851. Majumdar, A., Ajay, A., Zhang, X., Putta, P., Yenamandra, S., Henaff, M., Silwal, S., Mcvay, P., Maksymets, O., Arnaud, S., Yadav, K., Li, Q., Newman, B., Sharma, M., Berges, V.-P., Zhang, S., Agrawal, P., Bisk, Y., Batra, D., Kalakrishnan, M., Meier, F., Paxton, C., Sax, A., and Rajeswaran, A. Openeqa: Embodied question answering in the era of foundation models. 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 16488 16498, 2024. URL https: //api.semanticscholar.org/Corpus ID:268066655. Masry, A., Long, D. X., Tan, J. Q., Joty, S. R., and Hoque, E. Chartqa: A benchmark for question answering about charts with visual and logical reasoning. Ar Xiv, abs/2203.10244, 2022. URL https://api. semanticscholar.org/Corpus ID:247593713. Mathew, M., Karatzas, D., Manmatha, R., and Jawahar, C. V. Docvqa: A dataset for vqa on document images. 2021 IEEE Winter Conference on Applications of Computer Vision (WACV), pp. 2199 2208, 2020. URL https:// api.semanticscholar.org/Corpus ID:220280200. Mathew, M., Bagal, V., Tito, R. P., Karatzas, D., Valveny, E., and Jawahar, C. Infographicvqa. 2022 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), pp. 2582 2591, 2021. URL https://api. semanticscholar.org/Corpus ID:233394125. Open AI. Gpt-4 technical report. 2023a. URL https: //api.semanticscholar.org/Corpus ID:257532815. Open AI. Gpt-4 technical report. ar Xiv, 2023b. Open AI. Gpt-4o: A multimodal, multilingual generative pre-trained transformer. https://openai.com/index/ hello-gpt-4o/, 2024. Accessed: 2024-06-03. Podell, D., English, Z., Lacey, K., Blattmann, A., Dockhorn, T., Müller, J., Penna, J., and Rombach, R. Sdxl: Improving latent diffusion models for high-resolution image synthesis. ar Xiv preprint ar Xiv:2307.01952, 2023. Prabhu, V., Yenamandra, S., Chattopadhyay, P., and Hoffman, J. Lance: Stress-testing visual models by generating language-guided counterfactual images. Advances in Neural Information Processing Systems, 36:25165 25184, 2023. Radford, A., Kim, J. W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., Sastry, G., Askell, A., Mishkin, P., Clark, J., Krueger, G., and Sutskever, I. Learning transferable visual models from natural language supervision. In International Conference on Machine Learning, 2021. URL https://api.semanticscholar.org/Corpus ID: 231591445. Rafailov, R., Sharma, A., Mitchell, E., Manning, C. D., Ermon, S., and Finn, C. Direct preference optimization: Your language model is secretly a reward model. Advances in Neural Information Processing Systems, 36, 2024. Ramakrishnan, S., Agrawal, A., and Lee, S. Overcoming language priors in visual question answering with adversarial regularization. In Neural Information Processing Systems, 2018. URL https://api.semanticscholar. org/Corpus ID:52946763. Ramesh, A., Pavlov, M., Goh, G., Gray, S., Voss, C., Radford, A., Chen, M., and Sutskever, I. Zero-shot text-toimage generation. In International conference on machine learning, pp. 8821 8831. Pmlr, 2021. Ren, T., Liu, S., Zeng, A., Lin, J., Li, K., Cao, H., Chen, J., Huang, X., Chen, Y., Yan, F., Zeng, Z., Zhang, H., Li, F., Yang, J., Li, H., Jiang, Q., and Zhang, L. Grounded sam: Assembling open-world models for diverse visual tasks, 2024. Rohrbach, A., Hendricks, L. A., Burns, K., Darrell, T., and Saenko, K. Object hallucination in image captioning. In Conference on Empirical Methods in Natural Language Processing, 2018. URL https://api. semanticscholar.org/Corpus ID:52176506. Rombach, R., Blattmann, A., Lorenz, D., Esser, P., and Ommer, B. High-resolution image synthesis with latent diffusion models. In CVPR, 2022. Shen, S., Li, L. H., Tan, H., Bansal, M., Rohrbach, A., Chang, K.-W., Yao, Z., and Keutzer, K. How much can clip benefit vision-and-language tasks? Ar Xiv, abs/2107.06383, 2021. URL https://api. semanticscholar.org/Corpus ID:235829401. Singh, A., Natarajan, V., Shah, M., Jiang, Y., Chen, X., Batra, D., Parikh, D., and Rohrbach, M. Towards vqa models that can read. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 8309 Probing Visual Language Priors in VLMs 8318, 2019a. URL https://api.semanticscholar. org/Corpus ID:85553602. Singh, A., Natarajan, V., Shah, M., Jiang, Y., Chen, X., Batra, D., Parikh, D., and Rohrbach, M. Towards vqa models that can read. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 8317 8326, 2019b. Singh, A., Hu, R., Goswami, V., Couairon, G., Galuba, W., Rohrbach, M., and Kiela, D. Flava: A foundational language and vision alignment model. 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 15617 15629, 2021. URL https://api. semanticscholar.org/Corpus ID:244954250. Soldaini, L., Kinney, R., Bhagia, A., Schwenk, D., Atkinson, D., Authur, R., Bogin, B., Chandu, K., Dumas, J., Elazar, Y., Hofmann, V., Jha, A. H., Kumar, S., Lucy, L., Lyu, X., Lambert, N., Magnusson, I., Morrison, J., Muennighoff, N., Naik, A., Nam, C., Peters, M. E., Ravichander, A., Richardson, K., Shen, Z., Strubell, E., Subramani, N., Tafjord, O., Walsh, P., Zettlemoyer, L., Smith, N. A., Hajishirzi, H., Beltagy, I., Groeneveld, D., Dodge, J., and Lo, K. Dolma: An Open Corpus of Three Trillion Tokens for Language Model Pretraining Research. ar Xiv preprint, 2024. URL https://arxiv.org/abs/2402.00159. Sterz, H., Pfeiffer, J., and Vuli c, I. Dare: Diverse visual question answering with robustness evaluation. 2024. URL https://api.semanticscholar.org/Corpus ID: 272910642. Tan, H. H. and Bansal, M. Lxmert: Learning cross-modality encoder representations from transformers. In Conference on Empirical Methods in Natural Language Processing, 2019. URL https://api.semanticscholar.org/ Corpus ID:201103729. Team, G., Anil, R., Borgeaud, S., Wu, Y., Alayrac, J.-B., Yu, J., Soricut, R., Schalkwyk, J., Dai, A. M., Hauth, A., et al. Gemini: a family of highly capable multimodal models. ar Xiv preprint ar Xiv:2312.11805, 2023. Thrush, T., Jiang, R., Bartolo, M., Singh, A., Williams, A., Kiela, D., and Ross, C. Winoground: Probing vision and language models for visio-linguistic compositionality. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 5238 5248, 2022. Tong, S., Brown, E., Wu, P., Woo, S., Middepogu, M., Akula, S. C., Yang, J., Yang, S., Iyer, A., Pan, X., et al. Cambrian-1: A fully open, vision-centric exploration of multimodal llms. ar Xiv preprint ar Xiv:2406.16860, 2024. Touvron, H., Lavril, T., Izacard, G., Martinet, X., Lachaux, M.-A., Lacroix, T., Rozière, B., Goyal, N., Hambro, E., Azhar, F., Rodriguez, A., Joulin, A., Grave, E., and Lample, G. Llama: Open and efficient foundation language models. Ar Xiv, abs/2302.13971, 2023a. URL https: //api.semanticscholar.org/Corpus ID:257219404. Touvron, H., Martin, L., Stone, K. R., Albert, P., Almahairi, A., Babaei, Y., Bashlykov, N., Batra, S., Bhargava, P., Bhosale, S., Bikel, D. M., Blecher, L., Ferrer, C. C., Chen, M., Cucurull, G., Esiobu, D., Fernandes, J., Fu, J., Fu, W., Fuller, B., Gao, C., Goswami, V., Goyal, N., Hartshorn, A. S., Hosseini, S., Hou, R., Inan, H., Kardas, M., Kerkez, V., Khabsa, M., Kloumann, I. M., Korenev, A. V., Koura, P. S., Lachaux, M.-A., Lavril, T., Lee, J., Liskovich, D., Lu, Y., Mao, Y., Martinet, X., Mihaylov, T., Mishra, P., Molybog, I., Nie, Y., Poulton, A., Reizenstein, J., Rungta, R., Saladi, K., Schelten, A., Silva, R., Smith, E. M., Subramanian, R., Tan, X., Tang, B., Taylor, R., Williams, A., Kuan, J. X., Xu, P., Yan, Z., Zarov, I., Zhang, Y., Fan, A., Kambadur, M., Narang, S., Rodriguez, A., Stojnic, R., Edunov, S., and Scialom, T. Llama 2: Open foundation and fine-tuned chat models. Ar Xiv, abs/2307.09288, 2023b. URL https: //api.semanticscholar.org/Corpus ID:259950998. Wang, F., Zhou, W., Huang, J. Y., Xu, N., Zhang, S., Poon, H., and Chen, M. mdpo: Conditional preference optimization for multimodal large language models. ar Xiv preprint ar Xiv:2406.11839, 2024a. Wang, W., Bao, H., Dong, L., and Wei, F. Vlmo: Unified vision-language pre-training with mixture-of-modalityexperts. Ar Xiv, abs/2111.02358, 2021a. URL https: //api.semanticscholar.org/Corpus ID:241035439. Wang, X., Chen, J., Wang, Z., Zhou, Y., Zhou, Y., Yao, H., Zhou, T., Goldstein, T., Bhatia, P., Huang, F., and Xiao, C. Enhancing visual-language modality alignment in large vision language models via self-improvement. Ar Xiv, abs/2405.15973, 2024b. URL https://api. semanticscholar.org/Corpus ID:270062295. Wang, X., Chen, J., Wang, Z., Zhou, Y., Zhou, Y., Yao, H., Zhou, T., Goldstein, T., Bhatia, P., Huang, F., et al. Enhancing visual-language modality alignment in large vision language models via self-improvement. ar Xiv preprint ar Xiv:2405.15973, 2024c. Wang, Z., Yu, J., Yu, A. W., Dai, Z., Tsvetkov, Y., and Cao, Y. Simvlm: Simple visual language model pretraining with weak supervision. Ar Xiv, abs/2108.10904, 2021b. URL https://api.semanticscholar.org/Corpus ID: 237291550. Wu, Y., Zhao, Y., Zhao, S., Zhang, Y., Yuan, X., Zhao, G., and Jiang, N. Overcoming language priors in visual question answering via distinguishing superficially similar instances. In International Conference on Probing Visual Language Priors in VLMs Computational Linguistics, 2022. URL https://api. semanticscholar.org/Corpus ID:252367981. Xiao, W., Huang, Z., Gan, L., He, W., Li, H., Yu, Z., Jiang, H., Wu, F., and Zhu, L. Detecting and mitigating hallucination in large vision language models via fine-grained ai feedback. ar Xiv preprint ar Xiv:2404.14233, 2024. Xie, Y., Li, G., Xu, X., and Kan, M.-Y. V-dpo: Mitigating hallucination in large vision language models via visionguided direct preference optimization. ar Xiv preprint ar Xiv:2411.02712, 2024. Yu, T., Yao, Y., Zhang, H., He, T., Han, Y., Cui, G., Hu, J., Liu, Z., Zheng, H.-T., Sun, M., and Chua, T.-S. Rlhf-v: Towards trustworthy mllms via behavior alignment from fine-grained correctional human feedback. 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 13807 13816, 2023a. URL https://api.semanticscholar.org/Corpus ID: 265608723. Yu, T., Zhang, H., Yao, Y., Dang, Y., Chen, D., Lu, X., Cui, G., He, T., Liu, Z., Chua, T.-S., et al. Rlaif-v: Aligning mllms through open-source ai feedback for super gpt-4v trustworthiness. ar Xiv preprint ar Xiv:2405.17220, 2024. Yu, W., Yang, Z., Li, L., Wang, J., Lin, K., Liu, Z., Wang, X., and Wang, L. Mm-vet: Evaluating large multimodal models for integrated capabilities. Ar Xiv, abs/2308.02490, 2023b. URL https://api.semanticscholar.org/ Corpus ID:260611572. Yu, W., Yang, Z., Li, L., Wang, J., Lin, K., Liu, Z., Wang, X., and Wang, L. Mm-vet: Evaluating large multimodal models for integrated capabilities. ar Xiv preprint ar Xiv:2308.02490, 2023c. Yuan, W., Pang, R. Y., Cho, K., Sukhbaatar, S., Xu, J., and Weston, J. Self-rewarding language models. ar Xiv preprint ar Xiv:2401.10020, 2024. Yue, Z., Zhang, L., and Jin, Q. Less is more: Mitigating multimodal hallucination from an eos decision perspective. ar Xiv preprint ar Xiv:2402.14545, 2024. Zhang, P., Goyal, Y., Summers-Stay, D., Batra, D., and Parikh, D. Yin and yang: Balancing and answering binary visual questions. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 5014 5022, 2016. Zhao, Z., Wang, B., Ouyang, L., Dong, X., Wang, J., and He, C. Beyond hallucinations: Enhancing lvlms through hallucination-aware direct preference optimization. ar Xiv preprint ar Xiv:2311.16839, 2023a. Zhao, Z., Wang, B., Ouyang, L., wen Dong, X., Wang, J., and He, C. Beyond hallucinations: Enhancing lvlms through hallucination-aware direct preference optimization. Ar Xiv, abs/2311.16839, 2023b. URL https: //api.semanticscholar.org/Corpus ID:265466428. Zhou, Y., Cui, C., Rafailov, R., Finn, C., and Yao, H. Aligning modalities in vision large language models via preference fine-tuning. ar Xiv preprint ar Xiv:2402.11411, 2024a. Zhou, Y., Fan, Z., Cheng, D., Yang, S., Chen, Z., Cui, C., Wang, X., Li, Y., Zhang, L., and Yao, H. Calibrated self-rewarding vision language models. ar Xiv preprint ar Xiv:2405.14622, 2024b. Zhu, K., Zhao, L., Ge, Z., and Zhang, X. Self-supervised visual preference alignment. In Proceedings of the 32nd ACM International Conference on Multimedia, pp. 291 300, 2024a. Zhu, L., Ji, D., Chen, T., Xu, P., Ye, J., and Liu, J. Ibd: Alleviating hallucinations in large vision-language models via image-biased decoding. ar Xiv preprint ar Xiv:2402.18476, 2024b. Probing Visual Language Priors in VLMs Q:What is the color scheme of the train 𝐼! train The has a Q:What is the color scheme of the train blue and yellow color train The has a blue and yellow color Q:What is the color scheme of the train train The has a Q:What is the color scheme of the train blue and yellow color train The has a pink and red color Figure 6: Gradients difference between Image-DPO and Text-DPO. For Text-DPO (Rafailov et al., 2024), the model receives positive gradients (green arrows) for the preferred answer Aw and negative gradients Al (red arrows) for the dispreferred answer. In contrast, our proposed Image-DPO approach applies positive gradients when the preferred image Iw is input and negative gradients for the dispreferred image Il, both based on the same output answer. A. Image-DPO Mathematical Details In this section, we give the complete proof of Image DPO. Simlarly to DPO, we start from the RL fine-tuning objective and then derive its variant for image dpo. A.1. RL Fine-tuning for VLM SFT Given question Q, answer A and image I, we can train a SFT model πSF T with supervised learning on high-quality data. As the SFT model is a language generation model, it is still a model modeling the text outputs with question and image. πSF T (A|Q, I) Reward Modeling Phase In this stage, we construct a static dataset of comparisons S = {Ai, Qi, Ii w, Ii l }, and we present the QIA pairs (Q, Iw, A), (Q, Il, A) to human for preference. Following the idea of RL fine-tuning, the preference are assumed to be obtained from a a latent reward function r (Q, I, A) which are not tractable, and we we use BT model to represent the preference distribution p as: p ((Q, Iw, A) (Q, Il, A)) = exp (r (Q, Iw, A)) exp(r (Q, Iw, A) + exp(r (Q, Il, A)) (5) Now given the human labeled preference, we can try to optimize a reward model rϕ to estimate r by using maximum likelihood. Framing this as a binary classification, we can have this negative log-likelihood loss: LR(rϕ, S) = E(A,Q,Iw,Il) S [log σ(rϕ(Q, Iw, A) rϕ(Q, Il, A))] (6) Here σ is a logistic function. Basically, this reward function gives score jointly considering image, question and image quality. RL Fine-Tuning Phrase During the RL phase, the learned reward function is used to provide feeback to the VLM model. Following DPO Probing Visual Language Priors in VLMs paper (Rafailov et al., 2024), the optimization is formulated as : max πθ E(Q,I) S,A πθ(A|Q,I) [rϕ(Q, I, A)] βDKL [πθ(A|Q, I) πref(A|Q, I)] , (7) where β is a parameter controlling the deviation from the base reference policy πref, namely the initial SFT model πSFT. Due to the discrete nature of language generation, this object is also not differentiable and is typically optimized with reinforcement learning (Jaques et al., 2017; 2020). A.2. Image DPO and RL Fine-Tuning According to the DPO paper, a straightforward optimal solution to the KL-constrainted reward function maximization object in Eq. 6 is: πr(A|Q, I) = 1 Z(Q, I)πref(A|Q, I) exp( 1 β r(Q, I, A)) (8) where Z(Q, I) = P A πref(A|Q, I) exp( 1 β r(Q, I, A) is a partition function. Here r should be any reward function, which makes Z hard to tract. We provide the proof of this step in A.3. Taking the logarithm of both side, and with some algebra, we get r(Q, I, A) = β πr(A|Q, I) πref(A|Q, I) + β log Z(Q, I) (9) This parametrization could be applied to ground-truth reward r and the corresponding optimal model π . The BT model with the optimal policy is p ((Q, I1, A) (Q, I2, A)) = exp (r (Q, Iw, A)) exp (r (Q, Iw, A)) + exp (r (Q, Il, A)) (10) We plug Eq. 9 into the BT model, we have: p ((Q, I1, A) (Q, I2, A)) = exp β log π (A|Q,Iw) πref(A|Q,Iw) + β log Z(Q, Iw) β log π (A|Q,Iw) πref(A|Q,Iw) + β log Z(Q, Iw) + β log π (A|Q,Il) πref(A|Q,Il) + β log Z(Q, Il) 1 + exp(β log π (A|Il,Q) πref(A|Il,Q) β log π (A|Iw,Q) πref(A|Iw,Q) + β log Z(Il, Q) β log Z(Iw, Q)) = σ exp(β log π (A|Il, Q) πref(A|Il, Q) β log π (A|Iw, Q) πref(A|Iw, Q) + β log Z(Il, Q) β log Z(Iw, Q)) Now we have the probability of human preference data in terms of the optimal policy rather than the reward model, we can formulate a maximum likelihood objective for a policy πθ. Our policy objective is : L(πθ; πref) = E(Q,A,Iw,Il) S log σ β log πθ(A|Iw, Q) πref(A|Iw, Q) β log πθ(A|Il, Q) πref(A|Il, Q) + β log Z(Iw, Q) β log Z(Il, Q) As f(x) = log σ(x) is a convex function (σ is the sigmoid function), we can apply Jensen s inequality f( 1 2y) 1 2f(x) + 1 L(πθ; πref) E 1 2 log σ 2β log πθ(A|Iw, Q) πref(A|Iw, Q) 2β log πθ(A|Il, Q) πref(A|Il, Q) 2 log σ (2β log Z(Iw, Q) 2β log Z(Il, Q)) Probing Visual Language Priors in VLMs As log σ (Z(I, Q)) is not a function of πθ, the above objective is equivalent to the below Eq.13, where α = 2β. It is the same as our objective listed in Eq.4 of the main paper. L(πθ; πref) E(Q,Iw,Il,A) S log σ(α πθ(A | Q, Iw) πref(A | Q, Iw) α πθ(A | Q, Il) πref(A | Q, Il)) (13) In this sense, our optimization objective Eq.4 in main paper are optimizing the upper bound on the RL fine-tuning objective, i.e., Eq.7. A.3. Deriving the Optimum of the KL-Constrained Reward Maximization Objective In this appendix, we will derive Eq.8. Similarly to Eq.7, we optimize the following objective: max π E(Q,I) S,A π [r(Q, I, A)] βDKL [π(A|Q, I) πref(A|Q, I)] (14) under any reward function r(Q, I, A), reference model πref, and a general non-parametric policy class. We now have: max π E(Q,I) S,A π [r(Q, I, A)] βDKL [π(A|Q, I) πref(A|Q, I)] = max π E(Q,I) SEA π(A|Q,I) r(Q, I, A) β log π(A|Q, I) πref(A|Q, I) = min π E(Q,I) SEA π(A|Q,I) log π(A|Q, I) πref(A|Q, I) 1 β r(Q, I, A) = min π E(Q,I) SEA π(A|Q,I) log π(A|Q, I) 1 Z(Q,I)πref(A|Q, I) exp 1 β r(Q, I, A) log Z(Q, I) where we have the partition function: Z(Q, I) = X A πref(A|Q, I) exp 1 β r(Q, I, A) (16) Observe that the partition function depends solely on (Q, I) and the reference policy πref, and is independent of the policy π. We can now define the Equation 8. π (A|Q, I) = 1 Z(Q, I)πref(A|Q, I) exp 1 β r(Q, I, A) , (17) Probing Visual Language Priors in VLMs B. More details and comparisons of our benchmarks B.1. More data samples of Vi LP Four Six Eight Modern drones typically have four propellers. How many propellers does the drone in the picture have? Zebra Whale Seal Zebra has black stripes. Which animal with black stripes shown in the image? 1000 200 100 A kilometer equals to 1000 meters. How many meters are in a kilometer, as indicated in the image? Round Square Diamond Apple has round shape. What is the shape of the apple shown in the image? Two One Five Bike usually have two wheels. How many wheels does the bike in the center of the picture have? Kangaroo Koala Wombat The kangaroo is known for its excellent jumping. Which animal native to Australia is shown in the image demonstrating a jump? Person Cat Eraser People usually use mop to clean the floor. Who is using this mop to clean the floor in the picture? Soil Cloud Desert Orange trees grow on soil. Based on the picture, where is this orange tree growing? Fabric Wood Umbrellas are usually made of fabric. What material is used to make the umbrella in the image? Cow Whale Seal Cows could produce milk. Which animal does the image show produces milk for dairy products? Seattle Rome Paris The Space Needle is situated in Seattle. Based on the image, what city is the Space Needle in? Respond with one word or abbreviation, like NYC. Horse Rabbit Seal Horse is often used for riding. Which animal in the image is ridden by humans? Sunglasses Blender Wallet Sunglasses are worn to protect eyes from bright light. What accessory in the image shields the eyes from UV rays? Round Hexagon Octagon The moon is typically round. What shape is the moon in the image? Figure 7: Randomly sampled data from Vi LP. Probing Visual Language Priors in VLMs Lion Elephant Pig The lion is known as the king of the savannah. Can you identify the largest animal in the image wearing a crown? Map Pizza Pineapple A map shows oceans without water and lands without buildings. Which object in the picture have these characteristics? Ocean Toilet Sink Ships typically sail on the ocean. Based on the image, what is the ship sailing on? Worm Cat Book A common phrase is the early bird catches the worm . What does the early bird catch in the image? Seismograph Blender Umbrella A seismograph is used to measure earthquake intensity. Which instrument in the picture records ground vibrations? Gavel Drill Wrench A gavel is used by judges in court. From the image, which object symbolizes judicial authority? Parachute Guitar Umbrella A parachute slows descent through the air. Which equipment in the image is used for safe skydiving? Butterfly Moth Bee A caterpillar becomes a butterfly. What does the caterpillar become in the image? Anchor Spoon Toothbrush An anchor keeps a ship steady in the water. From the image, which item prevents boats from drifting? Triangle Square Round A pyramid consists of triangular faces. What shape are on the faces of the pyramid shown in the image? Key Crayon Carrot Pianos usually have black and white keys. What object in this picture is covering the keys on the piano? One Two Three There is only one moon orbiting the Earth. How many moons are orbiting the planet in this picture? Platypus Kangaroo Bear The platypus is a mammal that lays eggs. Which mammal in the image is shown an egg-layer? Leonardo Vincent Claude Leonardo da Vinci painted the Mona Lisa. Who is painting the Mona Lisa in the image? Please answer his/her first name. Figure 8: Randomly sampled data from Vi LP. Probing Visual Language Priors in VLMs B.2. More details in Vi LPbenchmark data generation Our proposed dataset introduces Question-Image-Answer (QIA) triplets designed to challenge state-of-the-art Vision Language Models (VLMs) against visual language priors. The construction process combines human-guided and automated efforts to ensure quality and alignment. Question-Answer (QA) Generation: Most QA pairs are authored by humans following the design principles in Section 3.1. Additionally, candidate QA pairs are generated using models like Open AI-O1 and Claude-3.5-Sonnet with carefully crafted prompts. One example prompts shown in Figure 10. These candidates undergo human review, where they are refined or removed to meet our quality standards. Image Generation: For each QA pair, we use GPT-4 to generate multiple descriptive image prompts (see Figure 9). These prompts are provided to image generation models, such as FLUX and DALL-E 3, to produce candidate images. Human reviewers then select the most suitable image or request re-generation as needed to maintain consistency with the QA context. Human Review and Testing: At every stage, human reviewers rigorously evaluate the generated outputs to ensure quality, clarity, and challenge level. In addition to filtering out low-quality or insufficiently challenging triplets, we dynamically test the QIAs to confirm that they remain intuitive for humans while being difficult for VLMs. Cost: The complexity of our data creation process leads to a significant average cost of approximately $2.50 per QIA triplet in Vi LP, excluding human labor costs. Task: Using the provided question and possible image-based answers, generate detailed text prompts for image generation. Each image prompt should reflect the question s context and incorporate one of the image-based answers. Question: Question Image-based Answer:[Answer1, Answer2, Answer3] For each possible image-based answer, create an image prompt that describes what the image might look like based on the question. Please be creativity. For example, if the question asks who is using this mop to clean the floor in the picture? and the answer is eraser. The image prompt should really describe the image of an eraser uses a mop to clean the floor. Format the output strictly as a JSON list, like this example: [ "prompt1": "Image Generation Prompt text here", "prompt2": "Image Generation Prompt text here", "prompt3": "Image Generation Prompt text here", ] Figure 9: The prompt we used for generating text-prompt for image generation. In the below, I try to propose questions along with three answers where the first answer is corresponding to the question text directly, while the other two are usual and counterintuitive, which could lead to wrongs of VLMs. Please help me generate more Question-3 answer pairs, which are different from what I have provided. - All the potential answers should a single world. - Help me generate a format where I can direct copy paste into Goole Sheet. Also, please a ; between question and each answers. - Please be very creative and different from my provided examples - the answer 2 & 3 should be very diverse and different compared to answer 1. - Every question contains a statement at the beginning which consists of the answer1 as part of it. - Please understand the principles and generate the QA very different from my provided examples Some Examples: - A screwdriver is used for tightening screws. From the image, which tool is used to turn screws? Screwdriver Hammer Scissors - A pen is a tool used for writing. Which object in the image is used to write on paper? Pen Hammer Shoe - Clocks are used to measure time. Can you identify the item in the image that is used to measure time? Clock Spoon Candle - A violin has four strings and is played using a bow. According to the image, which musical instrument is being played with a bow? Violin Guitar Saxophone - Camels have humps. Which animal in the image stores fat in its humps? Camel Horse Tiger - Honey is made by bees. Which insect in the image produces honey? Bee Ant Dragonfly - An anvil is a tool used by blacksmiths. What object in the image is used by blacksmiths to forge metal? Anvil Fork Wrench - A gavel is used by judges in court. From the image, which object symbolizes judicial authority? Gavel Hammer Wrench - A syringe is used to inject medicine. From the image, which tool is used for administering injections? Syringe Scissor Drill - An anchor keeps a ship steady in the water. From the image, which item prevents boats from drifting? Anchor Spoon Toothbrush - A chainsaw is a power tool for cutting wood. What device shown is typically used by lumberjacks to fell trees? Chainsaw Blender Stapler Figure 10: One prompt we used for potential QAs designs of Vi LP B.3. Ablation studies: realistic images Our benchmark data are currently generated by DALL E3 (Ramesh et al., 2021) and Flux, both of which produce cartoon-like, synthetic images rather than photorealistic ones. To assess the impact of image style, we regenerated a subset of 45 QIA pairs using GPT-4o s latest image generation model to enhance realism, as illustrated in Figure 11. We then measured changes in model correctness Probing Visual Language Priors in VLMs Figure 11: Realistic image comparison. Each image pair shows our original benchmark data on the left and a corresponding realistic example generated by GPT-4o on the right. when these more realistic images were used, with negative values indicating performance degradation. For definitions of these metrics, please refer to the beginning of Section 5. The results in Table 6 show that increased realism slightly reduces performance for the Score metric in most cases, while its effects on Prior are generally negligible. These ablation findings suggest that introducing more realistic images may increase the task s difficulty, highlighting an important direction for future research. V i LP F P rior V i LP F Score V i LP P P rior V i LP P Score GPT-4o 0 -2.2% 0 0 GPT-4o-mini 0 -1.1% 0 -2.2% Claude-Sonnet-3.5 -2.2% -4.4% 0 -3.3% Claude-Opus-3 2.2% -1.1% -2.2% -1.1% Table 6: Impacts of Realistic Styles. Each value represents the change in correctness when replacing the original images with realistic ones (Realistic - Original). Negative values indicate a drop in performance, suggesting increased task difficulty. The metric definitions are provided in the beginning of Section 5. B.4. Comparisons to other datasets In this section, we compare our benchmark to other benchmarks, including Winoground (Thrush et al., 2022), Whoops!(Bitton-Guetta et al., 2023a), and Hallusion- Bench(Guan et al., 2023). While these datasets are impactful, their evaluation perspectives differ from ours, covering a range from high-level design principles to low-level formats. B.4.1. COMPARE TO WINOGROUND (THRUSH ET AL., Winoground centers on vision-linguistic compositional reasoning by presenting models with two images and two captions that contain the same words arranged differently. The goal is to match each image to its correct caption based on the text s compositional structures and the visual content (as detailed in the Introduction and Sec. 3.1 of (Thrush et al., 2022)). However, Winoground s captions do not challenge language priors or introduce out-of-distribution visual information. Both captions adhere to common linguistic expectations, and there is no explicit misleading information provided to test resistance to language biases. Additionally, most of their images are typical internet images, featuring common visual patterns. Qualitative Comparison: As shown in Figure 12 consisting of Winoground examples, both captions and images are normal and satisfy common linguistic expectations and common sense. The evaluation focuses on whether the model can discern the compositional differences between the two images and two captions, then match them correctly. Comparing Figure 12 and Figure 1, 7, 8, You can discern the Probing Visual Language Priors in VLMs Figure 12: Winoground Data Example. Our benchmark is different from Winoground, as Winoground focuses on vision-linguistic compositional reasoning. Both captions and images are normal and satisfy common linguistic expectations and common sense. significant differences among the tested images. Quantitative Comparison: Both Vi LP; and Winoground benchmarks include paired textual information in their setups. In our benchmark, Vi LP; Prior QAs and Vi LP; Score QAs share the same question but differ in their answers. In Winoground, each example has two captions, and the task is to match each caption to its correct image. Setting. To demonstrate the differences, we use GPT to evaluate the commonness of these paired textual components. Specifically, GPT-4o rates the oddity of scenarios described in texts on a scale from 1 (very rare) to 10 (very common). The resulting scores are then compared. Results. In our benchmark, Prior QAs scored 9.37, indicating that these answers are designed to align with language priors and are highly common. Score QAs scored 1.65, showing that these QA pairs are rare, making them difficult to infer without the corresponding visual information. Notably, Prior and Score QAs share the same question but differ in their answers, and this significant contrast in scores showcases how we inject strong language priors to test a model s vulnerability to linguistic distractions. By comparison, Winoground s two captions scored 8.05 and 8.08, indicating two primary observations: (1) both captions align well with language priors, which means Winoground does not challenge language priors or evaluate out-of-distribution scenarios; (2) the minimal score difference between the two captions confirms there is no significant variance in language priors, as examining how VLM models react to different language priors is beyond the scope of Winoground. In contrast, that aspect is precisely our focus. B.4.2. COMPARE TO WHOOPS! (BITTON-GUETTA ET AL., 2023A) Whoops! is designed to evaluate a model s ability to detect weirdness in images, emphasizing tasks where images depict unusual or nonsensical scenarios. It heavily relies on common sense reasoning, requiring models to recognize visual elements and then identify subtle inconsistencies among them. For example, for the lit candle inside a tightly sealed glass jar on the homepage, models must realize that a candle needs a constant oxygen supply to burn, which would not exist in a sealed jar", making a burning candle inside a sealed jar unlikely. This benchmark thus focuses on common-sense reasoning rather than challenging visual language priors. Qualitative Comparison Although Whoops! also includes creative, out-ofdistribution images, it does not focus on using language Probing Visual Language Priors in VLMs Figure 13: Example of Whoops! Dataset.Whoops! also has creative images. While unlike ours, its questions are common questions without strong language priors. priors to test a model s susceptibility to linguistic distraction, as our benchmark does. In its QA mode (comparable to our task), the questions are straightforward and lack strong language priors. Some examples can be found in Figure 13. Additionally, Whoops! uses open-ended questions, offering greater freedom in answers while introducing potential ambiguity or divergence in responses. Quantitative Comparison Setting. Unlike Winoground and our benchmark, Whoops! does not provide control groups or textual components for comparison. To measure language priors, we analyze the suggestiveness of questions by evaluating GPT-4o s certainty when answering them without any visual context. A more suggestive question typically yields more determined and confident responses, whereas a less suggestive question produces more varied answers. We calculate how many unique answers GPT-4o provides over five attempts at temperature 1.0 to promote randomness. Semantic differences are normalized to exclude synonyms. Results. We find that Whoops! questions produce an average of 2.58 unique answers (out of five attempts) with a standard deviation of 1.48. For our benchmark, without facts, GPT4o provides an average of 1.53 unique answers (std 0.94), and with facts, 1.10 unique answers (std 0.42). Although both benchmarks use creative images, these results indicate that Whoops! questions remain more general and do not push GPT-4o toward stereotypical responses. In contrast, our benchmark deliberately uses suggestive ques- Probing Visual Language Priors in VLMs Figure 14: Example of Hallusion Bench. Hallusion Bench also has questions which can be answered without images. While it is based on facts instead of stereotypes like ours. Moreover, its images are limited to Chart, Map, OCR and Table. tions to elicit stereotype-consistent answers, reflecting our emphasis on testing language priors. B.4.3. COMPARE TO HALLUSIONBENCH (GUAN ET AL., Hallusion Bench has two components: Visual Dependent, which focuses on testing models general visual reasoning skills, and Visual Supplement, which examines a model s visual reasoning in tandem with its parametric memory. The Visual Supplement part is related to our benchmark, as its questions, like ours, can be answered without visual information. However, the key difference lies in their design. Hallusion Bench questions rely on parametric memory and strict factual knowledge (e.g., Which country has the most gold medals in the 2024 Olympics? ), whereas our benchmark questions are based on common stereotypes (e.g., A soccer ball is round. ). This distinction significantly constrains the diversity of Hallusion Bench s questions its official release contains only 50 question pairs while our approach allows for a broader range of scenarios, including texture, shape, conceptual combinations, hallucinated elements, and proverb-based contexts. Moreover, Hallusion Bench focuses on a narrow set of image types (charts, tables, maps, and OCR). It tests reasoning through subtle modifications (like changing one digit in a chart), which is fundamentally different from our approach of creating diverse visual contexts. Figure 14 showcases typical Hallusion Bench examples, illustrating the limited scope of both images and QAs, whereas our benchmark includes a variety of out-of-distribution images and more suggestive QAs (see Figure 1, 7, 8). B.5. Human Study For human evaluation, we hired Ph.D.-level candidates to participate in testing. They were asked to answer questions with a single image provided each time, and the QIAs were randomly shuffled to avoid any sequential context. To ensure efficiency, we conducted an oral test instead of a written one, recording their responses. After the test, we updated the synonym sets for the QIAs based on their answers. B.6. More failure cases analysis This subsection provides several typical failure cases observed in the inferred results of VLMs on Vi LP. We identified several consistent failure patterns, illustrated in Figure 15. For deeper analysis, we also prompted the VLMs to output their reasoning before finalizing answers, allowing us to better understand how these errors arise. Below, we Probing Visual Language Priors in VLMs Question: The moon is typically depicted as round. What shape is the moon in this night sky image? Reasoning: clear white edges surrounding the moon suggest augmented artistic aspects, possibly to emphasize the brightness or create a specific visual effect the main body of the moon retains its typical roundness, but the overall image is more dramatic and highlighted than a standard photograph of the moon Single word Answer: Round Correct Single word Answer: Octagon Question: Microscopes are used to view tiny organisms. In the image, what piece of equipment is viewing tiny organisms? Reasoning: based on the description and visual cues in the image, the main focus is on the object resembling a microscope, which is typically used to view tiny organisms this particular microscope is anthropomorphized with \"eyes\" and \"hands,\" creatively designed to look like a character but retaining the essential parts of a microscope (lenses and a stage) the context indicates that it is equipped to observe or interact with tiny specimens Single word Answer: Microscope Correct Single word Answer: Stapler Question: The Eiffel Tower is situated in Paris. From the image, can you identify the city where the Eiffel Tower is located?Please respond with a single word or abbreviation, like NYC for New York City.", Reasoning: first, examine the distinct landmarks in the image the eiffel tower is depicted, which is a globally recognized symbol of paris, france there are also architectural elements that resemble those found in beijing's tiananmen square however, since the question specifically asks about the eiffel tower, and it is exclusively located in paris, this can lead us to the conclusion Single word Answer: Paris Correct Single word Answer: Beijing Question: An umbrella could have a saxophone as its handle. Which object in this image represents the handle of the umbrella? Reasoning: observing the image, there is an object that uniquely resembles a musical instrument and is connected to the main structure of the umbrella, fulfilling the typical location and function of an umbrella handle Single word Answer: Saxophone Correct Single word Answer: Banana Question: Violin usually has 4 strings. How many strings does the violin have, as shown in the image? Reasoning: first, let us examine the image the object in question is a violin, and we need to count its strings typically, a violin has 4 strings. upon looking closely at the image provided, the violin depicted in the picture also has 4 strings Single word Answer: Four Correct Single word Answer: Three Question: A common phrase is the early bird catches the worm . What does the early bird catch in the image? Reasoning: the early bird catches a worm in the image Single word Answer: Worm Correct Single word Answer: Cat Figure 15: Failure cases analysis. We list six typical failure cases generated by VLMs on Vi LP. summarize the failure patterns in the same order as they appear in the figure, from left to right and top to bottom. - Shape recognition can fail in VLMs, causing them to revert to priors rather than accurately interpreting visual input. - Models sometimes struggle to count accurately. Instead of performing an actual count, they default to relying on learned priors to estimate quantities. - Models may refuse to accept visual information that contradicts their learned priors, whereas humans can comprehend hypothetical scenarios. For instance, the model recognizes the city as Beijing but rejects the correct answer because it expects the Eiffel Tower to be in Paris. - Sometimes the model overly relies on memorized proverbs, resulting in predictions that align with these proverbs rather than the actual content of the input image. - For images with creative concepts, the model may overly rely on its learned priors. As illustrated, a common prior is that microscopes are used to view organisms, leading the model to answer microscope rather than identifying the creatively depicted stapler. - For images with blended features, the model may rely mostly on text input while overlooking the visual cues. As illustrated, the VLM heavily depends on textual input leading to saxophone as the answer. Probing Visual Language Priors in VLMs C. Details in Image-DPO data generation and training Our image-DPO data generation pipeline consists of two stages. In the first stage, we leverage the VLM we aim to enhance to perform self-guided data generation with the aid of pre-trained image generative models. This stage produces a large number of new question-image-answer (QIA) triplets. In the second stage, we apply three types of image corruptions Gaussian blurring, pixelation, and semantic editing to generate good-bad QIA pairs, denoted as Iw (good) and Il (bad). Details of the hyperparameters used in the experiments are provided at the end of this section. C.1. VLM self-guided data generation VLM Seed Image Pre-trained Model Generated Image Tool : Stable Diffusion, Prompt : Generate a new image inspired by the green truck driving on the dirt road. The image should feature a futuristic cityscape with the truck driving through the streets. The truck should be the central focus. Instructions Justify: The response provides relevant information , addressing key aspects of the monster truck and its purpose. It demonstrates an accurate understanding of the image's details .... Score: 4 Rating of QA Q: What is the purpose of the green truck in the image? A: The green truck is a monster truck, , the truck's green color and its presence on a city street suggest that it could be participating in a promotional event or a show for the public. Generated QA Grounding DINO Instruct Pix2Pix Stable-Diffusion XL Figure 16: Overview of our data generation pipeline. We begin with an image only, from which instructions are derived using a VLM. These instructions guide the creation of a new image or the modification of the existing one. The generated image is then processed by the VLM to generate QA pairs. Both the QA pair and the image are subsequently input back into the VLM to assess the quality of the answers. No human-written in-context examples are used throughout this process. As illustrated in Figure 16, our data generation process begins by utilizing VLMs to suggest modifications or draw inspiration for input images without relying on any in-context examples. The used text prompt is shown in Figure 18. Subsequently, pre-trained models such as Stable Diffusion XL (Podell et al., 2023), Instruct-Pix2Pix (Brooks et al., 2023b), and Grounded-SAM (Ren et al., 2024) are employed to either modify existing images or generate entirely new ones. The altered or newly created images, along with the instructions that guided their generation, are then used by the same VLMs to produce corresponding question-answer pairs (QAs) based on the text prompt shown in Figure 19. An example of this process is provided in Figure 17. Importantly, all instructions, tool selections, and QA generation are autonomously handled by the same VLM we aimed to improve. In particular, Grounded-SAM requires the VLM to specify the object to be modified before generating images. To facilitate this, we use an additional text prompt (Figure 20) after the VLM generates the initial instructions (the pink region of Figure 17). To provide a better understanding of our generated QIAs, we randomly sampled and listed some examples of the generated QIA data, as shown in Figures 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, and 35. Probing Visual Language Priors in VLMs Tool : Stable Diffusion, Prompt : Generate an image of a snowy night in a city, with the train as the main subject. Include a few people walking around, some benches, and a few trees. Make the scene more artistic and visually engaging by adding a hint of surrealism. Instructions Tool : Instruct-Pix2Pix, Prompt : Transform the image to show a snowy landscape with a few people walking around and benches nearby. Adjust the colors to make the scene more vivid. Instructions Tool : Grounded-SAM, Prompt : Replace the people near the train with a robot or a humanoid figure. Instructions Q: How does the image convey the atmosphere of a winter night in the city? A: The image conveys the atmosphere of a winter night in the city through the combination of the yellow train on the tracks, the snow-covered ground, and the presence of benches Generated QA Q: How did the modification affect the color palette and overall visual appeal of the image? A: The modification enhanced the color palette and visual appeal of the image by adjusting the colors to create a more vivid and engaging scene Generated QA Q: What elements were added or removed in the transition from the original image to the new image, and why might these changes have been made? A: In the transition , people were removed, and a humanoid figure or robot were added. The removal of people and the addition of humanoid figures suggest that the creator wanted to shift the focus of the image from the human element to the technological or mechanical aspect. Generated QA Figure 17: Randomly sampled generation examples. Our proposed data generation pipeline utilizes pretrained model to generate diverse new images from existing seed image datasets. Given this image, please suggest a range of creative edits, tasks, or transformations that could be applied using advanced image processing tools. These tasks may include artistic transformations, vivid color adjustments, object detection and modification, or completely creating a new image inspired by the original. Specify which tool would be best suited for each task, choosing from Stable Diffusion for image generation, Instruct Pix2Pix for image modification, or Grounding DINO for object modification. Your recommendations should help in understanding the potential of the image and exploring creative possibilities. Expected Response Format: Item Number: 1 Tool Used: [Specify the tool - Stable Diffusion or Instruct Pix2Pix or Grounding DINO] Text Prompt for Processing: [Detailed description of the task or transformation to be performed. For image generation, please provide complete description based on the understanding of the provided images, since we only feed text prompt for this task.] Item Number: 2 Tool Used: [Specify the tool - Stable Diffusion or Instruct Pix2Pix or Grounding DINO] Text Prompt for Processing: [Detailed description of the task or transformation to be performed. For image generation, please provide complete description based on the understanding of the provided images, since we only feed text prompt for this task.] Item Number: 3 Tool Used: [Specify the tool - Stable Diffusion or Instruct Pix2Pix or Grounding DINO] Text Prompt for Processing: [Detailed description of the task or transformation to be performed. For image generation, please provide complete description based on the understanding of the provided images, since we only feed text prompt for this task.] Figure 18: The prompt for instruction generation. We ask the VLM to generate instructions for using pre-trained image models. Probing Visual Language Priors in VLMs Given this image, could you please generate a series of insightful and diverse question-answer pairs based on the image and its descriptions? We are interested in exploring various facets of the image, including: - Holistic styles and layouts: Questions that analyze the overall design, style, and layout of the image. - Object-specific details: Questions that delve into particular elements or objects within the image, discussing their characteristics or functions. - Background context: Questions that speculate about the background story or the setting of the image. - Overall themes: Questions that interpret the thematic elements and messages portrayed in the image. We encourage creative and thought-provoking questions that extend beyond the basics. Please generate questions that cover a broader range of observations and insights drawn from the image. Each question should be followed by a comprehensive answer, providing depth and context. Expected Multiple Response Format: Item Number: 1 Question: [Propose a unique and insightful question based on the descriptions and the images.] Answer: [Provide a comprehensive answer to the proposed question.] Item Number: 2 Question: [Propose a unique and insightful question based on the descriptions and the images.] Answer: [Provide a comprehensive answer to the proposed question.] Please ensure each question-answer pair is well-defined and informative. Please provide at least 5 question-answer pairs based on the input provided. Figure 19: The prompt for single-image QAs. We ask the VLM itself to generate single-image QAs based on the generated images by pre-trained models. Analyze the provided image and its accompanying modification instruction to identify the removed object description, the new object description, and the new image description. Modification Instructions: Expected Multiple Response Format: Item Number: 1 Removed Object Description: [Brief description of the object to be detected and removed] New Object Description: [Description of a new, different object to replace the removed one] New Image Description: [Description of the image after each object s removal, focusing on changes and remaining elements] Item Number: 2 Removed Object Description: [Brief description of the object to be detected and removed] New Object Description: [Description of a new, different object to replace the removed one] New Image Description: [Description of the image after each object s removal, focusing on changes and remaining elements] Figure 20: The prompt for instruction generation of Grounded-SAM. We ask the VLM to generate designated instructions to use Grounded-SAM. C.2. Image DPO data preparation and training details This section details the construction of good-bad question-image-answer (QIA) pairs (Iw, Il) based on the QIAs generated by the pipeline described in Appendix C.1. In brief, the data generation pipeline outlined in Appendix C.1 utilizes VLMs in conjunction with pre-trained image models to generate or modify images and create corresponding question-answer pairs. This process results in a collection of QIA triplets, as illustrated in the Figures 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, and 35. After generating the QIA triplets, we apply three image corruption methods Gaussian blurring, pixelation, and semantic editing to create good-bad QIA pairs for Image DPO training (Section 4.1), while keeping the QA components unchanged. For Gaussian blur, we use a kernel size of 40 for Cambrian and 80 for LLa VA, as the larger kernel size showed better performance for LLa VA. For pixelation, we apply block sizes of 32 for Cambrian and 64 for LLa VA. For these two types of corruption, we utilize data generated by all three pre-trained models: Stable Diffusion, Instruct-Pix2Pix, and Grounded-SAM. Probing Visual Language Priors in VLMs However, for semantic editing, we rely solely on data from Grounded-SAM, which modifies a single object in the image while leaving the rest unchanged. The object to be altered and the type of modification are determined by the VLMs based on the text prompt (Figure 20). For instance, the VLM decides to add a headset to the chef, as shown in Figure 21. Grounded-SAM then modifies the image accordingly by adding the headset, after which the VLM generates a single QA pair based on the text prompt (Figure 19). Note that, we retain only images where the altered region covers more than 10% of the image size, as this ensures a significant and distinct change. For model training, we constructed datasets comprising 800k and 400k DPO pairs to fine-tune LLa VA (7B and 13B) and Cambrian-8B, respectively. Batch sizes are set to 112 for LLa VA-7B, 80 for LLa VA-13B, and 8 (with 4 gradient accumulation steps) for Cambrian-8B. We employ Lo RA with a rank of 128, an alpha of 256, and a learning rate of 5e-7, training each model for 2 epochs. The GPUs we used are 8-L40S. Q: What is the significance of the headset in the man's work environment? A: The headset is likely used for communication purposes, such as coordinating with other kitchen staff, receiving instructions from a manager or client, or listening to music to create a more enjoyable working atmosphere. In a busy kitchen environment, clear communication and efficient coordination are essential to ensure smooth operations and maintain high standards of food quality and safety. Q: What are the potential benefits of the man wearing a headset while working in the kitchen? A: Wearing a headset while working in the kitchen can provide several benefits. First, it can help the man communicate more effectively with his colleagues or clients, ensuring that everyone is on the same page and any issues are addressed promptly. Second, it can help him focus on his tasks by minimizing distractions and external noise, allowing him to concentrate on his work. Finally, if the headset plays music, it can create a more pleasant and up Figure 21: Randomly sampled single-image QA generated by our data generation pipeline. Probing Visual Language Priors in VLMs Q: What is the purpose of the futuristic train depicted in the image? A: The futuristic train in the image serves as a mode of transportation, likely designed for high-speed travel and efficiency. The sleek, modern appearance of the train suggests that it is built for the future, with advanced technology and features to enhance the travel experience for passengers. Q: What message does the image of the futuristic train convey about the city's progress and development? A: The image of the futuristic train conveys a message of progress and development in the city. The sleek and modern design of the train suggests that the city is constantly evolving and improving its infrastructure to meet the needs and expectations of its residents and visitors. The train's presence also indicates that the city is investing in advanced transportation systems to enhance the quality Figure 22: Randomly sampled single-image QA generated by our data generation pipeline. Q: What is the purpose of the fence in the image? A: The fence in the image serves as a barrier between the sheep and the viewer, creating a sense of separation and emphasizing the boundary between the animals and the human observer. The fence may also serve a practical purpose, such as keeping the sheep contained within a specific area or protecting them from potential threats. Furthermore, the fence could be seen as a symbol of the relationship between humans and animals, with the fence acting as a physical representation of the distance that can exist between the two. Q: What is the significance of the sheep standing next to the fence in the image? A: The sheep standing next to the fence in the image could be interpreted as a representation of the boundary between nature and human-made structures. The fence may symbolize the barrier that separates the animals from their natural habitat, while the sheep represent the resilience of nature and the animals' ability to adapt to their environment. Additionally, the sheep's presence near the fence may also signify the potential for interaction between humans and nature, as well as the importance of maintaining a balance between the two. Figure 23: Randomly sampled single-image QA generated by our data generation pipeline. Probing Visual Language Priors in VLMs Q: What is the tablecloth's design and how does it complement the colorful bowls and plates? A: The tablecloth has a multicolored, patterned, and floral design. This design complements the colorful bowls and plates set atop the table, as they are visually appealing and create a harmonious aesthetic. The vibrant colors within the colorful bowls and plates, as well as the lively tablecloth, offer a lively and inviting ambiance, making the dining table appear more vibrant and engaging. This can contribute to creating a lighthearted and cheerful atmosphere for diners enjoying their meal at the table. Q: What is the purpose of the potted plants in the image? A: The potted plants in the image serve as decorative elements, adding a touch of greenery and natural beauty to the setting. They can also help create a more relaxed and welcoming atmosphere, making the dining experience more enjoyable for guests. Figure 24: Randomly sampled single-image QA generated by our data generation pipeline. Q: What design elements can be observed in the kitchen that contribute to its modern and futuristic appearance? A: The kitchen features a sleek and minimalist design, with white appliances and surfaces that create a clean and uncluttered look. The presence of a large island in the middle of the kitchen adds functionality and creates a focal point for the space. The lighting, particularly the presence of a light on the ceiling, enhances the futuristic atmosphere of the kitchen. The overall layout and design choices reflect a contemporary aesthetic, making the kitchen feel both stylish and efficient. Q: What is the significance of the presence of fruits in the kitchen? A: The presence of fruits in the kitchen suggests that the homeowner values healthy eating and incorporates fresh ingredients into their meals. Fruits are an excellent source of vitamins, minerals, and fiber, and their inclusion in the kitchen indicates a commitment to maintaining a balanced and nutritious diet. Additionally, the fruits add a touch of color and natural beauty to the kitchen, enhancing its overall aesthetic appeal. Figure 25: Randomly sampled single-image QA generated by our data generation pipeline. Probing Visual Language Priors in VLMs Q: How does the arrangement of the clocks and figures contribute to the overall design of the space? A: The arrangement of the clocks and figures in the image creates a visually striking and cohesive design. The clocks, with their different shapes and styles, are placed in a way that draws attention to each one, while the figures surrounding them add an element of balance and harmony to the composition. The overall effect is a space that is both functional and aesthetically pleasing, inviting viewers to appreciate the artistry and craftsmanship of the clocks and figures. Q: What is the context or setting of the image? A: Based on the image, it is difficult to determine the exact context or setting in which the clocks and figures are displayed. However, it can be inferred that the space is likely a public or semi-public area, such as a lobby, gallery, or museum, where art and design are showcased. The presence of multiple clocks and figures suggests that the space may be used to convey a message or theme related to time or human experience, or simply to provide an aesthetically pleasing and functional display. Figure 26: Randomly sampled single-image QA generated by our data generation pipeline. Q: What is the significance of the green-themed bathroom and the choice of plants used for decoration? A: The green-themed bathroom with green plants on the walls and a green bathtub suggests a connection to nature and a focus on creating a relaxing and calming atmosphere. The choice of greenery as decoration adds a touch of freshness and tranquility to the space, making it an ideal environment for unwinding and rejuvenation. Green plants are also known for their air-purifying properties, which can contribute to a healthier and more comfortable indoor environment. Q: How does the green color scheme and the use of plants affect the overall ambiance of the bathroom? A: The green color scheme and the use of plants in the bathroom contribute to a serene, soothing, and natural atmosphere. The green hue is often associated with tranquility, relaxation, and freshness, which are all essential elements in a bathroom setting. The incorporation of green plants adds a touch of nature and life to the space, making it feel more inviting and comfortable. Additionally, the plants can help create a more balanced and harmonious environment, promoting a sense of well-being and relaxation. Figure 27: Randomly sampled single-image QA generated by our data generation pipeline. Probing Visual Language Priors in VLMs Q: What is the purpose of the giraffes being in a line? A: The giraffes are lined up in the image, possibly for a variety of reasons. One possibility is that they are waiting for a treat or a meal, as giraffes often gather around food sources in the wild. Another possibility is that they are standing in a line as part of a social behavior or a natural formation that occurs within their herd. This line formation could also be a result of the giraffes being trained or directed by a zookeeper or wildlife caretaker. Q: What is the color palette used in this painting, and how does it contribute to the overall mood or atmosphere? A: The painting features a green and orange color palette, which is commonly associated with natural environments such as forests or grasslands. These colors evoke a sense of tranquility, warmth, and vitality, creating a vibrant and lively atmosphere that complements the giraffes' natural habitat. The use of these colors in the painting also helps to create a sense of depth and dimension, making the image more engaging and visually striking. Figure 28: Randomly sampled single-image QA generated by our data generation pipeline. Q: What is the significance of the teddy bear holding a cheerleader's baton? A: The image of the teddy bear holding a cheerleader's baton is likely symbolic or playful in nature. It could represent a celebration of team spirit, friendship, or support for a sports team or a school event. The combination of the stuffed animal and the baton might evoke feelings of nostalgia, innocence, and youthful enthusiasm. Q: How does the image of the teddy bear holding a cheerleader's baton relate to the theme of cheerleading? A: The image of the teddy bear holding a cheerleader's baton relates to the theme of cheerleading by incorporating a symbol of team spirit, support, and enthusiasm. The presence of the baton suggests that the teddy bear is a cheerleader, while the stuffed animal represents a more innocent, youthful, and friendly aspect of the cheerleading experience. The combination of these elements creates a visually appealing and emotionally evocative image that highlights the positive and uplifting aspects of cheerleading. Figure 29: Randomly sampled single-image QA generated by our data generation pipeline. Probing Visual Language Priors in VLMs Q: How is the man's eating experience different from a regular meal? A: The man's eating experience is different from a regular meal because the sandwich he is eating is significantly larger than what one would typically consume. This unusual size could make the eating experience more challenging and time-consuming, as he would have to consume the sandwich in smaller bites or even take it apart to eat it. It could also be a novelty or a fun experience for the man, as he attempts to eat the giant sandwich. Q: What is the significance of the man eating a giant sandwich in the image? A: The image of the man eating a giant sandwich could signify various things, such as the enjoyment of food, indulgence, or even a metaphor for taking on challenges or overcoming obstacles. The sandwich might also represent a cultural or personal preference for the individual. The image could be used for advertising or promoting a restaurant or food establishment, or it could be a piece of artwork meant to evoke a particular emotion or message. Figure 30: Randomly sampled single-image QA generated by our data generation pipeline. Q: How does the color of the grass in the image contribute to the overall mood and atmosphere? A: The lush green grass in the image creates a serene and peaceful atmosphere. The vibrant green color symbolizes growth, renewal, and harmony with nature. The presence of elephants walking across the grassy field adds to the tranquil and natural setting, evoking a sense of wonder and appreciation for the beauty of the natural world. Q: How does the image of elephants in a field relate to the broader topic of conservation and wildlife preservation? A: The image of elephants in a field serves as a reminder of the importance of conservation and wildlife preservation efforts. Elephants are considered vulnerable species, with their populations facing threats such as habitat loss, poaching, and human-wildlife conflict. The image highlights the need to protect and conserve these majestic creatures and their habitats, ensuring their survival and well-being for future generations. Figure 31: Randomly sampled single-image QA generated by our data generation pipeline. Probing Visual Language Priors in VLMs Q: What is the purpose of the motorcycle's kickstand? A: The kickstand is used to keep the motorcycle stable and upright when it is parked. Q: What is the overall design and style of the motorcycle? A: The motorcycle has a sleek, modern design, and it appears to be well-maintained and ready for use. Q: What is the significance of the motorcycle's location on the street? A: The motorcycle's location on the street suggests that it is either parked or waiting to be ridden, and it is likely that the owner is nearby. The motorcycle's position on the street could also indicate that the owner is taking a break from riding or has stopped for a specific purpose, such as visiting a nearby business or meeting someone. Figure 32: Randomly sampled single-image QA generated by our data generation pipeline. Q: What architectural style is the building with the clock tower? A: The building with the clock tower appears to be a church, featuring a clock tower and a cross. This suggests that the architectural style is likely influenced by Christian or Gothic styles, which are commonly found in churches. These styles often include pointed arches, ribbed vaults, flying buttresses, and intricate stained glass windows. Q: What is the color scheme of the church's clock tower and cross? A: The church's clock tower and cross are predominantly white, which is a common color choice in religious architecture to symbolize purity, innocence, and divinity. The white color may also serve to emphasize the contrast between the clock tower and the cross, highlighting the them. Figure 33: Randomly sampled single-image QA generated by our data generation pipeline. Probing Visual Language Priors in VLMs Q: What is the primary ingredient in the dish being cooked in the pot? A: The primary ingredient in the dish being cooked in the pot is broccoli. Q: What is the role of the wooden spoon in the cooking process? A: The wooden spoon is being used to stir the broccoli and other ingredients in the pot, ensuring even cooking and a well-mixed dish. Q: What is the dish being prepared in the pot? A: The dish being prepared in the pot is a stir-fry, which is a popular and quick cooking method that involves frying ingredients in a small amount of oil over high heat. Figure 34: Randomly sampled single-image QA generated by our data generation pipeline. Q: How does the image capture the essence of a typical day at a ski resort? A: The image captures the essence of a typical day at a ski resort by showcasing the various activities and elements that one would expect to see during a winter sports trip. There are people skiing and snowboarding down the slopes, which indicates that the resort offers different types of winter sports for visitors to enjoy. The presence of snow-covered pine trees in the background adds to the picturesque winter landscape, creating a serene and inviting atmosphere for guests. Additionally, the fact that the resort is bustling with activity suggests that it is a popular destination for winter sports enthusiasts, further emphasizing the essence of a typical day at a ski resort. Q: What is the significance of the snow-covered pine trees in the image? A: The snow-covered pine trees in the image serve as a beautiful and natural backdrop for the ski resort. They add to the overall wintery atmosphere and enhance the picturesque quality of the scene. Additionally, the presence of pine trees is indicative of the type of environment that ski resorts are typically located in, which is a mountainous region with a significant amount of snowfall during the winter months. The snow-covered pine trees also provide a sense of tranquility and harmony with nature, which can be appealing to visitors seeking a peaceful and serene winter experience. Figure 35: Randomly sampled single-image QA generated by our data generation pipeline.