# do_visionlanguage_models_really_understand_visual_language__e558b52c.pdf Do Vision-Language Models Really Understand Visual Language? Yifan Hou 1 Buse Giledereli 1 Yilei Tu 1 Mrinmaya Sachan 1 { yifan.hou, bgiledereli, yiletu, mrinmaya.sachan }@inf.ethz.ch Visual language is a system of communication that conveys information through symbols, shapes, and spatial arrangements. Diagrams are a typical example of a visual language depicting complex concepts and their relationships in the form of an image. The symbolic nature of diagrams presents significant challenges for building models capable of understanding them. Recent studies suggest that Large Vision-Language Models (LVLMs) can even tackle complex reasoning tasks involving diagrams. In this paper, we investigate this phenomenon by developing a comprehensive test suite to evaluate the diagram comprehension capability of LVLMs. Our test suite uses a variety of questions focused on concept entities and their relationships over a set of synthetic as well as real diagrams across domains to evaluate the recognition and reasoning abilities of models. Our evaluation of LVLMs shows that while they can accurately identify and reason about entities, their ability to understand relationships is notably limited. Further testing reveals that the decent performance on diagram understanding largely stems from leveraging their background knowledge as shortcuts to identify and reason about the relational information. Thus, we conclude that LVLMs have a limited capability for genuine diagram understanding, and their impressive performance in diagram reasoning is an illusion emanating from other confounding factors, such as the background knowledge in the models. 1Department of Computer Science, ETH Z urich. Correspondence to: Yifan Hou , Mrinmaya Sachan . Proceedings of the 42 nd International Conference on Machine Learning, Vancouver, Canada. PMLR 267, 2025. Copyright 2025 by the author(s). 1. Introduction Symbolic signals such as language serve as powerful tools in communication by abstracting and interpreting information. Visual language is a form of communication that uses symbols, shapes, and spatial arrangements to convey complex ideas (Greenspan & Shanker, 2009; Li, 2023). Diagrams, which encapsulate symbolic information in the visual stream, are a prime example of visual language (Zdebik, 2012; Anderson et al., 2011) that are extensively used in practice across various domains, e.g., mathematics (Seo et al., 2015), science (Lu et al., 2022), education (Kembhavi et al., 2016; 2017), and illustrations (Hiippala & Orekhova, 2018; Lu et al., 2021). Developing models capable of understanding symbolic information, e.g. in diagrams, is a critical milestone in advancing machine intelligence (Bauer & Johnson-Laird, 1993; de Rijke, 1999; Cromley et al., 2010). Even though recent Large Vision-Language Models (LVLMs, Open AI, 2023; Anil et al., 2023) have demonstrated some success on diagram-based visual reasoning tasks (Lu et al., 2023; Zhang et al., 2024; Chen et al., 2024), it remains unclear whether the performance on these tasks truly reflects the models ability to comprehensively understand the symbolic information in diagrams. For this purpose, we design a comprehensive test suite that investigates the ability of LVLMs to understand diagrams. As defined by Foucault (1977) and Deleuze (1986), diagrams are abstract tools that organize visual entities using relational information. Drawing inspiration from this, our test suite focuses on evaluating diagram understanding by assessing how well models can understand entities and relations in typical diagrams ( 2.1). We evaluate diagram understanding by defining two types of tasks pertaining to fast recognition of entities and relations and slow multi-step reasoning (Kahneman, 2011) over them ( 2.2). While we cannot cover every diagram type for practical reasons, we still cover diagrams across six domains. To ensure that our evaluation is both controlled and generalizable, our test suite includes both clean synthetic diagrams and 1, 001 annotated real diagrams carefully selected from existing datasets Krishnamurthy et al. (2016); Kembhavi et al. (2016) ( 2.3). Using our test suite, we conduct a detailed analysis of the Do Vision-Language Models Really Understand Visual Language? (a) Synthe c Diagram (b) Real Diagram Q: Which en ty connects to Gorilla? A: Wagon, B: Clock, C: Trash Bin D: Owl Q: Which creature is the predator of Krill? A: Phytoplankton, B: Whale, C: Fish, D: Krill GPT-4o: From the image, the predator of Krill is both the Fish and Squid, as indicated by the arrows poin ng from Krill towards them. So, the correct answer would be C: Fish. GPT-4o: From the diagram you provided, the en ty that directly connects to the gorilla is the clock (op on B). There is a visible arrow connec ng the clock to the gorilla. Accuracy (%) Synthe c Real (c) Results on Diagram QA Figure 1: The responses of GPT-4o to two diagram-related questions reveal a notable pattern. The model struggles to correctly answer the relation question in the simple synthetic diagram, yet it successfully understands the relationship in a complex real diagram. We demonstrate that this pattern occurs consistently (Tabs. 3 and 4). model s strengths and weaknesses in understanding diagrams and explore the following questions: Q1: Can LVLMs understand diagrams? To assess the basic capabilities of LVLMs, we first test them using clean synthetic diagrams, followed by evaluations in real-world scenarios for comparison. Our findings from these evaluations lead to three key observations: LVLMs can identify entities and reason about them. By generating synthetic diagrams to evaluate models from multiple perspectives, we observe that models consistently perform well on entity-related questions. They can accurately identify and reason about entities in synthetic diagrams, regardless of whether the entity is represented textually or visually ( 3.1). LVLMs struggle with identifying and reasoning about relations (Fig. 1a). In synthetic scenarios, the models exhibit significant difficulty in identifying relationships between depicted concepts and in performing reasoning tasks based on those relationships ( 3.2). This challenge persists across various diagram settings and prompting templates (Apps. F.2.3 and F.2.4). For real diagrams, LVLMs still understand entities and cannot reason about relations. But they can identify relations (Fig. 1b). We annotate real diagrams with questions from multiple aspects and evaluate models on them. Unexpectedly, we find that the models perform significantly better on relation recognition questions in real diagrams compared to those in synthetic diagrams ( 3.3). Q2: If LVLMs cannot identify relations in simple synthetic diagrams, how do they manage to answer complex questions in practice? (Fig. 1c) One potential hypothesis is that the models leverage their background knowledge as a shortcut to answer these questions. To test it, we explore the impact of knowledge on question-answering (QA) and draw the following observations: Knowledge enhances model in relation recognition. We construct synthetic diagrams that incorporate semantic knowledge and observe a notable improvement in relation recognition questions, suggesting that models perform better on knowledge-grounded diagrams ( 4.1). Additionally, in real diagrams, we categorize questions based on whether they require background knowledge (e.g., commonsense) or can be answered independently of it. Results show that models excel at answering questions that draw upon background knowledge ( 4.2). LVLMs only answer relation questions correctly for simple real diagrams. We classify diagrams by complexity that is based on the number of entities, and analyze QA performance for simple and complex diagrams. Results reveal that while accuracy on entity questions remains consistent, that on relation questions drops significantly with the increase of complexity. This indicates that the models seemingly good performance on relation questions is primarily driven by handling simpler diagrams, rather than by genuine relation understanding ( 5.1). LVLMs rely on learned knowledge to hallucinate relations. In the case study, we demonstrate that even when no relations are provided, LVLMs infer relations based on their learned knowledge. Furthermore, if provided relations contradict the models learned knowledge, they tend to disregard them and instead rely on their background knowledge to answer the questions ( 5.2). Do Vision-Language Models Really Understand Visual Language? Synthetic Diagram Question Example Entity QS(V |KF, NR) Which one of the entities exists in the diagram? QS(V |KF, NC) How many text labels are there in the diagram? Implicit Relation QS(E|KF, NR) Which one of the text labels is placed on the left of the entity cow? QS(E|KF, NC) How many text labels are placed on the left of the entity cow? Explicit Relation QS(E|KF, NR) Which one of the pairs is connected in the diagram? QS(E|KF, NC) How many entities are connected to cow? Real Diagram Question Example QR(V |KF, NR) Which entity is in the diagram? QR(V |KF, NC) How many entities are there in the diagram? QR(V |KR, NR) Which producer is in the diagram? QR(V |KR, NC) How many consumers are in the diagram? QR(E|KF, NR) Which entity is connected to Fish? QR(E|KF, NC) How many arrows are linked to Fish in the diagram? QR(E|KR, NR) Which is not the predator of Krill? QR(E|KR, NC) How many types of prey are consumed by Fish in the foodweb? Table 1: The template and example of question annotations. The underlined entity varies across diagrams. There are no KR questions for synthetic diagrams since they do not have background knowledge. Questions for real diagrams correspond to Fig. 1b. In terms of relations, Explicit Relation refers to relations that are clearly depicted through arrows or segments, while Implicit Relation refers to those that are conveyed indirectly, such as through relative position relationships. Findings. We summarize our research findings as follows: While LVLMs can recognize and reason about entities, they struggle with relations. The models do not engage in genuine diagram parsing or reasoning; rather, their seemingly strong performance on various benchmarks is an illusion created by their reliance on knowledge shortcuts. Specifically, these models identify the entities depicted in the diagrams and simply retrieve relevant pre-learned knowledge. 2. Test Suite Design In this section, we provide a definition of a diagram and provide a desiderata for our evaluation suite. 2.1. Diagrams As Graphs Diagrams work as an abstract tool to describe concepts and relationships (Foucault, 1977; Deleuze, 1986). In practice, we choose this representation as it is quite flexible and a broad set of diagrams can be represented in a format as shown in Fig. 1. For example, logical diagrams such as water cycles illustrate the process transitions (relations) among cycle stages (entities). Schematic diagrams such as circuits demonstrate the connections (relations) among electronic components (entities). Previous work (Song et al., 1995; Hiippala & Orekhova, 2018) also chose to annotate and model diagrams as concepts and their connections.1 Following that, we define a diagram as a graph G = {V, E}. Here, V is the set of entities (e.g., Squid in the example 1Given the variety of diagram types, we leave certain cases that are challenging to represent in this way for future work. diagram in Fig. 1b). Each entity V V could be represented in multiple ways, e.g., text and visuals in the diagram. Each relation E = (V , V ) E connects two entities. Relations are either explicitly represented by arrows or via implicit relationships (e.g., relative positioning of entities). 2.2. How Do We Evaluate The Model s Diagram Understanding Ability? In designing our test suite, we draw inspiration from (Kahneman, 2011) who argues that the thinking process can be naturally divided into two modes: System 1, which handles automatic, quick, and intuitive thinking (e.g., pattern recognition and everyday decisions), and System 2, which is responsible for deliberate, slow, and logical thinking (e.g., logical reasoning and critical analysis). We evaluate both the recognition and reasoning abilities of models via questionanswering (QA). We carefully design a set of questions posed in a multiple-choice QA format with each question having one correct answer and three incorrect options and simply use the model s accuracy in answering the questions as an evaluation of the model s ability on that skill. Overall, we denote the set of questions as Q. We denote the set of questions related to understanding entities as Q(V ), while those related to relations as Q(E). Additionally, we use subscripts to distinguish between different types of questions: questions on synthetic diagrams are denoted as QS, whereas questions for real diagrams are denoted as QR. Our test suite questions are categorized as follows: Recognition vs. Reasoning Questions. To measure the two key abilities of models in recognition of entities and Do Vision-Language Models Really Understand Visual Language? relations vs. reasoning, we design two types of questions: Name Recognition (NR) and Number Counting (NC). The NR questions measure the recognition ability of models by verifying the existence of specific entities or relations. In contrast, NC questions measure reasoning ability by asking for the number of certain types of entities or relations. We formally denote these question sets as Q( |NR) and Q( |NC). Knowledge-Required vs. Knowledge-Free. Next, we test if LVLMs use any knowledge shortcuts (Ye & Kovashka, 2021; Tang et al., 2023) to answer our questions without true diagram understanding. Diagrams often encode some background real-world knowledge, and the models may use their background knowledge as a shortcut to answer the questions. To further tease out the models true understanding of diagrams, we design questions that both do and do not require background knowledge, allowing us to test the models capabilities in each scenario. If a question requires the model to use background knowledge (e.g., semantic or commonsense), we call it a knowledge-required (KR) question, which is denoted as Q( |KR). On the other hand, questions that do not rely on such external knowledge are termed knowledge-free (KF) questions, denoted as Q( |KF). This distinction helps us clearly separate the model s capacity for pure visual processing from its ability to incorporate and utilize prior knowledge when answering questions. Templates and examples of each question type are presented in Tab. 1. Each question type targets a specific diagram component or model ability within the evaluation. For real diagrams, the question templates are tailored to the specific context of each domain. Details on the question design in each domain are in App. E.1. From these questions, we can derive the following intuition: Intuition 1. KR questions in real diagrams are generally more challenging than KF questions in synthetic diagrams. The reasons are: 1) Real diagrams are inherently more complex, containing a wider range of information compared to synthetic diagrams; 2) Beyond assessing basic abilities, answering KR questions also requires the integration of additional background knowledge. 2.3. Which Diagrams Do We Consider? Diagrams are extensively utilized in various domains, appearing in different forms and encompassing a wide range of information types. While evaluating models in a synthetic setting helps to reduce the impact of confounding factors, the resulting conclusions may not fully extend to real-world cases. Conversely, evaluating models on real diagrams allows for broader coverage of diagram types, though it may introduce biases due to the additional information or knowledge embedded in these diagrams. To address this challenge, our test suite incorporates both synthetic and real diagrams, providing a balanced and comprehensive evaluation. Synthetic diagram set. We generate synthetic diagrams in two steps. First, we randomly create between 2 to 9 entities represented by images or text. Then, we randomly establish between 1 to maximum relations among them using directed arrows.To ensure clarity in the synthetic diagrams, we carefully avoid situations when arrows cross over certain entities. In practice, we construct our entity set V by sampling from a pre-defined set containing 377 distinct entities provided by Lu et al. (2021). By default, we generate 1, 000 diagrams for each experiment. Real diagram set. We carefully filtered and curated a selection of 1, 001 real-world diagrams from Krishnamurthy et al. (2016); Kembhavi et al. (2016) to include in our test suite. These diagrams span a diverse range of domains, including ecology, biology, physics, astronomy, chemistry, and geology. This selection ensures that our test suite covers a broad spectrum of scientific disciplines, providing a comprehensive evaluation of the models capabilities. During the filtering process, diagrams are first categorized by domain. Subsequently, low-quality diagrams, along with those considered too simplistic or excessively complex, were removed to ensure reliable annotations. Example questions are given in Tab. 1. Detailed statistical information and annotation details of these diagrams can be found in in App. E. 3. Do LVLMs Understand Diagrams? In this section, we investigate whether LVLMs can understand entities ( 3.1) and relations ( 3.2) in synthetic diagrams. Additionally, we present the evaluation results on real diagrams ( 3.3). 3.1. Can LVLMs Understand Entities? We begin by evaluating whether LVLMs can identify and reason about entities represented by text boxes (i.e., text entities) or visual icons (i.e., visual entities). Additionally, we examine the models ability to correctly identify the spatial information (i.e., locations) of these entities in App. F.2.2. Preparation. We evaluate three open-source models: LLa MA-3.2-11B-Vision (i.e., LLa MA, Dubey et al., 2024)), LLa VA-One Vision-7B (i.e., LLa VA, Li et al., 2024a)), and Qwen2-VL-7B (i.e., Qwen2, Wang et al., 2024)), as well as three large models: GPT-4Vision (i.e., GPT-4V, Open AI, 2023)), GPT-4o (Open AI, 2024), and Gemini 1.5 Pro (i.e., Gemini, Anil et al., 2023)), where the evaluation takes place from June to September 2024. More details about the model configuration can be found in App. F.1. The prompting templates and demonstration examples for various models are given in Figs. 11 to 14 in App. G.1.1. Do Vision-Language Models Really Understand Visual Language? Acc (%) Text Entity Visual Entity QS(V |KF, NR) QS(V |KF, NC) QS(V |KF, NR) QS(V |KF, NC) LLa VA 38.9 26.6 46.4 30.8 Molmo 93.4 78.8 64.1 54.0 LLa MA 91.3 90.9 72.7 70.1 Qwen-2B 82.3 73.2 63.1 53.5 Qwen-7B 97.6 73.0 94.5 73.0 Qwen-72B 99.1 97.9 90.6 86.4 GPT-4V 97.8 99.6 85.7 93.7 GPT-4o 99.2 100.0 92.6 94.9 Gemini 88.1 95.8 87.7 86.5 Average 87.5 81.8 77.5 71.4 Table 2: Performance on QA in terms of entities in text labels or visual icons. LVLMs can always identify entities correctly, and can also reason about them effectively. Results. We evaluate LVLMs under the Chain-of-Thought prompting (Co T, Wei et al., 2022) as in Tab. 2. Results are consistent under the zero-shot prompting (ZS) setting (App. F.2.1). The results demonstrate that all LVLMs can easily recognize entities in both text boxes (> 85% accuracy) and visual icons ( 80% accuracy). The accuracies on NR questions, which assess entity recognition, remain consistently high. For the NC questions, which evaluate reasoning ability, we find that LVLMs can answer them pretty well, achieving 80% accuracy for text entities and 75% for visual entities. Our findings on entity recognition and reasoning can be summarized as follows: Observation 1 (Ability to understand entity). LVLMs can nearly perfectly identify entities in diagrams and demonstrate strong reasoning abilities regarding these entities. This ability is fundamental to various vision tasks. Our observation aligns with existing research, confirming that models possess the basic capability to perform simple object detection and count objects to some extent. Next, we turn our focus to complex relations to determine whether LVLMs can comprehend the intricate interactions between entities. 3.2. Can LVLMs Understand Relations? We categorize relations into two types for our research: implicit relations (e.g., relative positions of entities) and explicit relations (e.g., arrows or segments). Preparation. We generate synthetic diagrams following previous settings ( 3.1). To reduce errors contributed by entity understanding, here we represent entities by text, which yields the best performance on corresponding NR and NC questions (Tab. 2). Example questions are in Tab. 1. The prompting templates and demonstration examples are in Figs. 17 to 20 in App. G.1.2. Acc (%) Implicit Relation Explicit Relation QS(E|KF, NR) QS(E|KF, NC) QS(E|KF, NR) QS(E|KF, NC) LLa VA 30.2 27.5 35.1 28.3 Molmo 71.7 31.2 59.1 50.4 LLa MA 75.4 32.0 55.2 46.1 Qwen2-2B 63.3 29.8 44.0 33.1 Qwen2-7B 74.4 59.0 59.8 51.5 Qwen2-72B 77.9 67.1 70.3 63.8 GPT-4V 72.3 34.4 61.6 59.5 GPT-4o 77.3 55.3 76.6 70.2 Gemini 60.9 31.8 68.5 70.2 Average 67.0 40.9 58.9 52.6 Table 3: Performance on QA for relations. LVLMs struggle to identify both implicit and explicit relations and are unable to reason about them effectively. Results. Tab. 3 presents the accuracies for relation questions. Results indicate that all models generally struggle with relation recognition (NR questions), which leads to an average accuracy of around 65% and 60%. Models also have difficulty reasoning about relations (NC questions), with the average accuracy around 40% and 54%. Notably, even for GPT-4V, its performance on counting implicit relations (i.e., relative positions) is nearly equivalent to random guessing (34.4%), and it also shows significant difficulty in recognizing or reasoning about explicit relations. Consistency Verification. Before proceeding, we validate our findings on relations to ensure their reliability. We adjust the diagram generation settings (e.g., arrow features) and observe that the results remained consistent (App. F.2.3). Beyond diagram variations, we also test the robustness of our results by examining the consistency of LVLMs performance with different prompting templates. First, results of zero-shot prompting is consistent, which also support our findings are valid (App. F.2.1). Second, we run our evaluations under the in-context learning (ICL) setting. The findings indicate that ICL does not improve the models ability to identify or reason about relations (App. F.2.4). These results further confirm the reliability of our conclusions. Thus, we can summarize our observation as follows: Observation 2 (Ability to understand relations). LVLMs can barely identify both implicit and explicit relations, and they are unable to reason about them. This observation contradicts the remarkable success that LVLMs have demonstrated in understanding complex diagrams. To further investigate their failures, we evaluate them on real diagrams to determine whether they can effectively comprehend more complex, real-world scenarios. Do Vision-Language Models Really Understand Visual Language? Acc (%) Entity Relation QR(V |KR, NR) QR(V |KR, NC) QR(E|KR, NR) QR(E|KR, NC) LLa VA 56.5 37.3 45.1 30.2 Molmo 82.7 54.9 59.8 51.2 LLa MA 87.3 59.7 73.7 51.2 Qwen2-2B 66.6 40.7 45.7 39.3 Qwen2-7B 90.0 56.1 71.4 58.0 Qwen2-72B 93.7 77.8 79.4 69.8 GPT-4V 88.9 78.8 78.7 59.9 GPT-4o 93.1 82.3 84.1 72.9 Gemini 85.0 68.4 80.5 57.7 Average 82.6 61.8 68.7 54.5 Table 4: Performance on KR questions for real diagrams. Results indicate models continue to recognize and large models can reason about entities effectively, and they struggle with reasoning about relations. However, surprisingly, models (except LLa VA) can recognize relations in real diagrams pretty well. 3.3. Do LVLMs Understand Real Diagrams? Synthetic diagrams are used to evaluate the basic abilities. Next, we move to real diagrams to double-check how these models perform in practical diagram understanding. Preparation. The evaluation follows similar settings to that on synthetic diagrams. We provide the performance on KR questions for both entities and relations in real diagrams. Example questions are given in Tab. 1, and the question design for each domain is in App. E.1. Prompt templates and examples are in Figs. 27 to 30 in App. G.2. Results. Tab. 4 presents the performance of LVLMs on real diagrams. We observe that models continue to perform well in recognizing and reasoning about entities, with GPT4o achieving 93.1% accuracy in recognition and 82.3% accuracy in reasoning. Besides, models still struggle to reason about relations in real diagrams, similar to their performance on synthetic diagrams. Notably, though, we find that three large models can recognize relations quite well in real diagrams, with an average accuracy of 81.1%.From these results, we can obtain the following observation: Observation 3 (Performance on real diagrams). LVLMs struggle to recognize relations in simple synthetic diagrams, yet they can recognize relations in complex real diagrams. We substantiate that LVLMs cannot understand relations in synthetic diagrams, yet this finding reveals a contradiction. While the models do not inherently possess the ability to recognize relations, they are able to do so in real diagrams. This outcome contradicts our initial intuition (Intuition 1). Therefore, we further investigate these counterintuitive findings by examining the key difference between synthetic and real diagrams: the role of knowledge. While synthetic diagrams contain entities and relations that are random in Acc (%) QS(E|KF, NR) QS(E|KF, NC) Semantic Know. w/o w/ w/o w/ LLa VA 35.1 49.314.2 28.3 31.33.0 Molmo 59.1 69.810.7 50.4 55.04.6 LLa MA 55.2 73.318.1 46.1 54.48.3 Qwen2-2B 44.0 60.716.7 33.1 39.96.8 Qwen2-7B 59.8 74.214.4 51.5 42.59.0 Qwen2-72B 70.3 81.411.1 63.8 60.73.1 GPT-4V 61.6 74.412.8 59.5 57.91.6 GPT-4o 76.6 82.86.2 70.2 72.92.7 Gemini 68.5 72.23.7 70.2 64.85.4 Average 58.9 70.9 12.0 52.6 53.3 0.7 Table 5: Performance on synthetic diagrams without (Vanilla, w/o) and with semantic knowledge (Knowledge-Grounded, w/). The accuracy change from Vanilla to Knowledge-Grounded reveals that models better identify relations in knowledge-grounded diagrams while still struggling to reason about them effectively. their construction, real diagrams often portray entities and relations that agree with commonsense knowledge about the underlying concepts such as the stages of the water cycle or the predator-prey relations in a food chain. 4. Knowledge Shortcut: Quantitative Analysis Given that rich knowledge has been encoded into LVLMs during various training stages, a possible hypothesis emerges: models may not truly understand diagrams but instead rely on their ingrained knowledge as shortcuts to provide answers. In this section, we explore the impact of knowledge on the models ability to answer questions. 4.1. Knowledge Grounding of Diagrams Improves Relation Recognition To determine the effect of knowledge, we compare the models performance on diagrams with and without embedded knowledge. Our focus is on the explicit relation in synthetic diagrams. We construct relations that incorporate semantic knowledge to simulate practical conditions where models might use this knowledge as shortcuts. If our hypothesis is correct, we would expect to see an increase in accuracy on QA tasks for these specially constructed diagrams. Preparation. To construct diagrams grounded with semantic knowledge, we generate relations with real meaning behind them. Specifically, for each entity, we get its Word2Vec embedding (Mikolov et al., 2013) based on the text attribute, and use cosine similarity implemented by spa Cy (Honnibal et al., 2020). If the similarity between entity text is larger than 0.5, we regard that there exists a relation. We construct a semantic graph on all the entities from Lu et al. (2021) with 377 entities and 1901 relations. Do Vision-Language Models Really Understand Visual Language? Acc (%) QR(V |KF, NR) QR(V |KR, NR) QR(V |KF, NC) QR(V |KR, NC) LLa VA 40.6 56.515.9 31.0 37.36.3 Molmo 79.4 82.73.3 60.7 54.95.8 LLa MA 85.2 87.32.1 59.8 59.70.1 Qwen2-2B 62.2 66.64.4 40.2 40.70.5 Qwen2-7B 90.2 90.00.2 60.6 56.14.5 Qwen2-72B 92.5 93.71.2 77.8 77.80.0 GPT-4V 91.5 88.92.6 75.2 78.83.6 GPT-4o 93.6 93.10.5 79.1 82.33.2 Gemini 82.0 85.03.0 75.5 68.47.1 Average 79.7 82.6 2.9 62.2 61.8 0.4 Table 6: Performance for questions on entities in real diagrams. The accuracy gap (from answering KF questions to KR questions) indicates that models perform similarly on entity questions, regardless of whether or not they require background knowledge. We randomly generate diagrams following the settings mentioned in 2.3, but we constrain the generated diagram as the subgraph in the constructed semantic graph. See Figs. 25 and 26 in App. G.1.5 for the prompting templates and demonstration examples. Results. Tab. 5 presents the evaluation accuracies on diagrams without and with semantic knowledge. For comparison, we include the performance changes relative to the original results shown in Tab. 3. Overall, models exhibit improved relation recognition in diagrams containing semantic knowledge. The average accuracy improvement for NR questions is 11.6%. However, for NC questions, the performance remains largely unchanged. The findings are consistent with our hypothesis: models utilize knowledge in diagrams as relation recognition shortcuts. With these results, we have below observation: Observation 4 (Knowledge in diagrams). LVLMs are more effective at recognizing relations in diagrams that relate to background knowledge, as they can use it as a shortcut. This observation indicates that even when questions do not explicitly require background knowledge to answer them, the presence of knowledge in the diagram can still enhance the performance of relation recognition. Next, we investigate whether questions that require models to actively use knowledge could further improve their performance. 4.2. Relation Recognition Questions Requiring Knowledge Show Improvements We then evaluate the models using questions that do not require background knowledge (KF questions) and compare their performance with questions that do require such knowledge (KR questions). Evaluation on entity questions We follow the same settings as in previous experiments. Tab. 6 illustrates the role of Acc (%) QR(E|KF, NR) QR(E|KR, NR) QR(E|KF, NC) QR(E|KR, NC) LLa VA 44.7 45.10.4 29.3 30.20.9 Molmo 54.4 59.85.4 49.4 51.21.8 LLa MA 63.4 73.710.3 56.7 51.25.5 Qwen2-2B 42.6 45.73.1 32.9 39.36.4 Qwen2-7B 58.8 71.412.6 53.5 58.04.5 Qwen2-72B 66.7 79.412.7 62.7 69.87.1 GPT-4V 69.7 78.79.0 57.1 59.92.8 GPT-4o 73.9 84.110.2 65.0 72.97.9 Gemini 66.3 80.514.2 61.2 57.73.5 Average 60.0 68.7 8.7 52.0 54.5 2.5 Table 7: Performance for questions on relations in real diagrams. The accuracy gap (from answering KF questions to KR questions) suggests that models perform better on NR questions when these questions require knowledge to answer. However, there is no significant improvement in accuracy when it comes to NC questions. knowledge in questions related to entities. The results show that the gap between KF and KR questions for entities in real diagrams is negligible. The average accuracy increase in recognition is only smaller than 3% (except LLa VA achieves better performance when knowledge is required). The increase in reasoning is similarly minimal at 0.2%. Evaluation on relation questions. Similarly, Tab. 7 illustrates the impact of knowledge on questions related to relations. The results show that when questions require knowledge, models are better at recognizing relations (except LLa VA achieves roughly the same performance).2 However, their reasoning ability remains largely unchanged. Specifically, the average accuracy increase gap in recognition is 9.5%, while the gap in reasoning is only 1.2%. Thus, we have the observation below: Observation 5 (Knowledge shortcuts help). LVLMs are better at recognizing relations in real diagrams when the question requires them to use background knowledge. However, whether or not a question requires knowledge does not significantly impact the models performance in entity recognition, entity reasoning, or reasoning about relations. Observations 4 and 5 lead to conclusions that are entirely contrary to our initial intuition (Intuition 1). In this section, we use quantitative analysis to support our hypothesis that knowledge acts as a shortcut for recognizing relations. In the next section, we provide qualitative analysis to further substantiate the validity of this hypothesis. 5. Knowledge Shortcut: Qualitative Analysis We provide further evidence to confirm that LVLMs can only recognize and reason about entities but not relations in 2We presume that this is because LLa VA is weaker than other models such that it cannot handle relations in real diagrams well. Do Vision-Language Models Really Understand Visual Language? 4 5-6 7-8 9-10 11 QR(V |KF, NR) QR(V |KF, NC) QR(V |KR, NR) QR(V |KR, NC) (a) Acc.: entity questions 4 5-6 7-8 9-10 11 QR(E|KF, NR) QR(E|KF, NC) QR(E|KR, NR) QR(E|KR, NC) (b) Acc.: relation questions Figure 2: Performance of LVLMs (Co T) on answering questions for real diagrams with different complexities (i.e., the number of entities in the diagram, |V|). Results show that models can always answer questions on entity well but cannot handle questions on relations if the diagram is complex. real diagrams ( 5.1). The ability to recognize relations in real diagrams appears to be an illusion driven by knowledge shortcuts rather than genuine understanding ( 5.2). 5.1. LVLMs Cannot Recognize and Reason About Relations in Real Diagrams Preparation. We use the number of entities in a diagram as an indicator of its complexity, with the answers to QR(V |KF, NC) providing the entity count, as introduced in Tab. 1. We then divide all real diagrams into five bins based on their entity count, ensuring that each bin contains more than 100 diagrams (detailed statistics are provided in Fig. 9). For each subset of diagrams, we report the average accuracy for all questions under the Co T setting. Results. Fig. 2 present the average accuracy of three models on all questions (detailed results are in Fig. 10 in App. F.3). Overall, the performance on entity questions remains consistent across different levels of diagram complexity, while there is a noticeable decline in accuracy for relation questions as diagram complexity increases. These results further support that LVLMs can understand entities but struggle with understanding relations. Additionally, they suggest that the models apparent success in recognizing and reasoning about relations in real diagrams is largely due to their performance on simpler diagrams, rather than a true comprehension of complex relational structures. 5.2. LVLMs Hallucinate Relations in Real Diagrams Next, we provide an intuitive and vivid case study to demonstrate that LVLMs hallucinate when interpreting relations in diagrams. Using the example diagram from Fig. 1, we construct two special cases for comparison: in the first case, we remove all relations from the diagram, and in the second case, we replace the relations with random ones. The goal is to test the GPT-4o s response when there is either no relational information or when the relations presented conflict with background knowledge. For evaluation, we pose an annotated question (QR(E|KR, NR)) and a complex reasoning question involving food chain counting and use the configuration in App. F.1. In Fig. 3, when comparing the responses in the left subfigure (vanilla) with those in the middle subfigure (w/o relation), we observe that even in the absence of explicit relational information, the model still identifies the correct predator. Additionally, for the food chain counting question, the model continues to provide the original answers. This indicates that the model has pre-existing knowledge and it can use the knowledge as a shortcut for answering questions. Similarly, when comparing the responses in the left subfigure (vanilla) with those in the right subfigure (with random relations), we find that the model provides the same answers despite the introduction of random relations. The new correct predator could be A) or C) , and the new correct food chain count is 4. This demonstrates that the model relies on learned knowledge rather than parsing the diagram itself. With this additional evidence, we can reasonably conclude the following finding: Finding: Current LVLMs can recognize and reason about entities in diagrams but struggle with understanding relations. However, they manage to answer diagram-related complex questions by identifying entities and leveraging relevant learned knowledge as a shortcut. 6. Related Work LVLM Evaluation. Recent benchmarks evaluate Large Vision-Language Models (LVLMs) on perception, knowledge, hallucination, and reasoning (Fu et al., 2023; Yu et al., 2023; Yue et al., 2024; Liu et al., 2023). These benchmarks use captioning or visual question answering to probe model capabilities, but often use natural or synthetic images that Do Vision-Language Models Really Understand Visual Language? Q: Which creature is the predator of Krill? A: Phytoplankton, B: Whale, C: Fish, D: Krill GPT-4o: From the image, ... the correct answer would be C: Fish. Q: Think step by step before answering the ques on. How many food chains are there? A) 6 B) 4 C) 7 D) 5 GPT-4o: Let's carefully analyze the diagram Answer: A) 6 GPT-4o: Based on ... the correct answer is: C: Fish. GPT-4o: Based on ... the best answer would be: C: Fish GPT-4o: To approach this problem answer is: A) 6 GPT-4o: Let s analyze the image answer is: A) 6 Figure 3: The model response on the example diagram and its variants. Results suggest that the model relies on background knowledge as a shortcut rather than accurately recognizing and reasoning about relations. differ from symbolic visual languages like diagrams. Diagram QA and Reasoning. Prior DQA work spans synthetic datasets like NLVR and CLEVR (Suhr et al., 2017; Johnson et al., 2017), textbook diagrams like AI2D (Kembhavi et al., 2016), and statistical graphics (Masry et al., 2022). While some focus on layout or procedure understanding, others probe commonsense reasoning. Our work uniquely investigates the gap between synthetic and real diagrams, attributing LVLMs diagram performance to knowledge shortcuts. LVLM Capabilities and Limits. While LVLMs show strong performance on QA and image reasoning (Chen et al., 2023; Fatemi et al., 2024), recent studies highlight their limitations in basic reasoning and perception (Hu et al., 2024; Yang et al., 2024). Our work builds on this view, showing how diagram understanding is often an illusion created by leveraging pre-trained knowledge. 7. Conclusion We evaluate three LVLMs on diagram understanding using our test suite, including synthetic and real diagrams. Our findings reveal that while these models can perfectly recognize and reason about entities depicted in the diagrams, they struggle with recognizing the depicted relations. Furthermore, we demonstrate that the models rely on knowledge shortcuts when answering complex diagram reasoning questions. These results suggest that the apparent successes of LVLMs in diagram reasoning tasks create a misleading impression of their true diagram understanding capabilities. Impact Statement Our study reveals a critical gap in the current evaluation of Large Vision-Language Models (LVLMs): their apparent success in diagram reasoning tasks often stems from leveraging background knowledge rather than genuine diagram comprehension. By introducing a comprehensive test suite that decouples visual recognition, relational reasoning, and knowledge reliance, we provide a robust framework for analyzing the symbolic reasoning capabilities of LVLMs. This insight challenges prevailing assumptions about multimodal understanding and offers practical implications for future benchmarks, model development, and safety-sensitive applications where precise symbolic reasoning is essential. Acknowledgment We thank the reviewers for their constructive feedback. We also thank Chenxi Pang, Jingwei Ni, and Shaobo Cui for their valuable input during the early stages of this work. Yifan Hou is supported by the Swiss Data Science Center Ph D Grant (P22-05). Do Vision-Language Models Really Understand Visual Language? Anderson, M., Meyer, B., and Olivier, P. Diagrammatic representation and reasoning. Springer Science & Business Media, 2011. Anil, R., Borgeaud, S., Wu, Y., Alayrac, J., Yu, J., Soricut, R., Schalkwyk, J., Dai, A. M., Hauth, A., Millican, K., Silver, D., Petrov, S., Johnson, M., Antonoglou, I., Schrittwieser, J., Glaese, A., Chen, J., Pitler, E., Lillicrap, T. P., Lazaridou, A., Firat, O., Molloy, J., Isard, M., Barham, P. R., Hennigan, T., Lee, B., Viola, F., Reynolds, M., Xu, Y., Doherty, R., Collins, E., Meyer, C., Rutherford, E., Moreira, E., Ayoub, K., Goel, M., Tucker, G., Piqueras, E., Krikun, M., Barr, I., Savinov, N., Danihelka, I., Roelofs, B., White, A., Andreassen, A., von Glehn, T., Yagati, L., Kazemi, M., Gonzalez, L., Khalman, M., Sygnowski, J., and et al. Gemini: A family of highly capable multimodal models. Co RR, abs/2312.11805, 2023. doi: 10.48550/ARXIV.2312.11805. Bauer, M. I. and Johnson-Laird, P. N. How diagrams can improve reasoning. Psychological Science, 4(6):372 378, 1993. doi: 10.1111/j.1467-9280.1993.tb00584.x. Chen, L., Li, B., Shen, S., Yang, J., Li, C., Keutzer, K., Darrell, T., and Liu, Z. Large language models are visual reasoning coordinators. In Oh, A., Naumann, T., Globerson, A., Saenko, K., Hardt, M., and Levine, S. (eds.), Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, Neur IPS 2023, New Orleans, LA, USA, December 10 - 16, 2023, 2023. Chen, L., Li, J., Dong, X., Zhang, P., Zang, Y., Chen, Z., Duan, H., Wang, J., Qiao, Y., Lin, D., and Zhao, F. Are we on the right way for evaluating large vision-language models? Co RR, abs/2403.20330, 2024. doi: 10.48550/ ARXIV.2403.20330. Cromley, J. G., Snyder-Hogan, L. E., and Luciw-Dubas, U. A. Cognitive activities in complex science text and diagrams. Contemporary Educational Psychology, 35(1): 59 74, 2010. ISSN 0361-476X. doi: https://doi.org/10. 1016/j.cedpsych.2009.10.002. de Rijke, M. Logical reasoning with diagrams, gerard allwein and jon barwise, eds. J. Log. Lang. Inf., 8(3):387 390, 1999. doi: 10.1023/A:1008348918681. Deleuze, G. Foucault. Univ of Minnesota Press, 1986. Dubey, A., Jauhri, A., Pandey, A., Kadian, A., Al-Dahle, A., Letman, A., Mathur, A., Schelten, A., Yang, A., Fan, A., Goyal, A., Hartshorn, A., Yang, A., Mitra, A., Sravankumar, A., Korenev, A., Hinsvark, A., Rao, A., Zhang, A., Rodriguez, A., Gregerson, A., Spataru, A., Rozi ere, B., Biron, B., Tang, B., Chern, B., Caucheteux, C., Nayak, C., Bi, C., Marra, C., Mc Connell, C., Keller, C., Touret, C., Wu, C., Wong, C., Ferrer, C. C., Nikolaidis, C., Allonsius, D., Song, D., Pintz, D., Livshits, D., Esiobu, D., Choudhary, D., Mahajan, D., Garcia-Olano, D., Perino, D., Hupkes, D., Lakomkin, E., Al Badawy, E., Lobanova, E., Dinan, E., Smith, E. M., Radenovic, F., Zhang, F., Synnaeve, G., Lee, G., Anderson, G. L., Nail, G., Mialon, G., Pang, G., Cucurell, G., Nguyen, H., Korevaar, H., Xu, H., Touvron, H., Zarov, I., Ibarra, I. A., Kloumann, I. M., Misra, I., Evtimov, I., Copet, J., Lee, J., Geffert, J., Vranes, J., Park, J., Mahadeokar, J., Shah, J., van der Linde, J., Billock, J., Hong, J., Lee, J., Fu, J., Chi, J., Huang, J., Liu, J., Wang, J., Yu, J., Bitton, J., Spisak, J., Park, J., Rocca, J., Johnstun, J., Saxe, J., Jia, J., Alwala, K. V., Upasani, K., Plawiak, K., Li, K., Heafield, K., Stone, K., and et al. The llama 3 herd of models. Co RR, abs/2407.21783, 2024. doi: 10.48550/ARXIV.2407.21783. Fatemi, B., Halcrow, J., and Perozzi, B. Talk like a graph: Encoding graphs for large language models. In The Twelfth International Conference on Learning Representations, ICLR 2024, Vienna, Austria, May 7-11, 2024. Open Review.net, 2024. Foucault, M. Discipline and punish: The birth of the prison.[translated from the french by alan m. sheridan]. London: Allen Lane.[Original work published in 1975], 1977. Fu, C., Chen, P., Shen, Y., Qin, Y., Zhang, M., Lin, X., Qiu, Z., Lin, W., Yang, J., Zheng, X., Li, K., Sun, X., and Ji, R. MME: A comprehensive evaluation benchmark for multimodal large language models. Co RR, abs/2306.13394, 2023. doi: 10.48550/ARXIV.2306.13394. Greenspan, S. I. and Shanker, S. The first idea: How symbols, language, and intelligence evolved from our primate ancestors to modern humans. Da Capo Press, 2009. Gupta, T. and Kembhavi, A. Visual programming: Compositional visual reasoning without training. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2023, Vancouver, BC, Canada, June 17-24, 2023, pp. 14953 14962. IEEE, 2023. doi: 10.1109/ CVPR52729.2023.01436. Hiippala, T. and Orekhova, S. Enhancing the AI2 diagrams dataset using rhetorical structure theory. In Calzolari, N., Choukri, K., Cieri, C., Declerck, T., Goggi, S., Hasida, K., Isahara, H., Maegaard, B., Mariani, J., Mazo, H., Moreno, A., Odijk, J., Piperidis, S., and Tokunaga, T. (eds.), Proceedings of the Eleventh International Conference on Language Resources and Evaluation, LREC 2018, Miyazaki, Japan, May 7-12, 2018. European Language Resources Association (ELRA), 2018. Do Vision-Language Models Really Understand Visual Language? Honnibal, M., Montani, I., Van Landeghem, S., and Boyd, A. spa Cy: Industrial-strength Natural Language Processing in Python. 2020. doi: 10.5281/zenodo.1212303. Hu, Y., Tang, X., Yang, H., and Zhang, M. Case-based or rule-based: How do transformers do the math? Co RR, abs/2402.17709, 2024. doi: 10.48550/ARXIV.2402. 17709. Islam, M. S., Rahman, R., Masry, A., Laskar, M. T. R., Nayeem, M. T., and Hoque, E. Are large vision language models up to the challenge of chart comprehension and reasoning. In Al-Onaizan, Y., Bansal, M., and Chen, Y.-N. (eds.), Findings of the Association for Computational Linguistics: EMNLP 2024, pp. 3334 3368, Miami, Florida, USA, November 2024. Association for Computational Linguistics. Johnson, J., Hariharan, B., van der Maaten, L., Fei-Fei, L., Zitnick, C. L., and Girshick, R. B. CLEVR: A diagnostic dataset for compositional language and elementary visual reasoning. In 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, July 21-26, 2017, pp. 1988 1997. IEEE Computer Society, 2017. doi: 10.1109/CVPR.2017.215. Kahneman, D. Thinking, fast and slow. Farrar, Straus and Giroux, New York, 2011. ISBN 9780374275631 0374275637. Kembhavi, A., Salvato, M., Kolve, E., Seo, M. J., Hajishirzi, H., and Farhadi, A. A diagram is worth a dozen images. In Leibe, B., Matas, J., Sebe, N., and Welling, M. (eds.), Computer Vision - ECCV 2016 - 14th European Conference, Amsterdam, The Netherlands, October 11-14, 2016, Proceedings, Part IV, volume 9908 of Lecture Notes in Computer Science, pp. 235 251. Springer, 2016. doi: 10.1007/978-3-319-46493-0\ 15. Kembhavi, A., Seo, M., Schwenk, D., Choi, J., Farhadi, A., and Hajishirzi, H. Are you smarter than a sixth grader? textbook question answering for multimodal machine comprehension. 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5376 5384, 2017. Krishnamurthy, J., Tafjord, O., and Kembhavi, A. Semantic parsing to probabilistic programs for situated question answering. In Su, J., Duh, K., and Carreras, X. (eds.), Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, pp. 160 170, Austin, Texas, November 2016. Association for Computational Linguistics. doi: 10.18653/v1/D16-1016. Kuhnle, A. and Copestake, A. Shapeworld - a new test methodology for multimodal language understanding. Co RR, abs/1704.04517, 2017. Li, B., Zhang, Y., Guo, D., Zhang, R., Li, F., Zhang, H., Zhang, K., Li, Y., Liu, Z., and Li, C. Llava-onevision: Easy visual task transfer. Co RR, abs/2408.03326, 2024a. doi: 10.48550/ARXIV.2408.03326. Li, F.-F. The Worlds I See: Curiosity, Exploration, and Discovery at the Dawn of AI. Flatiron books: a moment of lift book, 2023. Li, Y., Du, Y., Zhou, K., Wang, J., Zhao, W. X., and Wen, J. Evaluating object hallucination in large visionlanguage models. In Bouamor, H., Pino, J., and Bali, K. (eds.), Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, EMNLP 2023, Singapore, December 6-10, 2023, pp. 292 305. Association for Computational Linguistics, 2023. doi: 10.18653/V1/2023.EMNLP-MAIN.20. Li, Y., Wang, L., Hu, B., Chen, X., Zhong, W., Lyu, C., Wang, W., and Zhang, M. A comprehensive evaluation of gpt-4v on knowledge-intensive visual question answering, 2024b. Liu, F., Guan, T., Li, Z., Chen, L., Yacoob, Y., Manocha, D., and Zhou, T. Hallusionbench: You see what you think? or you think what you see? an image-context reasoning benchmark challenging for gpt-4v(ision), llava-1.5, and other multi-modality models. Co RR, abs/2310.14566, 2023. doi: 10.48550/ARXIV.2310.14566. Lu, P., Qiu, L., Chen, J., Xia, T., Zhao, Y., Zhang, W., Yu, Z., Liang, X., and Zhu, S. Iconqa: A new benchmark for abstract diagram understanding and visual language reasoning. In Vanschoren, J. and Yeung, S. (eds.), Proceedings of the Neural Information Processing Systems Track on Datasets and Benchmarks 1, Neur IPS Datasets and Benchmarks 2021, December 2021, virtual, 2021. Lu, P., Mishra, S., Xia, T., Qiu, L., Chang, K., Zhu, S., Tafjord, O., Clark, P., and Kalyan, A. Learn to explain: Multimodal reasoning via thought chains for science question answering. In Koyejo, S., Mohamed, S., Agarwal, A., Belgrave, D., Cho, K., and Oh, A. (eds.), Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, Neur IPS 2022, New Orleans, LA, USA, November 28 - December 9, 2022, 2022. Lu, P., Bansal, H., Xia, T., Liu, J., Li, C., Hajishirzi, H., Cheng, H., Chang, K., Galley, M., and Gao, J. Mathvista: Evaluating math reasoning in visual contexts with gpt4v, bard, and other large multimodal models. Co RR, abs/2310.02255, 2023. doi: 10.48550/ARXIV.2310. 02255. Mao, C., Teotia, R., Sundar, A., Menon, S., Yang, J., Wang, X., and Vondrick, C. Doubly right object recognition: Do Vision-Language Models Really Understand Visual Language? A why prompt for visual rationales. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2023, Vancouver, BC, Canada, June 17-24, 2023, pp. 2722 2732. IEEE, 2023. doi: 10.1109/CVPR52729. 2023.00267. Masry, A., Do, X. L., Tan, J. Q., Joty, S., and Hoque, E. Chart QA: A benchmark for question answering about charts with visual and logical reasoning. In Muresan, S., Nakov, P., and Villavicencio, A. (eds.), Findings of the Association for Computational Linguistics: ACL 2022, pp. 2263 2279, Dublin, Ireland, May 2022. Association for Computational Linguistics. doi: 10.18653/v1/2022. findings-acl.177. Mikolov, T., Chen, K., Corrado, G., and Dean, J. Efficient estimation of word representations in vector space. In Bengio, Y. and Le Cun, Y. (eds.), 1st International Conference on Learning Representations, ICLR 2013, Scottsdale, Arizona, USA, May 2-4, 2013, Workshop Track Proceedings, 2013. Open AI. GPT-4 technical report. Co RR, abs/2303.08774, 2023. doi: 10.48550/ARXIV.2303.08774. Open AI. Hello GPT-4o. 2024. Pan, H., Zhang, Q., Caragea, C., Dragut, E., and Latecki, L. J. Flowlearn: Evaluating large vision-language models on flowchart understanding. In Endriss, U., Melo, F. S., Bach, K., Diz, A. J. B., Alonso-Moral, J. M., Barro, S., and Heintz, F. (eds.), ECAI 2024 - 27th European Conference on Artificial Intelligence, 19-24 October 2024, Santiago de Compostela, Spain - Including 13th Conference on Prestigious Applications of Intelligent Systems (PAIS 2024), volume 392 of Frontiers in Artificial Intelligence and Applications, pp. 73 80. IOS Press, 2024. doi: 10.3233/FAIA240473. Qin, C., Zhang, A., Zhang, Z., Chen, J., Yasunaga, M., and Yang, D. Is chatgpt a general-purpose natural language processing task solver? In Bouamor, H., Pino, J., and Bali, K. (eds.), Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, EMNLP 2023, Singapore, December 6-10, 2023, pp. 1339 1384. Association for Computational Linguistics, 2023. doi: 10.18653/V1/2023.EMNLP-MAIN.85. Seo, M., Hajishirzi, H., Farhadi, A., Etzioni, O., and Malcolm, C. Solving Geometry Problems: Combining Text and Diagram Interpretation. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pp. 1466 1476, Lisbon, Portugal, September 2015. Association for Computational Linguistics. doi: 10.18653/v1/D15-1171. Singh, S., Chaurasia, P., Varun, Y., Pandya, P., Gupta, V., Gupta, V., and Roth, D. Flow VQA: Mapping multimodal logic in visual question answering with flowcharts. In Ku, L.-W., Martins, A., and Srikumar, V. (eds.), Findings of the Association for Computational Linguistics: ACL 2024, pp. 1330 1350, Bangkok, Thailand, August 2024. Association for Computational Linguistics. doi: 10.18653/v1/2024.findings-acl.78. Song, I.-Y., Evans, M., and Park, E. K. A comparative analysis of entity-relationship diagrams. Journal of Computer and Software Engineering, 3(4):427 459, 1995. Suhr, A., Lewis, M., Yeh, J., and Artzi, Y. A Corpus of Natural Language for Visual Reasoning. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pp. 217 223, Vancouver, Canada, 2017. Association for Computational Linguistics. doi: 10.18653/v1/P17-2034. Tang, R., Kong, D., Huang, L., and Xue, H. Large language models can be lazy learners: Analyze shortcuts in in-context learning. In Rogers, A., Boyd-Graber, J. L., and Okazaki, N. (eds.), Findings of the Association for Computational Linguistics: ACL 2023, Toronto, Canada, July 9-14, 2023, pp. 4645 4657. Association for Computational Linguistics, 2023. doi: 10.18653/V1/2023. FINDINGS-ACL.284. Wang, P., Bai, S., Tan, S., Wang, S., Fan, Z., Bai, J., Chen, K., Liu, X., Wang, J., Ge, W., Fan, Y., Dang, K., Du, M., Ren, X., Men, R., Liu, D., Zhou, C., Zhou, J., and Lin, J. Qwen2-vl: Enhancing vision-language model s perception of the world at any resolution. Co RR, abs/2409.12191, 2024. doi: 10.48550/ARXIV.2409. 12191. Wei, J., Wang, X., Schuurmans, D., Bosma, M., Ichter, B., Xia, F., Chi, E. H., Le, Q. V., and Zhou, D. Chain-ofthought prompting elicits reasoning in large language models. In Koyejo, S., Mohamed, S., Agarwal, A., Belgrave, D., Cho, K., and Oh, A. (eds.), Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, Neur IPS 2022, New Orleans, LA, USA, November 28 - December 9, 2022, 2022. Yang, H., Meng, F., Lin, Z., and Zhang, M. Parrot mind: Towards explaining the complex task reasoning of pretrained large language models with template-content structure, 2024. Ye, K. and Kovashka, A. A case study of the shortcut effects in visual commonsense reasoning. In Thirty-Fifth AAAI Conference on Artificial Intelligence, AAAI 2021, Thirty Third Conference on Innovative Applications of Artificial Do Vision-Language Models Really Understand Visual Language? Intelligence, IAAI 2021, The Eleventh Symposium on Educational Advances in Artificial Intelligence, EAAI 2021, Virtual Event, February 2-9, 2021, pp. 3181 3189. AAAI Press, 2021. doi: 10.1609/AAAI.V35I4.16428. Yu, W., Yang, Z., Li, L., Wang, J., Lin, K., Liu, Z., Wang, X., and Wang, L. Mm-vet: Evaluating large multimodal models for integrated capabilities. Co RR, abs/2308.02490, 2023. doi: 10.48550/ARXIV.2308.02490. Yue, X., Ni, Y., Zhang, K., Zheng, T., Liu, R., Zhang, G., Stevens, S., Jiang, D., Ren, W., Sun, Y., Wei, C., Yu, B., Yuan, R., Sun, R., Yin, M., Zheng, B., Yang, Z., Liu, Y., Huang, W., Sun, H., Su, Y., and Chen, W. Mmmu: A massive multi-discipline multimodal understanding and reasoning benchmark for expert agi. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 9556 9567, June 2024. Zdebik, J. Deleuze and the diagram. Deleuze and the Diagram, pp. 1 256, 2012. Zhang, R., Jiang, D., Zhang, Y., Lin, H., Guo, Z., Qiu, P., Zhou, A., Lu, P., Chang, K., Gao, P., and Li, H. Mathverse: Does your multi-modal LLM truly see the diagrams in visual math problems? Co RR, abs/2403.14624, 2024. doi: 10.48550/ARXIV.2403.14624. Do Vision-Language Models Really Understand Visual Language? A. Limitations Our work has two main limitations. First, we focus exclusively on diagrams that depict various entities and relationships. There may be other specialized types of diagrams that are not well-suited to this representation. We encourage future research to explore and analyze model performance on such diagrams. Second, while we demonstrate that LVLMs have limited diagram understanding capabilities and that their strong performance is largely due to knowledge shortcuts, we do not offer insights on how to address this issue. Future work could focus on developing strategies to enhance LVLMs true diagram understanding abilities. Besides, we use the simple prompt to make sure the comparison is fair. Future work may explore more complex prompt templates and better answer (i.e., model output) parsers for better performance. B. Extended Discussion on Related Work Evaluation of LVLMs. A growing number of benchmarks assess the capabilities of multimodal models, particularly LVLMs, across various dimensions such as perception (Fu et al., 2023), factual knowledge (Yu et al., 2023; Li et al., 2024b), hallucination (Liu et al., 2023; Li et al., 2023), and reasoning (Yue et al., 2024). These benchmarks use image captioning or visual question answering (VQA) as evaluation formats, using images sourced from natural image datasets (Li et al., 2023), or generated specifically for probing model abilities (Liu et al., 2023; Zhang et al., 2024). While informative, they primarily focus on natural images, and thus may not reflect models performance on more abstract visual representations like diagrams. Diagram Question Answering (DQA). DQA studies have explored a range of tasks, from spatial reasoning with simple synthetic diagrams (e.g., NLVR (Suhr et al., 2017), Shape World (Kuhnle & Copestake, 2017)) to understanding textbookstyle illustrations (e.g., Food Web (Krishnamurthy et al., 2016), AI2D (Kembhavi et al., 2016), and TQA (Kembhavi et al., 2017)). While CLEVR (Johnson et al., 2017) and Icon QA (Lu et al., 2021) involve spatial and logical reasoning over diagrams, many real-world datasets rely heavily on commonsense knowledge rather than visual reasoning. Recent work on statistical diagram understanding (e.g., charts and flowcharts) highlights LVLMs limitations in parsing basic visual elements and relations (Masry et al., 2022; Islam et al., 2024; Singh et al., 2024; Pan et al., 2024). Our work aligns with these findings in synthetic settings and further explains why LVLMs excel in real-world diagram tasks by leveraging prior knowledge instead of true diagram parsing. Capabilities and Limitations of LVLMs. LVLMs are often praised for their impressive performance on multimodal reasoning tasks, including diagram understanding (Chen et al., 2024), complex question answering (Qin et al., 2023), and visual reasoning (Chen et al., 2023; Gupta & Kembhavi, 2023). However, a contrasting body of work points to fundamental limitations, such as hallucination, shortcut learning, and poor generalization on tasks requiring perception and logical consistency (Yang et al., 2024; Hu et al., 2024; Mao et al., 2023). Notably, LVLMs often fail in basic reasoning tasks like counting (Fu et al., 2023) and relation tracking (Yue et al., 2024). Our work reinforces this perspective by demonstrating that LVLMs diagram reasoning is often an illusion driven by background knowledge rather than actual visual-symbolic reasoning. C. Ethical Considerations Our paper focuses on evaluation using synthetic and public datasets, and we thereby do not foresee any ethical issues originating from this work. The license of Food Web (Krishnamurthy et al., 2016) and AI2D (Kembhavi et al., 2016) is BSD-2-Clause and Apache-2.0 respectively. D. Reproducibility We have tried our best to ensure the reproducibility of our work. The datasets and models that we use in this paper are publicly available. We present all the details about our evaluation and data including the annotation in the Appendix. At the same time, we have uploaded our evaluation code, generated synthetic diagram data, annotated real diagram data, as well as the responses of LVLMs in supplementary files. Every stage of our work ranging from code to results is introduced thoroughly and can be easily reproduced. Do Vision-Language Models Really Understand Visual Language? E. Supplementary Test Suite Details We provide additional details about our test suite here. E.1. Question Annotation Details For questions on synthetic diagrams, we annotate them as described in Tab. 1. Regarding real diagrams, we introduce their annotation details concerning the 6 domains respectively. For each domain, we choose a representative diagram to show how we annotate, as shown in Fig. 4. For each question in real diagram set, we provide four options. We first annotate the correct option. The three incorrect options are sampled uniformly without replacement from the pool of correct answers for questions of the same type, excluding the current correct answer. Furthermore, we manually verify (and edit if necessary) the sampled negative options to ensure that they are not inadvertently correct answers. The four options are randomly shuffled before feeding into LVLMs. Tab. 8 records some statistics of our real diagram set. Domains Ecology Biology Physics Astronomy Chemistry Geology Num. 462 205 77 101 54 102 Entity Rep. Text & Visual Text & Visual Visual Text & Visual Text & Visual Text & Visual Relation Rep. Explicit Explicit Explicit Implicit Explicit Implicit Topics Food Chain Life Cycle Circuit Solar System Water Cycle Planet Structure Food Web Satellite System Carbon Cycle Star Structure Table 8: Details of the real diagram set. Entity Rep. and Relation Rep. refer to the way that entities and relations are represented. Topics introduces the typical types of diagrams in that domain. E.1.1. ECOLOGY - FOOD WEB; FOOD CHAIN Food web diagrams illustrate predatory relationships among animals (and between animals and plants or detritus) in the same environment (e.g. prairie, forests, sea, etc.). Any possible path from a plant to a top animal is a single food chain. QR(V |KF, NR). This question type is consistently annotated as Which entity is in the diagram? . For Fig. 4a, the correct option is Sparrow . QR(V |KF, NC). This question type is consistently annotated as How many entities are in the diagram? . For Fig. 4a, the correct option is 10 . QR(V |KR, NR). The annotation of the question type is determined by the number of producers in the diagram: (a) If the diagram contains three or fewer producers, the question type is annotated as Which producer is in the diagram? . (b) If the diagram includes more than three producers, the annotation changes to Which producer is not in the diagram? . (c) Only in cases where the food web diagram contains no producers (i.e. in the detritus environment) do we annotate the question as Which consumer is not in the diagram? . Fig. 4a contains two producers, Grass and Blueberry Bush , which is applicable to category (a). The correct option we annotate is Blueberry Bush . The (a)(b)(c) question subtypes account for 85.71%, 10.60%, and 3.68% respectively. As the LVLM must comprehend the concepts of producer and consumer to correctly answer these questions, this question type is classified as requiring knowledge. QR(V |KR, NC). This question type is consistently annotated as How many distinct consumers are there in the food web? , which represents a subset of QR(V |KF, NC). For Fig. 4a, the correct option is 8 . Since the LVLM needs to comprehend the meaning of consumer, this question type is classified as requiring knowledge. QR(E|KF, NR). Depending on the number of arrows linked to an entity, for an entity with numerous connections, the question is annotated as Which entity is not connected to Entity? For an entity with fewer connections, the question is annotated as Which entity is connected to Entity? . These two subtypes account for 48.05% and 51.95% respectively. For Fig. 4a, we ask Which entity is connected to Shrew , and the annotated correct option is Great Horned Owl . This Do Vision-Language Models Really Understand Visual Language? (a) Ecology (b) Biology (c) Physics (d) Astronomy (e) Chemistry (f) Geology Figure 4: The representative diagrams of 6 domains. Do Vision-Language Models Really Understand Visual Language? question type only involves recognizing the relations between an entity of interest and others, and thus no knowledge is required. QR(E|KF, NC). This question type is consistently annotated as How many arrows are linked to Entity in the diagram? . For each diagram, we randomly select an entity that has both an arrow pointing to it and an arrow pointing out of it as the entity of interest. For example, in Fig. 4a, we annotate How many arrows are linked to Skunk in the diagram? with the correct option 3 This question type only involves counting the relations (represented by arrows) between an entity of interest and other entities, and thus no knowledge is required. QR(E|KR, NR). Corresponding to QR(E|KF, NR), we translate Which entity is not connected to Entity? to Which is not the predator of Entity? or Which is not the prey of Entity? . Similarly, we translate Which entity is connected to Entity? to Which is the predator of Entity? or Which is the prey of Entity? . Entity of interest is the same as QR(E|KF, NR). These four question subtypes account for 3.98%, 38.94%, 30.75%, and 26.33% respectively. Refer to our annotation of Fig. 4a QR(E|KF, NR), for this question type, we annotate question as Which is the predator of Shrew? with the correct option Great Horned Owl . Since the LVLM needs to comprehend the meaning of prey (arrow points to the entity) and predator (arrow points out of the entity), this question type requires knowledge. QR(E|KR, NC). This question type is consistently annotated as How many types of prey are consumed by Entity in the foodweb? with the same entity of interest as QR(E|KR, NR). Accordingly, for Fig. 4a, we annotate How many types of prey are consumed by Skunk in the foodweb? with the correct option 2 . Similarly, since the LVLM needs to comprehend the meaning of prey, this question type requires knowledge. E.1.2. BIOLOGY - LIFE CYCLE In contrast to E.1.1, this domain delineates the developmental morphology of a single species across its life stages (e.g., for insects in diagrams like Fig. 4b, stages such as egg, pupa, and adult are depicted). These morphological stages are interconnected via arrows, forming a directed cyclic graph. Notably, those diagrams include arrows pointing from the mature form (adult) to its offspring (egg, embryo, or seed). We claim that the mature form represents the final stage of the species life cycle, while the offspring represents the first stage. QR(V |KF, NR). This question type is consistently annotated as Which entity is in the diagram? . For Fig. 4b, the correct option is Larva . QR(V |KF, NC). This question type is consistently annotated as How many entities are in the diagram? . For Fig. 4b, the correct option is 4 . QR(V |KR, NR). This question type is annotated as What is the [first/last] life stage for the creature in the diagram? , where we choose one of two words manually, resulting in 47.80% for first and 52.20% for last . For Fig. 4b, we annotate first with the correct option Egg . Determining the first and last stage of a lifecycle requires background knowledge in biology. QR(V |KR, NC). This question type is consistently annotated as How many life stages are after the stage Entity in the diagram? . For Fig. 4b, we replace Entity with Larva and the correct option is 2 Since the LVLM needs to comprehend the meaning of life stage and the arrow direction, this question type requires knowledge. QR(E|KF, NR). This question type is consistently annotated as Which entity is connected to the Entity? For Fig. 4b, we replace Entity with Larva and the correct option is Egg . QR(E|KF, NC). This question type is consistently annotated as How many arrows are in the diagram? . In Fig. 4b, the correct option is 4 . QR(E|KR, NR). This question type is consistently annotated as Which stage is after the Entity stage in the diagram? . For Fig. 4b, we replace Entity with Larva and the correct option is Pupa . This question type requires LVLMs to understand the meaning of stage and arrow directions, and thus knowledge is required. Do Vision-Language Models Really Understand Visual Language? QR(E|KR, NC). This question type is consistently annotated as How many stages can the creature change in the diagram? Same as QR(E|KF, NC), in Fig. 4b, the correct option is 4 . Similarly, since the LVLM needs to comprehend the meaning of stage and creature, this question type requires knowledge. A few diagrams use the same lifecycle paradigm to describe multiple species, e.g. egg tadpole/chick frog/chicken. The correct answer to QR(E|KR, NC) in this case is 3 while that to QR(E|KF, NC) is 6. E.1.3. PHYSICS - CIRCUIT This domain contains simple middle-to-high-school circuit diagrams, typically containing only power, switches, wires, and a few appliances (e.g., light bulbs), and does not contain circuit diagrams for complex electronics that require knowledge beyond high school. Circuit diagrams can be abstracted as undirected cyclic graphs. Note that in this domain, we do not regard wires as entities even if there is explicit text in the diagram. Instead, we consider them to be the representations of relations. QR(V |KF, NR). This question type is consistently annotated as Which entity is in the diagram? . For Fig. 4c, the correct option is Bulb . QR(V |KF, NC). This question type is consistently annotated as How many entities are in the diagram? . For Fig. 4c, the correct option is 3 . QR(V |KR, NR). This question type is annotated as Which electronic component is in the diagram? . For Fig. 4c, the correct option is Battery . Knowledge is required to understand the meaning of electronic component. QR(V |KR, NC). This question type is consistently annotated as How many bulbs in the diagram will glow when the switch is closed? In Fig. 4c, the correct option is 1 Since the LVLM needs to understand the function of switch as well as determine whether it is a closed circuit, this question type is classified as requiring knowledge. All the diagrams contain switches, but some do not contain light bulbs or they are intentionally short-circuited. In such cases, the correct answer is 0. QR(E|KF, NR). This question type is consistently annotated as Which entity is connected to the Entity by the line? . For Fig. 4c, we replace Entity with Bulb and the correct option is Battery . QR(E|KF, NC). This question type is consistently annotated as How many line segments are in the diagram? . For Fig. 4c, the correct option is 3 . QR(E|KR, NR). This question type is consistently annotated as Which electronic component is connected to the Entity by the wire? Same as QR(E|KF, NR), in Fig. 4c, we replace Entity with Bulb and the correct option is Battery . This question type requires LVLMs to understand the meaning of electronic component and wire, and thus knowledge is required. QR(E|KR, NC). This question type is consistently annotated as How many wires are in the diagram? . Same as QR(E|KF, NC), in Fig. 4c, the correct option is 3 . Similarly, since the LVLM needs to comprehend the meaning of wire, this question type requires knowledge. E.1.4. ASTRONOMY - SOLAR SYSTEM; SATELLITE SYSTEM The subject of the diagrams in this domain encompasses seasonal changes caused by the Earth s revolution around the Sun, moon phase changes caused by the rotation of satellites around the planets, and planetary revolutions in the solar system. Diagrams describing phase changes can be regarded as directed cyclic graphs. A diagram depicting the solar system uses relative positions to express the relation between astronomical objects without using arrows. QR(V |KF, NR). This question type is consistently annotated as Which entity is in the diagram? . For Fig. 4d, the correct option is Sun . QR(V |KF, NC). This question type is consistently annotated as How many entities are in the diagram? . For Fig. 4d, the correct option is 5 . Do Vision-Language Models Really Understand Visual Language? QR(V |KR, NR). This question type is annotated as Which astronomical object is in the diagram? . For Fig. 4d, the correct option is Sun . Knowledge is required to understand the meaning of astronomical object. QR(V |KR, NC). This question type is consistently annotated as How many planets or satellites are in the diagram? For Fig. 4d, the correct option is 4 . Since the LVLM needs to understand the meaning of planets and satellites, this question type requires background knowledge. QR(E|KF, NR). This question type is consistently annotated as Which entity is connected to the Entity in the diagram? . For Fig. 4d, we replace Entity with September and the correct option is December . QR(E|KF, NC). This question type is consistently annotated as How many arrows are in the diagram? . For Fig. 4d, the correct option is 5 . Some of the diagrams rely on relative positions rather than arrows to represent relations between entities, in which case the answer is 0. QR(E|KR, NR). This question type is consistently annotated as What is the next phase after the Entity in the diagram? . For Fig. 4d, we replace Entity with Earth in Summer and the correct option is Earth in Fall This question type requires LVLMs to understand the meaning of phase, and thus knowledge is required. QR(E|KR, NC). This question type is consistently annotated as How many times that the planets or satellites can change in the diagram? . For Fig. 4d, the correct option is 4 . Similarly, since the LVLM needs to comprehend the meaning of planets and satellites, this question type requires knowledge. E.1.5. CHEMISTRY - WATER CYCLE; CARBON CYCLE This domain includes topics such as the water cycle and the carbon cycle where various plants and animals participate, and photosynthesis and transpiration of a specific plant. Water or carbon enjoy multiple pathways to transfer between two phases, so that substantial diagrams can be viewed as directed multigraphs. Other diagrams that describe a single cyclic pathway (such as photosynthesis in a single plant) can be viewed as directed cyclic graphs. For example, the visual entities in the Fig. 4e are Sun, House Emissions, Carbon Dioxide, Cow, Ground (Soil), Tree, Worm, and Cloud. Among these, Smoke, Cow, Ground, Tree, and Worm are actively involved in the depicted carbon cycle. These entities are called cycle stages. Entities such as soil, lakes, and forests are, as per image segmentation principles, considered as single entities despite their spatial extent. QR(V |KF, NR). This question type is consistently annotated as Which entity is in the diagram? . For Fig. 4e, the correct option is Cow . QR(V |KF, NC). This question type is consistently annotated as How many entities are in the diagram? . As mentioned in the domain summary, there are 8 visual entities in Fig. 4e. QR(V |KR, NR). This question type is annotated as Which cycle stage is described in the diagram? . For Fig. 4e, the correct option is Animal . QR(V |KR, NC). This question type is consistently annotated as How many different cycle stages are in the diagram? As mentioned before, there are 5 cycle stages in Fig. 4e. Since the LVLM needs to understand the function of cycle stages, this question type requires knowledge. QR(E|KF, NR). This question type is consistently annotated as Which entity is connected to the Entity by the arrow? . For Fig. 4e, we replace Entity with Cow and the correct option is Carbon Dioxide QR(E|KF, NC). This question type is consistently annotated as How many arrows are in the diagram? . In Fig. 4e, there are 1 arrows. Do Vision-Language Models Really Understand Visual Language? QR(E|KR, NR). This question type is consistently annotated as Which cycle stage will happen after the Entity in the diagram? . For Fig. 4e, we replace Entity with Animal and the correct option is Carbon Dioxide This question type requires LVLMs to understand the meaning of cycle stage, and thus knowledge is required. QR(E|KR, NC). This question type is consistently annotated as How many different processes of transitions are in the diagram? For Fig. 4c, the correct option is 5 . Similarly, since the LVLM needs to comprehend the meaning of processes of transitions, this question type requires knowledge. E.1.6. GEOLOGY - PLANET STRUCTURE; STAR STRUCTURE Diagrams in this domain depict the geological structure of the Earth or other astronomical objects, showing the relationship between geological strata through their relative positions (typically nested structures) rather than explicit arrows. QR(V |KF, NR). This question type is consistently annotated as Which layer is included in the diagram? . For Fig. 4f, the correct option is Crust . QR(V |KF, NC). This question type is consistently annotated as How many layers are in the diagram? . For Fig. 4f, the correct option is 4 . QR(V |KR, NR). This question type is annotated as Which stratification that is outside the core is included in the diagram? . For Fig. 4f, the correct option is Crust Understanding the meaning of stratification and determining the containing relations between stratigraphic layers requires the use of background knowledge. QR(V |KR, NC). This question type is consistently annotated as How many stratifications are outside the core in the diagram? . For Fig. 4f, the correct option is 2 ( Mantle and Crust ). Same as above, answering this question also requires background knowledge. QR(E|KF, NR). This question type is consistently annotated as Which layer is next to the Layer in the diagram? . For Fig. 4f, we replace Layer with Crust and the correct option is Mantle . This question type only involves recognizing the adjacency between a layer of interest and others, and thus no knowledge is required. QR(E|KF, NC). This question type is consistently annotated as How many layer boundaries are in the diagram? . For Fig. 4f, the correct option is 3 . We regard the boundary between layers as the representation of relation. This question only involves counting the number of boundaries, so no background knowledge is required. QR(E|KR, NR). This question type is consistently annotated as Which stratification is the next outside layer of the Layer in the diagram? . For Fig. 4f, we replace Layer with Mantle and the correct option is Crust . This question type requires LVLMs to understand the meaning of stratification and the containing relations between layers, hence knowledge is required. QR(E|KR, NC). This question type is consistently annotated as How many transition zones of the structure are in the diagram? . For Fig. 4f, the correct option is 3 . Similarly, since the LVLM needs to comprehend the meaning of transition zones, this question type requires knowledge. F. Supplementary Results We introduce the details of the models we evaluate and provide additional results in this section. F.1. Model Configurations Generally, we spend around 800$ for all experiments. The LVLM models are provided with the system message: You are a visual assistant answering multiple choice questions about diagrams. Read the question, inspect the diagram, and answer with the correct choice in the following format: A) 0 . Do Vision-Language Models Really Understand Visual Language? GPT-4V. Model is used with the key gpt-4-vision-preview with the Open AI API, Chat Completions. The temperature parameter is set to 0 to ensure deterministic outputs and a seed is given to the model to help with reproducibility. The max tokens is limited to 600. GPT-4o. Model is used with the key gpt-4o with the Open AI API, Chat Completions. The temperature parameter is set to 0 to ensure deterministic outputs and a seed is given to the model to help with reproducibility. The max tokens is limited to 600. Gemini. Model is used with the key gemini-1.5-pro with the Vertex AI API, Generative Models. The temperature parameter is set to 0 to ensure deterministic outputs. The max output tokens is limited to 600. F.2. Synthetic Diagram The additional results on the synthetic diagrams are given in this subsection. F.2.1. ZERO-SHOT PROMPTING Zero-Shot Accuracy (%) LLa VA Molmo LLa MA Qwen-2B Qwen-7B Qwen-72B GPT-4V GPT-4o Gemini Average Text Entity QS(V |KF, NR) 35.5 73.0 87.5 97.4 97.5 98.3 97.4 91.6 86.9 85.0 QS(V |KF, NC) 24.0 76.2 80.2 40.0 76.4 84.2 50.6 64.1 71.9 63.1 Visual Entity QS(V |KF, NR) 43.1 56.0 71.0 87.5 91.8 85.1 83.4 87.5 90.2 77.3 QS(V |KF, NC) 34.1 62.5 77.5 42.9 68.5 81.7 32.4 46.7 67.2 57.1 Implicit Relation QS(E|KF, NR) 27.5 74.7 79.8 71.2 78.4 80.7 75.9 72.5 58.5 68.8 QS(E|KF, NC) 26.3 29.5 28.8 32.7 35.0 44.9 30.4 37.0 30.4 32.8 Explicit Relation QS(E|KF, NR) 31.1 55.0 49.7 49.7 58.2 68.3 57.6 61.8 61.0 54.7 QS(E|KF, NC) 28.2 52.3 45.9 35.0 49.3 56.3 49.6 57.6 69.6 49.3 w/ Semantic Know. QS(E|KF, NR) 33.2 67.0 72.9 69.4 72.5 79.2 74.2 77.9 72.2 68.7 QS(E|KF, NC) 25.2 51.6 50.7 34.0 45.9 51.8 55.8 60.7 68.1 49.3 QR(V |KF, NR) 32.0 77.7 85.6 69.0 81.4 92.2 91.9 88.3 88.1 78.5 QR(V |KR, NR) 43.7 59.1 81.9 70.4 71.0 89.5 84.3 87.9 87.3 75.0 QR(V |KF, NC) 16.8 48.8 61.6 15.5 48.3 69.6 38.6 45.9 58.4 44.8 QR(V |KR, NC) 18.7 50.7 58.9 25.7 51.4 71.2 49.4 61.4 53.5 49.0 QR(E|KF, NR) 31.6 47.2 60.0 44.2 56.0 70.8 67.0 69.2 68.0 57.1 QR(E|KR, NR) 34.8 51.5 69.6 52.3 66.3 79.5 77.5 81.6 80.9 66.0 QR(E|KF, NC) 20.2 55.1 57.8 27.7 47.8 61.4 38.3 50.0 53.4 45.7 QR(E|KR, NC) 22.7 53.2 55.6 28.8 45.5 61.8 42.1 50.4 51.4 45.7 Table 9: All the evaluation results under the zero-shot prompting setting. We can find that these results are consistent with those under the Co T prompting setting, and our conclusions are also supported by them. We report all the results together that are obtained under the zero-shot prompt setting in Tab. 9. We can find that the conclusions are consistent. F.2.2. ENTITY POSITION AND SPATIAL RELATION Synthetic Diagram Question Question Example (with Answer Options) Entity Position QS(V |KF, NR) Which one of the text labels exists in the top row of the diagram? QS(V |KF, NC) How many text labels are there in the top row of the diagram? Table 10: The template and example of questions for the evaluation of entity position and spatial relation in synthetic diagrams. The text with underline (e.g., top row) is specific and varies across diagrams. Preparation. We generate synthetic diagrams similar to the previous settings (Tab. 1). We introduce a grid structure to describe the absolute positions of entities. Specifically, we use a 3 3 grid and gridlines to define the compartments in the Do Vision-Language Models Really Understand Visual Language? canvas. The entities are placed in the center of them with generated arrows connecting them. We represent the entity via text. As depicted in the example in Tab. 10, we use top/center/bottom row/column to describe the entity location. Accuracy (%) QS(V |KF, NR) QS(V |KF, NC) LLa VA (ZS/Co T) 31.6 / 39.1 27.4 / 28.9 Molmo (ZS/Co T) 84.3 / 77.5 52.2 / 61.1 LLa MA (ZS/Co T) 74.0 / 71.0 39.1 / 37.2 Qwen2-2B (ZS/Co T) 64.4 / 66.6 34.8 / 24.1 Qwen2-7B (ZS/Co T) 63.3 / 74.4 43.8 / 59.0 Qwen2-72B (ZS/Co T) 89.4 / 83.5 65.3 / 72.2 GPT-4V (ZS/Co T) 75.3 / 77.7 41.8 / 64.0 GPT-4o (ZS/Co T) 78.1 / 89.5 50.3 / 79.5 Gemini (ZS/Co T) 63.6 / 64.8 66.8 / 74.0 Average (ZS/Co T) 69.3 / 71.6 46.8 / 55.6 Table 11: Performance of LVLMs on entity position QA. LVLMs can capture part of the position information and struggle with entity identification and reasoning. Results. LVLMs start to struggle with identifying the entity s position attribute (Tab. 11). Even with Co T prompting, the average score of NR questions is 77.33%, which is worse compared to the entity text recognition accuracy in Tab. 2 (i.e., 95.02%). Since LVLMs only identify the position partially, the average accuracy on NC questions (i.e., 72.50% with Co T prompting) is also worse than that of text entity (i.e., 98.46%), where Gemini performs much worse than the other two models. Center Row Bottom Row Left Column Center Column Right Column GPT-4V (Co T) GPT-4o (Co T) Gemini (Co T) (a) QS(V |KF, NR) Center Row Bottom Row Left Column Center Column Right Column GPT-4V (Co T) GPT-4o (Co T) Gemini (Co T) (b) QS(V |KF, NC) Figure 5: Accuracies of LVLMs on QS(V |KF, NR) and QS(V |KF, NC) with entities located in different positions (top row, center row, bottom row, left column, center column, and right column). Analysis. We further analyze the results for more insights by visualizing the performance of LVLMs with entities in different compartments. Results (Fig. 5) show that LVLMs can answer NR and NC questions much better if the entities are not in the center area, where this phenomenon is more obvious for two GPT models. F.2.3. CONSISTENCY: DIAGRAM VARIATION Preparation. We change the relation attributes, i.e., the arrow features in synthetic diagrams, to see if LVLMs can understand relations better. Specifically, we randomly change the arrowhead size to its 1.5, or 2 times, change the line width to its 0.5, 2, or 4 times, and the arrow color to black, red, or blue. Other settings remain the same as in 2.3. Then, we ask the same questions on these newly generated diagrams to observe how performance changes. For simplicity, we only consider the Co T prompting setting since it achieves better results. The prompting templates as well as demonstration examples are shown in Figs. 21 and 22. Do Vision-Language Models Really Understand Visual Language? Accuracy (%) Vanilla New-Arrows (a) Q(E, [KF, NR]) Vanilla New-Arrows (b) Q(E, [KF, NC]) Figure 6: Accuracies of LVLMs on the diagrams with modified arrow features (denoted by New-Arrows). New results are consistent with our previous findings (denoted by Vanilla) as in 2.3. Results. We visualize the results in Fig. 6 comparing with the results in Tab. 3. We find that changing the relation attributes does not yet improve the accuracy of QA for both NR questions and NC questions. The maximum improvement is only 1%, while the average accuracies remain roughly the same (1.74% lower for NR questions and 0.3% higher for NC questions). These results further support that our findings on relations are valid and that LVLMs indeed do not understand relations even with different types of relations. F.2.4. CONSISTENCY: PROMPT VARIATION Preparation. We follow the settings in Li et al. (2024b) to construct ICL prompts. We randomly select 4 examples and concatenate these diagrams as well as questions and answers as few-shot examples (represented by image). Then, we modify the template to adapt to the ICL examples under the setting of Co T prompting for evaluation. See Figs. 23 and 24 for the prompting templates and demonstration examples. Accuracy (%) Co T Co T with ICL (a) Q(E, [KF, NR]) Co T Co T with ICL (b) Q(E, [KF, NC]) Figure 7: Accuracies of LVLMs with 4 ICL examples (with Co T prompting). Results are also consistent with our previous findings in 2.3. Results. We observe that ICL does not help with the relation identification (Fig. 7a) and reasoning (Fig. 7b) at all. Overall the average scores decrease, dropping 3% for NR questions and dropping 6% on NC questions. The findings further support that LVLMs can neither identify nor reason about relations even when provided a few examples with answers in context. F.3. Real Diagram We provide the supplementary results for evaluations on real diagrams. Do Vision-Language Models Really Understand Visual Language? F.3.1. PRIOR KNOWLEDGE DOES NOT HELP QS(E|KF,NR) QS(E|KF,NC) QR(V |KF,NR) QR(V |KR,NR) QR(V |KF,NC) QR(V |KR,NC) QR(E|KF,NR) QR(E|KR,NR) QR(E|KF,NC) QR(E|KR,NC) Accuracy (%) of GPT-4o (Co T) Original No-Prior Figure 8: Test accuracies of GPT-4o under the Co T settings on both synthetic and real diagrams for the original system message (i.e., Original ) and no-prior knowledge-required system message (i.e., No-Prior ). Results show that asking the model do not use prior knowledge could not help the model better perform the tasks. To test if the prior knowledge in the model affect the performance, we adjust our system prompts to clearly let the model ignore its prior knowledge when answering the questions. The original system message (i.e., prompt instructions) as well as the new one are listed below. Original system message: You are a visual assistant answering multiple choice questions about diagrams. Read the question, inspect the diagram, and answer with the correct choice in the following format: A) 0 . New system message that asks the model to ignore its prior knowledge: You are a visual assistant answering multiple choice questions about diagrams. Read the question, only inspect the diagram but do not use your prior knowledge, and answer with the correct choice in the following format: A) 0 . We test GPT-4o model with these two prompts under the Co T setting. Fig. 8 presents the test accuracies on both system messages are more or less the same, while our original one can achieve slightly better overall performance. F.3.2. DIAGRAM DISTRIBUTION ON ENTITY NUMBER 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 18 14 10 10 5 6 1 2 1 1 0 1 Entity Number Figure 9: The distribution of the number of diagrams with respect to the number of entities in them. For our 1, 001 real diagrams, we annotate diagrams with the number of entities in them. Thus, we visualize the distribution in Fig. 9. We can find that it is similar to the long-tail distribution, and most diagrams have 3 10 entities. Do Vision-Language Models Really Understand Visual Language? 4 5-6 7-8 9-10 11 GPT-4V (Co T) GPT-4o (Co T) Gemini (Co T) (a) QR(V |KF, NR) 4 5-6 7-8 9-10 11 GPT-4V (Co T) GPT-4o (Co T) Gemini (Co T) (b) QR(V |KF, NC) 4 5-6 7-8 9-10 11 GPT-4V (Co T) GPT-4o (Co T) Gemini (Co T) (c) QR(V |KR, NR) 4 5-6 7-8 9-10 11 GPT-4V (Co T) GPT-4o (Co T) Gemini (Co T) (d) QR(V |KR, NC) 4 5-6 7-8 9-10 11 GPT-4V (Co T) GPT-4o (Co T) Gemini (Co T) (e) QR(E|KF, NR) 4 5-6 7-8 9-10 11 GPT-4V (Co T) GPT-4o (Co T) Gemini (Co T) (f) QR(E|KF, NC) 4 5-6 7-8 9-10 11 GPT-4V (Co T) GPT-4o (Co T) Gemini (Co T) (g) QR(E|KR, NR) 4 5-6 7-8 9-10 11 GPT-4V (Co T) GPT-4o (Co T) Gemini (Co T) (h) QR(E|KR, NC) Figure 10: Accuracies of LVLMs on all questions with respect to the number of entities in the diagram. Do Vision-Language Models Really Understand Visual Language? F.3.3. SUPPLEMENTARY RESULTS FOR STATISTICAL ANALYSIS We provide all the detailed accuracies for three models on QA with respect to the number of entities (Fig. 10). Generally, all these three models have similar tendencies, and the accuracy tendencies are similar to their average as in Fig. 2. G. Examples of Prompts and Responses G.1. Synthetic Diagrams G.1.1. ENTITY Prompting templates and demonstration examples for text entities and visual entities are shown in Figs. 11 to 16. G.1.2. RELATION Prompting templates and demonstration examples of explicit relations E (i.e., arrows) and implicit relations EP (i.e., spatial relations) are shown in Figs. 17 to 20. G.1.3. DIAGRAM GENERATION Prompting templates and demonstration examples of synthetic diagrams with different arrow features are shown in Figs. 21 and 22. G.1.4. IN-CONTEXT LEARNING (ICL) PROMPTING TEMPLATE Four-shot ICL Prompting templates and demonstration examples are shown in Figs. 23 and 24. G.1.5. DIAGRAMS WITH SEMANTIC KNOWLEDGE Prompting templates and demonstration examples of synthetic diagrams containing semantic knowledge are shown in Figs. 25 and 26. G.2. Real Diagrams Prompting templates and demonstration examples of real diagrams (Astronomy - Solar System; Satellite System) for different question types are shown in Figs. 27 to 34. Do Vision-Language Models Really Understand Visual Language? QS(V |KF, NR) Question Prompt: Think step by step before answering the question and show your reasoning. Which one of the entities exists in the diagram? A) bus B) hexagon C) horse D) piggy bank GPT-4V: The diagram contains labels for various objects. I will look for the entities listed in the multiplechoice options within the diagram. A) bus - This label is not present in the diagram. B) hexagon - There is no label for a hexagon in the diagram. C) horse - The label horse is not visible in the diagram. D) piggy bank - The label piggy bank is present in the top left corner of the diagram. The correct answer is: D) piggy bank GPT-4o: To determine which entity exists in the diagram, I will examine each option and check if it is present in the diagram. A) bus - There is no mention of a bus in the diagram. B) hexagon - There is no mention of a hexagon in the diagram. C) horse - There is no mention of a horse in the diagram. D) piggy bank - The term piggy bank is present in the diagram. Therefore, the correct answer is: D) piggy bank Gemini 1.5 Pro: The diagram contains the word chicken . The answer is: C) chicken. Figure 11: An existence question example for text entity. Do Vision-Language Models Really Understand Visual Language? QS(V |KF, NC) Question Prompt: Think step by step before answering the question and show your reasoning. How many text labels are there in the diagram? A) 2 B) 3 C) 4 D) 5 GPT-4V: To determine the number of text labels in the diagram, I will count each instance of text: 1. turkey 2. bird 3. chair There are three text labels in total. The correct answer is: B) 3 GPT-4o: To determine the number of text labels in the diagram, let s count each distinct text label present: 1. turkey 2. bird 3. chair There are three text labels in the diagram. Therefore, the correct answer is: B) 3 Gemini 1.5 Pro: The text labels in this image are turkey , bird , and chair . There are 3 text labels. **C) 3** Figure 12: A count question example for text entity. Do Vision-Language Models Really Understand Visual Language? QS(V |KF, NR) Question Prompt: Think step by step before answering the question and show your reasoning. Which one of the entities exists in the diagram? A) lion B) tree house C) fence D) hay bale GPT-4V: To answer the question, I will inspect each item in the diagram and compare it to the entities listed in the options: - A) lion: There is no image of a lion in the diagram. - B) tree house: The image in the top right corner resembles a structure with multiple windows and a roof, which could be interpreted as a tree house. - C) fence: There is no image of a fence in the diagram. - D) hay bale: There is no image that resembles a hay bale in the diagram. Based on the inspection, the entity that exists in the diagram that most closely resembles one of the options is the tree house. B) tree house GPT-4o: To determine which entity exists in the diagram, let s examine each option and compare it to the images present in the diagram: A) Lion - There is no image of a lion in the diagram. B) Tree house - There is no image of a tree house in the diagram. C) Fence - There is no image of a fence in the diagram. D) Hay bale - There is no image of a hay bale in the diagram. The images present in the diagram are: 1. Sushi 2. A cauldron 3. A building 4. A table tennis paddle and ball None of these images correspond to a lion, tree house, fence, or hay bale. Therefore, none of the entities listed in the options exist in the diagram. Gemini 1.5 Pro: The image contains a fence. C) fence. Figure 13: An existence question example for the visual entity. Do Vision-Language Models Really Understand Visual Language? QS(V |KF, NC) Question Prompt: Think step by step before answering the question and show your reasoning. How many icons are there in the diagram? A) 3 B) 7 C) 4 D) 9 GPT-4V: To determine the number of icons in the diagram, I will count each distinct icon: 1. The first icon at the top left appears to be a loaf of bread. 2. The second icon at the top middle is a keyboard. 3. The third icon at the top right is a shopping cart with books. 4. The fourth icon at the bottom left is a bag of coffee beans. 5. The fifth icon at the bottom middle is a pair of tongs. 6. The sixth icon at the bottom right is a pair of flip-flops. 7. The seventh icon in the middle is a sign with arrows pointing up. There are a total of 7 icons in the diagram. The correct answer is: B) 7 GPT-4o: To determine the number of icons in the diagram, let s count each distinct icon: 1. Bread icon (top left) 2. Keyboard icon (top center) 3. Book icon (top right) 4. Milk and cookies icon (bottom left) 5. Fork icon (center) 6. Road sign icon (bottom center) 7. Sandals icon (bottom right) There are 7 distinct icons in the diagram. Therefore, the correct answer is: B) 7 Gemini 1.5 Pro: Let s count the icons in the image: bread, keyboard, book, fork, sign, sandals, milk and cookies. There are 8 icons. As 8 is not an option, the closest one is 9. D) 9 Figure 14: A count question example for the visual entity. Do Vision-Language Models Really Understand Visual Language? QS(V |KF, NR) Question Prompt: Think step by step before answering the question and show your reasoning. Which one of the text labels exists in the bottom row of the diagram? A) pool B) ice skate C) juice D) ruler GPT-4V: To find the text label that exists in the bottom row of the diagram, I will look at the labels positioned at the bottom of the diagram. There are three labels in the bottom row: paintbrush, juice, and blanket. Now, I will compare these labels with the options provided: A) pool - This label is in the top row of the diagram. B) ice skate - This label is also in the top row of the diagram. C) juice - This label is indeed in the bottom row of the diagram. D) ruler - This label is in the middle row of the diagram. The correct answer is the one that matches a label in the bottom row, which is: C) juice GPT-4o: To determine which text label exists in the bottom row of the diagram, let s examine the positions of the labels: 1. The diagram is divided into a grid with three rows. 2. The bottom row contains three labels: paintbrush, juice, and blanket. Now, let s match these labels with the options provided: A) pool - This label is in the top row. B) ice skate - This label is in the top row. C) juice - This label is in the bottom row. D) ruler - This label is in the middle row. Therefore, the correct answer is: C) juice Gemini 1.5 Pro: The text labels in the bottom row are paintbrush , juice , and blanket . Of these, blanket is one of the choices. So the answer is D) ruler. Figure 15: An existence question example to evaluate the position of entities. Do Vision-Language Models Really Understand Visual Language? QS(V |KF, NC) Question Prompt: Think step by step before answering the question and show your reasoning. How many text labels are there in the top row of the diagram? A) 3 B) 1 C) 2 D) 0 GPT-4V: To determine the number of text labels in the top row of the diagram, I will inspect the top row from left to right. - The first label on the left is candy. - The second label in the middle is paper. - There is no third label in the top row. Therefore, there are two text labels in the top row of the diagram. C) 2 GPT-4o: To determine the number of text labels in the top row of the diagram, follow these steps: 1. Identify the top row of the diagram. The top row is the uppermost horizontal section of the grid. 2. Count the text labels within this top row. In the top row, we see the following text labels: - candy - paper - castle There are three text labels in the top row. Therefore, the correct answer is: A) 3 Gemini 1.5 Pro: The top row contains the words candy , paper , and castle . There are 3 words in the top row. A) 3. Figure 16: A count question example to evaluate the position of entities. Do Vision-Language Models Really Understand Visual Language? QS(E|KF, NR) Question Prompt: Think step by step before answering the question and show your reasoning. Which one of the text labels is placed on the right of the entity fishing pole? A) hummingbird B) nail C) van D) square GPT-4V: To find the text label that is placed on the right of the entity labeled fishing pole, I will look at the diagram and locate the fishing pole label. Then, I will identify which label is directly to its right. Upon inspecting the diagram, I see that the fishing pole label is located towards the left side of the image. To the right of the fishing pole label, there is the label van. Therefore, the correct answer is: C) van GPT-4o: To determine which text label is placed to the right of the entity fishing pole, let s analyze the diagram step by step: 1. Locate the fishing pole label in the diagram. It is positioned towards the left side of the diagram. 2. Identify the labels that are to the right of the fishing pole label. From the diagram: - van is above and to the right of fishing pole. - finger puppet is directly to the right of fishing pole. - fork is further to the right and slightly below fishing pole. - square is below and to the left of fishing pole. Now, let s match the options with the labels: A) hummingbird - Not present in the diagram. B) nail - Not present in the diagram. C) van - Present in the diagram, above and to the right of fishing pole. D) square - Present in the diagram, but to the left of fishing pole. The correct answer is: C) van Gemini 1.5 Pro: The text label finger puppet is on the right of the entity fishing pole . The answer is **D) square**. Figure 17: An existence question example from the Spatial Relation subtask. Do Vision-Language Models Really Understand Visual Language? QS(E|KF, NC) Question Prompt: Think step by step before answering the question and show your reasoning. How many text labels are placed on the left of the entity seal? A) 2 B) 0 C) 3 D) 1 GPT-4V: To determine how many text labels are placed on the left of the entity seal, I will look at the positions of the text labels relative to the word seal in the diagram. 1. The text label hockey ball is placed above and to the left of the word seal. 2. The text label sauce is placed below the word seal and does not appear to be on the left or right, but rather directly below. Based on the positions of the text labels, there is only one text label ( hockey ball ) that is placed on the left of the entity seal. D) 1 GPT-4o: To determine how many text labels are placed on the left of the entity seal, let s examine the diagram step by step: 1. Identify the position of the entity seal in the diagram. 2. Look for any text labels that are positioned to the left of seal. From the diagram: - seal is located towards the left side of the diagram. - The text labels hockey ball and sauce are both positioned to the right of seal. There are no text labels to the left of seal. Therefore, the correct answer is: B) 0 Gemini 1.5 Pro: The text label seal is on the left of the entity seal . There are no other text labels on the left of seal . Therefore, the answer is D) 1. D) 1 Figure 18: A count question example from the Spatial Relation subtask. Do Vision-Language Models Really Understand Visual Language? QS(E|KF, NR) Question Prompt: Think step by step before answering the question and show your reasoning. Which one of the pairs are connected in the diagram? A) gorilla and sandwich B) penguin and sandwich C) award ribbon and sandwich D) award ribbon and penguin GPT-4V: To determine which pairs are connected in the diagram, let s look at each pair in the multiple-choice options: A) gorilla and sandwich - There is no direct line connecting the gorilla and the sandwich in the diagram. B) penguin and sandwich - There is no direct line connecting the penguin and the sandwich in the diagram. C) award ribbon and sandwich - There is no direct line connecting the award ribbon and the sandwich in the diagram. D) award ribbon and penguin - There is a direct line connecting the award ribbon and the penguin in the diagram. Based on the connections shown in the diagram, the correct answer is: D) award ribbon and penguin GPT-4o: To determine which pairs are connected in the diagram, we need to look at the arrows and see which items are directly linked by them. 1. **Gorilla and Sandwich**: There is no direct arrow connecting gorilla and sandwich . 2. **Penguin and Sandwich**: There is a direct arrow from penguin to sandwich . 3. **Award Ribbon and Sandwich**: There is no direct arrow connecting award ribbon and sandwich . 4. **Award Ribbon and Penguin**: There is no direct arrow connecting award ribbon and penguin . Based on the diagram, the correct answer is: B) penguin and sandwich Gemini 1.5 Pro: The award ribbon points to the gorilla and the penguin points to the gorilla. D) award ribbon and penguin Figure 19: An existence question example from the Explicit Relation subtask. Do Vision-Language Models Really Understand Visual Language? QS(E|KF, NC) Question Prompt: Think step by step before answering the question and show your reasoning. How many entities are connected to ring? A) 6 B) 1 C) 3 D) 2 GPT-4V: To determine how many entities are connected to ring, I will inspect the diagram and count the lines that connect directly to the word ring. Upon inspection, I can see that there are three lines connecting to ring. These lines connect ring to stop light, circle, and pencil. Therefore, the correct answer is: C) 3 GPT-4o: To determine how many entities are connected to ring, we need to count the number of arrows pointing to or from ring. 1. There is an arrow connecting ring to stop light. 2. There is an arrow connecting ring to circle. 3. There is an arrow connecting ring to pencil. 4. There is an arrow connecting ring to train. 5. There is an arrow connecting ring to birthday candle. 6. There is an arrow connecting ring to tree house. So, there are 6 entities connected to ring. The correct answer is: A) 6 Gemini 1.5 Pro: Ring is connected to stop light and pencil. Therefore, there are 2 entities connected to ring. **D) 2** Figure 20: A count question example from the Explicit Relation subtask. Do Vision-Language Models Really Understand Visual Language? QS(E|KF, NR) Question Prompt:Think step by step before answering the question and show your reasoning. Which one of the pairs are connected in the diagram? A) arrow and bird B) bird and domino C) arrow and river D) filing cabinet and river GPT-4V: To determine which pairs are connected in the diagram, let s look at each pair in the multiple-choice options: A) arrow and bird - There is no line connecting the arrow to the bird. B) bird and domino - There is no line connecting the bird to the domino. C) arrow and river - There is a line connecting the arrow to the river. D) filing cabinet and river - There is no line connecting the filing cabinet to the river. Based on the connections shown in the diagram, the correct answer is: C) arrow and river GPT-4o: To determine which pairs are connected in the diagram, let s examine each pair one by one: A) Arrow and bird: - The arrow labeled arrow is pointing towards the dog . - The bird is connected to the domino . - Therefore, the arrow and bird are not connected. B) Bird and domino: - The bird is connected to the domino with a direct line. - Therefore, the bird and domino are connected. C) Arrow and river: - The arrow is pointing towards the dog . - The river is not connected to the arrow . - Therefore, the arrow and river are not connected. D) Filing cabinet and river: - The filing cabinet is connected to the dog and dice . - The river is not connected to the filing cabinet . - Therefore, the filing cabinet and river are not connected. The correct answer is: B) bird and domino Gemini 1.5 Pro: The arrow points to the dog and the filing cabinet points to the dog. The answer is D) filing cabinet and river. Figure 21: An existence question example from the Explicit Relation subtask for a different arrow feature. Do Vision-Language Models Really Understand Visual Language? QS(E|KF, NC) Question Prompt: Think step by step before answering the question and show your reasoning. How many entities are connected to apricot? A) 4 B) 0 C) 3 D) 2 GPT-4V: To determine how many entities are connected to apricot, I will count the lines that connect directly to the word apricot in the diagram. 1. There is one line connecting apricot to rubber duck. 2. There is another line connecting apricot to sugar. 3. There is a third line connecting apricot to parachute. No other lines are connected to apricot. Therefore, there are three entities connected to apricot. C) 3 GPT-4o: To determine how many entities are connected to apricot, we need to count the number of arrows pointing to and from apricot. 1. There is an arrow from apricot to rubber duck. 2. There is an arrow from apricot to sugar. 3. There is an arrow from sugar to apricot. 4. There is an arrow from parachute to apricot. So, apricot is connected to four entities: rubber duck, sugar, sugar (again, but counted once), and parachute. Therefore, the correct answer is: A) 4 Gemini 1.5 Pro: Apricot has connections to rubber duck, parachute, and sugar. There are 3 connections in total. Therefore, the answer is **C) 3**. Figure 22: A count question example from the Explicit Relation subtask for a different arrow feature. Do Vision-Language Models Really Understand Visual Language? QS(E|KF, NR) Question Prompt: Following the in-context examples in the first image, answer the following question about the second image. Think step by step before answering the question and show your reasoning. Which one of the pairs are connected in the diagram? A) clock tower and flower B) swim fin and flower C) clock tower and piano D) flower and piano GPT-4V: To answer the question, I will examine the second image to determine which pairs of items are connected by lines. - The clock tower is connected to the piano by a line. ... Based on these observations, the correct answer is: B) swim fin and flower GPT-4o: To determine which pairs are connected in the diagram, let s examine the connections: 1. **Clock tower**: - Connected to **piano**. ... Since we need to find the correct pair, the answer is: B) Swim fin and flower Gemini 1.5 Pro: The swim fin and clock tower are connected by a line. The answer is **C) clock tower and piano**. Figure 23: An existence question example from the Explicit Relation subtask with few-shot examples. Do Vision-Language Models Really Understand Visual Language? QS(E|KF, NC) Question Prompt: Following the in-context examples in the first image, answer the following question about the second image. Think step by step before answering the question and show your reasoning. How many entities are connected to rectangle? A) 0 B) 3 C) 2 D) 1 GPT-4V: In the second image, the entity labeled rectangle is connected to two other entities: pie and motorcycle. Therefore, the correct answer to the question is: C) 2 GPT-4o: To determine how many entities are connected to the rectangle in the second image, follow these steps: 1. Identify the rectangle in the diagram. 2. Count the number of arrows pointing to or from the rectangle. Upon inspection: - The rectangle has two arrows pointing to it from motorcycle. Therefore, the number of entities connected to the rectangle is 2. The correct answer is: C) 2 Gemini 1.5 Pro: The entities connected to rectangle are motorcycle and salad. Motorcycle is connected with a single line, and salad is connected with a line that goes through motorcycle. Therefore, there are **2** entities connected to rectangle. **C) 2** Figure 24: A count question example from the Explicit Relation subtask with few-shot examples. Do Vision-Language Models Really Understand Visual Language? QS(E|KR, NR) Question Prompt:Think step by step before answering the question and show your reasoning. Which one of the pairs are connected in the diagram? A) butterfly and skunk B) bug and skunk C) snake and goldfish D) caterpillar and snake GPT-4V: To determine which pair of words are connected in the diagram, I will look at each pair in the multiple-choice options and see if there is a line connecting them in the diagram. A) butterfly and skunk - There is no line directly connecting butterfly and skunk in the diagram. B) bug and skunk - There is no line directly connecting bug and skunk in the diagram. C) snake and goldfish - There is no line directly connecting snake and goldfish in the diagram. D) caterpillar and snake - There is a line directly connecting caterpillar and snake in the diagram. The correct answer is: D) caterpillar and snake GPT-4o: To determine which pair is connected in the diagram, we need to follow the arrows between the pairs mentioned in the options. A) Butterfly and skunk: - There is no direct arrow connecting butterfly and skunk. B) Bug and skunk: - There is no direct arrow connecting bug and skunk. C) Snake and goldfish: - There is no direct arrow connecting snake and goldfish. D) Caterpillar and snake: - There is a direct arrow from caterpillar to snake. Therefore, the correct answer is: D) caterpillar and snake Gemini 1.5 Pro: The caterpillar has an arrow pointing to the butterfly. The butterfly has an arrow pointing to the snake. This means the caterpillar and snake are connected. D) caterpillar and snake **D) caterpillar and snake** Figure 25: An existence question example for a diagram generated from a knowledge graph. Do Vision-Language Models Really Understand Visual Language? QS(E|KR, NC) Question Prompt: Think step by step before answering the question and show your reasoning. How many entities are connected to doll? A) 2 B) 0 C) 1 D) 3 GPT-4V: To determine how many entities are connected to doll, let s look at the diagram: - There is one line connecting doll to toy. There are no other lines connecting doll to any other entities in the diagram. Therefore, doll is connected to only one entity. The correct answer is: C) 1 GPT-4o: To determine how many entities are connected to doll, let s examine the diagram step by step: 1. Identify the entities in the diagram: cylinder, rectangle, doll, and toy. 2. Look for arrows pointing to or from doll. From the diagram: - There is one arrow pointing from toy to doll. Therefore, doll is connected to one entity. The correct answer is: C) 1 Gemini 1.5 Pro: Doll is connected to toy. There is an arrow pointing from toy to doll. Therefore, the answer is C) 1. Figure 26: A count question example for a diagram generated from a knowledge graph. Do Vision-Language Models Really Understand Visual Language? QR(V |KR, NR) Question Prompt: Think step by step before answering the question and show your reasoning. Which astronomical object is in the diagram? A) earth B) moon C) belt D) jupiter GPT-4V: The diagram shows a central yellow sphere representing the Sun and a blue-green sphere in four different positions around it, representing the Earth at various points in its orbit. The diagram includes labels for the equinoxes and solstices, which are events that occur due to the Earth s tilt and orbit around the Sun. The Earth is depicted with a tilt, and the different positions show how the sunlight hits the Earth differently throughout the year, leading to the changing seasons. Given the options provided and the context of the diagram... The correct answer is: A) earth GPT-4o: To determine which astronomical object is in the diagram, let s analyze the elements present: 1. The diagram shows a central yellow object, which is likely the Sun. 2. Surrounding the Sun, there are four blue and green spheres with landmasses that resemble continents on Earth. 3. The diagram labels specific dates: March 20 or 21 (equinox), June 20 or 21 (solstice), September 22 or 23 (equinox), and December 21 or 22 (solstice). These dates correspond to significant points in Earth s orbit around the Sun, marking the equinoxes and solstices. Given these observations, the object in the diagram is Earth. Therefore, the correct answer is: A) earth Gemini 1.5 Pro: The diagram shows the earth revolving around the sun. So the answer is A) Figure 27: A knowledge-required entity recognition question example for a real diagram. Do Vision-Language Models Really Understand Visual Language? QR(V |KR, NC) Question Prompt: Think step by step before answering the question and show your reasoning. How many planets or satellites are in the diagram? A) 3 B) 8 C) 10 D) 4 GPT-4V: The diagram shows four distinct celestial bodies: one is the Sun, and the other three are representations of Earth at different times of the year, indicating the position of Earth relative to the Sun during the equinoxes and solstices. These are not three different planets or satellites, but rather one planet, Earth, shown at three different points in its orbit around the Sun. Therefore, the answer to how many planets or satellites are in the diagram is: D) 4 GPT-4o: To determine the number of planets or satellites in the diagram, let s carefully analyze the image step by step: 1. **Identify the objects in the diagram:** - There is a central yellow object, which represents the Sun. - There are four blue and green objects surrounding the Sun, which represent the Earth at different positions in its orbit around the Sun. 2. **Count the distinct objects:** - The Sun is one object. - The Earth is shown in four different positions, but it is still the same planet. 3. **Summarize the count:** - Total number of distinct objects: 1 (Sun) + 1 (Earth) = 2. Since the question asks for the number of planets or satellites, and we only have the Earth (a planet) and the Sun (which is a star, not a planet or satellite), the correct count is 1 planet. Therefore, the correct answer is: D) 4 Gemini 1.5 Pro: There are four images of the Earth and one of the Sun. The Earth is a planet and the Sun is a star. A) 4 Figure 28: A knowledge-required entity reasoning question example for a real diagram. Do Vision-Language Models Really Understand Visual Language? QR(E|KR, NR) Question Prompt: Think step by step before answering the question and show your reasoning. What is the next stage after the earth in summer in the diagram? A) full moon B) perigee C) earth in fall D) earth in june GPT-4V: The diagram shows the Earth at different points in its orbit around the Sun, corresponding to the seasons. The Earth in summer is labeled June 20 or 21 (solstice). Following the direction of the arrows, the next stage after the Earth in summer would be the Earth in fall. The Earth in fall is labeled September 22 or 23 (equinox). Therefore, the correct answer is: C) earth in fall GPT-4o: To determine the next stage after the Earth in summer in the diagram, let s follow these steps: 1. Identify the current position of the Earth in summer. The diagram shows that the Earth in summer is labeled as June 20 or 21 (solstice) . 2. Follow the direction of the arrows to see the sequence of the Earth s positions around the Sun. 3. The arrow from June 20 or 21 (solstice) points to September 22 or 23 (equinox) . Therefore, the next stage after the Earth in summer is earth in fall . The correct answer is: C) earth in fall Gemini 1.5 Pro: The diagram shows the earth at different points in its orbit around the sun. The earth in summer (June 20 or 21) is followed by the earth in fall (September 22 or 23). **C) earth in fall** Figure 29: A knowledge-free relation recognition question example for a real diagram. Do Vision-Language Models Really Understand Visual Language? QR(E|KR, NC) Question Prompt: Think step by step before answering the question and show your reasoning. How many times that the planets or satellites can change in the diagram? A) 0 B) 12 C) 4 D) 8 GPT-4V: The diagram shows the Earth at four different points in its orbit around the Sun, corresponding to the equinoxes and solstices. These are the positions where the tilt of Earth s axis is oriented such that the sun s rays are directly over the equator (equinoxes) or over the Tropic of Cancer or Capricorn (solstices), leading to the start of the different seasons. The question seems to be asking how many times the planets or satellites can change positions in the diagram. Since the diagram is specifically showing Earth s position at four distinct times of the year, the planets or satellites (in this case, Earth) change positions four times as it moves through its orbit from one equinox or solstice to the next. Therefore, the answer is: C) 4 GPT-4o: To determine how many times the planets or satellites can change in the diagram, let s analyze the diagram step by step: 1. The diagram shows the Earth at four different positions in its orbit around the Sun, corresponding to the solstices and equinoxes. 2. The positions are: - March 20 or 21 (equinox) - June 20 or 21 (solstice) - September 22 or 23 (equinox) - December 21 or 22 (solstice) 3. The arrows indicate the transition from one position to the next. Each transition from one position to the next represents a change. There are four positions, and each position transitions to the next: - March to June - June to September - September to December - December to March There are 4 transitions in total. Therefore, the number of times the planets or satellites can change in the diagram is: C) 4 Gemini 1.5 Pro: The diagram shows the Earth at four different points in its orbit around the sun. Therefore, the planets or satellites in the diagram can change 4 times. **C) 4** Figure 30: A knowledge-required relation reasoning question example for a real diagram. Do Vision-Language Models Really Understand Visual Language? QR(V |KF, NR) Question Prompt: Think step by step before answering the question and show your reasoning. Which entity is in the diagram? A) earth B) belt C) jupiter D) moon GPT-4V: The diagram shows a central yellow sphere representing the Sun and four different positions of a blue and green sphere representing a planet as it orbits around the Sun. The four positions correspond to different dates and are labeled with specific astronomical events: two equinoxes (March 20 or 21 and September 22 or 23) and two solstices (June 20 or 21 and December 21 or 22)... Given the options provided: A) Earth - This is the correct answer because the diagram depicts Earth s orbit around the Sun and the changing seasons, which are specific to Earth. B) Belt - This is not correct; there is no belt depicted in the diagram. C) Jupiter - This is not correct; the diagram does not represent Jupiter s orbit or its relation to the Sun. D) Moon - This is not correct; the diagram does not show the Moon or its phases. The correct answer is: A) Earth GPT-4o: To determine which entity is in the diagram, let s analyze the elements present: 1. The diagram shows a central yellow circle, which represents the Sun. 2. Surrounding the Sun, there are four blue and green spheres with landmasses that resemble Earth s continents. 3. The labels indicate specific dates related to equinoxes and solstices, which are events that occur due to Earth s orbit around the Sun. Given these observations, the entity in the diagram is: A) earth Gemini 1.5 Pro: The diagram shows the earth at different points in its orbit around the sun. **A)** Figure 31: A knowledge-free entity recognition question example for a real diagram. Do Vision-Language Models Really Understand Visual Language? QR(V |KF, NC) Question Prompt: Think step by step before answering the question and show your reasoning. How many entities are in the diagram? A) 15 B) 5 C) 12 D) 2 GPT-4V: To determine the number of entities in the diagram, let s count each distinct item: 1. The Sun (in the center) 2. Earth at the top (March equinox) 3. Earth on the right (December solstice) 4. Earth at the bottom (September equinox) 5. Earth on the left (June solstice) Adding these up, we have a total of 5 entities. The correct answer is: B) 5 GPT-4o: To determine the number of entities in the diagram, let s identify and count each distinct object: 1. The Sun in the center. 2. Four Earth positions around the Sun, each representing different times of the year (March equinox, June solstice, September equinox, December solstice). So, we have: - 1 Sun - 4 Earth positions Adding these together, we get a total of 5 entities. Therefore, the correct answer is: B) 5 Gemini 1.5 Pro: There are 4 Earth illustrations and 1 Sun illustration. 4 + 1 = 5 **B) 5** Figure 32: A knowledge-free entity reasoning question example for a real diagram. Do Vision-Language Models Really Understand Visual Language? QR(E|KF, NR) Question Prompt: Think step by step before answering the question and show your reasoning. Which entity is connected to the june in the diagram? A) waning crescent B) summer C) waxing crescent D) september GPT-4V: The diagram shows the Earth at four key points in its orbit around the Sun, corresponding to the equinoxes and solstices. The date June 20 or 21 is labeled with (solstice), which refers to the summer solstice in the Northern Hemisphere. This is the time when the Northern Hemisphere is tilted most directly towards the Sun and experiences the longest day of the year, marking the beginning of summer. Therefore, the entity connected to June in the diagram is: B) summer GPT-4o: To determine which entity is connected to June in the diagram, let s analyze the diagram step by step: 1. The diagram shows the Earth s position relative to the Sun at different times of the year. 2. The positions are labeled with specific dates and events... 3. The question asks for the entity connected to June. In the diagram, June 20 or 21 is labeled as a solstice. 4. The solstice in June is known as the summer solstice in the Northern Hemisphere, marking the beginning of summer. Now, let s match this information with the given options: A) waning crescent - This is a phase of the moon, not related to the solstice. B) summer - This is correct, as the June solstice marks the beginning of summer in the Northern Hemisphere. C) waxing crescent - This is a phase of the moon, not related to the solstice. D) September - This is another month, not directly connected to June. Therefore, the correct answer is: B) summer Gemini 1.5 Pro: The arrow points from June to September. **D) september** Figure 33: A knowledge-free relation recognition question example for a real diagram. Do Vision-Language Models Really Understand Visual Language? QR(E|KF, NC) Question Prompt: Think step by step before answering the question and show your reasoning. How many arrows are in the diagram? A) 4 B) 1 C) 13 D) 5 GPT-4V: To determine the number of arrows in the diagram, I will count each arrow that is depicted. Starting from the top right and moving clockwise: 1. There is one large arrow pointing from the globe labeled December 21 or 22 (solstice) to the globe labeled March 20 or 21 (equinox) . 2. Another large arrow points from the globe labeled March 20 or 21 (equinox) to the globe labeled June 20 or 21 (solstice) . 3. A third large arrow points from the globe labeled June 20 or 21 (solstice) to the globe labeled September 22 or 23 (equinox) . 4. Finally, a fourth large arrow points from the globe labeled September 22 or 23 (equinox) back to the globe labeled December 21 or 22 (solstice) . There are no other arrows visible in the diagram. Therefore, the total count is four arrows. The correct answer is: A) 4 GPT-4o: To determine the number of arrows in the diagram, let s carefully count each arrow step by step: 1. Starting from the top Earth (March 20 or 21), there is an arrow pointing to the right towards the Earth labeled December 21 or 22. 2. From the Earth labeled December 21 or 22, there is an arrow pointing downwards towards the Earth labeled September 22 or 23. 3. From the Earth labeled September 22 or 23, there is an arrow pointing to the left towards the Earth labeled June 20 or 21. 4. From the Earth labeled June 20 or 21, there is an arrow pointing upwards towards the Earth labeled March 20 or 21. So, there are a total of 4 arrows in the diagram. The correct answer is: A) 4 Gemini 1.5 Pro: There is one arrow connecting each globe. There are four globes. Therefore, there are four arrows. **A) 4** Figure 34: A knowledge-free relation reasoning question example for a real diagram.