# pixelworld_towards_perceiving_everything_as_pixels__fd96018a.pdf Pixel World: How Far Are We from Perceiving Everything as Pixels? Zhiheng Lyu1,2 Xueguang Ma1 Wenhu Chen1,2 1University of Waterloo 2Vector Institute, Toronto {z63lyu,x93ma,wenhuchen}@uwaterloo.ca Recent agentic language models increasingly need to interact with real-world environments that contain tightly intertwined visual and textual information, often through raw camera pixels rather than separately processed images and tokenized text. This shift highlights the need for a unified perception paradigm. To investigate this idea, we explore Perceive Everything as Pixels (PEAP) and introduce Pixel World, a benchmark that renders natural-language, tabular, mathematical, and diagrammatic inputs into a shared pixel space. Experiments across multiple benchmarks show that PEAP achieves comparable performance to token-based approaches on semantic understanding tasks, suggesting that vision transformers can partially capture global textual semantics without explicit tokenization. In contrast, reasoning-intensive tasks such as mathematics and code show notable performance degradation, although Chain-of-Thought prompting helps mitigate this gap by compensating for missing symbolic structure. We further find that when visual and textual information are closely integrated, representing everything as pixels simplifies preprocessing and avoids cross-modal misalignment. Pixel World thus provides a systematic and practical framework for evaluating unified vision language models and facilitates further exploration of pixel-based multimodal learning. 1 Introduction In recent years, large vision language models (L-VLMs) (Wang et al., 2024a; Open AI, 2025; Gemini, 2024) have achieved remarkable progress across diverse real-world tasks. However, these models still rely on distinct processing pipelines for different modalities treating images as pixels and text as discrete tokens. Such disjoint tokenization introduces a representation mismatch between vision and language, which hampers unified multimodal understanding and complicates system design. As recent works (Zheng et al., 2024; Koh et al., 2024; Tellex et al., 2020; Driess et al., 2023) push toward agentic systems capable of perceiving and acting in complex environments from physical navigation (Elnoor et al., 2024) and travel booking (Chen et al., 2024a) to code repair on Git Hub (Yang et al., 2024) this mismatch becomes increasingly consequential. In such intertwined visual textual settings, maintaining separate tokenization and perception modules not only incurs high preprocessing overhead (Xie et al., 2024; Koh et al., 2024) but also leads to information loss and layout inconsistencies (Dagan et al., 2024; Chai et al., 2024), ultimately limiting the scalability and robustness of multimodal agents. To address these limitations, we explore a unified perception approach: Perceive Everything as Pixels (PEAP). Building on earlier efforts that considered pixel-based representations (Singh et al., 2024; Zhang et al., 2024), we systematically examine how representing both text and visual inputs uniformly in pixel space affects model behavior. In this paradigm, a vision language model (VLM) jointly models multimodal inputs without requiring separate tokenization or modality-specific encoders. To better understand the benefits and challenges of this approach, we introduce Pixel World, a comprehensive benchmark suite designed to evaluate how well existing VLMs perform under the PEAP setting. Language Model Text Instruction Output PEAP Framework Tradition Framework Language Model OCR Text Instruction Vision Language Model Figure 1: Overview of the PEAP Framework. PEAP (Perceive Everything as Pixels) unifies text, structural, and visual inputs into a single pixel space, where a Vision Transformer (Vi T) encodes the pixels and a language decoder performs reasoning. Both components are enclosed within the dashed box to indicate that they jointly form a vision language model (VLM). By eliminating modality-specific preprocessing such as OCR and tokenization, PEAP better aligns with human perception and reduces cross-modal engineering overhead. Insight 1: Performance trend via Modality ( 3) PEAP Performance gain (Gemini-Flash) Acc(pixel_input)-Acc(token_input) Text-only Structural Multimodal (MMLUPro, -15.5) (Table Bench, -4.6) (Slides VQA, 34.2) Semantic Data (GLUE, -3.9) (Super GLUE, -6.5) (Table Bench-FC, -2.75) (Slides VQA, 34.2) Coding Data Reasoning Data (MMLUPro, -15.5) (Mathverse, -10.7) Insight 4: Attention Patten Analysis ( 4.1); PEAP-Sparse ( 4.2) Plain PEAP PEAP-Sparse (MBPP, -16.8) (Table Bench-Vis, -11.0) Insight 3: Transferability through Model Scale ( 3) Result from Gemini-Flash Model Scale PEAP Performance gain Acc(pixel_input)-Acc(token_input) (Qwen2VL-7B, -3.7) (Qwen2VL-2B, -21.7) (GPT4o, -0.6) Slides VQA (GPT4o, 29.2) (Qwen2VL-7B, 23.6) (Qwen2VL-2B, 26.1) Insight 2: Degradation via Task Complexity ( 3) Similar Attention Figure 2: Key Findings on the Pixel World Benchmark. Evaluated across text-only, structural, and multimodal settings ( 2, 3), PEAP shows four major insights: (1) Modality Trend: consistent gains on layout-heavy and multimodal tasks such as websites, slides, and documents; (2) Task Complexity: performance degradation on reasoningand code-centric benchmarks (see 3.1 3.2); (3) Transferability by Scale: larger VLMs (e.g., GPT-4o, Gemini-Flash) exhibit smaller pixel token gaps; and (4) Attention and Efficiency: text and image inputs show similar global attention patterns, while the proposed PEAP-Fast reduces up to 80% of computation overhead ( 4.2). In Pixel World, we select 10 widely used benchmarks covering a diverse range of modalities and task scenarios. For each dataset, we construct both traditional token-based and pixel-based (PEAP) input formats using image synthesis and OCR techniques (see Table 1). We then evaluate vision language models of varying scales, from Qwen2VL-2B to GPT-4o. Cross-modal evaluation in Section 3 reveals three overarching insights: Insight 1: In intrinsically multimodal settings such as website rendering, slide comprehension, and document understanding, PEAP consistently mitigates OCR noise and yields stronger performance; Insight 2: For reasoning-intensive tasks such as mathematics and code, pixelization leads to noticeable accuracy drops, though the gap narrows as model capacity increases suggesting that scale plays a key role in enabling cross-modal transfer; and Insight 3: Larger models exhibit more robust instruction-following and long-context reasoning across modalities, whereas smaller models struggle, highlighting the importance of scale-aware training under the pixel-based paradigm. To further interpret these findings, we conduct three complementary analyses. (1) Representation analysis: We visualize the attention patterns of Qwen2VL-7B and observe broadly similar global structures between Dataset Name Size Task Modality Transfer Split GLUE (Wang, 2018) 59,879 Natural language understanding Synthesis test Super GLUE (Sarlin et al., 2020) 19,294 Natural language understanding Synthesis test MMLU-Pro (Wang et al., 2024b) 12,032 Domain knowledge and reasoning Synthesis test ARC (Clark et al., 2018) 3,548 Science question answering Synthesis test GSM8K (Cobbe et al., 2021) 1,319 Math problem solving Synthesis test MBPP (Austin et al., 2021) 757 Programming tasks Synthesis test Table Bench (Wu et al., 2024) 888 Table data understanding and analysis Synthesis test Math Verse (Zhang et al., 2025) 788 Math and visual reasoning Natural test MMMU-Pro (Yue et al., 2024) 1,730 Multimodal reasoning Synthesis test Slides VQA (Tanaka et al., 2023) 2,136 Multimodal question answering OCR test Wiki-SS (Ma et al., 2024) 3,000 Multimodal retrieval question answering OCR train Table 1: Overview of datasets categorized by modality, usage, size, and split. Modality Transfer means the method to adopt the dataset into counterpart modality. For OCR, we adopt the result from the origin datasets. For Wiki SS-QA, since the positive document of the test set is not released, we subsample 3,000 training data points randomly to evaluate. tokenand pixel-based inputs, suggesting that certain aspects of language modeling behavior may transfer into the visual space, though not implying full equivalence. (2) Efficiency optimization: We measure inference latency and find that while PEAP increases computational cost due to larger input resolution, our proposed PEAP-Fast algorithm effectively prunes blank patches, achieving up to 80% speedup with negligible accuracy degradation. (3) Prompt sensitivity: We study prompting strategies and find that Chain-of-Thought (Co T) reasoning yields more consistent gains under the pixel-based representation compared to standard direct prompting, indicating potential synergies between reasoning supervision and visual encoding. In summary, our contributions are as follows: 1. Pixel World Benchmark: We present a unified benchmark that transforms text, structural, and multimodal datasets into pixel space, offering a direct and reproducible framework to evaluate the trade-offs between pixeland token-based modeling. The benchmark and code are publicly released to facilitate standardized comparison and future research on multimodal perception. 2. Task scale insights: Through large-scale evaluation, we show that PEAP improves layout-heavy or intrinsically multimodal tasks (e.g., website and document understanding) while reducing accuracy on reasoningor code-centric tasks. The performance gap consistently narrows with model scale, underscoring the role of capacity in enabling cross-modal transfer. 3. Efficiency & interpretability: We propose PEAP-Fast, an inference-time pruning strategy that removes blank pixel patches, achieving up to a 3 latency reduction with minimal loss in accuracy. Attention visualizations reveal partially shared global structures across modalities, providing an interpretable perspective on how visual encoders approximate token-level reasoning behavior. Several representative datasets covering different skill domains are selected, as shown in Table 1. We primarily utilize the prompts provided by the datasets. If no prompts are available, we apply a default prompt. By default, we employ Direct Prompting; however, for more complex and mathematical datasets such as MBPP (Austin et al., 2021), MMLU-Pro (Wang et al., 2024b), and Math Verse (Zhang et al., 2025), we adopt Chain-of-Thought (Co T) prompting to enhance performance. All evaluations are conducted in a zero-shot Figure 3: The performance of text-only datasets. The comparison is made between text input and synthesized image input. Most models demonstrate comparable performance on language understanding datasets such as Super GLUE, GLUE, and ARC. However, notable performance disparities emerge between text-based input and synthesized image input on mathematical reasoning tasks (e.g., MMLU-Pro, GSM8K) and programming tasks (e.g., MBPP). Phi-3.5-Vision exhibits consistently poor performance across all vision tasks, primarily due to its insufficient instruction-following capabilities. manner to mitigate potential performance degradation caused by the sensitivity of instruction-tuned large models to few-shot prompting. To evaluate both Token-based and Pixel-based methods, we require paired Text-input and Image-input prompts. We adopted modality transfer strategies to reduce reliance on the information modality provided by existing datasets, as detailed in Table 1. For datasets categorized as Text-Only and Structured, all data is originally in plain text format, necessitating image synthesis prior to evaluation. For Multimodal datasets, textual content embedded in images is extracted using OCR, or the textual components provided by the original datasets are directly utilized for evaluation. Notably, the Math Verse dataset (Zhang et al., 2025) inherently includes a Text-Only modality, offering detailed textual descriptions of image-based information. Image Data Synthesis For text-only and structured datasets, we developed an image data synthesis pipeline to generate diverse image inputs for evaluation. Image widths were adaptively adjusted between 512 and 1024 pixels based on text length, with a fixed height of 256 pixels. Font sizes ranged from 15 to 25 points, and padding varied from 5 to 30 pixels. To enhance robustness, we applied various types of noise, including radial, horizontal, vertical, and Multi-Gaussian noise, as well as high-frequency Gaussian noise to simulate distortions commonly introduced by real-world cameras. For structured datasets, such as tables, data was rendered as images using the Python package dataframe_image. Example inputs from different tasks are provided in Appendix A. 3 Experiments In this section, we will detail our baseline, metrics and models. The experimental results will be organized by Text Input , Structued Input and Multimodal Input . Baseline We establish the baseline by using the same VLMs with text-only prompts. To ensure fairness, we employ identical prompts and add the instruction Please follow the instruction in the image when applying PEAP. This ensures that the VLMs can correctly process instructions embedded within images. Ideally, the baseline and PEAP should yield equivalent performance. This comparison helps identify areas for improvement in existing VLMs. Figure 4: The performance of the structured dataset. We report all the subsets for the Table Bench. In the semi setting, questions were presented as text, while tables were rendered as synthetic images. We observed that for tasks involving reasoning (numerical reasoning) and coding (visualization subset), synthetic images yielded inferior performance compared to text. However, for tasks emphasizing semantic understanding, such as data analysis and fact checking, synthetic images achieved performance comparable to or even surpassing text. Additionally, we found that the semi approach often performed worse than either text or synthetic images individually, providing insights into potential limitations and future directions for leveraging visionlanguage models (VLMs). Metrics For QA tasks (Wiki SS-QA, Slides VQA, Table Bench), we use ROUGE-L, which measures the longest common subsequence between prediction and reference to approximate answer overlap. We choose it for convenience and comparability, and expect other semantic metrics (e.g., BERTScore, LLM judges) to show similar trends. For classification benchmarks, including MMLU-Pro, GLUE, Super GLUE, ARC, and Math Verse, we use accuracy, which directly reflects the model s performance in selecting correct options. For GLUE and Super GLUE, we follow their standard evaluation protocols, utilizing task-specific metrics such as Matthews correlation, F1 score, and Pearson correlation. For the code generation task MBPP, we evaluate performance using the pass@1 rate, which measures whether the generated code successfully passes all test cases. For the mathematical reasoning dataset GSM8K, we employ exact match accuracy, as these problems require precise numerical answers. For the visualization subtask of Table Bench, following the original codebase, we treat it as a code generation task and evaluate the correctness of the generated visualizations. Model Selection To validate Pixel World, we selected a diverse set of vision-language models (VLMs) with varying scales to ensure the robustness and generalizability of our findings. It also allowed us to analyze the behavior of models across different sizes. We evaluated several widely used vision-language models (VLMs), including Qwen2VL-2B (Wang et al., 2024a), Phi-3.5-3.2B (Abdin et al., 2024), Qwen2VL7B (Wang et al., 2024a), Gemini-Flash (Gemini, 2024)1, and GPT-4o (Open AI, 2025). 3.1 Text Input Figure 3 reports model accuracy on text-only datasets (e.g., ARC, MMLU-Pro, GLUE, GSM8K, Super GLUE, MBPP). Two major insights emerge: Better Transferability in Larger Models Larger language models (e.g., GPT-4o, Gemini-Flash) exhibit better transferability between text and image-based performance, while smaller models struggle with both transferability and instruction following. For instance, on the ARC dataset, GPT-4o s performance declines by only 0.59 points when transitioning from text to synthetic images, whereas the smaller Qwen2-VL-2B suffers a substantial 21.73-point drop (from approximately 68.61 to 46.88). This trend suggests that more 1We use gemini-1.5-flash-002, which was the latest available version during this study. Figure 5: The performance of the multimodal dataset (MMMU-Pro). We adopt the result reported by the origin paper. We can observe that strong models perform better in PEAP. capable models preserve their reasoning abilities across modalities, while smaller models face greater difficulty. Additionally, smaller models (e.g., Phi-3.5-vision) not only show weaker overall performance on standard benchmarks but also struggle significantly when instructions are presented as images. Their performance consistently lags behind that of larger models, particularly on tasks like MBPP. This supports Insight 3 in Figure 2. Performance Degradation with More Complex Tasks We observe significant drops on benchmarks requiring advanced reasoning, such as mathematical, coding or domain-specific tasks. For example, when moving from text to image inputs on the MMLU-Pro dataset, GPT-4o exhibits a drop of more than 25 points. In contrast, on GLUE and Super GLUE, the decline remains under 5 points. These findings indicate that while existing large models achieve comparable performance between text and visual modalities on simpler tasks, a gap still exists at a deeper level in visual-based and text-based understanding, demonstrating room for improvement in modality adaptation training. 3.2 Structured Input Figure 4 summarizes model performance on four Table Bench subsets: Fact Checking, Data Analysis, Numerical Reasoning, and Visualization. Reasoning Complexity Impacts Performance Fact Checking and Data Analysis show moderate performance drops, as they rely on semantic understanding. In contrast, Numerical Reasoning and Visualization requiring more intricate reasoning and coding exhibit larger declines when switching to synthetic images. Combined with Performance Degradation with More Complex Tasks in Section 3.1, this supports Insight 2 in Figure 2. Smaller Performance Gaps with Structured Data Compared to text-only tasks, structured tasks show smaller performance gaps between text and image inputs. Notably, Qwen2VL-2B even outperforms its textbased results on Fact Checking, suggesting robust visual representations can aid semantic tasks in smaller models. Challenges with Mixed-Modality Inputs The semi format where tables appear as images while questions remain text-based performs worse than either fully text-based or fully image-based formats. This suggests that conventional VQA approaches, which process text and images using separate encoders, may be more susceptible to performance bottlenecks. As multimodal scenarios become increasingly prevalent, PEAP is expected to demonstrate superior performance compared to mixed-modality methods. Figure 6: The performance of the multimodal datasets (except MMMU-Pro). We compare text-only and vision-only subsets in Mathverse, while Slides VQA and Wiki SS-QA are evaluated as VQA tasks. Larger models perform better on text-based tasks with more modalities. GPT-4o tends to generate longer responses in long-context QA, leading to performance degradation on Wiki SS-QA. 3.3 Multimodal Input Figure 6 presents model performance on multimodal datasets, including text-only and vision-only subsets of Mathverse and VQA tasks like Slides VQA and Wiki SS-QA. Results on MMMU-Pro (Figure 5) use reported values from the original paper. Three key observations emerge: Image Inputs Enhance Disambiguation Incorporating images improves performance by reducing ambiguity compared to text-only benchmarks. In Slides VQA, all models outperform their text-only baselines, while in Wiki SS-QA and MMLU-Pro, visual context provides clarifying information, leading to accuracy gains in larger models. Combined with Smaller Performance Gaps with Structured Data in Section 3.2, this supports Insight 1 in Figure 2. Challenges in Complex Reasoning While multimodal inputs aid basic tasks, complex reasoning remains a bottleneck. In Mathverse, visual cues help but fail to support multi-step logical deductions. Even Gemini Flash shows accuracy drops on intricate reasoning tasks. Additionally, Wiki SS-QA poses challenges due to its long-context nature. Smaller models struggle with PEAP, and GPT-4o underperforms in token-based tasks, highlighting difficulties in processing extended contextual dependencies. This aligns with Sections 3.1 and 3.2. Larger Models Benefit More from Multimodal Data Larger models gain more from multimodal inputs. On Slides VQA, Gemini_Flash improves by 34.24 points, compared to Qwen2-VL-7B s 23.55-point boost. This suggests that larger models, with more extensive prior knowledge and advanced architectures, leverage multimodal data more effectively than smaller models. 4 Discussion 4.1 Q1: Does PEAP have the same attention? To examine whether VLMs attend to similar regions when processing textual and image inputs, we visualize the average attention map of the final layer in Qwen2-VL-7B using a heatmap (Figure 7). Specifically, we analyze the model s behavior on a Bool Q example from Super GLUE, comparing its attention patterns under text-based and image-based inference. Similar attention behaviors are observed across different datasets; more examples are shown in Appendix C. Figure 7: Last Layer Attention Heatmap on Qwen2VL-7B between token-based (left) and pixel-based (right) inference. Although the overall attention intensity on image inputs is generally lower, both modalities exhibit highly similar attention patterns. Computation of Heatmaps. The attention heatmaps are computed during greedy decoding by averaging the last-layer attention across tokens. For multi-head attention, we apply a simple mean across heads. Formally, the attention weight for position i is given by: Heatmap(i) = 1 H A(L) h [t, i] , i [s, e) where H denotes the number of attention heads, L is the index of the last layer, A(L) h [t, i] represents the attention weight from token t to token i in head h, and [s, e) corresponds to the decoding range during greedy decoding. Observations. As shown in Figure 7, Qwen2-VL-7B consistently focuses on task-relevant elements such as the question prompt ( will there be a sequel ... ), salient passage keywords (e.g., film , starring , Alice ), and the expected answer format ( Answer: True/False ). This pattern remains stable across both textual and visual representations, suggesting that the model exhibits largely comparable attention behavior regardless of input modality. However, we also observe that certain blank patches in the image-based inputs occasionally receive disproportionately high attention weights, indicating that while the visual encoder aligns closely with the text encoder, it still introduces redundant activations. 4.2 Q2: How to make PEAP more efficient? As a trade-off for generalization, image-based inference often requires significantly more computational resources than text-based inference. This is partly due to the additional overhead from the Vi T backbone and higher redundancy in image tokens. To estimate the performance gap quantitatively, we conducted experiments on Super GLUE (Table 2). The results show that inference latency for image-based inputs can exceed text-based methods by 150% to 250%. Super GLUE Evaluation Results Task Text PEAP PEAP-Fast Bool Q 79.69% 82.11% 80.89% CB 67.70% 40.77% 39.57% COPA 93.00% 91.00% 86.00% Multi RC 65.90% 61.28% 60.80% Re Co RD 12.54% 5.94% 6.08% RTE 82.31% 72.92% 77.26% Wi C 53.29% 55.80% 55.64% WSC 63.46% 65.38% 59.62% Final Score 64.74% 59.40% 58.23% Table 2: Performance of Qwen2VL-7B on Super GLUE dataset by Text, PEAP and PEAP-Fast. We can observe the comparable performance between PEAP and PEAP-Fast. Inference Time (s) Overhead (%) Subset Text PEAP PEAP-Fast PEAP PEAP-Fast Bool Q 369 1,381 906 274.80 145.55 CB 8 22 15 175.00 87.50 COPA 39 38 22 -2.56 -43.59 Multi RC 609 3,861 2,550 534.80 318.71 Re Co RD 7,016 19,012 14,288 171.01 103.72 RTE 68 117 92 72.06 35.29 Wi C 69 224 157 224.64 127.54 WSC 11 36 27 227.27 145.45 Total 8,089 24,690 18,051 205.27 123.19 Table 3: Inference Time (s) of Qwen2VL-7B on Super GLUE dataset with single A100 server by PEAP and PEAP-Fast. We can observe a 82.08% overhead reduce on PEAP-Fast method. Overhead is calculated as the percentage increase in time relative to the text method. To reduce redundancy in visual inputs, we propose PEAP-Fast, which first identifies empty patches via a simple variance-based threshold if the pixel-value variance in a patch is lower than a preset threshold, that patch is treated as empty and is pruned from all attention computations. Crucially, we preserve the original positional embeddings for the remaining tokens, ensuring no loss of spatial layout perception. This strategy aligns with how humans naturally focus on salient regions rather than blank spaces, thereby significantly reducing context length without sacrificing structural information. Testing PEAP-Fast on Super GLUE reveals a minor accuracy drop of only 1.17% (Table 2). More importantly, the average overhead decreases from 205.27% to 123.19%, yielding an 82.98% reduction (Table 3). These results demonstrate that removing empty patches offers substantial computational savings while maintaining strong performance, making imagebased inference more practical for real-world deployments. Attention heatmap between PEAP and PEAPFast are shown in Appendix B. 4.3 Q3: Is PEAP sensitive to the prompting method? In Section 3, results on purely textual synthesis tasks show that image-based inputs consistently underperform text inputs, likely due to dataset domain gaps and weaker instruction following on image representations. To address this, we applied Co T-style prompts to the Super GLUE dataset to enhance cross-domain instruction following (Table 4). Notably, Qwen2VL-7B showed significant improvements in tasks where image input underperformed compared to text input, such as CB and RTE. Overall, Co T prompts improved image input performance by 2.58%, surpassing the 0.3% improvement observed for text input. Metric Direct Co T Improve (Co T - Direct) Text PEAP Text PEAP Text PEAP Bool Q 79.88% 81.71% 81.13% 80.73% 1.25% -0.98% CB 67.70% 34.78% 81.04% 59.57% 13.34% 24.79% COPA 93.00% 87.00% 89.00% 83.00% -4.00% -4.00% Multi RC 65.73% 62.28% 69.08% 60.41% 3.35% -1.87% Re Co RD 12.50% 5.88% 6.37% 4.66% -6.13% -1.22% RTE 82.31% 72.92% 83.03% 77.26% 0.72% 4.34% Wi C 52.82% 54.39% 54.39% 53.92% 1.57% -0.47% WSC 65.38% 61.54% 57.69% 61.54% -7.69% 0.00% Overall 64.92% 57.56% 65.22% 60.14% 0.30% 2.58% Table 4: Comparison of Direct and Co T performance across Text and Image modalities, along with their respective improvements (Co T - Direct), presented as percentages. 5 Related Work Multimodal Large Language Models and Benchmarks Recent progress in multimodal AI has led to the development of models like GPT-4o (Open AI, 2025), Gemini (Gemini, 2024), and Claude-3.5 (Anthropic, 2025), which integrate vision-based training to improve instruction-following capabilities. Benchmarks for these models have evolved from task-specific datasets, such as VQA (Agrawal et al., 2016) and Doc VQA (Mathew et al., 2021), to more comprehensive evaluations, including MMMU-Pro (Yue et al., 2024), MMBench (Liu et al., 2024), and Mega Bench (Chen et al., 2024b). However, most current research focuses on the semantic understanding of visual content, with only a few benchmarks such as Math Verse (Zhang et al., 2025) and MMMU-Pro (Yue et al., 2024) addressing text recognition and comprehension within images. Our work shifts the focus towards evaluating how well large language models understand language through visual input compared to traditional token-based input. Screenshot LMs Recent studies have demonstrated that pretraining on synthetic screenshots can enable vision-language models (VLMs) to achieve performance comparable to that of BERT on language modeling tasks (Lee et al., 2022; Rust et al., 2023; Gao et al., 2024). This approach allows models to better capture text structures without relying on OCR-based methods. Furthermore, our analysis highlights a performance gap between existing VLMs on vision-based tasks and their text-only counterparts, particularly in the absence of relevant pretraining. Interestingly, in certain scenarios, VLMs perform as well as or even better than text-only models, underscoring the potential of this research direction. In the context of document retrieval, recent advancements (Faysse et al., 2024; Ma et al., 2024) have shown that large-scale pretraining on screenshots can outperform traditional OCR-based methods, further reinforcing the advantages of vision-language pretraining. Language Tokenization Tokenization methods, such as Byte Pair Encoding (BPE) (Shibata et al., 1999; Sennrich et al., 2016), are widely used in language modeling, but recent studies suggest that they may not always be optimal. For instance, Mega Byte (Yu et al., 2023) demonstrated that fixed-length tokenization can improve both computational efficiency and cross-modal capabilities. Similarly, BLT (Pagnoni et al., 2024) proposed entropy-based tokenization, while LCM (team et al., 2024) emphasized the benefits of processing higher-level semantic concepts rather than individual tokens. Inspired by these approaches, we explore whether adaptive image patches can effectively infer textual meaning. At a higher level, we investigate the unification of text and image inputs into a shared representation space, enabling reasoning through abstract semantic concepts rather than traditional token-based methods. 6 Conclusion We present Pixel World, a benchmark that renders text, tables, code, and images as pixels, enabling direct evaluation of the Perceive Everything as Pixels (PEAP) paradigm. Experiments yield three main takeaways. (1) Semantic understanding: PEAP performs on par with token-based baselines on sentenceand paragraph-level tasks, and its patch-level attention exhibits similar global structures to token attention, suggesting partial transfer of language modeling behavior into the visual domain. (2) Reasoning: Performance drops on math, logic, and program-repair benchmarks, though Chain-of-Thought prompting mitigates but does not close this gap, highlighting the continued need for explicit reasoning structure. (3) Multimodal perception: Pixel-based inputs outperform OCR pipelines on websites, slides, and documents by preserving spatial context and avoiding recognition noise. To alleviate the higher computational cost of pixel inputs, we propose PEAP-Fast, which prunes blank patches and achieves up to a 3 inference speedup with minimal accuracy loss. Overall, these findings illustrate both the potential and limitations of PEAP, positioning Pixel World as a diagnostic and reproducible benchmark for studying unified vision language representations and guiding future work on efficiency and reasoning in multimodal agents. Marah Abdin, Jyoti Aneja, Hany Awadalla, Ahmed Awadallah, Ammar Ahmad Awan, Nguyen Bach, Amit Bahree, Arash Bakhtiari, Jianmin Bao, Harkirat Behl, Alon Benhaim, Misha Bilenko, Johan Bjorck, Sébastien Bubeck, Martin Cai, Qin Cai, Vishrav Chaudhary, Dong Chen, Dongdong Chen, Weizhu Chen, Yen-Chun Chen, Yi-Ling Chen, Hao Cheng, Parul Chopra, Xiyang Dai, Matthew Dixon, Ronen Eldan, Victor Fragoso, Jianfeng Gao, Mei Gao, Min Gao, Amit Garg, Allie Del Giorno, Abhishek Goswami, Suriya Gunasekar, Emman Haider, Junheng Hao, Russell J. Hewett, Wenxiang Hu, Jamie Huynh, Dan Iter, Sam Ade Jacobs, Mojan Javaheripi, Xin Jin, Nikos Karampatziakis, Piero Kauffmann, Mahoud Khademi, Dongwoo Kim, Young Jin Kim, Lev Kurilenko, James R. Lee, Yin Tat Lee, Yuanzhi Li, Yunsheng Li, Chen Liang, Lars Liden, Xihui Lin, Zeqi Lin, Ce Liu, Liyuan Liu, Mengchen Liu, Weishung Liu, Xiaodong Liu, Chong Luo, Piyush Madan, Ali Mahmoudzadeh, David Majercak, Matt Mazzola, Caio César Teodoro Mendes, Arindam Mitra, Hardik Modi, Anh Nguyen, Brandon Norick, Barun Patra, Daniel Perez-Becker, Thomas Portet, Reid Pryzant, Heyang Qin, Marko Radmilac, Liliang Ren, Gustavo de Rosa, Corby Rosset, Sambudha Roy, Olatunji Ruwase, Olli Saarikivi, Amin Saied, Adil Salim, Michael Santacroce, Shital Shah, Ning Shang, Hiteshi Sharma, Yelong Shen, Swadheen Shukla, Xia Song, Masahiro Tanaka, Andrea Tupini, Praneetha Vaddamanu, Chunyu Wang, Guanhua Wang, Lijuan Wang, Shuohang Wang, Xin Wang, Yu Wang, Rachel Ward, Wen Wen, Philipp Witte, Haiping Wu, Xiaoxia Wu, Michael Wyatt, Bin Xiao, Can Xu, Jiahang Xu, Weijian Xu, Jilong Xue, Sonali Yadav, Fan Yang, Jianwei Yang, Yifan Yang, Ziyi Yang, Donghan Yu, Lu Yuan, Chenruidong Zhang, Cyril Zhang, Jianwen Zhang, Li Lyna Zhang, Yi Zhang, Yue Zhang, Yunan Zhang, and Xiren Zhou. Phi-3 technical report: A highly capable language model locally on your phone, 2024. URL https://arxiv.org/abs/2404.14219. Aishwarya Agrawal, Jiasen Lu, Stanislaw Antol, Margaret Mitchell, C. Lawrence Zitnick, Dhruv Batra, and Devi Parikh. Vqa: Visual question answering, 2016. URL https://arxiv.org/abs/1505.00468. Anthropic. Claude 3.5: A sonnet of progress, 2025. URL https://www.anthropic.com/news/ claude-3-5-sonnet. Accessed: 2025-01-13. Jacob Austin, Augustus Odena, Maxwell Nye, Maarten Bosma, Henryk Michalewski, David Dohan, Ellen Jiang, Carrie Cai, Michael Terry, Quoc Le, et al. Program synthesis with large language models. ar Xiv preprint ar Xiv:2108.07732, 2021. Yekun Chai, Yewei Fang, Qiwei Peng, and Xuhong Li. Tokenization falling short: On subword robustness in large language models. In Findings of the Association for Computational Linguistics: EMNLP 2024, pp. 1582 1599, 2024. Aili Chen, Xuyang Ge, Ziquan Fu, Yanghua Xiao, and Jiangjie Chen. Travelagent: An ai assistant for personalized travel planning. ar Xiv preprint ar Xiv:2409.08069, 2024a. Jiacheng Chen, Tianhao Liang, Sherman Siu, Zhengqing Wang, Kai Wang, Yubo Wang, Yuansheng Ni, Wang Zhu, Ziyan Jiang, Bohan Lyu, Dongfu Jiang, Xuan He, Yuan Liu, Hexiang Hu, Xiang Yue, and Wenhu Chen. Mega-bench: Scaling multimodal evaluation to over 500 real-world tasks, 2024b. URL https://arxiv.org/abs/2410.10563. Peter Clark, Isaac Cowhey, Oren Etzioni, Tushar Khot, Ashish Sabharwal, Carissa Schoenick, and Oyvind Tafjord. Think you have solved question answering? try arc, the ai2 reasoning challenge. ar Xiv:1803.05457v1, 2018. Karl Cobbe, Vineet Kosaraju, Mohammad Bavarian, Mark Chen, Heewoo Jun, Lukasz Kaiser, Matthias Plappert, Jerry Tworek, Jacob Hilton, Reiichiro Nakano, Christopher Hesse, and John Schulman. Training verifiers to solve math word problems. ar Xiv preprint ar Xiv:2110.14168, 2021. Gautier Dagan, Gabriel Synnaeve, and Baptiste Roziere. Getting the most out of your tokenizer for pretraining and domain adaptation. In Forty-first International Conference on Machine Learning, 2024. Danny Driess, Fei Xia, Mehdi SM Sajjadi, Corey Lynch, Aakanksha Chowdhery, Brian Ichter, Ayzaan Wahid, Jonathan Tompson, Quan Vuong, Tianhe Yu, et al. Palm-e: an embodied multimodal language model. In Proceedings of the 40th International Conference on Machine Learning, pp. 8469 8488, 2023. Mohamed Elnoor, Kasun Weerakoon, Gershom Seneviratne, Ruiqi Xian, Tianrui Guan, Mohamed Khalid M Jaffar, Vignesh Rajagopal, and Dinesh Manocha. Robot navigation using physically grounded visionlanguage models in outdoor environments. ar Xiv preprint ar Xiv:2409.20445, 2024. Manuel Faysse, Hugues Sibille, Tony Wu, Bilel Omrani, Gautier Viaud, Céline Hudelot, and Pierre Colombo. Colpali: Efficient document retrieval with vision language models, 2024. URL https://arxiv.org/abs/ 2407.01449. Tianyu Gao, Zirui Wang, Adithya Bhaskar, and Danqi Chen. Improving language understanding from screenshots, 2024. URL https://arxiv.org/abs/2402.14073. Gemini. Gemini: A family of highly capable multimodal models, 2024. URL https://arxiv.org/abs/ 2312.11805. Jing Yu Koh, Robert Lo, Lawrence Jang, Vikram Duvvur, Ming Chong Lim, Po-Yu Huang, Graham Neubig, Shuyan Zhou, Ruslan Salakhutdinov, and Daniel Fried. Visualwebarena: Evaluating multimodal agents on realistic visual web tasks. In ICLR 2024 Workshop on Large Language Model (LLM) Agents, 2024. Kenton Lee, Mandar Joshi, Iulia Turc, Hexiang Hu, Fangyu Liu, Julian Eisenschlos, Urvashi Khandelwal, Peter Shaw, Ming-Wei Chang, and Kristina Toutanova. Pix2struct: Screenshot parsing as pretraining for visual language understanding, 2022. URL https://arxiv.org/abs/2210.03347. Yuan Liu, Haodong Duan, Yuanhan Zhang, Bo Li, Songyang Zhang, Wangbo Zhao, Yike Yuan, Jiaqi Wang, Conghui He, Ziwei Liu, Kai Chen, and Dahua Lin. Mmbench: Is your multi-modal model an all-around player?, 2024. URL https://arxiv.org/abs/2307.06281. Xueguang Ma, Sheng-Chieh Lin, Minghan Li, Wenhu Chen, and Jimmy Lin. Unifying multimodal retrieval via document screenshot embedding. In Yaser Al-Onaizan, Mohit Bansal, and Yun-Nung Chen (eds.), Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, pp. 6492 6505, Miami, Florida, USA, November 2024. Association for Computational Linguistics. doi: 10.18653/v1/2024. emnlp-main.373. URL https://aclanthology.org/2024.emnlp-main.373/. Minesh Mathew, Dimosthenis Karatzas, and C. V. Jawahar. Docvqa: A dataset for vqa on document images, 2021. URL https://arxiv.org/abs/2007.00398. Open AI. Hello gpt-4o, 2025. URL https://openai.com/index/hello-gpt-4o/. Accessed: 2025-01-13. Artidoro Pagnoni, Ram Pasunuru, Pedro Rodriguez, John Nguyen, Benjamin Muller, Margaret Li, Chunting Zhou, Lili Yu, Jason Weston, Luke Zettlemoyer, et al. Byte latent transformer: Patches scale better than tokens. ar Xiv preprint ar Xiv:2412.09871, 2024. Phillip Rust, Jonas F. Lotz, Emanuele Bugliarello, Elizabeth Salesky, Miryam de Lhoneux, and Desmond Elliott. Language modelling with pixels. In The Eleventh International Conference on Learning Representations, 2023. URL https://openreview.net/forum?id=Fk Sp8VW8Rj H. Paul-Edouard Sarlin, Daniel De Tone, Tomasz Malisiewicz, and Andrew Rabinovich. Superglue: Learning feature matching with graph neural networks. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 4938 4947, 2020. Rico Sennrich, Barry Haddow, and Alexandra Birch. Neural machine translation of rare words with subword units. In Katrin Erk and Noah A. Smith (eds.), Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 1715 1725, Berlin, Germany, August 2016. Association for Computational Linguistics. doi: 10.18653/v1/P16-1162. URL https://aclanthology. org/P16-1162/. Yusuxke Shibata, Takuya Kida, Shuichi Fukamachi, Masayuki Takeda, Ayumi Shinohara, Takeshi Shinohara, and Setsuo Arikawa. Byte pair encoding: A text compression scheme that accelerates pattern matching. 1999. Ayush Singh, Mansi Gupta, Shivank Garg, Abhinav Kumar, and Vansh Agrawal. Beyond captioning: Taskspecific prompting for improved vlm performance in mathematical reasoning, 2024. URL https://arxiv. org/abs/2410.05928. Ryota Tanaka, Kyosuke Nishida, Kosuke Nishida, Taku Hasegawa, Itsumi Saito, and Kuniko Saito. Slidevqa: A dataset for document visual question answering on multiple images, 2023. URL https://arxiv.org/ abs/2301.04883. LCM team, Loïc Barrault, Paul-Ambroise Duquenne, Maha Elbayad, Artyom Kozhevnikov, Belen Alastruey, Pierre Andrews, Mariano Coria, Guillaume Couairon, Marta R. Costa-jussà, David Dale, Hady Elsahar, Kevin Heffernan, João Maria Janeiro, Tuan Tran, Christophe Ropers, Eduardo Sánchez, Robin San Roman, Alexandre Mourachko, Safiyyah Saleem, and Holger Schwenk. Large concept models: Language modeling in a sentence representation space, 2024. URL https://arxiv.org/abs/2412.08821. Stefanie Tellex, Nakul Gopalan, Hadas Kress-Gazit, and Cynthia Matuszek. Robots that use language. Annual Review of Control, Robotics, and Autonomous Systems, 3(1):25 55, 2020. Alex Wang. Glue: A multi-task benchmark and analysis platform for natural language understanding. ar Xiv preprint ar Xiv:1804.07461, 2018. Peng Wang, Shuai Bai, Sinan Tan, Shijie Wang, Zhihao Fan, Jinze Bai, Keqin Chen, Xuejing Liu, Jialin Wang, Wenbin Ge, Yang Fan, Kai Dang, Mengfei Du, Xuancheng Ren, Rui Men, Dayiheng Liu, Chang Zhou, Jingren Zhou, and Junyang Lin. Qwen2-vl: Enhancing vision-language model s perception of the world at any resolution. ar Xiv preprint ar Xiv:2409.12191, 2024a. Yubo Wang, Xueguang Ma, Ge Zhang, Yuansheng Ni, Abhranil Chandra, Shiguang Guo, Weiming Ren, Aaran Arulraj, Xuan He, Ziyan Jiang, et al. Mmlu-pro: A more robust and challenging multi-task language understanding benchmark. ar Xiv preprint ar Xiv:2406.01574, 2024b. Xianjie Wu, Jian Yang, Linzheng Chai, Ge Zhang, Jiaheng Liu, Xinrun Du, Di Liang, Daixin Shu, Xianfu Cheng, Tianzhen Sun, et al. Tablebench: A comprehensive and complex benchmark for table question answering. ar Xiv preprint ar Xiv:2408.09174, 2024. Tianbao Xie, Danyang Zhang, Jixuan Chen, Xiaochuan Li, Siheng Zhao, Ruisheng Cao, Toh Jing Hua, Zhoujun Cheng, Dongchan Shin, Fangyu Lei, Yitao Liu, Yiheng Xu, Shuyan Zhou, Silvio Savarese, Caiming Xiong, Victor Zhong, and Tao Yu. Osworld: Benchmarking multimodal agents for open-ended tasks in real computer environments, 2024. John Yang, Carlos E Jimenez, Alex L Zhang, Kilian Lieret, Joyce Yang, Xindi Wu, Ori Press, Niklas Muennighoff, Gabriel Synnaeve, Karthik R Narasimhan, et al. Swe-bench multimodal: Do ai systems generalize to visual software domains? ar Xiv preprint ar Xiv:2410.03859, 2024. Figure 8: An example input of GSM8K dataset, using Direct Prompt. Lili Yu, Dániel Simig, Colin Flaherty, Armen Aghajanyan, Luke Zettlemoyer, and Mike Lewis. Megabyte: Predicting million-byte sequences with multiscale transformers. Advances in Neural Information Processing Systems, 36:78808 78823, 2023. Xiang Yue, Tianyu Zheng, Yuansheng Ni, Yubo Wang, Kai Zhang, Shengbang Tong, Yuxuan Sun, Botao Yu, Ge Zhang, Huan Sun, Yu Su, Wenhu Chen, and Graham Neubig. Mmmu-pro: A more robust multidiscipline multimodal understanding benchmark, 2024. URL https://arxiv.org/abs/2409.02813. Renrui Zhang, Dongzhi Jiang, Yichi Zhang, Haokun Lin, Ziyu Guo, Pengshuo Qiu, Aojun Zhou, Pan Lu, Kai-Wei Chang, Yu Qiao, et al. Mathverse: Does your multi-modal llm truly see the diagrams in visual math problems? In European Conference on Computer Vision, pp. 169 186. Springer, 2025. Ruohong Zhang, Bowen Zhang, Yanghao Li, Haotian Zhang, Zhiqing Sun, Zhe Gan, Yinfei Yang, Ruoming Pang, and Yiming Yang. Improve vision language model chain-of-thought reasoning, 2024. URL https: //arxiv.org/abs/2410.16198. Boyuan Zheng, Boyu Gou, Jihyung Kil, Huan Sun, and Yu Su. Gpt-4v (ision) is a generalist web agent, if grounded. In Forty-first International Conference on Machine Learning, 2024. A Example Input Figure 8 and Figure 9 gives two examples about the vision input. B Attention Heatmap before and after Image Fast Method Figure 10 presents a heatmap comparison between PEAP and PEAP-Fast. PEAP-Fast effectively reduces redundant patches while preserving attention on key regions. C Attention Heatmap Comparison Between Datasets We provide additional attention visualizations on three representative datasets MBPP, MMLU-Pro, and Math Verse to illustrate how attention patterns vary across program synthesis, reasoning, and STEM tasks. Figure 9: An example input of Table Bench dataset, using Direct Prompt. D Broader Impact Statement This work explores a unified pixel-based perception paradigm that eliminates the need for separate text and image tokenization. While such unification can simplify multimodal pipelines and reduce reliance on OCR systems, it also introduces new risks. The computational cost of pixel-based models remains significantly higher than text-based counterparts, which may limit accessibility and increase the carbon footprint of largescale training and deployment. Furthermore, because pixel inputs may contain sensitive visual information, researchers must ensure that data synthesis and collection comply with privacy and ethical standards. On the positive side, the Pixel World benchmark provides a transparent and reproducible foundation for assessing multimodal understanding, encouraging fair comparisons across models and modalities. By highlighting where pixel-based representations succeed and fail, this work aims to guide the community toward more efficient and interpretable multimodal systems, fostering broader exploration of unified perception without compromising ethical or environmental considerations. Figure 10: Last Layer Attention Heatmap on Qwen2VL-7B between PEAP (left) and PEAP-Fast (right). Figure 11: Example 1: Last Layer Attention Heatmap on Qwen2VL-7B from MBPP. Figure 12: Example 2: Last Layer Attention Heatmap on Qwen2VL-7B from MMLU-Pro. Figure 13: Example 3: Last Layer Attention Heatmap on Qwen2VL-7B from Math Verse.