# multimodal_web_navigation_with_instructionfinetuned_foundation_models__bdbd8445.pdf
Published as a conference paper at ICLR 2024
MULTIMODAL WEB NAVIGATION WITH INSTRUCTIONFINETUNED FOUNDATION MODELS
Hiroki Furuta1,2 Kuang-Huei Lee2 Ofir Nachum2 Yutaka Matsuo1
Aleksandra Faust2 Shixiang Shane Gu1,2 Izzeddin Gur2
1The University of Tokyo 2Google Deep Mind furuta@weblab.t.u-tokyo.ac.jp
The progress of autonomous web navigation has been hindered by the dependence on billions of exploratory interactions via online reinforcement learning, and domain-specific model designs that make it difficult to leverage generalization from rich out-of-domain data. In this work, we study data-driven offline training for web agents with vision-language foundation models. We propose an instruction-following multimodal agent, Web GUM, that observes both webpage screenshots and HTML pages and outputs web navigation actions, such as click and type. Web GUM is trained by jointly finetuning an instruction-finetuned language model and a vision encoder with temporal and local perception on a large corpus of demonstrations. We empirically demonstrate this recipe improves the agent s ability of grounded multimodal perception, HTML comprehension, and multi-step reasoning, outperforming prior works by a significant margin. On the Mini Wo B, we improve over the previous best offline methods by more than 45.8%, even outperforming online-finetuned So TA, humans, and GPT-4-based agent. On the Web Shop benchmark, our 3-billion-parameter model achieves superior performance to the existing So TA, Pa LM-540B. Furthermore, Web GUM exhibits strong positive transfer to the real-world planning tasks on the Mind2Web. We also collect 347K high-quality demonstrations using our trained models, 38 times larger than prior work, and make them available to promote future research in this direction.
1 INTRODUCTION
Web navigation is a class of sequential decision making problems where agents interact with web interfaces following user instructions (Shi et al., 2017; Liu et al., 2018; Gur et al., 2019). Common web navigation tasks include, for example, form filling (Diaz et al., 2013), information retrieval (Nogueira & Cho, 2016; Adolphs et al., 2022), or sending emails via a sequence of interactions with computer interface such as click or type (Figure 1). Recently, there has been a growing interest in developing agents to automate these actions and free humans from repetitive interactions (Mazumder & Riva, 2020; Li et al., 2020; Shvo et al., 2021).
Most prior works studied web navigation problems as online RL to learn the optimal action distribution with task-specific models from scratch (Liu et al., 2018; Gur et al., 2019; Jia et al., 2019; Humphreys et al., 2022). However, online RL requires massive trials-and-errors and is often infeasible in practice since the failure in web navigation would result in undesirable consequences; for instance, wrong password may lead to account freeze, and sending email to the wrong person could be problematic in a business scene. In contrast, offline training from the static dataset enables safe development of web agents, but the performance has been sub-optimal compared to those online RL counterparts (Humphreys et al., 2022; Gur et al., 2022). Furthermore, many of the prior works was unable to leverage rich out-of-domain data for generalization, as they usually use specialized models to explicitly handle the hierarchical structures of document object model (DOM) and their dependencies, for example, with LSTM (Gur et al., 2019; 2021), self-attention (Liu et al., 2018), or GNN (Jia et al., 2019). And many of them only output a fixed set of categorical actions (Humphreys et al., 2022), which is unfavorable for truly open-ended web navigation in the real world.
*Work done as Student Researcher at Google.
Published as a conference paper at ICLR 2024
Instruction: Find Gisele s email and forward it to Siana, please.
{'action': 'click', 'ref': '8'} {'action': 'type', 'ref': '59', 'text': 'Siana'} {'action': 'click', 'ref': '50'} {'action': 'click', 'ref': '55'}
Step 2 Step 3 Step 4
Figure 1: Example episode on Mini Wo B++ (Shi et al., 2017; Liu et al., 2018) (email-inbox-forward-nl). The agent clicks the email from the proper sender, and types the correct receiver to forward that email, to satisfy the given instruction (e.g. Find Gisele s email and forward it to Siana, please). Web GUM makes use of both HTML and image screenshot information to adapt a pre-trained instruction-finetuned foundation model to solve challenging web-based tasks.
Recently, foundation models (Bommasani et al., 2021), especially large language models (LLM) (Brown et al., 2020; Chowdhery et al., 2022), have demonstrated superior performance in commonsense, symbolic, arithmetic, and multi-step logical reasoning (Wei et al., 2022b;c; Kojima et al., 2022). These models enable transformative generalization and are capable of solving wide ranges of interactive decision making problems in the wild, including but not limited to task planning in robotics (Huang et al., 2022a;b; Shah et al., 2022; Ahn et al., 2022), board game (Meta Fundamental AI Research Diplomacy Team et al., 2022), web-based retrieval and browser crawling (Nakano et al., 2021; Yao et al., 2022b; Zaheer et al., 2022).
In this work, we leverage pre-trained vision and language foundation models and introduce a competitive offline learning recipe for autonomous web agents: First, we hypothesize that grounded spatial understanding is important for web navigation (Humphreys et al., 2022; Toyama et al., 2021) and thus enables our agent to observe both HTML and screenshots by combining a language model and a Vi T (Dosovitskiy et al., 2020), from semantically rich multimodal tokens that perceive local and temporal information. Second, we observe that web navigation tasks are by nature instructionfollowing and thus base the language model on an instruction-tuned LLM (Wei et al., 2022a; Chung et al., 2022; Ouyang et al., 2022; Iyer et al., 2022) instead of self-supervisedly pre-trained LLMs (Raffel et al., 2020; Brown et al., 2020) as in Gur et al. (2022). Third, we collect a large multimodal corpus, with both HTML and screenshots, to finetune the language model and Vi T jointly. Fourth, our model outputs action in free-form text. These four key pieces together give us a multimodal web agent, which we call Web navigation via Grounded Understanding Models or Web GUM in short. As shown in Figure 1, our model takes in a command for a web-based task via a natural language instruction (e.g., in an email client, Find Gisele s email and forward it to Siana, please.) and uses multimodal observations of the computer interface to complete the task via a sequence of computer actions.
On Mini Wo B++ (Shi et al., 2017; Liu et al., 2018), a simulated web navigation environment benchmark, Web GUM outperforms previous best offline approaches trained with HTML inputs (Gur et al., 2022) by 45.8%, and even the best existing online RL approaches (Humphreys et al., 2022), despite being trained fully offline with much fewer experiences. Web GUM also shows better performance than humans and private-LLM-based agents (Kim et al., 2023; Sun et al., 2023). We perform extensive ablations and analysis in Section 5 to demonstrate Web GUM s advantages in (1) temporal and local multimodal perception, (2) dataset and model size scaling, (3) better HTML understanding, and (4) ability of multi-step reasoning. Web GUM grounds vision and HTML understanding on the computer interface, which is critical for solving multi-step tasks with dynamic page transitions or tasks that require visual contexts, such as booking flights (+50%), shape recognition (+22%), or crawling social media (+21%). Using instruction-finetuned language models (Chung et al., 2022), compared to using vanilla models (Raffel et al., 2020), improves the success rate on Mini Wo B++ by 25%, and is especially adept at handling the unknown composition of the tasks or out-of-distribution HTML inputs synthesized with realistic perturbations. On the Web Shop benchmark (Yao et al., 2022a), we demonstrate that the capability of multi-step reasoning (Wei et al., 2022c) in language models enables better performance than existing state-of-the-art few-shot Pa LM-540B (Yao et al., 2022b; Chowdhery et al., 2022), while our model only has 3 billion parameters. Web GUM exhibits strong positive transfer to the real-world action prediction tasks on the Mind2Web while surpassing
Published as a conference paper at ICLR 2024
T5 Encoder Tansformer T5 Decoder Tansformer
{'action': 'type', 'ref': '5', 'text': 'Olin'} Use the textbox to enter "Olin" and press "Search", then find and click the 7th search result.
Image Observations Action history, Instruction, HTML
Action {'action': click, 'ref': '6'}
Vi T Tokenization & Embedding Mini Wo B++
HTML Tokens Temporal Tokens
Local Tokens
Figure 2: Overview of Web GUM, our multimodal encoder-decoder model. It takes screenshots, action history, instruction, and HTML as inputs. Image observations are embedded to tokens via pre-trained vision transformer (Vi T) (Dosovitskiy et al., 2020). Visual tokens contain rich temporal information from recent H-step (H = 2) and local information from 16 16-size patches. Multimodal language-image tokens are fed into pre-trained T5 encoder-decoder models (Raffel et al., 2020), and are jointly trained to predict executable actions in text formats.
GPT-4. Finally, we collect 347K multimodal expert demonstrations on Mini Wo B++, 38 times larger than the existing unimodal dataset (Liu et al., 2018), and make these publicly available for future research 1. We believe that incorporating foundation models for efficient offline training is a scalable approach towards real-world web automation where online interactions are prohibitively costly.
2 RELATED WORK
Web Navigation Among many proposed benchmarks for autonomous web navigation (Toyama et al., 2021; Burns et al., 2022; Yao et al., 2022a), one of the most inclusive and representative benchmark to test the capability of autonomous agents is Mini Wo B++ (Shi et al., 2017; Liu et al., 2018), which consists of a set of simulated websites with various user instructions from primitive tasks to complex multi-step decision making tasks, such as sending emails or booking flights. Prior works have tried to solve this benchmark using a variety of techniques; Liu et al. (2018) and Gur et al. (2019; 2021) leverage the guidance during online RL from high-level workflow (Liu et al., 2018) or curriculum learning (Gur et al., 2019; 2021), which should be, however, designed per task, and then would not be scalable methods. Other approaches have employed supervised learning (SL) with a large million-scale dataset and following RL-finetuning (Humphreys et al., 2022), or SL with LLM-based agents (Gur et al., 2022). Offline SL agents often suffer from sub-optimal behavior, and online RL with tremendous exploratory experiences has been critical for proficient navigation on the web (Humphreys et al., 2022), which is, however, difficult to conduct in real websites as there is typically no reward signal and interactions are prohibitively costly. As shown in Appendix I, many of these approaches depend on task-specific hierarchical structures of DOM (Jia et al., 2019; He et al., 2020), tailored architectures to encode their dependencies such as LSTM (Gur et al., 2019; 2021), self-attention (Liu et al., 2018), or GNN (Jia et al., 2019), and task-dependent categorical output space (Humphreys et al., 2022), which could not handle open-ended multi-task settings similar to real world, or incorporate pre-trained models. In contrast, we remove such web-specific architectures and convert web navigation into visual question-answering format (text, image text), which allows us to leverage pre-trained foundation models (Chung et al., 2022; Dosovitskiy et al., 2020) as rich prior knowledge on the web, and then to learn the capable agents even with offline training.
Large Language Models for Web Navigation Concurrently, private-LLM-based agents, such as Instruct GPT (text-davinci-003) (Ouyang et al., 2022) and GPT-3.5-turbo, have achieved competitive performance to RL-fintuned models and humans by leveraging a handful of few-shot demonstrations with self-improvement (Kim et al., 2023), code generation (Sun et al., 2023), and structured prompts (Zheng et al., 2023). In contrast, Web GUM focuses on multimodality and finetuning with domain-specific data. With those, we show very competitive performance compared to Pa LM-540B with only 3 billion parameters. Web GUM can also handle long HTML observation tasks, such as book-flight or choose-date-hard, where agents that rely on in-context few-shot learning tend to run out of input tokens. In addition, our models do not requires ad-hoc prompt engineering.
1https://console.cloud.google.com/storage/browser/gresearch/webllm
Published as a conference paper at ICLR 2024
Methods Modality Pre-trained Models Offline Dataset Success Rate
CC-Net (SL) DOM+Image Res Net " 2400K 32.0% Web N-T5 HTML T5-XL " 12K 48.4%
Web GUM (Ours) HTML+Image Flan-T5-Base,Vi T-B16 " 2.8K 61.1% HTML Flan-T5-XL " 401K 88.7% HTML+Image Flan-T5-XL,Vi T-B16 " 401K 94.2%
WGE DOM % 12K+ 64.6% CC-Net (SL+RL) DOM+Image Res Net % 2400K+ 93.5% Human 93.5%
RCI HTML GPT-3.5-turbo ICL 0.1K 90.6% Ada Planner HTML text-davinci-003 ICL 0.1K 92.9% RCI HTML GPT-4 ICL 0.1K 94.0% Synapse HTML GPT-3.5-turbo ICL 0.1K 98.5%
Table 1: Average success rate on Mini Wo B++. We refer to Zheng et al. (2023) for the baseline performances. See Appendix G for the detailed scores. Web GUM outperforms the previous finetuned-LLM with 3B parameters (Gur et al., 2022), which is the best among offline methods, even with 2.8K dataset and Base-size model (310M parameters). Scaling dataset and model size, Web GUM beats the online RL-finetuned state-of-the-art (Humphreys et al., 2022) despite fully offline training, and exceeds humans or LLM-based agents with GPT-4 (Kim et al., 2023). + in Dataset column means extra billions of frames are required during the online RL phase.
In Appendix B, We discuss additional related works on multimodal large-scale models and foundation models for decision making.
3 PRELIMINARIES
We formulate autonomous web navigation as a deterministic sequential decision making problem; composed of a state space S, action space A, deterministic transition function T : S A S, instruction space G, reward function (or episodic success criteria) r : S G A {0, 1}. At each time step t, the agent follows a parameterized policy conditioned on previous states and actions π : S S | {z } t A A | {z } t G A, and transits to the next state: st+1 = T(st, at). This
process continues until the agent reaches the terminal state (e.g. Submit button is clicked) or the max time step is exceeded. An episode is treated as a success if given instruction g is satisfied (i.e. r(st, g, at) = 1), and as a failure if the agent takes a invalid action or reaches a wrong terminal state.
In autonomous web navigation, the state st S is a web page consisting of the raw HTML as a text sequence and a screenshot as an image. Following prior works (Shi et al., 2017; Liu et al., 2018; Gur et al., 2019; 2021), we assume the constraint action space: function(selector, text). function is either click or type, selector is an integer index that can uniquely specify the element, and text is a text input for type function.
Figure 1 presents an example episode of Mini Wo B (Shi et al., 2017), which involves multi-step decision making. To meet the given instruction, the agent clicks an email from the proper sender and types the correct receiver to forward that email. Mini Wo B also has primitive behavioral tasks such as clicking buttons or entering texts. For the examples of Web Shop (Yao et al., 2022a), see Appendix L.
4.1 MULTIMODAL TRANSFORMER MODELS WITH TEMPORAL AND LOCAL PERCEPTION
In this work, we follow Gur et al. (2022) to use T5 (Raffel et al., 2020), an encoder-decoder architecture, for HTML-based web navigation, as its bi-directional nature could be a good fit for the tree structure of HTML and the architecture has been shown to scale well. We combine T5 with a vision transformer (Vi T) (Dosovitskiy et al., 2020) for multimodality as illustrated in Figure 2. Specifically, we use the Vi T to map image observations (screenshots) into image tokens. The Vi T is pre-trained on Image Net-21K classification (Deng et al., 2009). The T5 encoder then consumes both visual and HTML tokens in a unified manner, and the decoder predicts actions in text. See Appendix C for more implementation details.
Published as a conference paper at ICLR 2024
Methods Modality Success Rate
Web GUM HTML 88.7% Web GUM (white) HTML+Image 90.9% Web GUM (random) HTML+Image 92.2% Web GUM HTML+Image 94.2%
choose-date-easy
search-engine
use-spinner
choose-date-medium
tic-tac-toe
guess-number
identify-shape
social-media-all
click-shape
book-flight
choose-list
click-shades
Performance Improvement (%)
Figure 3: (Left) Average success rate with white/random image inputs. The results imply that Web GUM successfully leverages multimodal information from temporal and local perception tokens. (Right) Top-10 performance improvement among Mini Wo B++ by adding image modality to HTML. We subtract the success rates to compute absolute improvement: (SR of Web GUM(HTML+Image)) - (SR of Web GUM(HTML)). Image modality is leveraged for multi-step tasks with dynamic page transitions or tasks that require visual concept understanding (e.g. book-flight or click-shape). See Appendix G and L for the details.
Encoding Temporal and Local Visual Tokens For language models to be aware of task temporal information and local scene recognition, the encoder considers multimodal tokens extracted from a history of patched screenshots (H = 2 steps). Temporal visual tokens contribute to predict the consistent actions in a multi-step tasks. To better extract spatial and semantic information across the local parts of websites, our Vi T encodes one local token per patch rather than global one per image (i.e. CLS-token). We divide an input image into 16 16 patches giving a total of 14 14 (number of patches) 2 (temporal window) = 392 visual tokens. We crop the screenshots of Mini Wo B++ to remove the yellow instruction part, and the image size becomes 160 160. We pad cropped images with white pixels to fit them into 224 224; the default input size for Vi T.
4.2 INSTRUCTION-FINETUNED LARGE LANGUAGE MODELS
We base our language model on Flan-T5 (Chung et al., 2022), an instruction-finetuned T5, as opposed to using a vanilla pre-trained T5 as in Gur et al. (2022). Flan-T5 is finetuned with largescale instruction-following format problems and chain-of-thought examples across a variety of domains, including reasoning or programming. Considering that web navigation is inherently an instruction-following task, we hypothesize that carefully trained instruction-finetuned models could generalize well to enhance the alignment with user instruction and zero-shot reasoning in the webnavigation, interactive decision making context. For the same reason, we also hypothesize that these high-performing instruction-finetuned models enable better sample efficiency and downstream performance, and thus are well-suited for offline learning. We further finetune the Flan-T5 language model and the Vi T vision encoder jointly (Figure 2) on a large corpus of instruction-following multimodal web navigation data, which we describe in Section 4.3. In Section 5, we empirically demonstrate that this instruction-finetuned recipe improves HTML comprehension, multi-step reasoning and decision making significantly.
4.3 LARGE-SCALE DATA COLLECTION WITH LANGUAGE MODEL AGENTS
Recent successes of foundation models are largely powered by internet-scale data (Brown et al., 2020; Radford et al., 2021; Chen et al., 2022; Wang et al., 2023). While large amount of data is critical, for web navigation domain, there is only a small public dataset for Mini Wo B++, consisting of 12K episodes of human demonstration (Liu et al., 2018). Moreover, the dataset only consists of DOM observations and lacks any visual features, which might limit the fine spatial perception of the elements on the page. A large-scale multimodal dataset, including screenshots of websites, is required to build a better navigation policy at scale.
To collect a huge amount of multimodal behavioral dataset on Mini Wo B++, we leverage the finetuned LLM policy from Gur et al. (2022), instead of human demonstrators (Liu et al., 2018; Humphreys et al., 2022). This significantly reduces the cost to construct a new dataset by leveraging the prior success of autonomous agents. We first rollout a LLM policy with 100 episodes per task, which results in a 2.8K successful episodes. Then, we finetune Flan-T5-XL models with this small dataset and run those with 10,000 episodes per task. Lastly, we collect additional 54K demonstrations
Published as a conference paper at ICLR 2024
IN (Temporal)
IN (Temp+Local)
Average Success Rate
2.8K 68K 347K Dataset Size
Base Large XL XXL Model Size
Flan-Pa LM: 72.8
Web N-T5: 49.8
MM-Web N-T5: 55.6
HTML+Image +Visual Token Ablation +SSL Vi T HTML HTML (Decoder-only)
Figure 4: Average success rate of Web GUM with visual perception tokens and Vi T pre-training ablations (left), different dataset size (middle) and model architectures (right). In dataset and model size results, X-axis is a logarithmic scale. (left) While the effects of various pre-trained Vi T with different datasets or self-supervised objectives are marginal, employing both temporal and local perception tokens is critical for the performance. (middle & right) As for both HTML and multimodal models, we could observe the scaling effect: the larger the dataset and model size are, the higher the success rates are. The results also prove that decoder-only Flan-Pa LM-8B is not as good as similar-size encoder-decoder models.
with Synapse (Zheng et al., 2023), a private-LLM-based agents with prompting, for the tasks where the finetuned-LLM may not complete well. Such efforts result in a multi-task dataset with 401K (347+54K) episodes including HTML and screenshots at each step. See Appendix F for more details.
We test our method on Mini Wo B++ (Shi et al., 2017; Liu et al., 2018) with 100 evaluation episodes per task, taking the average success rate over 56 tasks taken from Gur et al. (2022). Table 1 shows that Web GUM, with a small 2.8K dataset and Base-size model (310M parameters), significantly outperforms previous offline methods for web navigation (Humphreys et al., 2022; Gur et al., 2022). While they used 2.4 million episodes or 3 billion parameters, Web GUM could improve the data and parameter efficiency to achieve superior performance in offline regime, which is realized by the problem simplification of web navigation in order to leverage temporal-local visual perception and instruction-finetuned LLMs as strong inductive bias on web environments. In addition, scaling dataset and model size, Web GUM achieves 94.2% success rate2, exceeding the previous best offline model, Web N-T5 (Gur et al., 2022), by over 45.8% and even surpassing the online RL-finetuned So TA, CC-Net (Humphreys et al., 2022) (+0.7%), despite our fully offline training and much fewer data. Moreover, Web GUM surpasses humans and recent LLM-based agents, such as RCI (Kim et al., 2023) and Ada Planner (Sun et al., 2023), even with GPT-4 (Open AI, 2023). The per-task comparison and error analysis (Appendix G, L) imply that there is room for improvement in complex reasoning tasks requiring memory such as guess-number.
In the following sections, we perform extensive and precise ablations of Web GUM to clearly identify the source of improvement. Especially, we will demonstrate the contribution of (1) temporal and local multimodal perception (Section 5.1), architectures and pre-trained models, and (2) dataset and model size scaling (Section 5.2). We will also point out (3) better HTML comprehension (Section 5.3) and (4) capability of multi-step reasoning (Section 5.4) from instruction-finetuned LLMs. Furthermore, we prove that Web GUM can be transferable to the real-world tasks (Section 5.5).
5.1 TEMPORAL AND LOCAL VISUAL PERCEPTION FOR GROUNDED WEB NAVIGATION
To verify the importance of image modality, we design three ablations: (i) input replacement, (ii) removing visual perception tokens, and (iii) employing different pre-trained Vi T. We first replace image observations with completely white images, and with randomly sampled Mini Wo B++ screenshots taken in the initial states at test time. For visual token and pre-trained Vi T ablations, we prepare various pre-trained weights with Image Net-21K (IN) + Aug Reg (Steiner et al., 2022), JFT-300M (Sun et al., 2017), or JFT-3B (Zhai et al., 2022), and with self-supervised objectives such as CLIP (Radford et al., 2021), MAE (He et al., 2021), or DINO (Caron et al., 2021), and then finetune Base-size models as a proxy of larger-size models (Hoffmann et al., 2022) to reduce the computational costs.
2Videos are available at https://sites.google.com/view/mm-webnav/
Published as a conference paper at ICLR 2024
click-link click-button click-link_click-button
Methods Modality Success Rate
Web N-T5 (Gur et al., 2022) HTML 51.0% Synapse (Zheng et al., 2023) HTML 73.8%
Web GUM HTML 74.2% Web GUM HTML+Image 78.5%
Figure 5: (Left) Example of compositional evaluation on Mini Wo B++. We combine two different tasks (click-link and click-button) into a single sequential task (click-link_click-button) at test time (see Appendix H). (Right) Average success rate on 6 compositional Mini Wo B tasks. Web GUM generalizes combinational tasks better than Gur et al. (2022) and Zheng et al. (2023), a So TA LLM-agent in Mini Wo B++.
Add Coordinates Add extra HTML at the bottom Add extra HTML at the top Methods Modality Perturbation Success Rate
Web N-T5 HTML Top 24.7% (Gur et al., 2022) Bottom 42.8% Coordinates 6.4%
Web GUM HTML Top 53.6% Bottom 48.0% Coordinates 39.8%
Web GUM HTML+Image Top 71.8% Bottom 64.7% Coordinates 62.6%
Figure 6: (Left) Example of input perturbation for Mini Wo B++ evaluation, taken from click-button. We prepare three different types of perturbations at test time: adding extra HTML at the top of the original input HTML (left) or at the bottom (middle), and adding task-irrelevant attributes in each element (right) such as coordinate information (left, right, top, bottom). (Right) Average success rate of perturbation evaluation on Mini Wo B++. The results reveal that while all the methods are affected by input corruptions to some extent, Web GUM, especially with multimodality, achieves significantly better performances than previous method.
In Figure 3 (left), the performance of the model with white images is comparable to the unimodal model. Presumably because the model with randomly-taken images may accidentally contain the images from the target task, Web GUM (random) slightly surpasses Web GUM (white). These results prove Web GUM successfully obtains grounded vision and HTML understanding by leveraging temporal and local fine perception. In the visual token ablation, Figure 4 (left) shows that combining both temporal and local visual tokens (66.1%) improves the performance than temporal (64.2%) or local tokens only (64.0%). Interestingly, the effects of different pre-trained Vi T are marginal, compared to visual tokens, which highlights our contribution on designing suitable architecture for multimodal web navigation.
We also compare per-task performance gaps caused by adding vision modality to language models. Figure 3 (right) presents top-10 absolute performance improvement, suggesting Web GUM leverages visual inputs for multi-step tasks with dynamic page transitions (e.g. book-flight; +50%) or tasks requiring visual context understanding (e.g. click-shape; +22%) (see Appendix G and L).
5.2 SCALING EFFECT IN DATASET AND MODEL SIZE
In this section, we show the importance of scaling up the dataset and model size in Web GUM, similar to the observations in the language and vision domain (Shoeybi et al., 2019; Kaplan et al., 2020; Rae et al., 2021; Wei et al., 2022b; Chowdhery et al., 2022). To investigate data scaling, we prepare three dataset: minimal 2.8K demonstrations, 347K demonstrations, and its 20%-size demonstrations (68K), and then finetune Flan-T5-Base with them. Figure 4 (middle) proves that increasing dataset size leads to the improvement of success rate. Because multimodal models benefit from the scaling more, the larger dataset size might be more crucial in multimodal models, which also supports our attempts to construct large-scale multimodal dataset for web navigation. Notably, Base-size Web GUM with 2.8K episodes already achieves 55.7%/66.1%, surpassing previous best SL models (49.8%/55.6% we trained with 347K episodes). This surprising data efficiency comes from the sufficient inductive bias and alignment with the user intentions in instruction-finetuned LLMs.
In addition to dataset size, Figure 4 (right) shows that the performance of Web GUM improves as the number of parameters in T5 model increases from Base (220M) to XXL (11B). These results also reveal that scaling the models might be more important than the dataset; the low-capacity model may cap the performance at a lower level. In contrast, decoder-only Flan-Pa LM-8B only
Published as a conference paper at ICLR 2024
achieves 72.8% success, comparable to Web GUM-Large (770M), which emphasizes the advantage of encoder-decoder models in web navigation. See Appendix D for further details.
5.3 BETTER HTML COMPREHENSION FROM INSTRUCTION-FINETUNED LLMS
We have demonstrated that instruction-finetuned LLMs outperforms vanilla LLMs in web navigation. To analyze the effect of instruction-finetuning more precisely, we here focus on the capability of HTML understanding. Since instruction-finetuned LLMs perform well on many NLP tasks with content comprehension (Chung et al., 2022; Iyer et al., 2022), web navigation should also benefit from them. As a test bed for HTML comprehension, we investigate (1) generalization to unseen compositions of known tasks, and (2) robustness to the realistic input perturbations, which are also important challenges for the web agents to be deployed on the real-world internet. We also provide the base language model comparison on a standard HTML comprehension benchmark, Web SRC (Chen et al., 2021d) in Appendix E, where Flan-T5 achieves better EM/F1 scores than T5 after finetuning.
For the compositional tasks, we pick up 4 clicksomething (link, button, checkboxes, dialog) tasks and make 6 combinations of these by naively stitching with 2 or 3 tasks (e.g. Figure 5). See Appendix H for further details. The results show that Web GUM with HTML and image inputs outperforms prior finetuned-LLM (Gur et al., 2022) and Synapse (Zheng et al., 2023), a So TA LLM agent in Mini Wo B++, which implies Web GUM has obtained better reading skills for web navigation and could transfer them to handle unseen HTML in compositional tasks robustly.
To test the robustness against input corruptions, we test three different realistic perturbations; adding extra HTML at the top or bottom of the original HTML, and adding attributes of coordinates (left, right, top, bottom; they are unrelated to solving the tasks) in each element of HTML at test time. These perturbations often happen in the real world due to the renewal or API changes, not to mention unknown websites, but rule-based pre-processing may not fully cover them. The results show that while all the methods are affected by the input corruptions to some extent, Web GUM, with both HTML and HTML plus image modalities, achieves significantly better performances than Gur et al. (2022). Notably, Web GUM outperforms prior finetuned LLM (+ 56.2% in multimodal and +33.4% in unimodal models) even when extra distracted attributes are added to HTML. They support our hypothesis: instruction-finetuning imporves HTML comprehension in LLMs, which enables the downstream agents to deal with out-of-distribution inputs or tasks robustly.
5.4 ABILITY OF MULTI-STEP REASONING AS A PRIOR FOR INTERACTIVE DECISION MAKING
Methods Training Models Score Success Rate
Rule 45.6 9.6% IL SL BART, BERT 59.9 29.1% IL+RL SL+RL BART, BERT 62.4 28.7% Act In-context Pa LM-540B 62.3 30.1% Re Act In-context Pa LM-540B 66.6 40.0%
Web N-T5 SL T5-XL 61.0 29.8% Web GUM SL Flan-T5-XL 67.5 45.0%
Table 2: Average score and success rate on Web Shop (Yao et al., 2022a). Web GUM achieves 45.0% success, outperforming baseline methods including Re Act, a prompted Pa LM-540B. We refer Yao et al. (2022b) for the baselines.
Another notable feature in instruction-finetuned LLMs is an ability of multi-step reasoning (Chung et al., 2022). We hypothesize this reasoning capability would play an important role as a prior for interactive decision making. To decouple the evaluation of reasoning capability from visual page perception, HTML understanding, and the benchmark simulator (Mini Wo B++), we extensively evaluate our Web GUM on Web Shop (Yao et al., 2022a), another online-shopping website simulator with a large amount of real-world product data. Because it requires complex multi-step decisions considering previous contexts for item comparison, Web Shop is suitable for investigating the capability of multi-step reasoning from instruction-finetuned LLM in depth (Yao et al., 2022a;b). Web Shop provides a user instruction that describes the features of item (e.g. I need a long clip-in hair extension which is natural looking, and price lower than 20.00 dollars). The agents should search, compare and choose a proper product that matches the given instruction. The performance score is evaluated by the percentage of required attributes covered by the chosen product, and if the product meets all the requirements, that episode is labeled a success. See Appendix K for further details.
Table 2 shows that Web GUM achieves 45.0% success, significantly outperforming not only simple baselines, such as supervised imitation learning (IL), IL plus RL-finetuing and Web N-T5 (by more than 15%), but also recent prompt-based LLM agents, including Re Act (Yao et al., 2022b) (i.e. Pa LM-540B (Chowdhery et al., 2022) with one-shot prompt and reasoning annotations), while our
Published as a conference paper at ICLR 2024
Cross-Task Cross-Website Cross-Domain
Train Ele. Acc Op. F1 Step SR SR Ele. Acc Op. F1 Step SR SR Ele. Acc Op. F1 Step SR SR
GPT-4 ICL 41.6 60.6 36.2 2.0 35.8 51.1 30.1 2.0 37.1 46.5 26.4 2.0 Mind Act-Large SL 53.4 75.7 50.3 7.1 39.2 67.1 35.3 1.1 39.7 67.2 37.3 2.7 Mind Act-XL SL 55.1 75.7 52.0 5.2 42.0 65.2 38.9 5.1 42.1 66.5 39.6 2.9
Web GUM-Large (ours) SL 55.3 78.9 51.9 7.5 43.6 70.3 39.3 5.1 42.8 70.6 40.2 2.9 Web GUM-XL (ours) SL 57.2 80.3 53.7 8.5 45.3 70.9 41.6 5.2 43.9 72.2 41.4 3.2
Table 3: Action prediction evaluation in real-world Mind2Web dataset. We adopt the top-50 candidate generation results and direct QA formulation by following Deng et al. (2023). Web GUM, transferred from Mini Wo B, demonstrates superior performance to Mind Act and GPT-4 across task/website/domain generalization.
model only has 3 billion parameters. Due to the consistent reasoning and enhanced alignment with user intentions, Web GUM could compare the products with backtracking, and choose proper options (see Appendix L). Our results imply that ability of multi-step reasoning in Flan-T5 works as strong and transferable prior knowledge for downstream decision making.
5.5 STRONG TRANSFER TO REAL-WORLD ACTION PREDICTION
Lastly, we demonstrate the applicability of Web GUM to real-world problems. We test Web GUM on Mind2Web (Deng et al., 2023), a real-world demonstration dataset with about 2K instructions on 137 websites. In the action prediction tasks, we transfer Web GUM finetuned for Mini Wo B++ with 401K dataset into real-world Mind2Web by further finetuning with the training set. Web GUM takes top-50 relevant HTML snippet candidates, instructions, and action history as inputs and outputs next actions by predicting the element id, operations (e.g. click, type), and values. Table 3 reveals that Web GUM, transferred from Mini Wo B, achieves superior performance to Mind Act-Large/XL and even GPT-4 in all the categories (cross-task/website/domain). Because both Mind Act and Web GUM are based on Flan-T5, these results support that Web GUM exhibits strong positive transfer to real-world tasks.
6 DISCUSSION AND LIMITATION
Throughout the paper, we present an effective and practical methodology to simplify web navigation into offline training in order to leverage the inductive bias of web environments in instruction-finetuned LLMs. While Web GUM exhibits positive transferability to real-world problems in Mind2Web, we leave it as future work to scale multimodal foundation models into the deployment for real-world web navigation (Gur et al., 2023).
We collect and release a multimodal expert dataset with 347K episodes on Mini Wo B++. However, this is still far from internet-scale dataset that is necessary for generalist models. Collecting behavioral data at scale by iterative data-collection and deployment (Ghosh et al., 2021; Matsushima et al., 2021; Li et al., 2022a) might be a key for practical interactive agents. Since our approach taking raw HTML and screenshots as inputs and predicting parsable actions in text only has minimal assumptions which constraint model architectures, it might be applicable to any advanced LLMs or open-ended situations. While Web GUM could deal with out-of-distribution compositional and perturbed tasks in a robust manner, human-level broader generalization to the diverse real websites or instructions is still a hard problem to be resolved.
7 CONCLUSION
We develop Web navigation via Grounded Understanding Models (Web GUM), learning an instructionfollowing visual language foundation model for web navigation. Web GUM significantly improves the success rate on Mini Wo B, compared to previous offline-trained So TA from 48.4% to 94.2%. Our detailed ablations show that temporal and local visual tokens capture dynamic transition and visual context of the page, and that instruction-finetuned language models significantly improves web navigation performance due to the better HTML comprehension and capability of multi-step reasoning. Multi-step reasoning enables more robust generalization to out-of-distribution tasks, and outperforms Pa LM-540B in Web Shop. Web GUM also demonstrates strong positive transfer to real-world action prediction tasks in Mind2Web. Furthermore, we scale the existing Mini Wo B dataset into multimodal 347K expert demonstrations, about 38 times larger than before. We believe that our work is an significant step towards building more capable and scalable models for autonomous web navigation.
Published as a conference paper at ICLR 2024
ACKNOWLEDGEMENTS
HF was supported by JSPS KAKENHI Grant Number JP22J21582. We thank Yusuke Iwasawa, Mustafa Safdari, Austin Huang, Heiga Zen for helpful feedback on this work, and Shunyu Yao for setting up Web Shop experiments.
Leonard Adolphs, Benjamin Boerschinger, Christian Buck, Michelle Chen Huebscher, Massimiliano Ciaramita, Lasse Espeholt, Thomas Hofmann, Yannic Kilcher, Sascha Rothe, Pier Giuseppe Sessa, and Lierni Sestorain Saralegui. Boosting search engines with interactive agents. In Transactions on Machine Learning Research, 2022.
Armen Aghajanyan, Dmytro Okhonko, Mike Lewis, Mandar Joshi, Hu Xu, Gargi Ghosh, and Luke Zettlemoyer. Htlm: Hyper-text pre-training and prompting of language models. ar Xiv preprint ar Xiv:2107.06955, 2021.
Armen Aghajanyan, Bernie Huang, Candace Ross, Vladimir Karpukhin, Hu Xu, Naman Goyal, Dmytro Okhonko, Mandar Joshi, Gargi Ghosh, Mike Lewis, and Luke Zettlemoyer. Cm3: A causal masked multimodal model of the internet. ar Xiv preprint ar Xiv:2201.07520, 2022.
Michael Ahn, Anthony Brohan, Noah Brown, Yevgen Chebotar, Omar Cortes, Byron David, Chelsea Finn, Chuyuan Fu, Keerthana Gopalakrishnan, Karol Hausman, Alex Herzog, Daniel Ho, Jasmine Hsu, Julian Ibarz, Brian Ichter, Alex Irpan, Eric Jang, Rosario Jauregui Ruano, Kyle Jeffrey, Sally Jesmonth, Nikhil J Joshi, Ryan Julian, Dmitry Kalashnikov, Yuheng Kuang, Kuang-Huei Lee, Sergey Levine, Yao Lu, Linda Luu, Carolina Parada, Peter Pastor, Jornell Quiambao, Kanishka Rao, Jarek Rettinghouse, Diego Reyes, Pierre Sermanet, Nicolas Sievers, Clayton Tan, Alexander Toshev, Vincent Vanhoucke, Fei Xia, Ted Xiao, Peng Xu, Sichun Xu, Mengyuan Yan, and Andy Zeng. Do as i can, not as i say: Grounding language in robotic affordances. ar Xiv preprint arxiv:2204.01691, 2022.
Jean-Baptiste Alayrac, Jeff Donahue, Pauline Luc, Antoine Miech, Iain Barr, Yana Hasson, Karel Lenc, Arthur Mensch, Katie Millican, Malcolm Reynolds, Roman Ring, Eliza Rutherford, Serkan Cabi, Tengda Han, Zhitao Gong, Sina Samangooei, Marianne Monteiro, Jacob Menick, Sebastian Borgeaud, Andrew Brock, Aida Nematzadeh, Sahand Sharifzadeh, Mikolaj Binkowski, Ricardo Barreira, Oriol Vinyals, Andrew Zisserman, and Karen Simonyan. Flamingo: a visual language model for few-shot learning. ar Xiv preprint arxiv:2204.14198, 2022.
Rohan Anil, Andrew M. Dai, Orhan Firat, Melvin Johnson, Dmitry Lepikhin, Alexandre Passos, Siamak Shakeri, Emanuel Taropa, Paige Bailey, Zhifeng Chen, Eric Chu, Jonathan H. Clark, Laurent El Shafey, Yanping Huang, Kathy Meier-Hellstern, Gaurav Mishra, Erica Moreira, Mark Omernick, Kevin Robinson, Sebastian Ruder, Yi Tay, Kefan Xiao, Yuanzhong Xu, Yujing Zhang, Gustavo Hernandez Abrego, Junwhan Ahn, Jacob Austin, Paul Barham, Jan Botha, James Bradbury, Siddhartha Brahma, Kevin Brooks, Michele Catasta, Yong Cheng, Colin Cherry, Christopher A. Choquette-Choo, Aakanksha Chowdhery, Clément Crepy, Shachi Dave, Mostafa Dehghani, Sunipa Dev, Jacob Devlin, Mark Díaz, Nan Du, Ethan Dyer, Vlad Feinberg, Fangxiaoyu Feng, Vlad Fienber, Markus Freitag, Xavier Garcia, Sebastian Gehrmann, Lucas Gonzalez, Guy Gur-Ari, Steven Hand, Hadi Hashemi, Le Hou, Joshua Howland, Andrea Hu, Jeffrey Hui, Jeremy Hurwitz, Michael Isard, Abe Ittycheriah, Matthew Jagielski, Wenhao Jia, Kathleen Kenealy, Maxim Krikun, Sneha Kudugunta, Chang Lan, Katherine Lee, Benjamin Lee, Eric Li, Music Li, Wei Li, Ya Guang Li, Jian Li, Hyeontaek Lim, Hanzhao Lin, Zhongtao Liu, Frederick Liu, Marcello Maggioni, Aroma Mahendru, Joshua Maynez, Vedant Misra, Maysam Moussalem, Zachary Nado, John Nham, Eric Ni, Andrew Nystrom, Alicia Parrish, Marie Pellat, Martin Polacek, Alex Polozov, Reiner Pope, Siyuan Qiao, Emily Reif, Bryan Richter, Parker Riley, Alex Castro Ros, Aurko Roy, Brennan Saeta, Rajkumar Samuel, Renee Shelby, Ambrose Slone, Daniel Smilkov, David R. So, Daniel Sohn, Simon Tokumine, Dasha Valter, Vijay Vasudevan, Kiran Vodrahalli, Xuezhi Wang, Pidong Wang, Zirui Wang, Tao Wang, John Wieting, Yuhuai Wu, Kelvin Xu, Yunhan Xu, Linting Xue, Pengcheng Yin, Jiahui Yu, Qiao Zhang, Steven Zheng, Ce Zheng, Weikang Zhou, Denny Zhou, Slav Petrov, and Yonghui Wu. Palm 2 technical report. ar Xiv preprint ar Xiv:2305.10403, 2023.
Published as a conference paper at ICLR 2024
Srikar Appalaraju, Bhavan Jasani, Bhargava Urala Kota, Yusheng Xie, and R. Manmatha. Docformer: End-to-end transformer for document understanding. In International Conference on Computer Vision, 2021.
Jacob Austin, Augustus Odena, Maxwell Nye, Maarten Bosma, Henryk Michalewski, David Dohan, Ellen Jiang, Carrie Cai, Michael Terry, Quoc Le, and Charles Sutton. Program synthesis with large language models. ar Xiv preprint ar Xiv:2108.07732, 2021.
Rishi Bommasani, Drew A. Hudson, Ehsan Adeli, Russ Altman, Simran Arora, Sydney von Arx, Michael S. Bernstein, Jeannette Bohg, Antoine Bosselut, Emma Brunskill, Erik Brynjolfsson, Shyamal Buch, Dallas Card, Rodrigo Castellon, Niladri Chatterji, Annie Chen, Kathleen Creel, Jared Quincy Davis, Dora Demszky, Chris Donahue, Moussa Doumbouya, Esin Durmus, Stefano Ermon, John Etchemendy, Kawin Ethayarajh, Li Fei-Fei, Chelsea Finn, Trevor Gale, Lauren Gillespie, Karan Goel, Noah Goodman, Shelby Grossman, Neel Guha, Tatsunori Hashimoto, Peter Henderson, John Hewitt, Daniel E. Ho, Jenny Hong, Kyle Hsu, Jing Huang, Thomas Icard, Saahil Jain, Dan Jurafsky, Pratyusha Kalluri, Siddharth Karamcheti, Geoff Keeling, Fereshte Khani, Omar Khattab, Pang Wei Kohd, Mark Krass, Ranjay Krishna, Rohith Kuditipudi, Ananya Kumar, Faisal Ladhak, Mina Lee, Tony Lee, Jure Leskovec, Isabelle Levent, Xiang Lisa Li, Xuechen Li, Tengyu Ma, Ali Malik, Christopher D. Manning, Suvir Mirchandani, Eric Mitchell, Zanele Munyikwa, Suraj Nair, Avanika Narayan, Deepak Narayanan, Ben Newman, Allen Nie, Juan Carlos Niebles, Hamed Nilforoshan, Julian Nyarko, Giray Ogut, Laurel Orr, Isabel Papadimitriou, Joon Sung Park, Chris Piech, Eva Portelance, Christopher Potts, Aditi Raghunathan, Rob Reich, Hongyu Ren, Frieda Rong, Yusuf Roohani, Camilo Ruiz, Jack Ryan, Christopher Ré, Dorsa Sadigh, Shiori Sagawa, Keshav Santhanam, Andy Shih, Krishnan Srinivasan, Alex Tamkin, Rohan Taori, Armin W. Thomas, Florian Tramèr, Rose E. Wang, William Wang, Bohan Wu, Jiajun Wu, Yuhuai Wu, Sang Michael Xie, Michihiro Yasunaga, Jiaxuan You, Matei Zaharia, Michael Zhang, Tianyi Zhang, Xikun Zhang, Yuhui Zhang, Lucia Zheng, Kaitlyn Zhou, and Percy Liang. On the opportunities and risks of foundation models. ar Xiv preprint ar Xiv:2108.07258, 2021.
Anthony Brohan, Noah Brown, Justice Carbajal, Yevgen Chebotar, Joseph Dabis, Chelsea Finn, Keerthana Gopalakrishnan, Karol Hausman, Alex Herzog, Jasmine Hsu, Julian Ibarz, Brian Ichter, Alex Irpan, Tomas Jackson, Sally Jesmonth, Nikhil J Joshi, Ryan Julian, Dmitry Kalashnikov, Yuheng Kuang, Isabel Leal, Kuang-Huei Lee, Sergey Levine, Yao Lu, Utsav Malla, Deeksha Manjunath, Igor Mordatch, Ofir Nachum, Carolina Parada, Jodilyn Peralta, Emily Perez, Karl Pertsch, Jornell Quiambao, Kanishka Rao, Michael Ryoo, Grecia Salazar, Pannag Sanketi, Kevin Sayed, Jaspiar Singh, Sumedh Sontakke, Austin Stone, Clayton Tan, Huong Tran, Vincent Vanhoucke, Steve Vega, Quan Vuong, Fei Xia, Ted Xiao, Peng Xu, Sichun Xu, Tianhe Yu, and Brianna Zitkovich. Rt-1: Robotics transformer for real-world control at scale. ar Xiv preprint ar Xiv:2212.06817, 2022.
Tom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel M. Ziegler, Jeffrey Wu, Clemens Winter, Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam Mc Candlish, Alec Radford, Ilya Sutskever, and Dario Amodei. Language models are few-shot learners. ar Xiv preprint ar Xiv:2005.14165, 2020.
Andrea Burns, Deniz Arsan, Sanjna Agrawal, Ranjitha Kumar, Kate Saenko, and Bryan A. Plummer. A dataset for interactive vision-language navigation with unknown command feasibility. In European Conference on Computer Vision, 2022.
Mathilde Caron, Hugo Touvron, Ishan Misra, Hervé Jégou, Julien Mairal, Piotr Bojanowski, and Armand Joulin. Emerging properties in self-supervised vision transformers. ar Xiv preprint ar Xiv:2104.14294, 2021.
Lili Chen, Kevin Lu, Aravind Rajeswaran, Kimin Lee, Aditya Grover, Michael Laskin, Pieter Abbeel, Aravind Srinivas, and Igor Mordatch. Decision transformer: Reinforcement learning via sequence modeling. In Advances in Neural Information Processing Systems, 2021a.
Mark Chen, Jerry Tworek, Heewoo Jun, Qiming Yuan, Henrique Ponde de Oliveira Pinto, Jared Kaplan, Harri Edwards, Yuri Burda, Nicholas Joseph, Greg Brockman, Alex Ray, Raul Puri,
Published as a conference paper at ICLR 2024
Gretchen Krueger, Michael Petrov, Heidy Khlaaf, Girish Sastry, Pamela Mishkin, Brooke Chan, Scott Gray, Nick Ryder, Mikhail Pavlov, Alethea Power, Lukasz Kaiser, Mohammad Bavarian, Clemens Winter, Philippe Tillet, Felipe Petroski Such, Dave Cummings, Matthias Plappert, Fotios Chantzis, Elizabeth Barnes, Ariel Herbert-Voss, William Hebgen Guss, Alex Nichol, Alex Paino, Nikolas Tezak, Jie Tang, Igor Babuschkin, Suchir Balaji, Shantanu Jain, William Saunders, Christopher Hesse, Andrew N. Carr, Jan Leike, Josh Achiam, Vedant Misra, Evan Morikawa, Alec Radford, Matthew Knight, Miles Brundage, Mira Murati, Katie Mayer, Peter Welinder, Bob Mc Grew, Dario Amodei, Sam Mc Candlish, Ilya Sutskever, and Wojciech Zaremba. Evaluating large language models trained on code. ar Xiv preprint ar Xiv:2107.03374, 2021b.
Tao Chen, Jie Xu, and Pulkit Agrawal. A system for general in-hand object re-orientation. In Conference on Robot Learning, 2021c.
Xi Chen, Xiao Wang, Soravit Changpinyo, AJ Piergiovanni, Piotr Padlewski, Daniel Salz, Sebastian Goodman, Adam Grycner, Basil Mustafa, Lucas Beyer, Alexander Kolesnikov, Joan Puigcerver, Nan Ding, Keran Rong, Hassan Akbari, Gaurav Mishra, Linting Xue, Ashish Thapliyal, James Bradbury, Weicheng Kuo, Mojtaba Seyedhosseini, Chao Jia, Burcu Karagol Ayan, Carlos Riquelme, Andreas Steiner, Anelia Angelova, Xiaohua Zhai, Neil Houlsby, and Radu Soricut. Pali: A jointlyscaled multilingual language-image model. ar Xiv preprint arxiv:2209.06794, 2022.
Xingyu Chen, Zihan Zhao, Lu Chen, Jia Bao Ji, Danyang Zhang, Ao Luo, Yuxuan Xiong, and Kai Yu. Web SRC: A dataset for web-based structural reading comprehension. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp. 4173 4185, 2021d.
Aakanksha Chowdhery, Sharan Narang, Jacob Devlin, Maarten Bosma, Gaurav Mishra, Adam Roberts, Paul Barham, Hyung Won Chung, Charles Sutton, Sebastian Gehrmann, Parker Schuh, Kensen Shi, Sasha Tsvyashchenko, Joshua Maynez, Abhishek Rao, Parker Barnes, Yi Tay, Noam Shazeer, Vinodkumar Prabhakaran, Emily Reif, Nan Du, Ben Hutchinson, Reiner Pope, James Bradbury, Jacob Austin, Michael Isard, Guy Gur-Ari, Pengcheng Yin, Toju Duke, Anselm Levskaya, Sanjay Ghemawat, Sunipa Dev, Henryk Michalewski, Xavier Garcia, Vedant Misra, Kevin Robinson, Liam Fedus, Denny Zhou, Daphne Ippolito, David Luan, Hyeontaek Lim, Barret Zoph, Alexander Spiridonov, Ryan Sepassi, David Dohan, Shivani Agrawal, Mark Omernick, Andrew M. Dai, Thanumalayan Sankaranarayana Pillai, Marie Pellat, Aitor Lewkowycz, Erica Moreira, Rewon Child, Oleksandr Polozov, Katherine Lee, Zongwei Zhou, Xuezhi Wang, Brennan Saeta, Mark Diaz, Orhan Firat, Michele Catasta, Jason Wei, Kathy Meier-Hellstern, Douglas Eck, Jeff Dean, Slav Petrov, and Noah Fiedel. Palm: Scaling language modeling with pathways. ar Xiv preprint ar Xiv:2204.02311, 2022.
Hyung Won Chung, Le Hou, Shayne Longpre, Barret Zoph, Yi Tay, William Fedus, Eric Li, Xuezhi Wang, Mostafa Dehghani, Siddhartha Brahma, Albert Webson, Shixiang Shane Gu, Zhuyun Dai, Mirac Suzgun, Xinyun Chen, Aakanksha Chowdhery, Sharan Narang, Gaurav Mishra, Adams Yu, Vincent Zhao, Yanping Huang, Andrew Dai, Hongkun Yu, Slav Petrov, Ed H. Chi, Jeff Dean, Jacob Devlin, Adam Roberts, Denny Zhou, Quoc V. Le, and Jason Wei. Scaling instruction-finetuned language models. ar Xiv preprint arxiv:2210.11416, 2022.
Wenliang Dai, Junnan Li, Dongxu Li, Anthony Meng Huat Tiong, Junqi Zhao, Weisheng Wang, Boyang Li, Pascale Fung, and Steven Hoi. Instructblip: Towards general-purpose vision-language models with instruction tuning. ar Xiv preprint ar Xiv:2305.06500, 2023.
Deep Mind Interactive Agents Team, Josh Abramson, Arun Ahuja, Arthur Brussee, Federico Carnevale, Mary Cassin, Felix Fischer, Petko Georgiev, Alex Goldin, Mansi Gupta, Tim Harley, Felix Hill, Peter C Humphreys, Alden Hung, Jessica Landon, Timothy Lillicrap, Hamza Merzic, Alistair Muldal, Adam Santoro, Guy Scully, Tamara von Glehn, Greg Wayne, Nathaniel Wong, Chen Yan, and Rui Zhu. Creating multimodal interactive agents with imitation and self-supervised learning. ar Xiv preprint ar Xiv:2112.03763, 2021.
Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In Conference on Computer Vision and Pattern Recognition, 2009.
Xiang Deng, Yu Gu, Boyuan Zheng, Shijie Chen, Samuel Stevens, Boshi Wang, Huan Sun, and Yu Su. Mind2web: Towards a generalist agent for the web. ar Xiv preprint ar Xiv:2306.06070, 2023.
Published as a conference paper at ICLR 2024
Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. Bert: Pre-training of deep bidirectional transformers for language understanding. ar Xiv preprint ar Xiv:1810.04805, 2019.
Oscar Diaz, Itziar Otaduy, and Gorka Puente. User-driven automation of web form filling. In International Conference on Web Engineering, 2013.
Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. ar Xiv preprint ar Xiv:2010.11929, 2020.
Linxi Fan, Guanzhi Wang, Yunfan Jiang, Ajay Mandlekar, Yuncong Yang, Haoyi Zhu, Andrew Tang, De-An Huang, Yuke Zhu, and Anima Anandkumar. Minedojo: Building open-ended embodied agents with internet-scale knowledge. ar Xiv preprint ar Xiv:2206.08853, 2022.
Hiroki Furuta, Yusuke Iwasawa, Yutaka Matsuo, and Shixiang Shane Gu. A system for morphology-task generalization via unified representation and behavior distillation. ar Xiv preprint ar Xiv:2211.14296, 2022a.
Hiroki Furuta, Yutaka Matsuo, and Shixiang Shane Gu. Generalized decision transformer for offline hindsight information matching. In International Conference on Learning Representations, 2022b.
Dibya Ghosh, Abhishek Gupta, Ashwin Reddy, Justin Fu, Coline Manon Devin, Benjamin Eysenbach, and Sergey Levine. Learning to reach goals via iterated supervised learning. In International Conference on Learning Representations, 2021.
Shixiang Shane Gu, Manfred Diaz, Daniel C. Freeman, Hiroki Furuta, Seyed Kamyar Seyed Ghasemipour, Anton Raichuk, Byron David, Erik Frey, Erwin Coumans, and Olivier Bachem. Braxlines: Fast and interactive toolkit for rl-driven behavior engineering beyond reward maximization. ar Xiv preprint ar Xiv:2110.04686, 2021a.
Xiuye Gu, Tsung-Yi Lin, Weicheng Kuo, and Yin Cui. Open-vocabulary object detection via vision and language knowledge distillation. ar Xiv preprint arxiv:2104.13921, 2021b.
Izzeddin Gur, Ulrich Rueckert, Aleksandra Faust, and Dilek Hakkani-Tur. Learning to navigate the web. In International Conference on Learning Representations, 2019.
Izzeddin Gur, Natasha Jaques, Yingjie Miao, Jongwook Choi, Manoj Tiwari, Honglak Lee, and Aleksandra Faust. Environment generation for zero-shot compositional reinforcement learning. In Advances in neural information processing systems, 2021.
Izzeddin Gur, Ofir Nachum, Yingjie Miao, Mustafa Safdari, Austin Huang, Aakanksha Chowdhery, Sharan Narang, Noah Fiedel, and Aleksandra Faust. Understanding html with large language models. ar Xiv preprint arxiv:2210.03945, 2022.
Izzeddin Gur, Hiroki Furuta, Austin Huang, Mustafa Safdari, Yutaka Matsuo, Douglas Eck, and Aleksandra Faust. A real-world webagent with planning, long context understanding, and program synthesis. ar Xiv preprint arxiv:2307.12856, 2023.
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Conference on Computer Vision and Pattern Recognition, 2016.
Kaiming He, Xinlei Chen, Saining Xie, Yanghao Li, Piotr Dollár, and Ross Girshick. Masked autoencoders are scalable vision learners. ar Xiv preprint ar Xiv:2111.06377, 2021.
Zecheng He, Srinivas Sunkara, Xiaoxue Zang, Ying Xu, Lijuan Liu, Nevan Wichers, Gabriel Schubiner, Ruby Lee, Jindong Chen, and Blaise Agüera y Arcas. Actionbert: Leveraging user actions for semantic understanding of user interfaces. ar Xiv preprint ar Xiv:2012.12350, 2020.
Jordan Hoffmann, Sebastian Borgeaud, Arthur Mensch, Elena Buchatskaya, Trevor Cai, Eliza Rutherford, Diego de Las Casas, Lisa Anne Hendricks, Johannes Welbl, Aidan Clark, Tom Hennigan, Eric Noland, Katie Millican, George van den Driessche, Bogdan Damoc, Aurelia Guy, Simon Osindero, Karen Simonyan, Erich Elsen, Jack W. Rae, Oriol Vinyals, and Laurent Sifre. Training compute-optimal large language models. ar Xiv preprint ar Xiv:2203.15556, 2022.
Published as a conference paper at ICLR 2024
Wenlong Huang, Pieter Abbeel, Deepak Pathak, and Igor Mordatch. Language models as zero-shot planners: Extracting actionable knowledge for embodied agents. ar Xiv preprint ar Xiv:2201.07207, 2022a.
Wenlong Huang, Fei Xia, Ted Xiao, Harris Chan, Jacky Liang, Pete Florence, Andy Zeng, Jonathan Tompson, Igor Mordatch, Yevgen Chebotar, Pierre Sermanet, Noah Brown, Tomas Jackson, Linda Luu, Sergey Levine, Karol Hausman, and Brian Ichter. Inner monologue: Embodied reasoning through planning with language models. ar Xiv preprint arxiv:2207.05608, 2022b.
Peter C Humphreys, David Raposo, Toby Pohlen, Gregory Thornton, Rachita Chhaparia, Alistair Muldal, Josh Abramson, Petko Georgiev, Alex Goldin, Adam Santoro, and Timothy Lillicrap. A data-driven approach for learning to control computers. In International Conference on Machine Learning, 2022.
Srinivasan Iyer, Xi Victoria Lin, Ramakanth Pasunuru, Todor Mihaylov, Daniel Simig, Ping Yu, Kurt Shuster, Tianlu Wang, Qing Liu, Punit Singh Koura, Xian Li, Brian O Horo, Gabriel Pereyra, Jeff Wang, Christopher Dewan, Asli Celikyilmaz, Luke Zettlemoyer, and Ves Stoyanov. Opt-iml: Scaling language model instruction meta learning through the lens of generalization. ar Xiv preprint ar Xiv:2212.12017, 2022.
Michael Janner, Qiyang Li, and Sergey Levine. Offline reinforcement learning as one big sequence modeling problem. In Advances in Neural Information Processing Systems, 2021.
Sheng Jia, Jamie Ryan Kiros, and Jimmy Ba. DOM-q-NET: Grounded RL on structured language. In International Conference on Learning Representations, 2019.
Yunfan Jiang, Agrim Gupta, Zichen Zhang, Guanzhi Wang, Yongqiang Dou, Yanjun Chen, Li Fei-Fei, Anima Anandkumar, Yuke Zhu, and Linxi Fan. Vima: General robot manipulation with multimodal prompts. ar Xiv preprint ar Xiv:2210.03094, 2022.
Aishwarya Kamath, Mannat Singh, Yann Le Cun, Gabriel Synnaeve, Ishan Misra, and Nicolas Carion. Mdetr modulated detection for end-to-end multi-modal understanding. ar Xiv preprint ar Xiv:2104.12763, 2021.
Jared Kaplan, Sam Mc Candlish, Tom Henighan, Tom B. Brown, Benjamin Chess, Rewon Child, Scott Gray, Alec Radford, Jeffrey Wu, and Dario Amodei. Scaling laws for neural language models. ar Xiv preprint ar Xiv:2001.08361, 2020.
Geunwoo Kim, Pierre Baldi, and Stephen Mc Aleer. Language models can solve computer tasks. ar Xiv preprint arxiv:2303.17491, 2023.
Takeshi Kojima, Shixiang Shane Gu, Machel Reid, Yutaka Matsuo, and Yusuke Iwasawa. Large language models are zero-shot reasoners. In Advances In Neural Information Processing Systems, 2022.
Taku Kudo and John Richardson. Sentencepiece: A simple and language independent subword tokenizer and detokenizer for neural text processing. ar Xiv preprint ar Xiv:1808.06226, 2018.
Kenton Lee, Mandar Joshi, Iulia Turc, Hexiang Hu, Fangyu Liu, Julian Eisenschlos, Urvashi Khandelwal, Peter Shaw, Ming-Wei Chang, and Kristina Toutanova. Pix2struct: Screenshot parsing as pretraining for visual language understanding. ar Xiv preprint ar Xiv:2210.03347, 2022a.
Kuang-Huei Lee, Ofir Nachum, Mengjiao Yang, Lisa Lee, Daniel Freeman, Winnie Xu, Sergio Guadarrama, Ian Fischer, Eric Jang, Henryk Michalewski, and Igor Mordatch. Multi-game decision transformers. ar Xiv preprint arxiv:2205.15241, 2022b.
Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Ves Stoyanov, and Luke Zettlemoyer. Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. ar Xiv preprint ar Xiv:1910.13461, 2019.
Bo Li, Yuanhan Zhang, Liangyu Chen, Jinghao Wang, Jingkang Yang, and Ziwei Liu. Otter: A multi-modal model with in-context instruction tuning. ar Xiv preprint ar Xiv:2305.03726, 2023.
Published as a conference paper at ICLR 2024
Chenliang Li, Bin Bi, Ming Yan, Wei Wang, Songfang Huang, Fei Huang, and Luo Si. Structurallm: Structural pre-training for form understanding. ar Xiv preprint arxiv:2105.11210, 2021a.
Junlong Li, Yiheng Xu, Lei Cui, and Furu Wei. Markuplm: Pre-training of text and markup language for visually-rich document understanding. ar Xiv preprint arxiv:2110.08518, 2021b.
Peizhao Li, Jiuxiang Gu, Jason Kuen, Vlad I. Morariu, Handong Zhao, Rajiv Jain, Varun Manjunatha, and Hongfu Liu. Selfdoc: Self-supervised document representation learning. In Conference on Computer Vision and Pattern Recognition, 2021c.
Shuang Li, Xavier Puig, Chris Paxton, Yilun Du, Clinton Wang, Linxi Fan, Tao Chen, De-An Huang, Ekin Akyürek, Anima Anandkumar, Jacob Andreas, Igor Mordatch, Antonio Torralba, and Yuke Zhu. Pre-trained language models for interactive decision-making. In Advances In Neural Information Processing Systems, 2022a.
Yang Li, Jiacong He, Xin Zhou, Yuan Zhang, and Jason Baldridge. Mapping natural language instructions to mobile ui action sequences. In Annual Conference of the Association for Computational Linguistics, 2020.
Yujia Li, David Choi, Junyoung Chung, Nate Kushman, Julian Schrittwieser, Remi Leblond, Tom Eccles, James Keeling, Felix Gimeno, Agustin Dal Lago, Thomas Hubert, Peter Choy, Cyprien de, Masson d Autume, Igor Babuschkin, Xinyun Chen, Po-Sen Huang, Johannes Welbl, Sven Gowal, Alexey Cherepanov, James Molloy, Daniel J. Mankowitz, Esme Sutherland Robson, Pushmeet Kohli, Nando de Freitas, Koray Kavukcuoglu, and Oriol Vinyals. Competition-level code generation with alphacode, 2022b.
Evan Zheran Liu, Kelvin Guu, Panupong Pasupat, and Percy Liang. Reinforcement learning on web interfaces using workflow-guided exploration. In International Conference on Learning Representations, 2018.
Haotian Liu, Chunyuan Li, Qingyang Wu, and Yong Jae Lee. Visual instruction tuning. ar Xiv preprint ar Xiv:2304.08485, 2023.
Jiasen Lu, Christopher Clark, Rowan Zellers, Roozbeh Mottaghi, and Aniruddha Kembhavi. Unifiedio: A unified model for vision, language, and multi-modal tasks. ar Xiv preprint ar Xiv:2206.11795, 2022.
Tatsuya Matsushima, Hiroki Furuta, Yutaka Matsuo, Ofir Nachum, and Shixiang Gu. Deploymentefficient reinforcement learning via model-based offline optimization. In International Conference on Learning Representations, 2021.
Sahisnu Mazumder and Oriana Riva. Flin: A flexible natural language interface for web navigation. ar Xiv preprint ar Xiv:2010.12844, 2020.
Meta Fundamental AI Research Diplomacy Team, Anton Bakhtin, Noam Brown, Emily Dinan, Gabriele Farina, Colin Flaherty, Daniel Fried, Andrew Goff, Jonathan Gray, Hengyuan Hu, Athul Paul Jacob, Mojtaba Komeili, Karthik Konath, Minae Kwon, Adam Lerer, Mike Lewis, Alexander H. Miller, Sasha Mitts, Adithya Renduchintala, Stephen Roller, Dirk Rowe, Weiyan Shi, Joe Spisak, Alexander Wei, David Wu, Hugh Zhang, and Markus Zijlstra. Human-level play in the game of diplomacy by combining language models with strategic reasoning. Science, 378(6624): 1067 1074, 2022.
Reiichiro Nakano, Jacob Hilton, Suchir Balaji, Jeff Wu, Long Ouyang, Christina Kim, Christopher Hesse, Shantanu Jain, Vineet Kosaraju, William Saunders, Xu Jiang, Karl Cobbe, Tyna Eloundou, Gretchen Krueger, Kevin Button, Matthew Knight, Benjamin Chess, and John Schulman. Webgpt: Browser-assisted question-answering with human feedback. ar Xiv preprint ar Xiv:2112.09332, 2021.
Rodrigo Nogueira and Kyunghyun Cho. End-to-end goal-driven web navigation. In Advances In Neural Information Processing Systems, 2016.
Open AI. Gpt-4 technical report. ar Xiv preprint ar Xiv:2303.08774, 2023.
Published as a conference paper at ICLR 2024
Long Ouyang, Jeff Wu, Xu Jiang, Diogo Almeida, Carroll L. Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, John Schulman, Jacob Hilton, Fraser Kelton, Luke Miller, Maddie Simens, Amanda Askell, Peter Welinder, Paul Christiano, Jan Leike, and Ryan Lowe. Training language models to follow instructions with human feedback. ar Xiv preprint arxiv:2203.02155, 2022.
Alec Radford, Jeff Wu, Rewon Child, David Luan, Dario Amodei, and Ilya Sutskever. Language models are unsupervised multitask learners, 2019.
Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, Gretchen Krueger, and Ilya Sutskever. Learning transferable visual models from natural language supervision. ar Xiv preprint ar Xiv:2103.00020, 2021.
Jack W. Rae, Sebastian Borgeaud, Trevor Cai, Katie Millican, Jordan Hoffmann, Francis Song, John Aslanides, Sarah Henderson, Roman Ring, Susannah Young, Eliza Rutherford, Tom Hennigan, Jacob Menick, Albin Cassirer, Richard Powell, George van den Driessche, Lisa Anne Hendricks, Maribeth Rauh, Po-Sen Huang, Amelia Glaese, Johannes Welbl, Sumanth Dathathri, Saffron Huang, Jonathan Uesato, John Mellor, Irina Higgins, Antonia Creswell, Nat Mc Aleese, Amy Wu, Erich Elsen, Siddhant Jayakumar, Elena Buchatskaya, David Budden, Esme Sutherland, Karen Simonyan, Michela Paganini, Laurent Sifre, Lena Martens, Xiang Lorraine Li, Adhiguna Kuncoro, Aida Nematzadeh, Elena Gribovskaya, Domenic Donato, Angeliki Lazaridou, Arthur Mensch, Jean-Baptiste Lespiau, Maria Tsimpoukelli, Nikolai Grigorev, Doug Fritz, Thibault Sottiaux, Mantas Pajarskas, Toby Pohlen, Zhitao Gong, Daniel Toyama, Cyprien de Masson d Autume, Yujia Li, Tayfun Terzi, Vladimir Mikulik, Igor Babuschkin, Aidan Clark, Diego de Las Casas, Aurelia Guy, Chris Jones, James Bradbury, Matthew Johnson, Blake Hechtman, Laura Weidinger, Iason Gabriel, William Isaac, Ed Lockhart, Simon Osindero, Laura Rimell, Chris Dyer, Oriol Vinyals, Kareem Ayoub, Jeff Stanway, Lorrayne Bennett, Demis Hassabis, Koray Kavukcuoglu, and Geoffrey Irving. Scaling language models: Methods, analysis &; insights from training gopher. ar Xiv preprint ar Xiv:2112.11446, 2021.
Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J. Liu. Exploring the limits of transfer learning with a unified text-to-text transformer. Journal of Machine Learning Research, 21(140):1 67, 2020.
Scott Reed, Konrad Zolna, Emilio Parisotto, Sergio Gomez Colmenarejo, Alexander Novikov, Gabriel Barth-Maron, Mai Gimenez, Yury Sulsky, Jackie Kay, Jost Tobias Springenberg, Tom Eccles, Jake Bruce, Ali Razavi, Ashley Edwards, Nicolas Heess, Yutian Chen, Raia Hadsell, Oriol Vinyals, Mahyar Bordbar, and Nando de Freitas. A generalist agent. ar Xiv preprint arxiv:2205.06175, 2022.
Adam Roberts, Hyung Won Chung, Anselm Levskaya, Gaurav Mishra, James Bradbury, Daniel Andor, Sharan Narang, Brian Lester, Colin Gaffney, Afroz Mohiuddin, Curtis Hawthorne, Aitor Lewkowycz, Alex Salcianu, Marc van Zee, Jacob Austin, Sebastian Goodman, Livio Baldini Soares, Haitang Hu, Sasha Tsvyashchenko, Aakanksha Chowdhery, Jasmijn Bastings, Jannis Bulian, Xavier Garcia, Jianmo Ni, Andrew Chen, Kathleen Kenealy, Jonathan H. Clark, Stephan Lee, Dan Garrette, James Lee-Thorp, Colin Raffel, Noam Shazeer, Marvin Ritter, Maarten Bosma, Alexandre Passos, Jeremy Maitin-Shepard, Noah Fiedel, Mark Omernick, Brennan Saeta, Ryan Sepassi, Alexander Spiridonov, Joshua Newlan, and Andrea Gesmundo. Scaling up models and data with t5x and seqio. ar Xiv preprint ar Xiv:2203.17189, 2022.
Dhruv Shah, Blazej Osinski, Brian Ichter, and Sergey Levine. Lm-nav: Robotic navigation with large pre-trained models of language, vision, and action. In Conference on Robot Learning, 2022.
Tianlin Shi, Andrej Karpathy, Linxi Fan, Jonathan Hernandez, and Percy Liang. World of bits: An open-domain platform for web-based agents. In International Conference on Machine Learning, 2017.
Mohammad Shoeybi, Mostofa Patwary, Raul Puri, Patrick Le Gresley, Jared Casper, and Bryan Catanzaro. Megatron-lm: Training multi-billion parameter language models using model parallelism. ar Xiv preprint ar Xiv:1909.08053, 2019.
Published as a conference paper at ICLR 2024
Mohit Shridhar, Lucas Manuelli, and Dieter Fox. Perceiver-actor: A multi-task transformer for robotic manipulation. In Conference on Robot Learning, 2022.
Maayan Shvo, Zhiming Hu, Rodrigo Toro Icarte, Iqbal Mohomed, Allan D. Jepson, and Sheila A. Mc Ilraith. Appbuddy: Learning to accomplish tasks in mobile apps via reinforcement learning. In Canadian Conference on Artificial Intelligence, 2021.
Andreas Steiner, Alexander Kolesnikov, Xiaohua Zhai, Ross Wightman, Jakob Uszkoreit, and Lucas Beyer. How to train your vit? data, augmentation, and regularization in vision transformers. ar Xiv preprint ar Xiv:2106.10270, 2022.
Chen Sun, Abhinav Shrivastava, Saurabh Singh, and Abhinav Gupta. Revisiting unreasonable effectiveness of data in deep learning era. ar Xiv preprint ar Xiv:1707.02968, 2017.
Haotian Sun, Yuchen Zhuang, Lingkai Kong, Bo Dai, and Chao Zhang. Adaplanner: Adaptive planning from feedback with language models. ar Xiv preprint ar Xiv:2305.16653, 2023.
Zineng Tang, Ziyi Yang, Guoxin Wang, Yuwei Fang, Yang Liu, Chenguang Zhu, Michael Zeng, Cha Zhang, and Mohit Bansal. Unifying vision, text, and layout for universal document processing. ar Xiv preprint arxiv:2212.02623, 2022.
Yi Tay, Mostafa Dehghani, Vinh Q. Tran, Xavier Garcia, Jason Wei, Xuezhi Wang, Hyung Won Chung, Dara Bahri, Tal Schuster, Huaixiu Steven Zheng, Denny Zhou, Neil Houlsby, and Donald Metzler. Ul2: Unifying language learning paradigms. ar Xiv preprint ar Xiv:2205.05131, 2022.
Daniel Toyama, Philippe Hamel, Anita Gergely, Gheorghe Comanici, Amelia Glaese, Zafarali Ahmed, Tyler Jackson, Shibl Mourad, and Doina Precup. Androidenv: A reinforcement learning platform for android. ar Xiv preprint ar Xiv:2105.13231, 2021.
Chengyi Wang, Sanyuan Chen, Yu Wu, Ziqiang Zhang, Long Zhou, Shujie Liu, Zhuo Chen, Yanqing Liu, Huaming Wang, Jinyu Li, Lei He, Sheng Zhao, and Furu Wei. Neural codec language models are zero-shot text to speech synthesizers. ar Xiv preprint ar Xiv:2301.02111, 2023.
Jiapeng Wang, Lianwen Jin, and Kai Ding. Li LT: A simple yet effective language-independent layout transformer for structured document understanding. In Annual Meeting of the Association for Computational Linguistics, 2022a.
Qifan Wang, Yi Fang, Anirudh Ravula, Fuli Feng, Xiaojun Quan, and Dongfang Liu. Webformer: The web-page transformer for structure information extraction. ar Xiv preprint ar Xiv:2202.00217, 2022b.
Jason Wei, Maarten Bosma, Vincent Y. Zhao, Kelvin Guu, Adams Wei Yu, Brian Lester, Nan Du, Andrew M. Dai, and Quoc V. Le. Finetuned language models are zero-shot learners. In International Conference on Learning Representations, 2022a.
Jason Wei, Yi Tay, Rishi Bommasani, Colin Raffel, Barret Zoph, Sebastian Borgeaud, Dani Yogatama, Maarten Bosma, Denny Zhou, Donald Metzler, Ed H. Chi, Tatsunori Hashimoto, Oriol Vinyals, Percy Liang, Jeff Dean, and William Fedus. Emergent abilities of large language models. ar Xiv preprint ar Xiv:2206.08853, 2022b.
Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Brian Ichter, Fei Xia, Ed Chi, Quoc Le, and Denny Zhou. Chain of thought prompting elicits reasoning in large language models. ar Xiv preprint ar Xiv:2201.11903, 2022c.
Yiheng Xu, Minghao Li, Lei Cui, Shaohan Huang, Furu Wei, and Ming Zhou. Layout LM: Pretraining of text and layout for document image understanding. ar Xiv preprint arxiv:1912.13318, 2019.
Shunyu Yao, Howard Chen, John Yang, and Karthik Narasimhan. Webshop: Towards scalable real-world web interaction with grounded language agents. ar Xiv preprint arxiv:2207.01206, 2022a.
Published as a conference paper at ICLR 2024
Shunyu Yao, Jeffrey Zhao, Dian Yu, Nan Du, Izhak Shafran, Karthik Narasimhan, and Yuan Cao. React: Synergizing reasoning and acting in language models. ar Xiv preprint ar Xiv:2210.03629, 2022b.
Manzil Zaheer, Kenneth Marino, Will Grathwohl, John Schultz, Wendy Shang, Sheila Babayan, Arun Ahuja, Ishita Dasgupta, Christine Kaeser-Chen, and Rob Fergus. Learning to navigate wikipedia by taking random walks. ar Xiv preprint ar Xiv:2211.00177, 2022.
Andy Zeng, Maria Attarian, Brian Ichter, Krzysztof Choromanski, Adrian Wong, Stefan Welker, Federico Tombari, Aveek Purohit, Michael Ryoo, Vikas Sindhwani, Johnny Lee, Vincent Vanhoucke, and Pete Florence. Socratic models: Composing zero-shot multimodal reasoning with language. ar Xiv preprint ar Xiv:2204.00598, 2022.
Xiaohua Zhai, Alexander Kolesnikov, Neil Houlsby, and Lucas Beyer. Scaling vision transformers. ar Xiv preprint ar Xiv:2106.04560, 2022.
Longtao Zheng, Rundong Wang, and Bo An. Synapse: Leveraging few-shot exemplars for humanlevel computer control. ar Xiv preprint ar Xiv:2306.07863, 2023.
Published as a conference paper at ICLR 2024
A BROADER IMPACTS
While Web GUM is evaluated only in realistic web simulators (Shi et al., 2017; Liu et al., 2018; Yao et al., 2022a), we should carefully conduct it if we deploy the autonomous web agent on the real-world Internet because of security and safety reasons. For instance, the wrong password may cause an account freeze, and emailing the wrong person is problematic in a business scene. Training with online RL may often be infeasible for this reason, while we demonstrate an alternative approach; data-driven, fully offline training by leveraging inductive bias in foundation models. Autonomous agents, well-grounded with the user s intention, should be helpful in our daily lives by reducing our burden on computer tasks. Because a part of our training corpus (54K) includes the demonstrations taken from the output of LLMs (Anil et al., 2023), we will exclude those from the dataset release and it will result in 347K episodes.
B EXTENDED RELATED WORKS
Foundation Models for Decision Making Recently, the ability of multi-step reasoning and inductive bias in foundation models have been leveraged to solve text-based interactive tasks via sequential decisions considering few-shot in-context examples (Ahn et al., 2022; Huang et al., 2022a;b; Zeng et al., 2022; Yao et al., 2022b; Meta Fundamental AI Research Diplomacy Team et al., 2022). Even in continuous control (Chen et al., 2021a; Janner et al., 2021; Furuta et al., 2022b; Brohan et al., 2022) or computer games (Reed et al., 2022; Lee et al., 2022b; Fan et al., 2022), high-capacity transformer models are trained with a large amount of diverse dataset via multi-task behavioral distillation (Chen et al., 2021c; Gu et al., 2021a; Deep Mind Interactive Agents Team et al., 2021; Furuta et al., 2022a; Shridhar et al., 2022; Jiang et al., 2022). To build autonomous web navigation agents, we also leverage pre-trained LLM (Raffel et al., 2020; Chung et al., 2022), by finetuning with massively-curated multimodal demonstrations, and we point out that the better content comprehension and multi-step reasoning abilities, obtained through instruction-finetuning of LLM (Chung et al., 2022), are essential for the notable performance on downstream decision making aligned with human instructions.
Multimodal Large-scale Models Large language models have demonstrated extraordinary emergent abilities on a variety of NLP tasks, such as commonsense question answering, arithmetic, logical reasoning, open-ended text generation (Radford et al., 2019; Brown et al., 2020; Chowdhery et al., 2022; Wei et al., 2022b; Tay et al., 2022), or code completion (Chen et al., 2021b; Austin et al., 2021; Li et al., 2022b). In addition, some works have investigated vision-and-language understanding to improve the accuracy of common vision-based tasks such as open-ended image/object classification (Radford et al., 2021; Gu et al., 2021b; Kamath et al., 2021), image captioning, or visual question-answering (Lu et al., 2022; Alayrac et al., 2022; Chen et al., 2022; Reed et al., 2022; Liu et al., 2023; Dai et al., 2023; Li et al., 2023). Several works also have tackled document understanding with (multimodal) transformer models (Xu et al., 2019; Li et al., 2021a;c; Appalaraju et al., 2021; Tang et al., 2022; Wang et al., 2022a;b), including markup languages such as HTML (Aghajanyan et al., 2021; 2022; Li et al., 2021b; Lee et al., 2022a) for summarization of the documents or question answering on the contents. Despite the great efforts on document understanding, these works are less connected to interactive decision making problems. Our model obtains not only a grounded understanding of websites in a multimodal manner but also the ability to decide the optimal actions to achieve given instructions in web navigation, helping multi-step decisions and visual context understanding.
C IMPLEMENTATION DETAILS
We adopt the encoder-decoder models proposed by Raffel et al. (2020) as multimodal transformers, and vision transformer (Dosovitskiy et al., 2020) pre-trained with Image Net-21K (Deng et al., 2009) as an image encoder for the visual tokens3. We especially use Vi T-B16, a small-size transformer with 86 million parameters, which divides an input image into 16 16-size patches. We use publicly
3https://github.com/google-research/scenic
Published as a conference paper at ICLR 2024
available checkpoints of T5 (Raffel et al., 2020)4, Flan-T5 (Chung et al., 2022)5, and T5-XL finetuned with Mini Wo B++ demonstrations (Gur et al., 2022)6 for the experiments. To construct the training pipeline, we leverage Seq IO (Roberts et al., 2022) library, and use Sentence Piece (Kudo & Richardson, 2018) vocabulary with 32K tokens from C4 dataset (Raffel et al., 2020) for text tokenization. The batch size for training is 128, and input sequence length is set to 4096 tokens. Due to the huge computational requirements, we run one seed to train each model throughout the paper (Humphreys et al., 2022; Gur et al., 2022). We use cloud TPU-v4, which has a 32 Gi B HBM memory space for the experiments. Base-size models require 256 cores and XL-size models do 512 cores, which takes 1-2 days for finetuning.
D DETAILS ON DATASET AND MODEL SIZE SCALING
We here present how critical it is to scale up the dataset and model size in Web GUM. For the dataset size ablation, we use Flan-T5-Base and Vi T-B16. As for both HTML and multimodal models, we could observe the scaling effects in web navigation: the larger the dataset (Table 4) and model (Table 5) size are, the higher the success rates are. Surprisingly, our approach even with only 2.8K HTML episodes (about 25% of the previous one curated by Liu et al. (2018)) and Base-size model (about 7.3% parameters) already achieves 55.7%, surpassing previous SL state-of-the-art (48.4% by Gur et al. (2022)). This surprising efficiency might come from the sufficient inductive bias and alignment with the user intentions in instruction-finetuned LLMs, and Web GUM could fully leverage them for web automation problems. The margin of improvement might be smaller than expected due to the limited capacity of transformer to obtain the grounded understanding of natural language instructions, HTML, and screenshots. In fact, the results also reveal that scaling the models might be more important than the dataset; the low-capacity model may cap the performance at a lower level.
Pre-Trained Models Modality Dataset Success Rate
T5-XL (Gur et al., 2022) HTML 12K 48.4% T5-XL HTML 347K 49.8%
Flan-T5-Base HTML 2.8K 55.7% Flan-T5-Base HTML 68K 56.3% Flan-T5-Base HTML 347K 57.2%
Flan-T5-Base, Vi T-B16 HTML+Image 2.8K 61.1% Flan-T5-Base, Vi T-B16 HTML+Image 68K 62.3% Flan-T5-Base, Vi T-B16 HTML+Image 347K 66.1%
Table 4: Average success rate of Web GUM with different dataset sizes. We observe the larger the dataset size is, the higher the success rate is. Surprisingly, our approach outperforms previous state-of-the-art by over 7.3% even with 2.8K-episode dataset (about 25% of the previous dataset curated by Liu et al. (2018)).
Pre-Trained Models # of Params Modality Success Rate
Flan-T5-Base 220M HTML 57.2% Flan-T5-Large 770M HTML 72.4% Flan-T5-XL 3B HTML 75.5% Flan-T5-XXL 11B HTML 79.0%
Flan-T5-Base, Vi T-B16 310M HTML+Image 66.1% Flan-T5-Large, Vi T-B16 860M HTML+Image 77.4% Flan-T5-XL, Vi T-B16 3B HTML+Image 80.3%
Table 5: Average success rate of Web GUM with different model sizes. As for both HTML-only and multimodal models, we could observe the performance increases as the model size does.
4https://github.com/google-research/t5x/blob/main/docs/models.md# t5-11-checkpoints
5https://github.com/google-research/t5x/blob/main/docs/models.md# flan-t5-checkpoints
6https://console.cloud.google.com/storage/browser/gresearch/webllm/ webn_t5_3b
Published as a conference paper at ICLR 2024
We extensively evaluate the capability of HTML comprehension in instruction-finetuned LLMs with Web SRC (Chen et al., 2021d) where the models are asked to solve contextual QA problems understanding a given HTML and its structure. Those problems are curated from real websites to include key-value extraction, entity comparison, and table understanding problems. The answer formats are either text span in HTML or binary (yes/no). Because the context length is insufficient for raw HTML, we preprocess context HTML by extracting a snippet that includes the answers in advance. We finetune both T5-XL and Flan-T5-XL with the training dataset. Table 6 shows that Flan-T5 records better HTML comprehension performance than T5, which may accelerates the web navigation performance on Mini Wo B++ and Mind2Web.
Models EM F1
T5-XL 63.85 71.44 Flan-T5-XL 68.91 78.48
Table 6: Base language model performance in Web SRC (Chen et al., 2021d). We finetune both T5 and Flan-T5 with trainng dataset. Flan-T5 achieves better performance in HTML comprehension than T5.
F DATASET DETAILS
To construct a large-scale multimodal behavioral dataset on Mini Wo B++, we leverage a public finetuned-LLM policy (Gur et al., 2022) trained with multi-task human demonstration dataset (Liu et al., 2018)7 as a demonstrator. We run such LLM policies with 10,000 episodes per task and only keep successful trajectories to maintain the quality of dataset, following Humphreys et al. (2022). Lastly, we collect additional 54K demonstrations with Synapse (Zheng et al., 2023)8, a private-LLMbased agents with prompting, for the tasks where finetuned-LLM may not complete well such as click-scroll-list and enter-time, and also write a scripted policy for book-flight. We use Pa LM 2 (Anil et al., 2023) as a base LLM for Synapse. Such efforts result in a multi-task dataset with 401K (347K+54K) episodes including HTML and screenshots at each time step. Table 7 shows the details of our multimodal dataset (347K), consisting of HTML, screenshots, actions, and instructions at each time step.
7https://github.com/stanfordnlp/miniwob-plusplus-demos 8https://github.com/ltzheng/synapse
Published as a conference paper at ICLR 2024
Task # of episodes # of steps Ratio (episode)
book-flight 9999 90177 2.88% choose-date 383 1508 0.11% choose-date-easy 3353 12946 0.97% choose-date-medium 2222 8733 0.64% choose-list 1861 3724 0.54% click-button 9782 9909 2.82% click-button-sequence 10000 20000 2.88% click-checkboxes 9761 28904 2.81% click-checkboxes-large 1962 19072 0.57% click-checkboxes-soft 9228 36384 2.66% click-checkboxes-transfer 10000 59793 2.88% click-collapsible 5947 13077 1.71% click-collapsible-2 2199 5627 0.63% click-color 2554 2554 0.74% click-dialog 10000 10000 2.88% click-dialog-2 3285 3285 0.95% click-link 9961 9961 2.87% click-menu 3238 3243 0.93% click-option 9998 20000 2.88% click-pie 3724 8548 1.07% click-scroll-list 0 0 0.00% click-shades 0 0 0.00% click-shape 6116 6117 1.76% click-tab 9978 13177 2.88% click-tab-2 1844 2109 0.53% click-tab-2-hard 1574 1916 0.45% click-test 10000 10000 2.88% click-test-2 10000 10000 2.88% click-widget 9963 9963 2.87% count-shape 5849 5893 1.69% email-inbox 5159 14258 1.49% email-inbox-forward-nl 9995 39980 2.88% email-inbox-forward-nl-turk 4900 20165 1.41% email-inbox-nl-turk 4346 11416 1.25% enter-date 10000 20000 2.88% enter-password 9980 29940 2.88% enter-text 10000 20000 2.88% enter-text-dynamic 9983 19966 2.88% enter-time 0 0 0.00% focus-text 10000 10000 2.88% focus-text-2 10000 10000 2.88% grid-coordinate 8353 8353 2.41% guess-number 1021 2042 0.29% identify-shape 9007 9010 2.60% login-user 9793 29379 2.82% login-user-popup 9786 39170 2.82% multi-layouts 10000 40000 2.88% multi-orderings 10000 40000 2.88% navigate-tree 9864 15140 2.84% search-engine 8872 35095 2.56% social-media 2631 4407 0.76& social-media-all 95 208 0.03% social-media-some 319 893 0.09& tic-tac-toe 3947 13773 1.14% use-autocomplete 3465 6930 1.00% use-spinner 530 532 0.15%
Total 346827 867277 100%
Table 7: Details of our multimodal dataset. It contains about 347K episodes in total.
Published as a conference paper at ICLR 2024
G PER-TASK PERFORMANCE OF MINIWOB++
In this section, we present per-task success rate on Mini Wo B++ (Table 8) and absolute performance improvement by adding image modality to HTML input for Web GUM (Figure 7).
As for Table 8, we refer to Gur et al. (2022) and Zheng et al. (2023) for the baseline performances. We use 56 tasks as benchmark, while removing some duplicated tasks (e.g. -nodelay tasks) from 62 tasks adopted in Gur et al. (2022). During the evaluation on Mini Wo B++, we ignore the time limit due to the computational constraints.
Figure 7 presents full results of the absolute performance improvement, subtracting the success rates: (Success Rate of Web GUM(HTML+Image)) - (Success Rate of Web GUM(HTML)). The results suggest Web GUM leverages visual inputs for multi-step tasks with dynamic page transitions (e.g. book-flight or search-engine) or the tasks that require global contexts of the page (e.g. tic-tac-toe or click-shape). See Appendix L for the visualization.
social-media-some
use-autocomplete
click-button
click-button-sequence
click-checkboxes
click-checkboxes-soft
click-checkboxes-transfer
click-collapsible
click-color
click-dialog
click-dialog-2
click-option
click-scroll-list
click-tab-2-hard
click-test-2
click-widget
email-inbox-forward-nl
email-inbox-forward-nl-turk
enter-password
enter-text-dynamic
focus-text-2
grid-coordinate
multi-layouts
multi-orderings
navigate-tree
social-media
click-collapsible-2
email-inbox
email-inbox-nl-turk
login-user-popup
choose-date
click-checkboxes-large
click-tab-2
count-shape
choose-date-easy
search-engine
use-spinner
choose-date-medium
tic-tac-toe
guess-number
identify-shape
social-media-all
click-shape
book-flight
choose-list
click-shades
Performance Improvement (%)
Figure 7: Performance improvement by adding image modality to HTML on 56 tasks from Mini Wo B++. We subtract the success rates: (Success Rate of Web GUM(HTML+Image)) - (Success Rate of Web GUM(HTML)).
Task Synapse Human Ada Planner RCI RCI (GPT4)
CCNet CCNet (SL)
WGE Web NT5 Web GUM (HTML) Web GUM
bisect-angle n/a 0.92 n/a n/a n/a 0.97 0.29 n/a n/a n/a n/a book-flight 0.76 0.87 n/a n/a n/a 0.87 0.00 0.00 0.00 0.48 0.98 chase-circle n/a 0.82 n/a n/a n/a 0.93 0.80 n/a n/a n/a n/a choose-date 1.00 0.97 n/a n/a n/a 0.97 0.12 0.00 0.00 0.98 1.00 choose-date-easy n/a 0.99 n/a n/a n/a 0.99 0.42 n/a 0.03 0.95 1.00 choose-date-medium n/a 0.98 n/a n/a n/a 0.99 0.26 n/a 0.00 0.94 1.00 choose-list 1.00 0.98 1.00 1.00 1.00 0.99 0.19 0.16 0.26 0.24 1.00 circle-center n/a 0.96 n/a n/a n/a 0.97 0.36 n/a n/a n/a n/a click-button 1.00 0.98 1.00 1.00 1.00 1.00 0.78 1.00 1.00 1.00 1.00 click-button-sequence 1.00 0.94 1.00 1.00 1.00 1.00 0.47 0.99 1.00 1.00 1.00 click-checkboxes 1.00 0.97 1.00 1.00 1.00 0.98 0.32 0.98 0.96 1.00 1.00 click-checkboxes-large 1.00 0.87 1.00 0.94 0.94 0.71 0.00 0.68 0.22 0.97 0.99 click-checkboxes-soft 1.00 0.73 0.80 0.72 0.96 0.95 0.04 0.51 0.54 1.00 1.00 click-checkboxes-transfer 1.00 0.98 0.98 1.00 1.00 0.99 0.36 0.64 0.63 1.00 1.00 click-collapsible 1.00 0.99 1.00 1.00 1.00 1.00 0.81 1.00 0.00 1.00 1.00 click-collapsible-2 0.96 0.97 0.84 0.62 1.00 0.98 0.17 0.65 0.00 0.94 0.95 click-color 1.00 0.97 1.00 1.00 1.00 1.00 0.82 1.00 0.27 1.00 1.00 click-dialog 1.00 1.00 1.00 1.00 1.00 1.00 0.95 1.00 1.00 1.00 1.00 click-dialog-2 1.00 0.99 1.00 1.00 1.00 1.00 0.88 1.00 0.24 1.00 1.00 click-link 1.00 0.99 0.98 1.00 1.00 0.99 0.59 1.00 1.00 1.00 1.00 click-menu 1.00 0.97 0.78 1.00 1.00 0.94 0.22 n/a 0.37 0.99 0.97 click-menu-2 n/a 0.98 n/a n/a n/a 0.83 0.52 n/a n/a n/a n/a click-option 1.00 0.99 1.00 1.00 1.00 0.99 0.21 1.00 0.87 1.00 1.00 click-pie 1.00 0.98 n/a n/a n/a 0.97 0.15 0.32 0.51 0.99 0.99 click-scroll-list 1.00 0.91 1.00 1.00 1.00 0.60 0.01 n/a 0.00 1.00 1.00 click-shades 1.00 0.91 1.00 1.00 1.00 1.00 0.04 0.22 0.00 0.05 1.00 click-shape 0.96 0.88 0.75 0.98 0.98 0.95 0.11 0.64 0.53 0.72 0.94 click-tab 1.00 0.99 1.00 1.00 1.00 1.00 0.95 0.55 0.74 1.00 1.00 click-tab-2 0.94 0.97 0.85 0.74 1.00 0.98 0.27 0.64 0.18 0.95 0.99 click-tab-2-easy n/a 0.99 n/a n/a n/a 0.99 0.61 n/a n/a n/a n/a click-tab-2-hard 0.96 0.96 0.78 0.76 0.98 0.98 0.19 n/a 0.12 0.95 0.95 click-tab-2-medium n/a 0.97 n/a n/a n/a 0.99 0.54 n/a n/a n/a n/a click-test 1.00 1.00 1.00 1.00 1.00 1.00 1.00 1.00 1.00 1.00 1.00 click-test-2 1.00 0.99 1.00 1.00 1.00 1.00 0.95 1.00 1.00 1.00 1.00
Published as a conference paper at ICLR 2024
click-test-transfer n/a 0.99 n/a n/a n/a 1.00 0.94 n/a n/a n/a n/a click-widget 1.00 0.83 1.00 0.98 0.98 1.00 0.56 0.93 1.00 1.00 1.00 copy-paste 1.00 0.94 n/a n/a n/a 0.79 0.04 n/a n/a n/a n/a copy-paste-2 1.00 0.94 n/a n/a n/a 0.63 0.01 n/a n/a n/a n/a count-shape 0.78 0.82 0.50 0.40 0.4 0.85 0.21 0.59 0.41 0.64 0.68 count-sides n/a 0.98 n/a n/a n/a 1.00 0.74 n/a n/a n/a n/a drag-box n/a 0.99 n/a n/a n/a 1.00 0.61 n/a n/a n/a n/a drag-cube n/a 0.99 n/a n/a n/a 0.79 0.23 n/a n/a n/a n/a drag-item n/a 0.98 n/a n/a n/a 1.00 0.61 n/a n/a n/a n/a drag-items n/a 0.93 n/a n/a n/a 0.99 0.13 n/a n/a n/a n/a drag-items-grid n/a 0.87 n/a n/a n/a 0.98 0.05 n/a n/a n/a n/a drag-shapes n/a 0.96 n/a n/a n/a 0.99 0.26 n/a n/a n/a n/a drag-sort-numbers n/a 0.92 n/a n/a n/a 0.97 0.11 n/a n/a n/a n/a email-inbox 1.00 0.96 0.98 0.98 0.98 1.00 0.09 0.43 0.38 0.99 1.00 email-inbox-delete n/a 0.99 n/a n/a n/a 1.00 0.22 n/a n/a n/a n/a email-inbox-forward n/a 0.96 n/a n/a n/a 1.00 0.01 n/a n/a n/a n/a email-inbox-forward-nl 1.00 0.91 1.00 1.00 1.00 1.00 0.00 n/a 0.60 1.00 1.00 email-inbox-forward-nl-turk 1.00 0.88 1.00 0.94 0.94 1.00 0.00 n/a 0.33 1.00 1.00 email-inbox-important n/a 0.99 n/a n/a n/a 1.00 0.30 n/a n/a n/a n/a email-inbox-nl-turk 1.00 0.93 0.90 0.98 0.98 1.00 0.05 0.77 0.23 0.99 1.00 email-inbox-noscroll n/a 0.96 n/a n/a n/a 1.00 0.13 n/a n/a n/a email-inbox-reply n/a 0.91 n/a n/a n/a 1.00 0.00 n/a n/a n/a n/a email-inbox-star-reply n/a 0.95 n/a n/a n/a 1.00 0.11 n/a n/a n/a n/a enter-date 1.00 0.97 1.00 0.96 0.96 1.00 0.02 0.00 0.00 1.00 1.00 enter-password 1.00 0.96 0.98 1.00 1.00 1.00 0.02 0.99 0.97 1.00 1.00 enter-text 1.00 0.98 0.98 1.00 1.00 1.00 0.35 1.00 0.89 1.00 1.00 enter-text-2 n/a 0.91 n/a n/a n/a 0.98 0.04 n/a n/a n/a n/a enter-text-dynamic 1.00 0.97 0.96 1.00 1.00 1.00 0.39 1.00 0.98 1.00 1.00 enter-time 0.98 0.98 0.96 1.00 1.00 0.97 0.04 0.52 0.00 1.00 1.00 find-midpoint n/a 0.94 n/a n/a n/a 0.97 0.35 n/a n/a n/a n/a find-word 0.84 0.96 n/a n/a n/a 0.88 0.05 n/a n/a n/a n/a focus-text 1.00 1.00 1.00 1.00 1.00 1.00 0.99 1.00 1.00 1.00 1.00 focus-text-2 1.00 0.99 0.94 1.00 1.00 1.00 0.96 1.00 1.00 1.00 1.00 grid-coordinate 1.00 0.87 1.00 1.00 1.00 1.00 0.66 1.00 0.49 1.00 1.00 guess-number 1.00 0.99 0.88 0.20 0.20 1.00 0.21 0.00 0.00 0.34 0.43 highlight-text n/a 0.97 n/a n/a n/a 1.00 0.51 n/a n/a n/a n/a highlight-text-2 n/a 0.97 n/a n/a n/a 1.00 0.40 n/a n/a n/a n/a identify-shape 1.00 0.98 0.96 0.76 1.0 1.00 0.68 0.90 0.88 0.90 1.00 login-user 1.00 0.96 1.00 1.00 1.0 1.00 0.00 0.99 0.82 1.00 1.00 login-user-popup 1.00 0.94 0.98 0.68 0.68 1.00 0.02 n/a 0.72 0.99 1.00 moving-items n/a 0.18 n/a n/a n/a 0.88 0.13 n/a n/a n/a n/a multi-layouts 0.94 0.95 0.84 0.72 0.96 1.00 0.00 0.99 0.83 1.00 1.00 multi-orderings 1.00 0.96 1.00 1.00 1.00 1.00 0.00 0.99 0.88 1.00 1.00 navigate-tree 0.96 0.98 0.82 0.86 1.00 0.99 0.32 0.99 0.91 1.00 1.00 number-checkboxes n/a 0.96 n/a n/a n/a 0.99 0.00 n/a n/a n/a n/a read-table 1.00 0.97 n/a n/a n/a 0.97 0.01 n/a n/a n/a n/a read-table-2 n/a 0.95 n/a n/a n/a 0.94 0.00 n/a n/a n/a n/a resize-textarea n/a 0.94 n/a n/a n/a 1.00 0.27 n/a n/a n/a n/a right-angle n/a 0.87 n/a n/a n/a 0.98 0.26 n/a n/a n/a n/a scroll-text n/a 0.97 n/a n/a n/a 0.96 0.04 n/a n/a n/a n/a scroll-text-2 n/a 0.97 n/a n/a n/a 1.00 0.88 n/a n/a n/a n/a search-engine 1.00 0.97 1.00 1.00 1.00 1.00 0.15 0.26 0.34 0.91 0.96 simon-says n/a 0.62 n/a n/a n/a 0.00 0.02 n/a n/a n/a n/a simple-algebra 1.00 0.86 0.82 1.00 1.00 0.75 0.03 n/a n/a n/a n/a simple-arithmetic 1.00 0.96 n/a n/a 1.00 0.86 0.38 n/a n/a n/a n/a social-media 1.00 0.96 0.82 0.98 0.98 0.90 0.03 0.39 0.21 1.00 1.00 social-media-all 1.00 0.89 1.00 1.00 1.00 0.75 0.00 0.01 0.00 0.31 0.52 social-media-some 1.00 0.91 0.90 0.90 0.96 0.85 0.01 0.01 0.02 0.89 0.73 terminal 0.98 0.88 0.98 1.00 1.00 0.00 0.00 n/a n/a n/a n/a text-editor n/a 0.88 n/a n/a n/a 0.98 0.11 n/a n/a n/a n/a text-transform 1.00 0.86 n/a 0.80 0.80 0.60 0.19 n/a n/a n/a n/a tic-tac-toe 1.00 0.71 0.48 0.56 0.56 0.83 0.32 0.37 0.48 0.50 0.56 unicode-test 1.00 0.99 n/a n/a n/a 1.00 0.86 n/a n/a n/a n/a use-autocomplete 0.98 0.98 0.88 0.58 0.58 1.00 0.07 0.78 0.22 1.00 0.98 use-colorwheel n/a 0.90 n/a n/a n/a 0.98 0.68 n/a n/a n/a n/a use-colorwheel-2 n/a 0.94 n/a n/a n/a 0.95 0.38 n/a n/a n/a n/a use-slider n/a 0.98 n/a n/a n/a 0.91 0.18 n/a n/a n/a n/a use-slider-2 n/a 0.97 n/a n/a n/a 0.95 0.03 n/a n/a n/a n/a use-spinner 1.00 0.98 0.90 0.88 0.96 1.00 0.47 0.04 0.07 0.06 0.11 visual-addition n/a 0.97 n/a n/a n/a 0.99 0.36 n/a n/a n/a n/a
Average 0.985 0.935 0.929 0.906 0.940 0.935 0.305 0.646 0.484 0.887 0.942 # of Tasks 63 104 53 54 54 104 104 48 56 56 56
Table 8: Per-task success rate on Mini Wo B++. We refer to Gur et al. (2022) and Zheng et al. (2023) for the baseline performances.
Published as a conference paper at ICLR 2024
H COMPOSITIONAL EVALUATION ON MINIWOB++
For the compositional evaluation, we pick up 4 clicksomething (link, button, checkboxes, dialog) tasks and make some combinations of those by naively stitching with 2 or 3 tasks. Then, we prepare the following 6 combinational tasks,
click-button_click-checkboxes click-button_click-dialog click-button_click-link click-link_click-button click-link_click-button_click-dialog click-link_click-dialog
These tasks should be resolved in order of the name: for instance, in click-link_click-button_click-dialog task, the agent should click the proper link, click the proper button, click the proper dialog, and then the task results in success. In click-button_click-link task, the agent should click the proper button, and then click the proper link. The instructions for compositional tasks are also simply combined among original task instructions in order of the name. This evaluation could test the ability to transfer primitive skills to control computers to solve unseen tasks. Table 9 shows the per-task average success rate among 6 combinations above. Web GUM can solve the compositional tasks much better than baselines (Gur et al., 2022; Zheng et al., 2023) .
Compositional Task Web N-T5 Synapse Web GUM (HTML) Web GUM (HTML+Image)
click-button_click-checkboxes 0.26 0.84 0.21 0.27 click-button_click-dialog 0.95 1.00 0.87 0.93 click-button_click-link 0.87 0.99 0.81 0.88 click-link_click-button 0.35 1.00 0.90 0.95 click-link_click-button_click-dialog 0.08 0.60 0.73 0.73 click-link_click-dialog 0.55 0.00 0.93 0.95
Ave. 0.510 0.738 0.742 0.785
Table 9: Per-task average success rate on 6 tasks from compositional Mini Wo B++.
click-link click-button click-link_click-button
Figure 8: Example of compositional evaluation on Mini Wo B++ (the same as Figure 5).
Published as a conference paper at ICLR 2024
I COMPARISON AGAINST PRIOR WEB NAVIGATION AGENTS
Methods Architecture Pre-trained Input Output Offline
WGE (Liu et al., 2018) LSTM, self-attention % DOM Logit of action % Co DE (Gur et al., 2019; 2021) Bi-LSTM % DOM Logit of action % DOM-Q-NET(Jia et al., 2019) GNN % DOM Logit of action % CC-Net (Humphreys et al., 2022) LSTM, Transformer, Res Net % DOM, Screenshot Logit of action % Web Shop (Yao et al., 2022a) BERT, BART " Text (from HTML) Logit of action % / "
Web GUM (Ours) T5 Transformer, Vi T " HTML, Screenshot Text "
Table 10: Prior works have studied web navigation problem as online RL to learn the optimal action distribution with task-specific model architectures from scratch ( or partially using pre-trained vision encoder). We omit the web-specialized architecture and input-output space, and convert web navigation into visual question-answering format (text, image text), which allows us to learn the agents offline by leveraging pre-trained foundation models (Raffel et al., 2020; Chung et al., 2022; Dosovitskiy et al., 2020) in vision or language domains as strong inductive bias for web environments.
J INPUT PERTURBATION EVALUATION ON MINIWOB++
Add Coordinates Add extra HTML at the bottom Add extra HTML at the top
Figure 9: Example of input perturbation for Mini Wo B++ evaluation (the same as Figure 6).
Published as a conference paper at ICLR 2024
K EVALUATION ON WEBSHOP
In addition to Mini Wo B++, we extensively evaluate our Web GUM on Web Shop (Yao et al., 2022a) benchmark, another online-shopping websites simulator with a large amount of real-world product data. Web Shop provides user instruction that describes the feature of items (e.g. I need a long clip-in hair extension which is natural looking, and price lower than 20.00 dollars). The agents should search, compare and choose a proper product that matches the given instruction. Since Web Shop requires complex multi-step reasoning considering previous contexts for comparison (Yao et al., 2022a;b), we can test the capability of instruction-finetuned LLM in decision making tasks in depth. The performance score is evaluated by the percentage of required attributes covered by the chosen product (from 0 to 100), and if the product meets all the requirements, that episode is labeled a success.
Because Web Shop does not have API to get the screenshot of rendered websites, we focus on Web GUM with text inputs, parsed from noisy HTML in the real world.9 We convert the actions from raw texts (e.g. search[a long clip-in hair extension] or click[- ]) to dictionary-like format (e.g. {"action": "search", "ref": "a long clip-in hair extension"} or {"action": "click", "ref": "
- "}), as we use in Mini Wo B++, to improve the prediction accuracy. We finetune Flan-T5-XL with about 1K human demonstrations curated by Yao et al. (2022a)10, using only high-score demonstrations. The score threshold is score 50 and we have 840 episodes in total (Table 12). We construct the model input with action history, instruction, and text observation, the same as Mini Wo B++ experiments. We evaluate our method with 500 user instructions in the test set.
Table 11 shows that Web GUM achieves 45.0% success, significantly outperforming not only simple baselines, such as supervised imitation learning (IL) and IL plus RL-finetuing (by more than 15%), but also recent prompt-based LLM agents, including Re Act (Yao et al., 2022b) (i.e. Pa LM-540B (Chowdhery et al., 2022) with one-shot prompt and reasoning annotations), while our model only has 3 billion parameters. IL and IL plus RL-finetuning baselines use BART (Lewis et al., 2019) model for the search policy, and BERT (Devlin et al., 2019) model for the click policy. The better performance of Web GUM proves the hypothesis that the ability of multi-step reasoning in instruction-finetuned language models works as a prior for decision making problems.
Methods Training Model Modality Score Success Rate
Rule Text 45.6 9.6% IL SL BART, BERT Text(+Image) 59.9 29.1% IL+RL SL+RL BART, BERT Text(+Image) 62.4 28.7% Act In-context Pa LM-540B Text 62.3 30.1% Re Act In-context Pa LM-540B Text 66.6 40.0%
Web N-T5 SL T5-XL Text 61.0 29.8% Web GUM SL Flan-T5-XL Text 67.5 45.0%
Human Text+Image 82.1 59.6%
Table 11: Average score and success rate on Web Shop (Yao et al., 2022a) benchmark. Web GUM based on Flan-T5-XL achieves 45.0% success, outperforming most baseline approaches including Re Act, a prompted Pa LM-540B with reasoning annotations. We refer Yao et al. (2022b) for the baselines.
Threshold # of Episodes Score Success Rate
score 0 1026 67.2 44.4% score 50 840 67.5 45.0% score = 100 497 65.3 44.4%
Table 12: Average score and success rate on Web Shop with different score thresholds. Because we should balance the dataset coverage and proficiency, we choose 50 as a threshold.
9Web Shop just provides visual features of item pictures when the agents reach the product page. These features are extracted by Res Net-50 (He et al., 2016), rather than raw images or screenshots of the website. Some baseline agents (IL and IL+RL) incorporate such embeddings. 10https://github.com/princeton-nlp/Web Shop/tree/master/baseline_models/ data
Published as a conference paper at ICLR 2024
L EXAMPLE EPISODES OF WEBGUM
click-pie tic-tac-toe
choose-date-easy grid-coordinate
click-checkboxes-large
search-engine
Figure 10: Example of successful episodes demonstrated by multimodal Web GUM on Mini Wo B++ (Shi et al., 2017; Liu et al., 2018). The time step goes from left to right. As discussed in Section 5.1, image modality seems to be leveraged for multi-step tasks with dynamic page transitions (e.g. search-engine, choose-date-easy) or tasks that require global visual contexts (e.g. tic-tac-toe).
Published as a conference paper at ICLR 2024
count-shape: How many small items are there?
social-media-all: Click the share button on all posts by @renda and then click Submit.
guess-number: Guess the number between 0-9 and press Submit.
Web GUM failed to count the number of items because of the ambiguity of instructions ( small ).
Web GUM clicked the button on the wrong posts (not by @renda). The repetitive structure of the page may cause confusion.
Web GUM repeated the wrong answers during multi-step QA.
Figure 11: Example failure episodes by multimodal Web GUM on Mini Wo B. The time step goes from left to right. For instance, Web GUM suffers from (1) ambiguous instruction, such as small items (count-shape), (2) confusion with the repetitive structure on the page (social-media-all), and (3) long-horizon (guess-number) tasks that may require memory.
Published as a conference paper at ICLR 2024
Instruction: i am looking for dark denim color ethylene Instruction: i would like a xlarge plus red camellia fleece vinyl ultra train of size 10, 3rd generation for men, jacket that can be machine washed, and price lower and price lower than 160.00 dollars than 130.00 dollars
Action: search
dark denim color ethylene vinyl ultra Action: search
xlarge plus red camellia fleece jacket that train of size 10, 3rd generation for men
can be machine washed
Observation: Observation: back to search
page 1 (total results: 50)
next back to search
page 1 (total results: 50)
b08b1mcfkl b09h2qsj5p
salewa ultra train 3 - men s men s thermal underwear pants usb heated warm cotton $104.0 to $182.99 with zone 3 heating long trousers leggings bottoms b07j5mc2x1
$56.77 skechers men s low-top trainers, us men
$53.39 to $140.35 men s thermal underwear pants usb heated warm cotton b07x5hr3tz
long trousers leggings bottoms for winter washable baselayer bruno marc men s linen canvas stretch loafer shoes $56.77 to $56.99 $39.99
premium soft and smooth faux fur loose plush vest jackets Action: click
fashion sexy fuzzy sleeveless outwear Observation: $9.47 to $12.47 back to search prev
size 7 7.5 8 8.5 9 9.5 10 10.5 11 11.5
Action: click
Observation: color ombre blue | red orange raw green | black out back to search
page 2 (total results: 50)
prev next dark denim | red orange
salewa ultra train 3 - men s
price: $104.0 to $182.99 king and queen write hoodie couple matching sweater rating: n.a. pullover hooded sweatshirt jacket description features reviews buy now
$26.99 to $29.99 b09jmvfpzg
Action: click
dark denim | red orange
mens sherpa quilted lined flannel jacket plus size plaid Observation: you have clicked dark denim | red orange. button down fleece shirts coats thicken warm cotton outwear Action: click
$36.79 to $55.99 Observation: you have clicked 10.
mankalvxy womens jacket cardigan flannel cowl neck Action: click
asymmetric loose outerwear plus size xs-xxl 0-22 Termination: Your score (min 0.0, max 1.0): 1.0 $20.3 to $26.85
Action: click
Observation: back to search prev
size king-small queen-small king-medium queen-medium king-large queen-large king-xxl king-xlarge queen-xlarge queen-xxlarge
king and queen write hoodie couple matching sweater pullover hooded sweatshirt jacket price: $26.99 to $29.99 rating: n.a. description features reviews buy now
Action: click
Observation: back to search
page 2 (total results: 50)
king and ... b09jmvfpzg
mens sherpa ... b09kxfgbn2
mankalvxy womens...
Action: click
Observation: back to search prev
color army green black coffee gray green red a-army green a-red
size small medium large x-large xx-large 3x-large
mens sherpa quilted lined flannel jacket plus size plaid button down fleece shirts coats thicken warm cotton outwear price: $36.79 to $55.99 rating: n.a. description features reviews buy now
Action: click
Observation: you have clicked red.
Action: click
Observation: you have clicked x-large.
Action: click
Termination: Your score (min 0.0, max 1.0): 1.0
Table 13: Example of successful episodes demonstrated by Web GUM on Web Shop (Yao et al., 2022a). Web GUM could search and choose the required product and choose proper options (left). In addition, Web GUM could also compare the products with browsing and backtracking (i.e. clicking next or prev buttons) during the episodes (right).
Published as a conference paper at ICLR 2024
Example 1 Example 2
HTML: HTML: