Explorer: Scaling Exploration-driven Web Trajectory Synthesis for Multimodal Web Agents
Abstract
Recent success in large multimodal models (LMMs) has sparked promising applications of agents capable of autonomously completing complex web tasks. While open-source LMM agents have made significant advances in offline evaluation benchmarks, their performance still falls substantially short of human-level capabilities in more realistic online settings. A key bottleneck is the lack of diverse and large-scale trajectory-level datasets across various domains, which are expensive to collect. In this paper, we address this challenge by developing a scalable recipe to synthesize the largest and most diverse trajectory-level dataset to date, containing over 94K successful multimodal web trajectories, spanning 49K unique URLs, 720K screenshots, and 33M web elements. In particular, we leverage extensive web exploration and refinement to obtain diverse task intents. The average cost is 28 cents per successful trajectory, making it affordable to a wide range of users in the community. Leveraging this dataset, we train Explorer, a multimodal web agent, and demonstrate strong performance on both offline and online web agent benchmarks such as Mind2Web-Live, Multimodal-Mind2Web, and MiniWob++. Additionally, our experiments highlight data scaling as a key driver for improving web agent capabilities. We hope this study makes state-of-the-art LMM-based agent research at a larger scale more accessible.
Community
Paper: Explorer: Scaling Exploration-driven Web Trajectory Synthesis for Multimodal Web Agents
Link: https://arxiv.org/pdf/2502.11357.
- We develop a scalable multi-agent pipeline that leverages exploration for diverse web agent trajectory synthesis.
- SOTA results on Mind2Web-Live for a 4B and 7B model.
- Also outperforms other GUI agent baselines on Multimodal-Mind2Web and MiniWoB++.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Towards Internet-Scale Training For Agents (2025)
- WebWalker: Benchmarking LLMs in Web Traversal (2025)
- InfiGUIAgent: A Multimodal Generalist GUI Agent with Native Reasoning and Reflection (2025)
- OS-Genesis: Automating GUI Agent Trajectory Construction via Reverse Task Synthesis (2024)
- QLASS: Boosting Language Agent Inference via Q-Guided Stepwise Search (2025)
- TRISHUL: Towards Region Identification and Screen Hierarchy Understanding for Large VLM based GUI Agents (2025)
- AppVLM: A Lightweight Vision Language Model for Online App Control (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper