Seeing is Believing: An Open Foundation for Web Agents

Author: Denis Avetisyan


Researchers have unveiled MolmoWeb, a new open-source toolkit and dataset designed to empower the development of web agents capable of sophisticated visual web browsing.

At each processing step, the agent synthesizes task instructions, visual screen content, and prior actions to generate a natural language rationale, subsequently informing its selection of a browser action-a process fundamentally grounded in translating observational input into deliberate, reasoned behavior.
At each processing step, the agent synthesizes task instructions, visual screen content, and prior actions to generate a natural language rationale, subsequently informing its selection of a browser action-a process fundamentally grounded in translating observational input into deliberate, reasoned behavior.

MolmoWeb provides performant vision-language models and a diverse dataset (MolmoWebMix) achieving state-of-the-art results on web browsing benchmarks.

Despite the promise of autonomous web agents, progress is hampered by reliance on proprietary models lacking transparency and community access. This work, ‘MolmoWeb: Open Visual Web Agent and Open Data for the Open Web’, addresses this challenge by introducing MolmoWeb, a fully open-source foundation for web agents comprising the [latex]100K+[/latex] trajectory MolmoWebMix dataset and performant vision-language models achieving state-of-the-art results on standard benchmarks, surpassing even larger closed-source alternatives. MolmoWeb agents operate directly from webpage screenshots, predicting browser actions without requiring access to underlying code or APIs, and demonstrate substantial gains with test-time scaling. Will this open foundation accelerate the development of truly intelligent and accessible web automation for all?


The Inherent Fragility of Conventional Web Interaction

Conventional web automation techniques, designed for static HTML, often falter when confronted with the intricate and ever-changing landscape of modern websites. These systems typically rely on identifying elements by their underlying code – a precarious approach given that websites frequently update their structure and content. Dynamic content loaded with JavaScript, AJAX requests, and constantly shifting layouts present a significant challenge, causing automation scripts to break easily. This fragility stems from a fundamental mismatch between the rigid nature of traditional automation and the fluid, user-centric design principles that characterize the contemporary web, necessitating more adaptable and intelligent approaches to reliably interact with online resources.

Current web automation techniques often falter when confronted with the inherent unpredictability of the modern web. Unlike static documents, websites are fluid entities, constantly updating content, rearranging layouts, and introducing novel interactive components. This dynamism presents a significant challenge; scripts designed to locate elements based on fixed identifiers or precise pixel coordinates become easily broken by even minor alterations. Robustness is further compromised by variations in website design across different browsers, devices, and user preferences. Consequently, automation systems frequently require constant maintenance and adaptation, limiting their scalability and reliability in handling the ever-evolving landscape of online content and interfaces. The fragility of these methods highlights the need for more adaptable approaches capable of interpreting web pages with a degree of flexibility akin to human perception.

Effective web interaction demands more than simply locating elements on a page; it necessitates a system that perceives the web as humans do. This involves integrating computer vision – allowing the system to ‘see’ and interpret visual cues like buttons, forms, and images – with natural language understanding. By processing the text surrounding these elements, the system can infer their function and context, much like a person reading instructions or interpreting a website’s purpose. This combined approach enables robust navigation, even when websites undergo changes in layout or content, because the system isn’t solely reliant on fixed coordinates or predictable structures; instead, it understands what an element is and how it’s meant to be used, allowing it to adapt and interact with the web in a far more flexible and human-like manner.

MolmoWebMix is a dataset designed to evaluate web agents by integrating GUI perception-including screenshot question answering and referring expression grounding-with both synthetic and human-generated task trajectories, which are further decomposed into a defined set of atomic skills.
MolmoWebMix is a dataset designed to evaluate web agents by integrating GUI perception-including screenshot question answering and referring expression grounding-with both synthetic and human-generated task trajectories, which are further decomposed into a defined set of atomic skills.

MolmoWeb: A Vision-Language Approach to Web Agency

MolmoWeb is a web agent distinguished by its architecture, leveraging a Vision-Language Model (VLM) to interpret and act upon web-based tasks. This VLM foundation allows MolmoWeb to accept input in two distinct modalities: visual screenshots of webpages and natural language instructions detailing desired actions. By processing both visual and textual data, the agent avoids limitations inherent in text-only or image-only approaches. The VLM component facilitates understanding of webpage elements through image analysis, while the language component parses instructions to determine the appropriate sequence of interactions. This dual-input capability is central to MolmoWeb’s functionality, enabling it to bridge the gap between human intent, expressed in language, and the visual complexity of web interfaces.

The Vision-Language Model (VLM) at the core of MolmoWeb utilizes screenshot analysis to discern webpage structure and functionality. This process involves identifying elements such as headings, paragraphs, images, and crucially, interactive components like buttons, forms, and links. Robustness is achieved through techniques enabling accurate detection even with varying webpage designs, dynamic content loading, and differing resolutions. The VLM doesn’t merely recognize visual features; it associates these with their semantic roles within the page, establishing a representation of the page’s interactive possibilities to facilitate task completion.

MolmoWeb’s integration of visual and textual inputs significantly improves task execution by leveraging complementary information sources. Traditional web automation often relies solely on identifying elements via HTML attributes or XPath, which can be brittle and susceptible to webpage changes. By analyzing screenshots, MolmoWeb gains a robust understanding of the page layout and interactive elements, independent of underlying code. This visual context, combined with natural language task instructions, enables more accurate identification of target elements and reduces ambiguity. Consequently, MolmoWeb demonstrates increased success rates and reduced completion times for complex web tasks compared to systems relying on single modalities.

MolmoWeb utilizes a process of converting user-provided task instructions into a sequence of discrete browser actions. This translation involves parsing the instruction to identify required webpage manipulations, such as clicking buttons, filling forms, or navigating to specific URLs. The agent then formulates these actions as programmatic commands targeting the browser’s automation API. These commands are executed sequentially, simulating user interaction with the webpage. Successful completion of each action is verified through webpage analysis, utilizing the underlying Vision-Language Model to assess changes in the page’s visual and textual content, and to adjust subsequent actions as needed to ensure task completion.

Pass@k accuracy on WebVoyager and Online-Mind2Web increases with larger values of k, demonstrating that the [latex]MolmoWeb-8B[/latex] model significantly outperforms [latex]MolmoWeb-4B[/latex], and utilizing 100 steps yields substantially better results, achieving 94.7% and 60.5% at Pass@4.
Pass@k accuracy on WebVoyager and Online-Mind2Web increases with larger values of k, demonstrating that the [latex]MolmoWeb-8B[/latex] model significantly outperforms [latex]MolmoWeb-4B[/latex], and utilizing 100 steps yields substantially better results, achieving 94.7% and 60.5% at Pass@4.

Constructing MolmoWebMix: A Dataset for Robust Web Agent Training

MolmoWebMix is a large-scale training dataset constructed to improve the performance of web-based agents. The dataset comprises over 200,000 human-annotated examples and 1.5 million trajectories demonstrating atomic skill execution. Data is collected from interactions with a diverse set of 137 websites, encompassing a wide range of layouts, content, and interactive elements. This scale and diversity are intended to facilitate robust learning and generalization, enabling the agent to effectively navigate and complete tasks on previously unseen web environments. The dataset also includes GUI perception data, providing the agent with visual information about the web pages it interacts with.

MolmoWebMix leverages a multi-faceted data collection strategy to maximize learning potential. Human annotation provides explicit labels and demonstrations of desired behaviors, while atomic skill trajectories capture fundamental web interaction primitives – such as clicking, typing, and scrolling – as sequential data. Complementing these are GUI perception data, which include visual representations of webpage elements and their associated metadata. The integration of these three data types – human guidance, procedural skill data, and visual scene understanding – creates a diverse and informative training signal, enabling the agent to learn both high-level task strategies and low-level interaction details.

Synthetic data generation plays a critical role in expanding the MolmoWebMix dataset beyond directly annotated examples. This process involves algorithmically creating new web interaction scenarios, including variations in website layouts, element properties, and task objectives. By generating data that doesn’t precisely match the training set, the agent is exposed to a wider range of possible web environments and user interface elements. This exposure enhances the agent’s ability to generalize its learned skills to previously unseen websites and tasks, improving performance and robustness in real-world applications where encountering novel web designs is common.

The MolmoWebMix training dataset incorporates substantial diversity to address the inherent variability of the World Wide Web. This diversity is achieved through variations in website layouts, content, task objectives, and interaction methods. Specifically, the dataset includes data generated from numerous distinct websites, encompassing a range of domains and design paradigms. Furthermore, task instructions are phrased in multiple ways, and agents are exposed to variations in GUI elements and their positioning. This breadth of data exposure is critical for developing a web agent capable of generalizing to novel websites and adapting to changes in existing site structures, ultimately improving performance and robustness in real-world scenarios.

The data generation pipeline combines task sampling, trajectory generation via human or [latex]	ext{AxTree}[/latex] agents, and success filtering to produce a dataset for training and evaluation.
The data generation pipeline combines task sampling, trajectory generation via human or [latex] ext{AxTree}[/latex] agents, and success filtering to produce a dataset for training and evaluation.

MolmoWeb’s Impact and the Future of Web Agency

MolmoWeb represents a significant advancement in web agent technology by building upon and substantially enhancing established benchmarks like WebVoyager, Online-Mind2Web, DeepShop, and WebTailBench. Existing benchmarks often present limitations in task complexity or data diversity; MolmoWeb addresses these by providing a more robust and challenging environment for evaluating agent capabilities. This isn’t simply a matter of increased scale, but a deliberate effort to push the boundaries of what’s possible in automated web interaction, requiring agents to demonstrate more sophisticated reasoning and adaptability. By extending these existing frameworks, MolmoWeb facilitates more meaningful comparisons between different models and accelerates progress towards truly intelligent web agents capable of handling real-world online tasks.

MolmoWeb establishes a new benchmark in web agent capabilities by achieving state-of-the-art performance despite relying solely on visual screenshots as input – a significant departure from agents that directly interact with a website’s underlying code. Remarkably, this open-weight model not only surpasses the performance of other publicly available models, but also outperforms considerably larger, proprietary agents built using set-of-marks approaches. This demonstrates that intelligent web interaction doesn’t necessarily require access to a website’s HTML; instead, effective visual understanding can be sufficient, unlocking opportunities for more adaptable and broadly applicable web automation tools and offering a compelling alternative to traditional methods that depend on detailed site-specific knowledge.

Evaluations demonstrate that MolmoWeb-8B, operating with a 100-step execution limit, achieves remarkably high success rates on challenging web-based tasks. Specifically, the model attains a Pass@4 score of 94.7% on the WebVoyager benchmark – indicating it successfully completes nearly all attempted tasks when given four attempts – and a 60.5% Pass@4 score on the more complex Online-Mind2Web. These results signify a substantial advancement in web agent capabilities, showcasing the model’s proficiency in navigating and interacting with web environments to fulfill user requests, and establishing a new performance baseline for open-weight models in this domain.

Performance gains of over twenty percentage points are achievable through a refined evaluation strategy employing four parallel rollouts, each assessed by a Large Language Model (LLM) judge. This approach moves beyond single-attempt evaluations, allowing MolmoWeb to explore multiple solution pathways concurrently. The LLM judge, acting as an objective arbiter, then analyzes these parallel attempts, identifying the most successful strategy and providing a more robust performance metric. This method not only enhances the accuracy of evaluation but also leverages the LLM’s reasoning abilities to discern subtle differences in approach, ultimately leading to significant improvements in task completion rates across complex web-based challenges.

MolmoWeb’s performance extends beyond simple web navigation, showcasing a remarkable capacity for tackling genuinely complex tasks encountered in real-world online scenarios. The system proficiently manages information seeking, adeptly identifying and extracting relevant data from diverse web sources. Equally notable is its success in e-commerce interactions, where it can autonomously complete tasks like product searches, comparisons, and even simulated purchases. This range of capabilities suggests a significant advancement in building web agents capable of handling not just pre-defined actions, but also the nuanced demands of dynamic, user-driven online experiences, moving closer to agents that can truly assist users across the spectrum of web-based activities.

The demonstrated efficacy of MolmoWeb underscores a significant advancement in the field of artificial intelligence: the power of multimodal learning for creating web agents capable of sophisticated interaction. By processing information directly from visual screenshots, rather than relying solely on textual data, the model exhibits a remarkable ability to navigate and complete complex tasks online. This approach allows MolmoWeb to overcome limitations inherent in text-based agents, which can struggle with visually-rich or poorly-structured web pages. The resulting gains in performance, exceeding those of larger, proprietary models, suggest that integrating visual input is crucial for building truly intelligent and adaptable agents capable of seamlessly interacting with the dynamic and often visually-driven landscape of the modern web. This success paves the way for future research focused on further refining multimodal techniques and expanding the scope of tasks these agents can handle.

Continued development of MolmoWeb centers on enhancing its capacity for complex reasoning and tackling the challenges presented by ever-changing web environments. Current research aims to equip the agent with more sophisticated planning and problem-solving skills, enabling it to navigate ambiguous tasks and adapt to unforeseen circumstances. Simultaneously, efforts are underway to improve MolmoWeb’s handling of dynamic content – websites that update frequently or rely heavily on JavaScript – through techniques like asynchronous rendering and improved state tracking. These advancements promise to unlock even greater potential for MolmoWeb, paving the way for web agents capable of seamless interaction with the modern, interactive web and ultimately, more effective automation of online tasks.

Our web browsing trajectory collection tool features a Chrome extension displaying the current webpage alongside annotations for the last captured screenshot, a detailed instruction breakdown, a note-taking input for incomplete steps, and the final answer.
Our web browsing trajectory collection tool features a Chrome extension displaying the current webpage alongside annotations for the last captured screenshot, a detailed instruction breakdown, a note-taking input for incomplete steps, and the final answer.

The pursuit of robust web agents, as demonstrated by MolmoWeb, echoes a fundamental tenet of computational elegance: provable correctness. The framework’s emphasis on a diverse dataset, MolmoWebMix, and performant vision-language models isn’t merely about achieving state-of-the-art benchmark results; it’s about building a foundation where agent behavior is demonstrably reliable. As Geoffrey Hinton once stated, “The problem with deep learning is that it’s a black box.” MolmoWeb attempts to mitigate this ‘black box’ effect by focusing on creating a system where performance is rooted in verifiable data and model architecture, ultimately striving for a level of algorithmic purity that transcends empirical success.

What Lies Ahead?

The presented work, while demonstrating proficiency in navigating the constructed digital landscape, merely formalizes the obvious. A system capable of ‘web browsing’ is not, in itself, intelligent. The true challenge resides not in automating existing processes, but in defining a provable framework for genuine agency within a fundamentally unstructured information space. The benchmarks, while useful for iterative improvement, remain simulations – elegant exercises in pattern matching, divorced from the messiness of real-world information seeking.

Future effort must address the inherent limitations of reliance on synthetic data. While convenient for initial training, such datasets inevitably encode the biases and simplifying assumptions of their creators. A robust agent requires exposure to the full spectrum of the open web – a chaotic, often contradictory, and occasionally nonsensical realm. This necessitates not simply more data, but methods for discerning signal from noise – a task demanding formal verification, not merely empirical success.

Ultimately, the pursuit of ‘web agents’ risks becoming a technological echo chamber. The enduring question is not whether a machine can mimic human browsing behavior, but whether it can formulate independent queries, evaluate information with logical consistency, and – crucially – justify its conclusions. Until that threshold is crossed, the elegance of the algorithm remains purely aesthetic.


Original article: https://arxiv.org/pdf/2604.08516.pdf

Contact the author: https://www.linkedin.com/in/avetisyan/

See also:

2026-04-10 09:26