Breaking

News

AI2's MolmoWeb Sees the Web Like You Do

A Fully Open Web Agent That Only Uses Its Eyes Every serious web agent you've heard of — the ones booking flights, filling out forms, scraping product listings — comes from a company that guards its training data like state secrets. OpenAI, Google, Anthropic: they've all built capable

AI2's MolmoWeb Sees the Web Like You Do
Daily Neural — Latest Artificial Intelligence News Today

A Fully Open Web Agent That Only Uses Its Eyes

Every serious web agent you've heard of — the ones booking flights, filling out forms, scraping product listings — comes from a company that guards its training data like state secrets. OpenAI, Google, Anthropic: they've all built capable browser-automation systems, but the actual recipe stays locked up. That information asymmetry is exactly what the Allen Institute for AI is trying to dismantle with MolmoWeb, a fully open web agent released this week with weights, training data, and evaluation tooling all available under Apache 2.0.

What makes MolmoWeb architecturally interesting isn't just its openness — it's the constraint it operates under. The agent never reads a webpage's source code or accesses the DOM. It looks at a screenshot, decides what to do, acts, takes another screenshot, and repeats. Click, scroll, type, navigate — all derived purely from pixels. The developers argue this makes the system more robust in practice: a site's visual layout changes far less often than its underlying HTML structure, which tends to get reorganized every time a dev team ships a redesign.

The model runs on the Molmo2 architecture, pairing Qwen3 as the language backbone with SigLIP2 as the vision encoder. Two sizes ship today: 4B and 8B parameters. Neither required reinforcement learning or distillation from proprietary systems — just supervised fine-tuning on 64 H100 GPUs.

The Real Unlock: A Training Dataset Built for Scale

If you want to understand why web agents have lagged behind chatbots and code assistants in the open-source world, the answer has always been data. Demonstrating how to use a browser is expensive to collect and messy to standardize. MolmoWebMix is AI2's direct answer to that problem.

The dataset combines three sources. First, crowdworkers completed real browsing tasks while every click and page transition was recorded — yielding roughly 36,000 complete task runs across more than 1,100 websites, which AI2 describes as the largest public collection of human web task execution available. Second, an automated pipeline extended coverage beyond what human annotation could economically produce: a Gemini 2.5 Flash planner breaks goals into sub-steps, an operator executes browser actions, and a GPT-4o verifier checks screenshots to confirm each sub-goal actually completed. Third, over 2.2 million screenshot-question-answer pairs train the model to read and understand web interfaces, supplemented by more than seven million UI element localization examples.

One finding from the accompanying paper cuts against intuition: synthetic runs outperform human demonstrations when both cover identical tasks. Human annotators tend to poke around unfamiliar sites, take wrong turns, and backtrack. Automated agents find more direct paths, and the model learns better from those cleaner trajectories. Even more striking — just ten percent of the full dataset delivers 85 to 90 percent of final benchmark performance.

photo: freepik

Benchmark Numbers That Should Worry Larger Players

On WebVoyager — a standard test that covers navigation across 15 real websites including GitHub and Google Flights — MolmoWeb-8B scores 78.2 percent. OpenAI's o3 sits at 79.3 percent. That's not a rounding error; that's a genuinely competitive result from an 8-billion-parameter open model trained without distilling from any proprietary system. It also outperforms agents built on the much larger GPT-4o that have access to both annotated screenshots and structured page data, which is a more direct apples-to-apples win.

On ScreenSpot benchmarks for UI element localization, a specialized 8B variant beats Anthropic's Claude 3.7 and OpenAI's CUA — though it's worth noting AI2 selected older proprietary checkpoints for comparison, so those margins should be read with some caution.

Where MolmoWeb trails is against its own "teacher" — a Gemini-based agent that can access page structure directly — by around five percentage points. That gap represents the real cost of screenshot-only operation: the model has to do its own text recognition rather than receiving it pre-parsed. The team has a direct answer for closing that gap at inference time: run the task multiple times and take the best result. WebVoyager accuracy jumps from 78.2 to 94.7 percent under that test-time compute scaling strategy. Throwing more inference budget at the problem pays off significantly.

This puts pressure on proprietary web agent providers specifically because MolmoWeb requires no special browser instrumentation. Any developer with a screenshot pipeline and a GPU can deploy it.

Building With It Today

For developers who want to start immediately, the practical path is straightforward. The 4B model fits in roughly 6 GB of VRAM under 4-bit NF4 quantization, making it accessible on consumer hardware and free-tier cloud GPUs. The prompting format is structured and explicit: a task description, an accumulated history of prior steps with thoughts and actions, and the current page context — URL, title, and position in the session.

The agent's action vocabulary is compact and parseable: goto(), click() with normalized coordinates, type(), scroll(), press(), switch_tab(), and send_msg() to return a final answer. Connecting it to a real browser via Playwright is the natural production path — a screenshot loop with action execution in between. The GitHub repository is well-documented for teams ready to integrate, and a hosted demo is available for quick evaluation. Model weights and collections are also available on Hugging Face.

For teams building agents on top of the model rather than deploying it directly, MolmoWebMix itself is arguably the more valuable asset. A grounding dataset of seven million UI localization examples in particular fills a gap that has limited fine-tuning work in this space.

What This Means

AI2 is making an explicit bet that openness accelerates safety and capability research simultaneously. The same argument they made for language models with OLMo — that you can't audit what you can't inspect — now applies to browser-operating agents that take real-world actions. That framing matters because web agents carry meaningfully higher risk profiles than text generators: they can click purchase buttons, submit forms, and interact with services that have side effects.

The team is candid about what remains unsolved. MolmoWeb misreads text in screenshots at a non-trivial rate, performance degrades with vague instructions, and the demo deliberately blocks password fields and payment flows. The harder questions — how agents should interpret terms of service, how to prevent irreversible actions, how to handle legally sensitive content — are left deliberately open, with the argument that more researchers having access to the full stack is how you get better answers.

  • For developers: A production-capable web agent now fits in 6 GB of VRAM with no proprietary dependencies. The Playwright integration path is well-documented and the action space is clean to work with.
  • For founders: Browser automation products built on closed APIs now have a credible open alternative to evaluate. The cost and control calculus just shifted.
  • For researchers: MolmoWebMix — especially the 36,000 human trajectory runs and 7M+ grounding examples — is likely the more durable contribution. Dataset quality has been the ceiling for open web agent work.
  • For the broader AI ecosystem: This follows OpenSeeker's similar move for AI search agents. A pattern is forming: fully open stacks as a deliberate counter to big tech data monopolies in agentic AI.

AI2 has been under real organizational pressure since Microsoft hired away several of its senior researchers to join Mustafa Suleyman's superintelligence team. MolmoWeb reads partly as a statement of institutional identity: whatever else is happening, AI2 is still the organization that ships the receipts.

Written by