Ego2Web

A Web Agent Benchmark Grounded in Egocentric Videos
CVPR 2026
1Google DeepMind    2UNC Chapel Hill

From real-world video perception to online web execution

Ego2Web is a new multimodal agent benchmark that bridges real-world video perception and online action. Instead of relying on text or screenshots, agents must ground fine-grained evidence from egocentric videos—such as objects, brands, or actions—and use it to complete downstream web tasks. The benchmark spans diverse workflows across E-Commerce, Media Retrieval, Knowledge Lookup, and Local/Maps. Tasks require visual understanding like identifying a specific snack, an exercise step, a car brand, or a university name—demanding strong temporal reasoning, object grounding, and cross-modal transfer to web queries. Unlike existing benchmarks, Ego2Web evaluates whether agents can correctly interpret the real visual world before acting online, not just execute browser actions.

Ego2Web overview examples across E-Commerce, Media Retrieval, Knowledge Lookup, and Local/Maps
Ego2Web tasks require agents to extract grounded evidence from egocentric videos and transfer it into downstream web actions across multiple domains.

Multimodal AI agents are increasingly automating complex real-world workflows that involve online web execution. However, current web-agent benchmarks suffer from a critical limitation: they focus entirely on web-based interaction and perception, lacking grounding in the user’s real-world physical surroundings. This limitation prevents evaluation in crucial scenarios, such as when an agent must use egocentric visual perception (e.g., via AR glasses) to recognize an object in the user’s surroundings and then complete a related task online (e.g., making a purchase). To address this gap, we introduce Ego2Web, the first benchmark designed to bridge egocentric video perception and web agent execution. Ego2Web pairs real- world first-person video recordings with web tasks that require visual understanding, web task planning, and interaction in an online environment for successful completion. We utilize an automatic data-generation pipeline combined with human verification and refinement to curate well-constructed, high-quality video- task pairs across diverse web task types, including e-commerce, navigation, media search, and so on. To facilitate a more accurate and scalable evaluation for our novel benchmark, we also develop a novel LLM-as-a-Judge automatic evaluation method, Ego2WebJudge, which achieves approximately 84% agreement with human judgment, substantially higher than existing evaluation methods. Experiments with diverse state-of-the-art agents show that their performance remains far from perfect, revealing a major performance gap. We also conduct a comprehensive ablation study on task design, highlighting the necessity of video perception in the proposed task and the limitations of current agents. We hope Ego2Web can be a critical new resource for developing truly capable AI assistants that can seamlessly see, understand, and act across the physical and digital worlds.

Ego2Web comparison overview
Ego2Web evaluation pipeline
The Ego2Web pipeline integrates egocentric video perception with web-based agent actions. Agents operate on both video and textual instructions, while Ego2WebJudge enables scalable evaluation by grounding judgments in visual evidence and web trajectories.

Semi-Automatic Data Generation & Automatic Evaluation

Ego2Web is the first benchmark that connects egocentric video perception with real-world web agent execution. We build the benchmark with a semi-automatic pipeline: first, a vision-language model parses first-person videos into structured clip-level captions and visual metadata; then, a large language model generates grounded web task instructions over active websites such as Amazon, YouTube, and Wikipedia; finally, human annotators verify and refine each example to ensure visual grounding, web feasibility, and instruction quality. This model-human pipeline produces diverse, high-quality video-task pairs that require agents to identify relevant evidence in the physical world and translate it into downstream digital actions.

To support scalable evaluation in live web environments, we further introduce Ego2WebJudge, an automatic LLM-as-a-Judge framework tailored for visually grounded web tasks. Given the task instruction, action trajectory, web screenshots, and annotated visual evidence from the egocentric video, Ego2WebJudge first extracts key success criteria, then selects the most relevant screenshots from the web trajectory, and finally determines whether the agent has completed the task correctly and consistently with the visual evidence from the real world. This enables reliable online evaluation for multimodal agents operating across both physical perception and web interaction.

Ego2Web overview examples across E-Commerce, Media Retrieval, Knowledge Lookup, and Local/Maps
Ego2Web tasks require agents to extract grounded evidence from egocentric videos and transfer it into downstream web actions across multiple domains.
Ego2Web overview examples across E-Commerce, Media Retrieval, Knowledge Lookup, and Local/Maps
(a) Task type distribution of Ego2Web, showing the high-level task category composition across major web platforms. (b) Fine-grained website domain distribution in Ego2Web (count>4), highlighting coverage across e-commerce, media retrieval, knowledge bases, local/map services and others (indicated by color coding).

Current web agents remain far from perfect on Ego2Web

We evaluate a diverse set of state-of-the-art web agents on Ego2Web. Despite strong capabilities in web interaction and vision-language tasks, all models exhibit substantial performance gaps, highlighting the difficulty of grounding real-world visual perception for downstream web execution.

Evaluation Base MLLM Claude 3.7 Claude 4.5 GPT-5.4 SeeAct BU-GPT-4.1 BU-Gemini-3-Flash
Ego2WebJudge Qwen3-VL-Flash 20.832.238.829.634.657.2
Gemini-2.5 Pro 17.824.823.625.234.648.2
GPT-4o 19.427.226.826.847.651.4
Human Eval -- 26.432.830.634.244.458.6
Overall success rate across different base multimodal models and agent frameworks. Ego2WebJudge achieves strong alignment with human evaluation while enabling scalable benchmarking.
Domains / Agents Claude 3.7 Claude 4.5 GPT 5.4 SeeAct BU-GPT-4.1 BU-Gemini-3-Flash Avg. SR
E-Commerce 13.018.214.319.526.938.221.7
Media Retrieval 19.626.529.524.230.350.730.1
Knowledge Lookup 33.645.639.143.463.075.050.0
Local / Maps 6.412.929.019.322.548.323.1
Others 0.06.66.620.040.013.314.4
Total 17.824.823.625.234.648.229.0
Fine-grained performance across task domains. Agents perform best on knowledge lookup, but struggle significantly with real-world grounded tasks such as e-commerce and local navigation, revealing challenges in perception-to-action transfer.

Raw video perception is essential, and current agents still fail on transferring visual grounding, temporal reasoning, and cross-modal retrieval to web tasks.

We study the role of egocentric visual perception through controlled input ablations and further analyze the main failure modes of strong web agents on Ego2Web. The results show a clear hierarchy: no visual input < detailed caption < raw video. While structured captions help, they remain an imperfect proxy for true visual grounding.

Raw Video Detailed Caption E-Commerce Media Retrieval Knowledge Lookup Local / Maps Others Total
2.6 7.5 5.4 3.2 0.0 4.4
13.0 29.5 39.1 38.7 6.6 23.6
38.2 50.7 75.0 48.3 13.3 48.2
Table. Ablation study on the impact of video perception in Ego2Web. We report Success Rate (SR). Detailed Caption uses Gemini-3.1-Pro to generate structured descriptions for egocentric videos. Results show that direct raw video input substantially outperforms caption-only perception and language-only settings.

Impact of Visual Perception

Without any visual input, the agent performs extremely poorly, achieving only 4.4% SR, indicating that language-only signals are insufficient for visually grounded web tasks. Providing detailed captions improves performance to 23.6% SR, showing that textual summaries can partially capture semantic information. However, raw video input leads to a much stronger result of 48.2% SR, more than doubling the caption-based setting.

The improvement is consistent across all domains, with particularly large gains in Knowledge Lookup and Local / Maps, where fine-grained spatial and temporal cues are especially important.

Error Analysis

We manually inspected 50 unsuccessful trajectories and identified several recurring failure modes:

  • 36% Object Misidentification — the agent fails to identify the correct target object from the egocentric video.
  • 18% Temporal and Action Misunderstanding — the agent confuses temporal order or user actions in the video.
  • 16% Failure in Cross-Modal Retrieval — the agent identifies the target but cannot retrieve or verify the required web information.
  • 12% Coarse-Grained Matching Errors — the agent retrieves semantically similar but incorrect results.
  • 18% Others — including instruction misunderstanding, planning inefficiency, or external barriers such as CAPTCHA/authentication.

Failure Case Visualization

A compositional failure occurs when errors in temporal grounding propagate to downstream web retrieval and verification. For example, the agent may incorrectly identify the second picked-up sauce in the video, navigate to a related Walmart product page, yet still fail to verify the required package-size attribute. This illustrates that solving Ego2Web requires not only accurate visual perception, but also precise temporal reasoning and robust cross-modal alignment between video evidence and web content.

Ego2Web overview examples across E-Commerce, Media Retrieval, Knowledge Lookup, and Local/Maps
Visualization of web agent (BU-Gemini-3-Flash) failure case. The agent is required to identify the second picked-up sauce from the egocentric video and retrieve its product page. The agent incorrectly identifies the target item due to temporal misunderstanding and fails to verify the required information on the webpage.

Citation

@article{yu2026ego2web,
  title={Ego2Web: Benchmarking Web Agents with Egocentric Video Grounding},
  author={Yu, Shoubin and Shu, Lei and Yang, Antoine and Fu, Yao and 
          Sunkara, Srinivas and Wang, Maria and Chen, Jindong and 
          Bansal, Mohit and Gong, Boqing},
  journal={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
  year={2026}
}