A benchmark that evaluates AI agents on multi-environment web challenges, testing navigation and task completion across diverse live web environments.
BenchLM is tracking MEWC in the local dataset, but exact-source verification records for these rows are still being attached. To avoid a blank benchmark page, BenchLM shows the current tracked rows below as a display-only reference table.
These tracked rows are useful for inspection and spot-checking, but until exact-source attachments are completed they should not be treated as fully verified public benchmark rows.
BenchLM mirrors the published tracked score view for MEWC. MiniMax M2.5 leads the public snapshot at 74.4%. BenchLM does not use these results to rank models overall.
Year
2026
Tasks
Web-agent tasks
Format
Browser task completion
Difficulty
Open-web agent workflows
MEWC is useful as an agentic browsing benchmark because it focuses on open-web interaction and multi-environment task execution rather than single-site scripted browsing.
Version
MEWC 2026
Refresh cadence
Quarterly
Staleness state
Current
Question availability
Public benchmark set
BenchLM uses freshness metadata to decide whether a benchmark should still be treated as a strong differentiator, a benchmark to watch, or a display-only reference. For the full scoring policy, see the BenchLM methodology page.
A benchmark that evaluates AI agents on multi-environment web challenges, testing navigation and task completion across diverse live web environments.
MiniMax M2.5 currently leads the published MEWC snapshot with a tracked score of 74.4%. BenchLM shows this benchmark for display only and does not use it in overall rankings.
1 AI models are included in BenchLM's mirrored MEWC snapshot, based on the public leaderboard captured on April 8, 2026.
Get notified when new models drop, benchmark scores change, or the leaderboard shifts. One email per week.
Free. No spam. Unsubscribe anytime. We only store derived location metadata for consent routing.