Multi-Environment Web Challenge (MEWC)

A benchmark that evaluates AI agents on multi-environment web challenges, testing navigation and task completion across diverse live web environments.

How BenchLM shows MEWC right now

BenchLM is tracking MEWC in the local dataset, but exact-source verification records for these rows are still being attached. To avoid a blank benchmark page, BenchLM shows the current tracked rows below as a display-only reference table.

These tracked rows are useful for inspection and spot-checking, but until exact-source attachments are completed they should not be treated as fully verified public benchmark rows.

1 tracked modelsLocal tracked rowsAwaiting exact-source attachmentsDisplay only

Tracked score on MEWC — April 8, 2026

BenchLM mirrors the published tracked score view for MEWC. MiniMax M2.5 leads the public snapshot at 74.4%. BenchLM does not use these results to rank models overall.

1 modelsAgenticCurrentDisplay onlyUpdated April 8, 2026

About MEWC

Year

2026

Tasks

Web-agent tasks

Format

Browser task completion

Difficulty

Open-web agent workflows

MEWC is useful as an agentic browsing benchmark because it focuses on open-web interaction and multi-environment task execution rather than single-site scripted browsing.

BenchLM freshness & provenance

Version

MEWC 2026

Refresh cadence

Quarterly

Staleness state

Current

Question availability

Public benchmark set

CurrentDisplay only

BenchLM uses freshness metadata to decide whether a benchmark should still be treated as a strong differentiator, a benchmark to watch, or a display-only reference. For the full scoring policy, see the BenchLM methodology page.

Tracked score table (1 models)

1
MiniMax M2.5minimax-m2-5
74.4%

FAQ

What does MEWC measure?

A benchmark that evaluates AI agents on multi-environment web challenges, testing navigation and task completion across diverse live web environments.

Which model leads the published MEWC snapshot?

MiniMax M2.5 currently leads the published MEWC snapshot with a tracked score of 74.4%. BenchLM shows this benchmark for display only and does not use it in overall rankings.

How many models are evaluated on MEWC?

1 AI models are included in BenchLM's mirrored MEWC snapshot, based on the public leaderboard captured on April 8, 2026.

Last updated: April 8, 2026 · mirrored from the public benchmark leaderboard

Weekly LLM Benchmark Digest

Get notified when new models drop, benchmark scores change, or the leaderboard shifts. One email per week.

Free. No spam. Unsubscribe anytime. We only store derived location metadata for consent routing.