Skip to main content

PinchBench

An OpenClaw agent benchmark from Kilo that measures successful task completion across standardized real-world agent workflows.

How BenchLM shows PinchBench

BenchLM mirrors the public PinchBench average-success-rate view using the official snapshot updated on 04/13/2026, 4:44 PM: 68 models and 860 runs. PinchBench grades runs with automated checks plus an LLM judge.

This benchmark is display only on BenchLM. It is excluded from BenchLM overall rankings, category rankings, and weighted scoring. The table below uses average scores only, matching the public PinchBench average view rather than the best-run view.

68 models860 runsAverage scores onlyOfficial runsDisplay only

Average success rate on PinchBench — 04/13/2026, 4:44 PM

BenchLM mirrors the published average success rate view for PinchBench. Trinity-Large-Thinking leads the public snapshot at 91.9% , followed by Qwen3.6 Plus (84.0%) and MiniMax M2.7 (82.8%). BenchLM does not use these results to rank models overall.

68 modelsAgenticCurrentDisplay onlyUpdated 04/13/2026, 4:44 PM

The published PinchBench snapshot is tightly clustered at the top: Trinity-Large-Thinking sits at 91.9%, while the third row is only 9.1 points behind. The broader top-10 spread is 11.5 points, so the benchmark still separates strong models even when the leaders cluster.

68 models have been evaluated on PinchBench. The benchmark falls in the Agentic category. This category carries a 22% weight in BenchLM.ai's overall scoring system. PinchBench is currently displayed for reference but excluded from the scoring formula, so it does not directly affect overall rankings.

About PinchBench

Year

2026

Tasks

23 OpenClaw agent tasks

Format

Average success rate from official runs

Difficulty

Long-horizon agent workflows

PinchBench publishes official OpenClaw runs across 23 tasks and grades results with automated checks plus an LLM judge. BenchLM mirrors the public average-score view as a display-only benchmark.

BenchLM freshness & provenance

Version

PinchBench 2026

Refresh cadence

Quarterly

Staleness state

Current

Question availability

Public benchmark set

CurrentDisplay only

BenchLM uses freshness metadata to decide whether a benchmark should still be treated as a strong differentiator, a benchmark to watch, or a display-only reference. For the full scoring policy, see the BenchLM methodology page.

Average success rate table (68 models)

1
Trinity-Large-Thinkingarcee-ai/trinity-large-thinking
91.9%
2
Qwen3.6 Plusqwen/qwen3.6-plus-preview
84.0%
3
MiniMax M2.7minimax/minimax-m2.7
82.8%
4
Claude Opus 4.6anthropic/claude-opus-4.6
81.6%
5
MiMo-V2-Omnixiaomi/mimo-v2-omni
81.1%
6
GLM-5.1z-ai/glm-5.1
80.9%
7
Qwen3.5-122B-A10Bqwen/qwen3.5-122b-a10b
80.8%
8
Claude Sonnet 4.6anthropic/claude-sonnet-4.6
80.7%
9
GLM-5z-ai/glm-5
80.6%
10
Qwen3.5 397Bqwen/qwen3.5-397b-a17b
80.4%
11
MiMo-V2-Proxiaomi/mimo-v2-pro
80.4%
12
GLM-5-Turboz-ai/glm-5-turbo
80.3%
13
Claude Sonnet 4.5anthropic/claude-sonnet-4.5
80.0%
14
Seed-2.0-Litebytedance-seed/seed-2.0-lite
79.8%
15
MiniMax M2.1minimax/minimax-m2.1
79.7%
16
GPT-5.4openai/gpt-5.4
79.4%
17
Qwen3.5 Plusqwen/qwen3.5-plus-02-15
79.1%
18
Qwen3 Coder Nextqwen/qwen3-coder-next
79.1%
19
Claude Opus 4.5anthropic/claude-opus-4.5
78.8%
20
Kimi K2.5moonshotai/kimi-k2.5
78.6%
21
Qwen3.5-27Bqwen/qwen3.5-27b
78.5%
22
MiniMax M2.5minimax/minimax-m2.5
78.1%
23
Gemini 3.1 Progoogle/gemini-3.1-pro-preview
77.5%
24
Claude Haiku 4.5anthropic/claude-haiku-4.5
77.4%
25
77.3%
26
77.3%
28
GLM-4.5-Airz-ai/glm-4.5-air
76.8%
29
Step 3.5 Flashstepfun/step-3.5-flash
76.6%
30
Gemini 3 Flashgoogle/gemini-3-flash-preview
75.3%
33
Nemotron 3 Super 120B A12Bnvidia/nemotron-3-super-120b-a12b
73.1%
34
Qwen3 Maxqwen/qwen3-max-thinking
71.8%
35
Qwen3.5-35B-A3Bqwen/qwen3.5-35b-a3b
71.7%
36
GPT-5.4 miniopenai/gpt-5.4-mini
71.4%
37
Grok 4.1 Fastx-ai/grok-4.1-fast
71.3%
39
Grok 4.20x-ai/grok-4.20
71.2%
41
Mercury 2inception/mercury-2
70.0%
42
MiMo-V2-Flashxiaomi/mimo-v2-flash
69.7%
44
GPT-5.4 nanoopenai/gpt-5.4-nano
69.5%
45
GPT-5 miniopenai/gpt-5-mini
69.0%
46
DeepSeek V3.2deepseek/deepseek-v3.2
67.7%
47
Gemini 3 Progoogle/gemini-3-pro-preview
67.7%
48
GLM-5V-Turboz-ai/glm-5v-turbo
67.0%
52
Gemini 2.5 Progoogle/gemini-2.5-pro
65.0%
53
Trinity-Large-Thinkingarcee-ai/trinity-large-preview
63.7%
54
GPT-4o miniopenai/gpt-4o-mini
63.6%
55
Qwen3.6 Plusqwen/qwen3.6-plus
63.3%
56
62.8%
57
Gemini 2.5 Flashgoogle/gemini-2.5-flash
57.2%
58
GPT-4oopenai/gpt-4o
56.6%
59
GPT-5 nanoopenai/gpt-5-nano
56.2%
60
GPT-OSS 120Bopenai/gpt-oss-120b
52.0%
61
GPT-OSS 20Bopenai/gpt-oss-20b
50.3%
62
34.8%
63
Llama 4 Maverickmeta-llama/llama-4-maverick
34.8%
65
Llama 3.1 70B Instructmeta-llama/llama-3.1-70b-instruct
22.7%
66
Gemini 2.5 Flash-Litegoogle/gemini-2.5-flash-lite
12.7%
67
GPT-5.4 Proopenai/gpt-5.4-pro
12.0%
68
Llama 4 Scoutmeta-llama/llama-4-scout
5.4%

FAQ

What does PinchBench measure?

An OpenClaw agent benchmark from Kilo that measures successful task completion across standardized real-world agent workflows.

Which model leads the published PinchBench snapshot?

Trinity-Large-Thinking currently leads the published PinchBench snapshot with a average success rate of 91.9%. BenchLM shows this benchmark for display only and does not use it in overall rankings.

How many models are evaluated on PinchBench?

68 AI models are included in BenchLM's mirrored PinchBench snapshot, based on the public leaderboard captured on 04/13/2026, 4:44 PM.

Last updated: 04/13/2026, 4:44 PM · mirrored from the public benchmark leaderboard

The AI models change fast. We track them for you.

For engineers, researchers, and the plain curious — a weekly brief on new models, ranking shifts, and pricing changes.

Free. No spam. Unsubscribe anytime.