Head-to-head comparison across 6benchmark categories. Overall scores shown here use BenchLM's provisional ranking lane.
GLM-5
77
Qwen3.5 397B
66
Verified leaderboard positions: GLM-5 #13 · Qwen3.5 397B #11
Pick GLM-5 if you want the stronger benchmark profile. Qwen3.5 397B only becomes the better choice if reasoning is the priority.
Agentic
Coding
+2.9 difference
Reasoning
+2.4 difference
Knowledge
+5.5 difference
Multilingual
+1.6 difference
Inst. Following
GLM-5
Qwen3.5 397B
$0 / $0
$0 / $0
74 t/s
96 t/s
1.64s
2.44s
200K
128K
Pick GLM-5 if you want the stronger benchmark profile. Qwen3.5 397B only becomes the better choice if reasoning is the priority.
GLM-5 is clearly ahead on the provisional aggregate, 77 to 66. The gap is large enough that you do not need to squint at the spreadsheet to see the difference.
GLM-5's sharpest advantage is in knowledge, where it averages 70.7 against 65.2. The single biggest benchmark swing on the page is HLE, 50.4% to 28.7%. Qwen3.5 397B does hit back in reasoning, so the answer changes if that is the part of the workload you care about most.
GLM-5 gives you the larger context window at 200K, compared with 128K for Qwen3.5 397B.
GLM-5 is ahead on BenchLM's provisional leaderboard, 77 to 66. The biggest single separator in this matchup is HLE, where the scores are 50.4% and 28.7%.
GLM-5 has the edge for knowledge tasks in this comparison, averaging 70.7 versus 65.2. Inside this category, HLE is the benchmark that creates the most daylight between them.
GLM-5 has the edge for coding in this comparison, averaging 63.2 versus 60.3. Inside this category, SWE-bench Pro is the benchmark that creates the most daylight between them.
Qwen3.5 397B has the edge for reasoning in this comparison, averaging 63.2 versus 60.8. Inside this category, AI-Needle is the benchmark that creates the most daylight between them.
GLM-5 and Qwen3.5 397B are effectively tied for agentic tasks here, both landing at 56.2 on average.
GLM-5 and Qwen3.5 397B are effectively tied for instruction following here, both landing at 92.6 on average.
Qwen3.5 397B has the edge for multilingual tasks in this comparison, averaging 84.7 versus 83.1. Inside this category, NOVA-63 is the benchmark that creates the most daylight between them.
For engineers, researchers, and the plain curious — a weekly brief on new models, ranking shifts, and pricing changes.
Free. No spam. Unsubscribe anytime.