Skip to main content

GPT-5.4 mini

OpenAICurrentReleased Mar 17, 2026
Overall Score
Unranked
Arena Elo
1456
Categories Ranked
1of 8
Price (1M tokens)
$0.75 in / $4.5 out
Speed
201tok/s
Context
400K
ProprietaryReasoning
Confidence
mini

BenchLM is tracking GPT-5.4 mini, but this profile is currently excluded from the public leaderboard because it still lacks enough non-generated benchmark coverage to rank safely. Only non-generated public benchmark rows appear below.

GPT-5.4 mini is a proprietary model with a 400K token context window. It uses explicit chain-of-thought reasoning, which typically improves performance on math and complex reasoning tasks at the cost of higher latency and token usage.

GPT-5.4 mini sits inside the GPT-5.4 family alongside GPT-5.4, GPT-5.4 Pro, GPT-5.4 nano. BenchLM links it directly to GPT-5 mini as the earlier related model in that lineage. This profile currently has 11 of 185 tracked benchmarks. BenchLM only exposes non-generated benchmark rows publicly, so missing categories stay blank until a sourced evaluation is available.

Its strongest category is Instruction Following (#43). This performance profile makes it a well-rounded choice across a range of tasks.

Ranking Distribution

Category rank across 7 benchmark categories — sorted by best rank

Category Performance

Scores across all benchmark categories (0-100 scale)

Category Breakdown

Agentic

76.8/ 100
Weight: 22%5 benchmarks
Terminal-Bench 2.0BrowseCompOSWorld-VerifiedGAIATAU-benchWebArena

Coding

78.2/ 100
Weight: 20%1 benchmark
SWE-bench VerifiedLiveCodeBenchSWE-bench ProSWE-RebenchSciCode

Reasoning

17.3/ 100
Weight: 17%0 benchmarks
MuSRLongBench v2MRCRv2ARC-AGI-2

Knowledge

82.3/ 100
Weight: 12%3 benchmarks
GPQASuperGPQAMMLU-ProHLEFrontierScienceSimpleQA

Math

91.8/ 100
Weight: 5%0 benchmarks
AIME 2025BRUMO 2025MATH-500FrontierMath

Multilingual

0.0/ 100
Weight: 7%0 benchmarks
MGSMMMLU-ProX

Multimodal

70.0/ 100
Weight: 12%2 benchmarks
MMMU-ProOfficeQA ProCharXivCharXiv w/o tools

Inst. Following

#43
73.9/ 100
Weight: 5%0 benchmarks
IFEvalIFBench

Chatbot Arena Performance

Text Overall1456CI: ±5.811,968 votes
Coding1505CI: ±11.22,770 votes
Math1434CI: ±21.0744 votes
Instruction Following1442CI: ±9.93,494 votes
Creative Writing1412CI: ±14.01,888 votes
Multi-turn1473CI: ±12.92,147 votes
Hard Prompts1479CI: ±7.36,938 votes
Hard Prompts (English)1475CI: ±10.23,359 votes
Longer Query1455CI: ±9.93,610 votes

Benchmark Details

Only benchmark rows with an attached exact-source record are shown here. Source-unverified manual rows and generated rows are hidden from model pages.

GPT-5.4 Family

Mini

Canonical Entry

GPT-5.4

Related Earlier Model

GPT-5 mini

Frequently Asked Questions

How does GPT-5.4 mini perform overall in AI benchmarks?

GPT-5.4 mini has 11 published benchmark scores on BenchLM, but it does not yet have enough non-generated coverage to receive a global overall rank.

Is GPT-5.4 mini good for knowledge and understanding?

GPT-5.4 mini has visible benchmark coverage in knowledge and understanding, but BenchLM does not currently assign it a global category rank there.

Is GPT-5.4 mini good for coding and programming?

GPT-5.4 mini has visible benchmark coverage in coding and programming, but BenchLM does not currently assign it a global category rank there.

Is GPT-5.4 mini good for agentic tool use and computer tasks?

GPT-5.4 mini has visible benchmark coverage in agentic tool use and computer tasks, but BenchLM does not currently assign it a global category rank there.

Is GPT-5.4 mini good for multimodal and grounded tasks?

GPT-5.4 mini has visible benchmark coverage in multimodal and grounded tasks, but BenchLM does not currently assign it a global category rank there.

Which sibling models are related to GPT-5.4 mini?

GPT-5.4 mini belongs to the GPT-5.4 family. Related variants on BenchLM include GPT-5.4, GPT-5.4 Pro, GPT-5.4 nano.

Does GPT-5.4 mini have full benchmark coverage on BenchLM?

Not yet. GPT-5.4 mini currently has 11 published benchmark scores out of the 185 benchmarks BenchLM tracks. BenchLM only exposes non-generated public benchmark rows, so missing categories stay blank until a sourced evaluation is available.

What is the context window size of GPT-5.4 mini?

GPT-5.4 mini has a context window of 400K, which determines how much text it can process in a single interaction.

Last updated: May 1, 2026 · Runtime metrics stay blank until BenchLM has a sourced snapshot.

Don't miss the next GPT moment

Which models moved up, what’s new, and what it costs. One email a week, 3-min read.

Free. One email per week.