A native OCR benchmark for reading text from images across multilingual scripts, low-quality scans, handwriting, structured layouts, charts, and screenshots.
BenchLM mirrors the published score view for OCRBench V2. Interfaze Beta leads the public snapshot at 70.7%. BenchLM does not use these results to rank models overall.
Year
2025
Tasks
Image OCR tasks
Format
Accuracy
Difficulty
Native visual text understanding
OCRBench V2 evaluates whether multimodal models can extract visual text directly from images before downstream reasoning or structure extraction. BenchLM stores Interfaze's reported score as a display-only OCR row.
Version
OCRBench V2 2025
Refresh cadence
Quarterly
Staleness state
Current
Question availability
Public benchmark set
BenchLM uses freshness metadata to decide whether a benchmark should still be treated as a strong differentiator, a benchmark to watch, or a display-only reference. For the full scoring policy, see the BenchLM methodology page.
A native OCR benchmark for reading text from images across multilingual scripts, low-quality scans, handwriting, structured layouts, charts, and screenshots.
Interfaze Beta by Interfaze currently leads with a score of 70.7% on OCRBench V2.
1 AI models have been evaluated on OCRBench V2 on BenchLM.
For engineers, researchers, and the plain curious — a weekly brief on new models, ranking shifts, and pricing changes.
Free. No spam. Unsubscribe anytime.