Benchmark analysis of LFM2.5-1.2B-Thinking by LiquidAI across 32 sourced tests on BenchLM.
According to BenchLM.ai, LFM2.5-1.2B-Thinking ranks #115 out of 123 models with an overall score of 33/100. While not a frontier model, it offers specific advantages depending on the use case.
LFM2.5-1.2B-Thinking is a proprietary model with a 32K token context window. It uses explicit chain-of-thought reasoning, which typically improves performance on math and complex reasoning tasks at the cost of higher latency and token usage.
LFM2.5-1.2B-Thinking sits inside the LFM2.5 1.2B family alongside LFM2.5-1.2B-Instruct.
Its strongest category is Instruction Following (#90), while its weakest is Knowledge (#118). This performance profile makes it a well-rounded choice across a range of tasks.
Creator
LiquidAI
Source Type
ProprietaryReasoning
ReasoningContext Window
32K
Overall Score
Arena Elo
1043
LFM2.5-1.2B-Thinking ranks #115 out of 123 models with an overall score of 33. It is created by LiquidAI and features a 32K context window.
LFM2.5-1.2B-Thinking ranks #118 out of 123 models in knowledge and understanding benchmarks with an average score of 27. There are stronger options in this category.
LFM2.5-1.2B-Thinking ranks #118 out of 123 models in coding and programming benchmarks with an average score of 8.2. There are stronger options in this category.
LFM2.5-1.2B-Thinking ranks #110 out of 123 models in mathematics benchmarks with an average score of 42.3. There are stronger options in this category.
LFM2.5-1.2B-Thinking ranks #113 out of 123 models in reasoning and logic benchmarks with an average score of 38.4. There are stronger options in this category.
LFM2.5-1.2B-Thinking ranks #105 out of 123 models in agentic tool use and computer tasks benchmarks with an average score of 34.1. There are stronger options in this category.
LFM2.5-1.2B-Thinking ranks #115 out of 123 models in multimodal and grounded tasks benchmarks with an average score of 32.4. There are stronger options in this category.
LFM2.5-1.2B-Thinking ranks #90 out of 123 models in instruction following benchmarks with an average score of 72. There are stronger options in this category.
LFM2.5-1.2B-Thinking ranks #97 out of 123 models in multilingual tasks benchmarks with an average score of 60.7. There are stronger options in this category.
LFM2.5-1.2B-Thinking belongs to the LFM2.5 1.2B family. Related variants on BenchLM include LFM2.5-1.2B-Instruct.
LFM2.5-1.2B-Thinking has a context window of 32K, which determines how much text it can process in a single interaction.
New model releases, benchmark scores, and leaderboard changes. Every Friday.
Free. Your signup is stored with a derived country code for compliance routing.