LFM2.5-1.2B-Thinking is a lightweight reasoning-focused model optimized for agentic tasks, data extraction, and RAG-while still running comfortably on edge devices. It supports long context (up to 32K tokens) and is designed to provide higher-quality "thinking" responses in a small 1.2B model.
Try Now32,768 tokens
32,768 tokens
$0
$0
1.2B thinking model; too small for production