LFM 2 24B A2B

liquid

LFM2-24B-A2B is the largest model in the LFM2 family of hybrid architectures designed for efficient on-device deployment. Built as a 24B parameter Mixture-of-Experts model with only 2B active parameters per token, it delivers high-quality generation while maintaining low inference costs. The model fits within 32 GB of RAM, making it practical to run on consumer laptops and desktops without sacrificing capability.

Try Now

Example Use Cases

On-device high-quality generation

Efficient moe inference on consumer hardware

Low-cost edge deployment

Technical Specifications

Context Window

32,768 tokens

Max Output

32,768 tokens

Pricing

Token Costs (per 1M tokens)

Cache Miss Input

$0.03

Non-Reasoning Output

$0.12

Tool Costs (per 1K calls)

Web Search

$15

Code Execution

$0.19

⚠️ Legacy

Made legacy on

Reason

Untested

Recommended Replacement

Qwen3.5 Plus