MiniMax-01

MiniMax

MiniMax-01 combines MiniMax-Text-01 for text generation and MiniMax-VL-01 for image understanding. It has 456 billion parameters, with 45.9 billion parameters activated per inference, and can handle a context of up to 4 million tokens. The text model adopts a hybrid architecture that combines Lightning Attention, Softmax Attention, and Mixture-of-Experts (MoE). The image model adopts the "ViT-MLP-LLM" framework and is trained on top of the text model.

Try Now

Capabilities

Image Input

Technical Specifications

Context Window

1,000,192 tokens

Max Output

1,000,192 tokens

Pricing

Token Costs (per 1M tokens)

Cache Miss Input

$0.20

Non-Reasoning Output

$1.10

Legacy

Made legacy on

Reason

Original MiniMax model; three generations behind

Recommended Replacement

MiniMax M2.7