Mixtral 8x22B is currently the most performant open model. A 22B sparse Mixture-of-Experts (SMoE). Uses only 39B active parameters out of 141B.
Try NowComplex task requiring large model capacity
Multilingual generation
Math and coding on a budget
65,536 tokens
65,536 tokens
$2 per 1M tokens
$6 per 1M tokens
$15 per 1K calls
$0.19 per 1K calls