A 7B sparse Mixture-of-Experts (SMoE). Uses 12.9B active parameters out of 45B total.
General text task at moderate cost
Multilingual content generation
Efficient inference at scale
32,768 tokens
$0.70
$15
$0.19
Untested