A highly performant 32B multilingual model designed to rival monolingual performance through innovations in instruction tuning with data arbitrage, preference training, and model merging. Serves 23 languages including Arabic, Chinese, Japanese, Korean, and major European languages. With 128K context window, it handles substantial multilingual workloads effectively.
Try NowMultilingual task across 23 languages
Cross-lingual reasoning and generation
Non-english content generation on a budget
128,000 tokens
4,000 tokens
$0.50 per 1M tokens
$1.50 per 1M tokens
$15 per 1K calls
$0.19 per 1K calls