An innovative model pioneering breakthrough efficiency in long-context processing. Using revolutionary Sparse Attention technology, this model handles massive contexts (128K tokens) with exceptional speed and minimal resource use while maintaining quality. With thinking capabilities and an impressive 65K output window, it excels at tasks requiring extensive context understanding. Perfect for processing large documents, codebases, or datasets where traditional models slow down. Exceptional value for long-context work.
Try Now128,000 tokens
65,536 tokens
$0.28 per 1M tokens
$0.42 per 1M tokens
Minimum plan required
$0 per month