An innovative model pioneering breakthrough efficiency in long-context processing. Using revolutionary Sparse Attention technology, this model handles massive contexts (128K tokens) with exceptional speed and minimal resource use while maintaining quality. With thinking capabilities and an impressive 65K output window, it excels at tasks requiring extensive context understanding. Perfect for processing large documents, codebases, or datasets where traditional models slow down. Exceptional value for long-context work.
Try Now163,842 tokens
8,000 tokens
$0.27
$0.40
$0.216
$15
$0.19