An innovative experimental model pioneering breakthrough efficiency in long-context processing. Using revolutionary Sparse Attention technology, this model handles massive contexts (164K tokens) with exceptional speed and minimal resource use while maintaining quality. With thinking capabilities and an impressive 65K output window, it excels at tasks requiring extensive context understanding. Perfect for processing large documents, codebases, or datasets where traditional models slow down. Exceptional value for long-context work.
Try NowProcessing very large documents
Need long output generation (50k+ tokens)
Long-context task on a budget
163,840 tokens
65,536 tokens
$0.27
$0.41
$15
$0.19