DeepSeek V3.2

DeepSeek

An innovative model pioneering breakthrough efficiency in long-context processing. Using revolutionary Sparse Attention technology, this model handles massive contexts (128K tokens) with exceptional speed and minimal resource use while maintaining quality. With thinking capabilities and an impressive 65K output window, it excels at tasks requiring extensive context understanding. Perfect for processing large documents, codebases, or datasets where traditional models slow down. Exceptional value for long-context work.

Try Now

Capabilities

Thinking

Tool Use

Technical Specifications

Context Window

163,842 tokens

Max Output

8,000 tokens

Pricing

Token Costs (per 1M tokens)

Cache Miss Input

$0.27

Non-Reasoning Output

$0.40

Cache Read Input

$0.216

Tool Costs (per 1K calls)

Web Search

$15

Code Execution

$0.19