by DeepSeek· 1 years ago
R1 reasoning distilled into Llama 3 architecture. Strong reasoning at minimal compute cost.
Context Window
128K
Input Price
$0.07/M tokens
Output Price
$0.14/M tokens
Performance Profile
Budget-friendly at just $0.07/M input tokens
128K token context window — handles lengthy documents with ease
Fully open source — self-host, fine-tune, and customize without restrictions
8B parameter architecture for deep reasoning
vs similar-tier models
| Model | Input | Output | Context | Avg Score |
|---|---|---|---|---|
DeepSeek-R1-Distill-Llama-8BCurrent DeepSeek | $0.07 | $0.14 | 128K | 72.7 |
Claude Haiku 3.5 Anthropic | $0.80 | $4.00 | 200K | 77.0 |
Mistral Small Mistral AI | $0.10 | $0.30 | 32K | 69.8 |
Summarize an email
<$0.001~300 word email → short summary
400 in · 100 out
Analyze a 1,000-word article
<$0.001Blog post or news article → detailed analysis
1,333 in · 500 out
Chatbot conversation (10 turns)
<$0.001Full customer support interaction
4,000 in · 2,000 out
Summarize a 50-page report
$0.0029Legal contract or research paper → key points
37,500 in · 2,000 out
Review a 5,000-line codebase
$0.0022Full code review with suggestions
25,000 in · 3,000 out
Process a full novel
$0.0091~90,000 words → detailed summary & analysis
120,000 in · 5,000 out
Email summaries
$1/mo
$0.04/day
Chat conversations
$17/mo
$0.56/day
Document analysis
$87/mo
$3/day
No ratings yet. Be the first to rate this model!
Sign in to rate this model and share your experience.
Sign in to leave a comment and join the discussion.
DeepSeek
Hybrid model combining V3 and R1 strengths. Improved reasoning with RL techniques from R1.
Input
$0.27/M
Output
$1.10/M
Context
128K
DeepSeek
DeepSeek's open-source MoE model rivaling frontier models at a fraction of the cost.
Input
$0.27/M
Output
$1.10/M
Context
128K
DeepSeek
DeepSeek's reasoning model with transparent chain-of-thought. Open-source and highly competitive.
Input
$0.55/M
Output
$2.19/M
Context
128K
Anthropic
Anthropic's fastest and most affordable model. Great for high-volume, low-latency tasks.
Input
$0.80/M
Output
$4.00/M
Context
200K
Mistral AI
Mistral's efficient model for everyday tasks. Fast and cost-effective.
Input
$0.10/M
Output
$0.30/M
Context
32K
OpenAI
A fast, affordable variant of GPT-4.1 for high-volume workloads.
Input
$0.40/M
Output
$1.60/M
Context
1.0M