All Models
DeepSeekactiveOpen Source
DeepSeek-R1-Distill-Qwen-32B
deepseek-r1-distill-qwen-32bR1 reasoning distilled into Qwen2.5-32B. Strong reasoning in a smaller model.
Context Window
131.1K
tokens
Max Output
8.2K
tokens
Input Price
—
per 1M tokens
Output Price
—
per 1M tokens
Details
Familydeepseek-r1
Parameters32B
Training Cutoff2024-12-01
ReleasedJanuary 20, 2025
Evaluation Scores(5 benchmarks)
MATH-500
82.8%
HumanEval
72.6%
MMLU-Pro
63.4%
GPQA Diamond
48.9%
Quick Access
curl pikaainews.com/api/models/deepseek-r1-distill-qwen-32bnpx pika-models info deepseek-r1-distill-qwen-32bGet API Access
Third-Party Providers & Aggregators
Cerebras
Wafer-scale inference. 1000+ tokens/sec for select models.
DeepInfra
Lowest per-token rates for open-source models.
Fireworks AI
Fastest inference engine. Multimodal support, HIPAA/SOC2.
Groq
Ultra-fast LPU inference. Best latency for real-time apps.
OpenRouter
500+ models, one API key. Pay-per-token, no minimums.
SiliconFlow
China-optimized inference. Strong Qwen/DeepSeek support.
Together AI
Fast open-source model inference. Sub-100ms latency.
