Qwen3-8B
qwen3-8bEfficient 8B model for cost-sensitive deployments.
Context Window
131.1K
tokens
Max Output
8.2K
tokens
Input Price
$0.06
per 1M tokens
Output Price
$0.24
per 1M tokens
Details
Capabilities
https://dashscope.aliyuncs.com/compatible-mode/v1/chat/completionsEvaluation Scores(4 benchmarks)
Quick Access
curl pikaainews.com/api/models/qwen-qwen3-8bnpx pika-models info qwen-qwen3-8bGet API Access
Third-Party Providers & Aggregators
Cerebras
Wafer-scale inference. 1000+ tokens/sec for select models.
DeepInfra
Lowest per-token rates for open-source models.
Fireworks AI
Fastest inference engine. Multimodal support, HIPAA/SOC2.
Groq
Ultra-fast LPU inference. Best latency for real-time apps.
OpenRouter
500+ models, one API key. Pay-per-token, no minimums.
SiliconFlow
China-optimized inference. Strong Qwen/DeepSeek support.
Together AI
Fast open-source model inference. Sub-100ms latency.
Other qwen3 models
Qwen3-Max-Thinking
qwen3-max-thinking
Qwen3-Next
qwen3-next
Qwen3-Max
qwen3-max
Qwen3-32B
qwen3-32b
Qwen3-4B
qwen3-4b
Qwen3-1.7B
qwen3-1.7b
Qwen3-14B
qwen3-14b
Qwen3-0.6B
qwen3-0.6b
Qwen3-235B-A22B
qwen3-235b-a22b
