QwQ-32B
Qwen/QwQ-32B
QwQ is the reasoning model of the Qwen series. Compared with conventional instruction-tuned models, QwQ, which is capable of thinking and reasoning, can achieve significantly enhanced performance in downstream tasks, especially hard problems. QwQ-32B is the medium-sized reasoning model, which is capable of achieving competitive performance against state-of-the-art reasoning models, e.g., DeepSeek-R1, o1-mini. The model incorporates technologies like RoPE, SwiGLU, RMSNorm, and Attention QKV bias, with 64 layers and 40 Q attention heads (8 for KV in GQA architecture)

Details
Model Provider
QwQ
Type
text
Sub Type
chat
Size
32
Publish Time
Mar 6, 2025
Input Price
$
0.15
/ M Tokens
Output Price
$
0.58
/ M Tokens
Context length
32768
Tags
Reasoning,32B,32K