Qwen/QwQ-32B

Qwen/QwQ-32B is a 32.5 billion parameter causal language model developed by Qwen, designed specifically for enhanced reasoning capabilities. This model utilizes a transformer architecture with RoPE, SwiGLU, and RMSNorm, and supports an extensive context length of 131,072 tokens. It achieves competitive performance against state-of-the-art reasoning models like DeepSeek-R1 and o1-mini, making it suitable for complex problem-solving and tasks requiring deep logical inference.

Warm
Public
32.8B
FP8
131072
License: apache-2.0
Hugging Face