unsloth/QwQ-32B

QwQ-32B is a 32.5 billion parameter causal language model from Qwen, part of the Qwen series, specifically designed for enhanced reasoning capabilities. It utilizes a transformer architecture with RoPE, SwiGLU, RMSNorm, and Attention QKV bias, and supports a full context length of 131,072 tokens with YaRN for long inputs. This model is optimized for complex problem-solving and achieves competitive performance against other state-of-the-art reasoning models.

Warm
Public
32B
FP8
32768
License: apache-2.0
Hugging Face