simplescaling/s1.1-32B

The simplescaling/s1.1-32B model is a 32.8 billion parameter language model developed by simplescaling, serving as the successor to s1-32B. It features a 131,072 token context length and is specifically optimized for enhanced reasoning performance. This model leverages reasoning traces from r1, distinguishing it from its predecessor which used Gemini traces, making it particularly strong in complex problem-solving and mathematical tasks.

Warm
Public
32.8B
FP8
131072
License: apache-2.0
Hugging Face

No reviews yet. Be the first to review!