simplescaling/s1.1-32B
The simplescaling/s1.1-32B model is a 32.8 billion parameter language model developed by simplescaling, serving as the successor to s1-32B. It features a 131,072 token context length and is specifically optimized for enhanced reasoning performance. This model leverages reasoning traces from r1, distinguishing it from its predecessor which used Gemini traces, making it particularly strong in complex problem-solving and mathematical tasks.
No reviews yet. Be the first to review!