unsloth/Qwen2.5-3B

unsloth/Qwen2.5-3B is a 3.09 billion parameter causal language model from the Qwen2.5 series, developed by Qwen. This base model features a 32,768-token context length and is designed for pretraining, offering significantly improved capabilities in coding, mathematics, instruction following, and generating structured outputs like JSON. It supports over 29 languages and is intended for further fine-tuning rather than direct conversational use.

Warm
Public
3.1B
BF16
32768
License: other
Hugging Face