unsloth/Qwen2.5-0.5B
unsloth/Qwen2.5-0.5B is a 0.49 billion parameter causal language model from the Qwen2.5 series, developed by Qwen. This base model features a 32,768 token context length and is designed for pretraining, offering significant improvements in knowledge, coding, mathematics, and multilingual support across 29 languages. It is optimized for further fine-tuning for specific applications rather than direct conversational use.