unsloth/Qwen2.5-7B
unsloth/Qwen2.5-7B is a 7.61 billion parameter causal language model from the Qwen2.5 series, developed by Qwen. This base model features a Transformer architecture with RoPE, SwiGLU, and RMSNorm, supporting a substantial 131,072 token context length. It offers significantly improved capabilities in coding, mathematics, instruction following, and long text generation, alongside robust multilingual support for over 29 languages.