unsloth/Qwen2.5-7B

unsloth/Qwen2.5-7B is a 7.61 billion parameter causal language model from the Qwen2.5 series, developed by Qwen. This base model features a Transformer architecture with RoPE, SwiGLU, and RMSNorm, supporting a substantial 131,072 token context length. It offers significantly improved capabilities in coding, mathematics, instruction following, and long text generation, alongside robust multilingual support for over 29 languages.

Warm
Public
7.6B
FP8
131072
License: apache-2.0
Hugging Face

No reviews yet. Be the first to review!