unsloth/Qwen2.5-14B
The unsloth/Qwen2.5-14B is a 14.7 billion parameter causal language model from the Qwen2.5 series, developed by Qwen. This base model features a transformer architecture with RoPE, SwiGLU, RMSNorm, and Attention QKV bias, supporting a context length of 131,072 tokens. It offers significantly improved capabilities in coding, mathematics, instruction following, and long text generation, with multilingual support for over 29 languages. It is intended for further fine-tuning rather than direct conversational use.
No reviews yet. Be the first to review!