unsloth/Qwen2.5-14B

The unsloth/Qwen2.5-14B is a 14.7 billion parameter causal language model from the Qwen2.5 series, developed by Qwen. This base model features a transformer architecture with RoPE, SwiGLU, RMSNorm, and Attention QKV bias, supporting a context length of 131,072 tokens. It offers significantly improved capabilities in coding, mathematics, instruction following, and long text generation, with multilingual support for over 29 languages. It is intended for further fine-tuning rather than direct conversational use.

Warm
Public
14.8B
FP8
131072
License: apache-2.0
Hugging Face

No reviews yet. Be the first to review!