fluently-lm/FluentlyLM-Prinum

FluentlyLM-Prinum is a 32.5 billion parameter causal language model developed by fluently-lm, based on the QwenForCausalLM architecture. It features a substantial 131,072 token context length and supports multiple languages including English, French, Spanish, Russian, Chinese, Japanese, and Persian. This model is designed for general-purpose language generation and understanding, demonstrating competitive performance on the Open LLM Leaderboard.

Warm
Public
32B
FP8
32768
License: mit
Hugging Face