fluently-lm/FluentlyLM-Prinum
FluentlyLM-Prinum is a 32.5 billion parameter causal language model developed by fluently-lm, based on the QwenForCausalLM architecture. It features a substantial 131,072 token context length and supports multiple languages including English, French, Spanish, Russian, Chinese, Japanese, and Persian. This model is designed for general-purpose language generation and understanding, demonstrating competitive performance on the Open LLM Leaderboard.