mlx-community/DeepSeek-R1-Distill-Qwen-14B

The mlx-community/DeepSeek-R1-Distill-Qwen-14B is a 14 billion parameter language model converted to MLX format from deepseek-ai/DeepSeek-R1-Distill-Qwen-14B. This model is designed for general language tasks and is optimized for efficient deployment and inference on Apple silicon via the MLX framework. It offers a 32768 token context length, making it suitable for processing extensive inputs and generating detailed responses.

Warm
Public
14B
FP8
32768
Hugging Face

No reviews yet. Be the first to review!