mlx-community/DeepSeek-R1-Distill-Qwen-14B
The mlx-community/DeepSeek-R1-Distill-Qwen-14B is a 14 billion parameter language model converted to MLX format from deepseek-ai/DeepSeek-R1-Distill-Qwen-14B. This model is designed for general language tasks and is optimized for efficient deployment and inference on Apple silicon via the MLX framework. It offers a 32768 token context length, making it suitable for processing extensive inputs and generating detailed responses.