mlx-community/DeepSeek-R1-Distill-Llama-8B
DeepSeek-R1-Distill-Llama-8B is an 8 billion parameter language model, converted to MLX format by Focused from the original DeepSeek-AI model. With a 32768 token context length, this model is designed for efficient deployment and inference within the Apple MLX ecosystem. It offers a robust foundation for various natural language processing tasks, leveraging its distilled architecture for performance.
No reviews yet. Be the first to review!