unsloth/DeepSeek-R1-Distill-Llama-70B
DeepSeek-R1-Distill-Llama-70B is a 70 billion parameter language model developed by DeepSeek AI, distilled from the larger DeepSeek-R1 model and based on Llama-3.3-70B-Instruct. This model is specifically fine-tuned using reasoning data generated by DeepSeek-R1, aiming to transfer advanced reasoning patterns to a smaller, dense architecture. It excels in complex reasoning tasks across math, code, and general English benchmarks, offering strong performance for applications requiring robust analytical capabilities.
No reviews yet. Be the first to review!