Prashasst/Luffy-DeepSeek-R1-Distill-Llama-8B-4-bit

Prashasst/Luffy-DeepSeek-R1-Distill-Llama-8B-4-bit is an 8 billion parameter language model developed by Prashasst Dongre. This model is a distilled version of DeepSeek-R1-Distill-Llama, leveraging a 4-bit quantization for efficient deployment. It is based on the Llama architecture and is designed for general language understanding and generation tasks. With a context length of 32768 tokens, it can process extensive inputs for various applications.

Warm
Public
8B
FP8
32768
License: mit
Hugging Face

No reviews yet. Be the first to review!