yil384/CodeV-R1-Distill-Qwen3-0.6b

The yil384/CodeV-R1-Distill-Qwen3-0.6b is a 0.8 billion parameter language model, fine-tuned from Qwen/Qwen3-0.6B. This model has a context length of 40960 tokens and is specifically optimized for tasks related to the codev_r1_sft dataset. It is designed for applications requiring a compact yet capable model for specialized code-related functions.

Warm
Public
0.8B
BF16
32768
License: other
Hugging Face

No reviews yet. Be the first to review!