fakeid/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-rugged_bipedal_antelope

fakeid/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-rugged_bipedal_antelope is a 0.5 billion parameter instruction-tuned language model, fine-tuned from unsloth/Qwen2.5-0.5B-Instruct. This model was trained using the GRPO method, which is designed to enhance mathematical reasoning capabilities. It is optimized for tasks requiring robust mathematical problem-solving and logical deduction, making it suitable for applications in scientific computing and data analysis.

Warm
Public
0.5B
BF16
131072
Hugging Face