akshayballal/Qwen2.5-1.5B-Instruct-SFT-MedQA-merged

The akshayballal/Qwen2.5-1.5B-Instruct-SFT-MedQA-merged model is a 1.5 billion parameter instruction-tuned Qwen2.5 variant, developed by akshayballal. It was fine-tuned using Unsloth and Huggingface's TRL library, resulting in faster training. This model is specifically optimized for instruction-following tasks, leveraging its Qwen2.5 architecture for efficient performance.

Warm
Public
1.5B
BF16
32768
License: apache-2.0
Hugging Face

No reviews yet. Be the first to review!