SakanaAI/TinySwallow-1.5B

TinySwallow-1.5B is a 1.5 billion parameter Japanese compact language model developed by Sakana AI and the Swallow Team. It was created using Temporally Adaptive Interpolated Distillation (TAID), a novel knowledge distillation method, with Qwen2.5-32B-Instruct as the teacher model and Qwen2.5-1.5B-Instruct as the student. The model has undergone further pre-training on Japanese text data, specifically enhancing its Japanese language capabilities for research and development purposes.

Warm
Public
1.5B
BF16
131072
License: apache-2.0
Hugging Face

No reviews yet. Be the first to review!