SakanaAI/TinySwallow-1.5B-Instruct

TinySwallow-1.5B-Instruct is a 1.5 billion parameter instruction-tuned causal language model developed by Sakana AI, specifically optimized for Japanese language tasks. It was created using Temporally Adaptive Interpolated Distillation (TAID) with Qwen2.5-32B-Instruct as the teacher model and Qwen2.5-1.5B-Instruct as the student. This model excels at following instructions and engaging in conversations in Japanese, leveraging its 131072 token context length for complex interactions.

Warm
Public
1.5B
BF16
131072
License: apache-2.0
Hugging Face