AI-Sweden-Models/gpt-sw3-1.3b-instruct

The AI-Sweden-Models/gpt-sw3-1.3b-instruct is a 1.4 billion parameter decoder-only transformer language model developed by AI Sweden in collaboration with RISE and WASP WARA for Media and Language. This instruction-tuned model is designed for generating coherent text in Swedish, Norwegian, Danish, Icelandic, English, and programming languages, and can perform various text tasks through instruction following. It was trained on a 320 billion token dataset, emphasizing Nordic languages and code, making it particularly suitable for multilingual applications in these regions.

Cold
Public
1.4B
BF16
2048
License: other
Hugging Face

No reviews yet. Be the first to review!