AI-Sweden-Models/gpt-sw3-20b-instruct

The AI-Sweden-Models/gpt-sw3-20b-instruct is a 20.9 billion parameter decoder-only transformer language model developed by AI Sweden in collaboration with RISE and WASP WARA for Media and Language. It was pretrained on 320 billion tokens across Swedish, Norwegian, Danish, Icelandic, English, and programming code, then fine-tuned on instruction data. This model is designed for generating coherent text in five languages and four programming languages, and can perform various text tasks through instruction-following.

Cold
Public
20.9B
FP8
2048
License: other
Hugging Face

No reviews yet. Be the first to review!