AI-Sweden-Models/gpt-sw3-6.7b

GPT-Sw3 6.7B is a 7.1 billion parameter decoder-only transformer language model developed by AI Sweden in collaboration with RISE and WASP WARA for Media and Language. It was pretrained on 320 billion tokens across Swedish, Norwegian, Danish, Icelandic, English, and programming code. This model is designed for generating coherent text in multiple Nordic languages and English, as well as handling various text generation tasks.

Cold
Public
7.1B
FP8
2048
License: other
Hugging Face

No reviews yet. Be the first to review!