AI-Sweden-Models/gpt-sw3-6.7b
GPT-Sw3 6.7B is a 7.1 billion parameter decoder-only transformer language model developed by AI Sweden in collaboration with RISE and WASP WARA for Media and Language. It was pretrained on 320 billion tokens across Swedish, Norwegian, Danish, Icelandic, English, and programming code. This model is designed for generating coherent text in multiple Nordic languages and English, as well as handling various text generation tasks.
No reviews yet. Be the first to review!