AI-Sweden-Models/gpt-sw3-356m
TEXT GENERATIONConcurrency Cost:1Model Size:0.5BQuant:BF16Ctx Length:2kPublished:Dec 14, 2022License:otherArchitecture:Transformer0.0K Loading
GPT-Sw3 356M is a 0.5 billion parameter decoder-only transformer language model developed by AI Sweden in collaboration with RISE and WASP WARA for Media and Language. Trained on 320 billion tokens across Swedish, Norwegian, Danish, Icelandic, English, and programming code, it generates coherent text in five languages and four programming languages. This model is part of a collection focused on advancing large language models for Nordic languages, with a context length of 2048 tokens.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
–
top_p
–
top_k
–
frequency_penalty
–
presence_penalty
–
repetition_penalty
–
min_p
–