NorGLM/NorGPT-369M
TEXT GENERATIONConcurrency Cost:1Model Size:0.5BQuant:BF16Ctx Length:2kPublished:Mar 8, 2024License:cc-by-nc-sa-3.0Architecture:Transformer0.0K Open Weights Cold

NorGPT-369M is a 369 million parameter generative pretrained transformer model developed by NorGLM, specifically designed for the Norwegian language. Based on the GPT-2 architecture, it is part of the NorGLM suite of language models. This model is primarily optimized for generating text in Norwegian, having been trained on a 200GB dataset including Norwegian, Danish, Swedish, German, and English text, making it suitable for non-commercial applications requiring Norwegian language generation.

Loading preview...

NorGPT-369M: A Norwegian Generative Language Model

NorGPT-369M is a 369 million parameter generative pretrained transformer model, part of the NorGLM suite, specifically engineered for the Norwegian language. Built upon the GPT-2 architecture, this model is designed to facilitate text generation and understanding in Norwegian.

Key Capabilities

  • Norwegian Language Generation: Optimized for producing coherent and contextually relevant text in Norwegian.
  • Multilingual Training: Trained on a substantial 200GB dataset, encompassing approximately 25 billion tokens, including Norwegian, Danish, Swedish, German, and English, which enhances its understanding of related languages.
  • GPT-2 Architecture: Leverages the robust and widely understood GPT-2 framework for its generative capabilities.

Good for

  • Non-Commercial Norwegian NLP Applications: Ideal for research and development in natural language processing tasks focused on Norwegian.
  • Text Generation in Norwegian: Suitable for applications requiring the creation of Norwegian text, such as content generation, chatbots, or creative writing tools.
  • Exploring Generative Models for Scandinavian Languages: Provides a strong baseline for projects involving Norwegian and related Nordic languages due to its diverse training data.