mhenrichsen/danskgpt-tiny

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:1.1BQuant:BF16Ctx Length:2kPublished:Jan 3, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

DanskGPT-tiny is a 1.1 billion parameter LLaMA-based large language model developed by mhenrichsen. It is a continuation of the TinyLLaMA training, specifically designed for Danish language processing. Trained on 8 billion tokens of synthetic Danish text, this foundation/completion model is optimized for generating Danish text rather than conversational chat.

Loading preview...

DanskGPT-tiny Overview

DanskGPT-tiny, developed by mhenrichsen, is a 1.1 billion parameter large language model built upon the LLaMA architecture. It represents a continued training effort from the TinyLLaMA project, specifically adapted for the Danish language.

Key Characteristics

  • Architecture: LLaMA-based, providing a robust foundation for language understanding and generation.
  • Parameter Count: Features 1.1 billion parameters, making it a compact yet capable model for specific linguistic tasks.
  • Training Data: Extensively trained on 8 billion tokens of synthetic Danish text, ensuring strong proficiency in the Danish language.
  • Model Type: Classified as a "foundation/completion" model, meaning its primary function is to generate coherent text completions based on given prompts.

Use Cases

  • Danish Text Generation: Ideal for tasks requiring the generation of Danish content, such as creative writing, content creation, or data augmentation in Danish.
  • Research and Development: Suitable for researchers and developers exploring language models tailored for less-resourced languages or specific linguistic domains.
  • Completion Tasks: Excels at completing sentences, paragraphs, or longer texts in Danish, making it useful for various text-completion applications.

It is important to note that DanskGPT-tiny is not designed for conversational chat applications but rather for text generation and completion within the Danish language.