BramVanroy/GEITje-7B-ultra-sft

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:8kPublished:Jan 22, 2024License:cc-by-nc-4.0Architecture:Transformer0.0K Open Weights Cold

BramVanroy/GEITje-7B-ultra-sft is a 7 billion parameter instruction-tuned causal language model developed by Bram Vanroy, based on Rijgersberg/GEITje-7B and Mistral 7B. Fine-tuned on 240M tokens of synthetic Dutch datasets, including GPT-3.5-turbo and GPT-4-turbo data, it excels in multi-turn conversations and code generation with an 8192-token context length. This model is specifically designed for conversational AI in Dutch, leveraging diverse synthetic data for robust interaction.

Loading preview...

GEITje-7B-ultra-sft: A Conversational Dutch LLM

BramVanroy/GEITje-7B-ultra-sft is a 7 billion parameter instruction-tuned model, building upon the Mistral 7B-based Rijgersberg/GEITje-7B, with further pretraining on Dutch data. This model is specifically fine-tuned for conversational use, leveraging a diverse set of synthetic datasets totaling approximately 240 million tokens, including data generated by GPT-3.5-turbo and GPT-4-turbo.

Key Capabilities & Training Insights

  • Dutch Conversational AI: Optimized for multi-turn conversations in Dutch, incorporating various user personas (e.g., language learners, experts, children) during training to enhance adaptability.
  • Synthetic Data Training: Trained on a unique blend of translated and newly generated Dutch datasets, with 85.42% from BramVanroy/ultrachat_200k_dutch (GPT-4-turbo, multi-turn) and significant contributions from StackOverflow, Alpaca, and Dolly datasets.
  • Context Length: Supports an 8192-token context length, enabling more extensive and coherent conversations.
  • Training Methodology: Trained in full (without LoRA) using bfloat16 and Flash Attention 2, following the Hugging Face alignment handbook.
  • System Message Compatibility: Utilizes the Zephyr chat template, allowing for the inclusion of system messages in conversations.

Important Considerations

  • Alignment: This model is an SFT (chat-tuned) version and has not been aligned with DPO or other reinforcement learning techniques. For aligned use, the DPO variant is recommended.
  • Commercial Use: Due to its training on synthetic data derived from OpenAI/Azure services, this model is not suitable for commercial purposes.
  • Limitations: As an unaligned model, it may generate inaccurate, misleading, or potentially offensive content. Users should exercise caution and use it at their own risk.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p