BramVanroy/GEITje-7B-ultra-sft
BramVanroy/GEITje-7B-ultra-sft is a 7 billion parameter instruction-tuned causal language model developed by Bram Vanroy, based on Rijgersberg/GEITje-7B and Mistral 7B. Fine-tuned on 240M tokens of synthetic Dutch datasets, including GPT-3.5-turbo and GPT-4-turbo data, it excels in multi-turn conversations and code generation with an 8192-token context length. This model is specifically designed for conversational AI in Dutch, leveraging diverse synthetic data for robust interaction.
Loading preview...
GEITje-7B-ultra-sft: A Conversational Dutch LLM
BramVanroy/GEITje-7B-ultra-sft is a 7 billion parameter instruction-tuned model, building upon the Mistral 7B-based Rijgersberg/GEITje-7B, with further pretraining on Dutch data. This model is specifically fine-tuned for conversational use, leveraging a diverse set of synthetic datasets totaling approximately 240 million tokens, including data generated by GPT-3.5-turbo and GPT-4-turbo.
Key Capabilities & Training Insights
- Dutch Conversational AI: Optimized for multi-turn conversations in Dutch, incorporating various user personas (e.g., language learners, experts, children) during training to enhance adaptability.
- Synthetic Data Training: Trained on a unique blend of translated and newly generated Dutch datasets, with 85.42% from BramVanroy/ultrachat_200k_dutch (GPT-4-turbo, multi-turn) and significant contributions from StackOverflow, Alpaca, and Dolly datasets.
- Context Length: Supports an 8192-token context length, enabling more extensive and coherent conversations.
- Training Methodology: Trained in full (without LoRA) using bfloat16 and Flash Attention 2, following the Hugging Face alignment handbook.
- System Message Compatibility: Utilizes the Zephyr chat template, allowing for the inclusion of system messages in conversations.
Important Considerations
- Alignment: This model is an SFT (chat-tuned) version and has not been aligned with DPO or other reinforcement learning techniques. For aligned use, the DPO variant is recommended.
- Commercial Use: Due to its training on synthetic data derived from OpenAI/Azure services, this model is not suitable for commercial purposes.
- Limitations: As an unaligned model, it may generate inaccurate, misleading, or potentially offensive content. Users should exercise caution and use it at their own risk.
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.