robinsmits/Qwen1.5-7B-Dutch-Chat
TEXT GENERATIONConcurrency Cost:1Model Size:7.7BQuant:FP8Ctx Length:32kPublished:Mar 29, 2024License:cc-by-nc-4.0Architecture:Transformer0.0K Open Weights Cold

The robinsmits/Qwen1.5-7B-Dutch-Chat is a 7.7 billion parameter DPO-aligned language model based on the Qwen1.5 architecture, fine-tuned specifically for the Dutch language. It leverages the Qwen/Qwen1.5-7B-Chat as its base model and was trained on the Dutch ultra_feedback_dutch_cleaned dataset. This model demonstrates performance in Dutch natural language understanding and generation tasks comparable to GPT-3.5, making it suitable for Dutch-centric conversational AI applications.

Loading preview...

robinsmits/Qwen1.5-7B-Dutch-Chat: A Dutch-Optimized LLM

This model is a 7.7 billion parameter language model, fine-tuned by robinsmits, specifically designed for Dutch language applications. It is built upon the robust Qwen1.5-7B-Chat base model and has undergone Direct Preference Optimization (DPO) using the Dutch BramVanroy/ultra_feedback_dutch_cleaned dataset.

Key Capabilities and Performance

  • Dutch Language Proficiency: Optimized for natural language understanding and generation in Dutch.
  • DPO Alignment: Enhanced through DPO finetuning for improved conversational quality and alignment with human preferences.
  • ScandEval Performance: Achieves scores on the Dutch Natural Language Understanding and Dutch Natural Language Generation leaderboards that are very close to GPT-3.5's performance.
  • Base Model Strengths: Inherits the general capabilities of the Qwen1.5-7B-Chat model, including a 32K context length.

Evaluation Highlights

While primarily focused on Dutch, the model also shows competitive English performance on the Open LLM Leaderboard, with an average score of 53.66. Notable scores include 76.03 on HellaSwag (10-Shot) and 62.38 on MMLU (5-Shot).

Intended Uses

This model is particularly well-suited for applications requiring high-quality Dutch language processing, such as chatbots, content generation, and language understanding tasks in Dutch. Users should be aware of potential biases and hallucinations inherent in LLMs and perform thorough validation for specific use cases.