Overview
Overview
This model, denial07/Qwen2-72B-Instruct-kor-dpo, is an enhanced version of the Qwen2-72B-Instruct model, specifically fine-tuned to improve its performance in the Korean language. It features 72.7 billion parameters and supports a substantial context length of 131,072 tokens, making it suitable for complex and lengthy Korean language tasks.
Key Capabilities and Performance
The model demonstrates strong performance across several Korean-specific benchmarks:
- LogicKor Benchmark: Achieves a total score of 8.79, ranking competitively among large language models. It shows notable strengths in Coding (9.71), Understanding (9.64), and Writing (9.14) within the Korean context, outperforming the base Qwen2-72B-Instruct model in Reasoning, Math, Writing, and overall score.
- KMMLU Benchmark: Maintains an overall accuracy of 0.58, performing on par with the base Qwen2-72B-Instruct and slightly better than Mistral-Large-Instruct-2407 across categories like HUMSS, STEM, Applied Science, and Other.
When to Use This Model
This model is particularly well-suited for applications requiring high-quality Korean language understanding and generation. Its improvements over the base Qwen2 model make it a strong candidate for tasks such as:
- Korean-centric content creation and summarization.
- Complex reasoning and problem-solving in Korean.
- Code generation and understanding for Korean-speaking developers.
- Instruction-following tasks where nuanced Korean language comprehension is critical.