denial07/Qwen2-72B-Instruct-kor-dpo

Warm
Public
72.7B
FP8
131072
License: other
Hugging Face
Overview

Overview

This model, denial07/Qwen2-72B-Instruct-kor-dpo, is an enhanced version of the Qwen2-72B-Instruct model, specifically fine-tuned to improve its performance in the Korean language. It features 72.7 billion parameters and supports a substantial context length of 131,072 tokens, making it suitable for complex and lengthy Korean language tasks.

Key Capabilities and Performance

The model demonstrates strong performance across several Korean-specific benchmarks:

  • LogicKor Benchmark: Achieves a total score of 8.79, ranking competitively among large language models. It shows notable strengths in Coding (9.71), Understanding (9.64), and Writing (9.14) within the Korean context, outperforming the base Qwen2-72B-Instruct model in Reasoning, Math, Writing, and overall score.
  • KMMLU Benchmark: Maintains an overall accuracy of 0.58, performing on par with the base Qwen2-72B-Instruct and slightly better than Mistral-Large-Instruct-2407 across categories like HUMSS, STEM, Applied Science, and Other.

When to Use This Model

This model is particularly well-suited for applications requiring high-quality Korean language understanding and generation. Its improvements over the base Qwen2 model make it a strong candidate for tasks such as:

  • Korean-centric content creation and summarization.
  • Complex reasoning and problem-solving in Korean.
  • Code generation and understanding for Korean-speaking developers.
  • Instruction-following tasks where nuanced Korean language comprehension is critical.