Saxo/Linkbricks-Horizon-AI-Korean-Gemma-2-sft-dpo-27B

Cold
Public
27B
FP8
32768
Aug 7, 2024
License: apache-2.0
Hugging Face
Overview

Model Overview

Saxo/Linkbricks-Horizon-AI-Korean-Gemma-2-sft-dpo-27B is a 27 billion parameter Korean language model developed by Yunsung Ji (Saxo), a data scientist at Linkbricks Horizon-AI. It is fine-tuned from the gemma-2-27b-it base model using Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO) techniques on H100-80G GPUs.

Key Capabilities

  • Multilingual Proficiency: Trained with cross-lingual data for Korean, Chinese, English, and Japanese, enhancing understanding across these languages.
  • Logical Reasoning: Specifically designed to handle complex Korean logical problems.
  • Enhanced Analysis: Strengthened for high-level analysis of customer reviews and social media postings.
  • Coding Support: Features enhanced capabilities for coding tasks.
  • Efficient Training: Utilizes Deepspeed Stage 3, rslora, and BAdam Layer Mode for efficient training.

Performance

This model achieved Rank-1 on the Open Ko LLM Leaderboard Season 2 (2024/11/01~2024/12/28) with an average score of 51.37. Notable benchmark scores include:

  • Ko-GPQA: 25.25
  • Ko-Winogrande: 68.27
  • Ko-GSM8k: 70.96
  • Ko-EQ Bench: 50.25
  • Ko-IFEval: 49.84

Use Cases

This model is well-suited for applications requiring advanced Korean language understanding, complex logical problem-solving, multilingual text processing, and specialized tasks like customer feedback analysis or code generation.