rtzr/ko-gemma-2-9b-it

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:9BQuant:FP8Ctx Length:16kPublished:Jul 11, 2024License:gemmaArchitecture:Transformer0.1K Warm

The rtzr/ko-gemma-2-9b-it is a 9 billion parameter, instruction-tuned, decoder-only large language model developed by Return Zero Team. Built upon Google's Gemma 2 architecture, this model is specifically fine-tuned for conversational tasks in Korean. It leverages Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO) on high-quality Korean datasets, making it proficient in generating Korean-language text responses to various prompts.

Loading preview...

Model Overview

rtzr/ko-gemma-2-9b-it is a 9 billion parameter conversational model developed by Return Zero Team, specifically designed for Korean language processing. It is built on the Gemma 2 architecture, a family of lightweight, open models from Google, known for their text-to-text, decoder-only capabilities.

Key Capabilities

  • Korean Language Proficiency: Fine-tuned extensively on curated high-quality Korean datasets, including translated portions of Orca-Math and dpo-mix-7k, to excel in Korean conversational tasks.
  • Instruction-Tuned: Utilizes Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO) for human feedback, enhancing its ability to follow instructions and generate relevant responses.
  • Conversational AI: Optimized for generating natural and coherent Korean text in response to questions, prompts, or documents.
  • Gemma 2 Base: Benefits from the research and technology behind Google's Gemini models, offering strong performance for text generation tasks like question answering, summarization, and reasoning.

Performance Highlights

Internal evaluations using LogicKor (assessed with GPT-4o) show competitive performance against other Korean LLMs. The model demonstrates strong scores across categories such as Reasoning, Writing, and Understanding, outperforming the base google/gemma-2-9b-it in several metrics and significantly surpassing other Korean Llama-based models.

Good for

  • Korean Chatbots and Conversational Agents: Its fine-tuning for conversational tasks makes it ideal for building interactive Korean-speaking AI.
  • Korean Content Generation: Suitable for generating various forms of Korean text, from answers to questions to summaries of documents.
  • Resource-Constrained Environments: As a Gemma-family model, its relatively small size allows for deployment in environments with limited resources, such as local machines or private cloud infrastructure.