GAI-LLM/ko-en-llama2-13b-mixed-v3

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kLicense:cc-by-nc-2.0Architecture:Transformer Open Weights Warm

GAI-LLM/ko-en-llama2-13b-mixed-v3 is a 13 billion parameter auto-regressive language model based on the LLaMA2 transformer architecture, developed by Donghoon Oh, Hanmin Myung, and Eunyoung Kim (SK C&C G.AI Eng). This model is fine-tuned for mixed Korean and English language tasks, leveraging a combination of Open Korean Datasets including Kopen-platypus and kaist_cot_deepL. It is designed to process text input and generate text output, excelling in bilingual Korean-English applications.

Loading preview...

Overview

GAI-LLM/ko-en-llama2-13b-mixed-v3 is a 13 billion parameter auto-regressive language model built upon the LLaMA2 transformer architecture. Developed by Donghoon Oh, Hanmin Myung, and Eunyoung Kim from SK C&C G.AI Eng, this model specializes in processing and generating text in both Korean and English.

Key Capabilities

  • Bilingual Proficiency: The model is specifically trained on a mixed dataset combining Open Korean Datasets (Kopen-platypus + kaist_cot_deepL) to enhance its performance in Korean and English language tasks.
  • LLaMA2 Architecture: Leverages the robust LLaMA2 base model, specifically hyunseoki/ko-en-llama2-13b, providing a strong foundation for language understanding and generation.
  • Text-to-Text Generation: Designed for general text input and text output applications.

Training Details

The model was trained using A100 GPU 80GB * 8, indicating a significant computational investment to achieve its bilingual capabilities.

Benchmarking

Performance can be tracked and compared on the Open KO-LLM LeaderBoard, providing transparency and a reference for its standing among other Korean LLMs.

Recommended Use Cases

This model is particularly well-suited for applications requiring strong performance in both Korean and English, such as bilingual chatbots, content generation, translation assistance, or any task where understanding and generating text in these two languages is crucial.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p