MLP-KTLim/llama-3-Korean-Bllossom-8B
Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Apr 25, 2024License:llama3Architecture:Transformer0.4K Warm

MLP-KTLim/llama-3-Korean-Bllossom-8B is an 8 billion parameter Korean-English bilingual language model developed by MLPLab at Seoultech, Teddysum, and Yonsei University. Based on the Llama 3 architecture, it features significant Korean vocabulary expansion (over 30,000 words) and enhanced Korean context processing, supporting up to 8192 tokens. This model is optimized for Korean language tasks, leveraging extensive Korean pre-training data (250GB) and instruction tuning with culturally relevant data, achieving state-of-the-art scores on the LogicKor Korean benchmark for models under 10B parameters.

Loading preview...

Overview

MLP-KTLim/llama-3-Korean-Bllossom-8B is an 8 billion parameter Korean-English bilingual language model built upon the Llama 3 architecture. Developed through a collaboration between MLPLab at Seoultech, Teddysum, and Yonsei University, this model is specifically designed to excel in Korean language understanding and generation, while also linking Korean and English knowledge.

Key Capabilities

  • Extensive Korean Vocabulary Expansion: Features over 30,000 expanded Korean vocabulary words, significantly enhancing Korean expressiveness.
  • Enhanced Korean Context Handling: Processes approximately 25% longer Korean contexts compared to base Llama 3, with a context length of 8192 tokens.
  • Bilingual Knowledge Linking: Utilizes Korean-English parallel corpora during pre-training to connect knowledge across both languages.
  • Culturally Aligned Instruction Tuning: Fine-tuned with custom instruction-following data crafted by linguists, considering Korean language and culture.
  • Reinforcement Learning (DPO): Incorporates human feedback through DPO for improved performance.
  • Strong Korean Performance: Achieved state-of-the-art scores on the LogicKor benchmark for Korean models under 10 billion parameters, demonstrating strong capabilities in reasoning, writing, and understanding.

Good For

  • Applications requiring robust Korean language processing.
  • Bilingual Korean-English tasks and knowledge transfer.
  • Developing custom models with commercial use cases, leveraging its optimized Korean foundation.
Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p