Bllossom/llama-3.2-Korean-Bllossom-3B

Warm
Public
3.2B
BF16
32768
License: llama3.2
Hugging Face
Overview

Overview

Bllossom/llama-3.2-Korean-Bllossom-3B is a 3.2 billion parameter bilingual language model developed by the Bllossom team. It is built upon the Llama 3.2 architecture and features a 32768 token context length. The model has undergone extensive additional pre-training with 150GB of refined Korean data, specifically to enhance its Korean language capabilities without compromising its original English performance. This makes it a fully bilingual model, capable of handling both Korean and English tasks effectively.

Key Capabilities

  • Bilingual Proficiency: Enhanced for both Korean and English, maintaining strong performance in both languages.
  • Extensive Korean Pre-training: Utilizes 150GB of high-quality, refined Korean data for additional pre-training.
  • Instruction Tuning: Benefits from carefully curated instruction tuning.
  • Commercial Use: The model is available for commercial applications.

Good For

  • Applications requiring robust Korean language understanding and generation.
  • Use cases where strong bilingual (Korean-English) performance is critical.
  • Developers looking for a base model to further fine-tune with methods like DPO for specific performance targets, as it was not trained to target specific benchmarks.