beomi/Llama-3-Open-Ko-8B

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Apr 22, 2024License:llama3Architecture:Transformer0.2K Warm

The beomi/Llama-3-Open-Ko-8B is an 8 billion parameter language model developed by Junbum Lee (Beomi), based on the Llama-3-8B architecture. This model is a continued pre-trained version, specifically trained on over 17.7 billion tokens of publicly available, deduplicated Korean text data. It is optimized for Korean language understanding and generation, making it suitable for various natural language processing tasks in Korean.

Loading preview...

Model Overview

The beomi/Llama-3-Open-Ko-8B is an 8 billion parameter language model developed by Junbum Lee (Beomi). It is a continued pre-trained model built upon the Meta Llama-3-8B architecture, specifically adapted for the Korean language. The model was trained using publicly available resources, incorporating over 60GB of deduplicated texts, totaling more than 17.7 billion tokens with the new Llama-3 tokenizer.

Key Capabilities

  • Korean Language Specialization: Pre-trained extensively on Korean text data, enhancing its performance for Korean NLP tasks.
  • Llama-3 Foundation: Benefits from the optimized transformer architecture of the Meta Llama 3 family.
  • Publicly Sourced Training: Utilizes a large corpus of publicly available Korean texts for its continued pre-training.

Good for

  • Korean NLP Applications: Ideal for tasks requiring strong Korean language understanding and generation.
  • Foundation for Instruction Tuning: The base model can serve as an excellent starting point for creating new Korean chat or instruction-following models, as demonstrated by the related Llama-3-Open-Ko-8B-Instruct-preview.
  • Research and Commercial Use: Intended for both commercial and research applications, particularly where Korean language proficiency is critical.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p