davidkim205/Ko-Llama-3-8B-Instruct

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kLicense:llama3Architecture:Transformer0.0K Warm

Ko-Llama-3-8B-Instruct is an 8 billion parameter instruction-tuned causal language model developed by davidkim (changyeon kim), based on Meta-Llama-3-8B-Instruct. This model is specifically researched and fine-tuned using a rejection sampling technique to enhance performance in Korean language tasks. It aims to improve the capabilities of large language models for Korean-specific applications, leveraging its 8192 token context length.

Loading preview...

Ko-Llama-3-8B-Instruct Overview

Ko-Llama-3-8B-Instruct is an 8 billion parameter language model developed by davidkim (changyeon kim), building upon the meta-llama/Meta-Llama-3-8B-Instruct base model. Its primary focus is to advance the performance of Korean language models. The model was trained using a Supervised Fine-Tuning (SFT) approach, where the dataset (sft_rs_140k) was created through a REJECTION SAMPLING technique.

Key Capabilities and Features

  • Korean Language Optimization: Specifically designed and fine-tuned to improve performance in Korean language understanding and generation tasks.
  • Instruction Following: Benefits from the instruction-tuned base model, enabling it to follow user prompts effectively.
  • Rejection Sampling for Data Quality: Utilizes rejection sampling to create a high-quality SFT dataset, aiming for better model responses.
  • Benchmarked Performance: Evaluated on Korean-specific benchmarks such as kollm_evaluation (achieving an average accuracy of 0.47) and KEval (scoring 5.59 average, compared to GPT-4's 6.79).

When to Use This Model

This model is particularly suitable for applications requiring strong Korean language capabilities, especially when leveraging the Llama 3 architecture. It's a good candidate for research and development in Korean NLP, offering a specialized alternative to general-purpose models. Developers can integrate it for tasks like Korean text generation, question answering, and conversational AI, where its fine-tuning for Korean can provide an advantage.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p