davidkim205/Ko-Llama-3-8B-Instruct
Ko-Llama-3-8B-Instruct is an 8 billion parameter instruction-tuned causal language model developed by davidkim (changyeon kim), based on Meta-Llama-3-8B-Instruct. This model is specifically researched and fine-tuned using a rejection sampling technique to enhance performance in Korean language tasks. It aims to improve the capabilities of large language models for Korean-specific applications, leveraging its 8192 token context length.
Loading preview...
Ko-Llama-3-8B-Instruct Overview
Ko-Llama-3-8B-Instruct is an 8 billion parameter language model developed by davidkim (changyeon kim), building upon the meta-llama/Meta-Llama-3-8B-Instruct base model. Its primary focus is to advance the performance of Korean language models. The model was trained using a Supervised Fine-Tuning (SFT) approach, where the dataset (sft_rs_140k) was created through a REJECTION SAMPLING technique.
Key Capabilities and Features
- Korean Language Optimization: Specifically designed and fine-tuned to improve performance in Korean language understanding and generation tasks.
- Instruction Following: Benefits from the instruction-tuned base model, enabling it to follow user prompts effectively.
- Rejection Sampling for Data Quality: Utilizes rejection sampling to create a high-quality SFT dataset, aiming for better model responses.
- Benchmarked Performance: Evaluated on Korean-specific benchmarks such as
kollm_evaluation(achieving an average accuracy of 0.47) andKEval(scoring 5.59 average, compared to GPT-4's 6.79).
When to Use This Model
This model is particularly suitable for applications requiring strong Korean language capabilities, especially when leveraging the Llama 3 architecture. It's a good candidate for research and development in Korean NLP, offering a specialized alternative to general-purpose models. Developers can integrate it for tasks like Korean text generation, question answering, and conversational AI, where its fine-tuning for Korean can provide an advantage.
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.