davidkim205/komt-mistral-7b-v1

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:8kPublished:Oct 24, 2023Architecture:Transformer0.0K Cold

The davidkim205/komt-mistral-7b-v1 is a 7 billion parameter language model fine-tuned from Mistral-7B-Instruct-v0.1 by davidkim (changyeon kim). It is specifically optimized for Korean language performance through multi-task instruction tuning using a Korean multi-task instruction dataset. This model addresses challenges in generating accurate Korean text and providing effective answers in Korean, making it suitable for Korean-centric natural language processing applications.

Loading preview...

Overview

The davidkim205/komt-mistral-7b-v1 is a 7-billion parameter language model developed by davidkim (changyeon kim). It is a fine-tuned version of the Mistral-7B-Instruct-v0.1 architecture, specifically designed to enhance Korean language performance. The model leverages a multi-task instruction tuning technique, utilizing supervised datasets from various tasks to create comprehensive training data for Large Language Models (LLMs) in Korean.

Key Capabilities

  • Improved Korean Language Performance: Addresses observed struggles of many LLMs in providing accurate answers and generating effective Korean text.
  • Multi-task Instruction Tuning: Utilizes a specialized Korean multi-task instruction dataset for training, enhancing its ability across diverse tasks.
  • Mistral-7B Base: Built upon the robust Mistral-7B-Instruct-v0.1, inheriting its strong foundational capabilities.

Performance

Evaluations, conducted using ChatGPT as a reference model (following methodologies from "Self-Alignment with Instruction Backtranslation" and "Three Ways of Using Large Language Models to Evaluate Chat"), show komt-mistral-7b-v1 achieving a score of 131 (average 3.54, 70.81%) in comparison to other open and closed-source models. This places it competitively among models specifically targeting Korean language tasks.

Good for

  • Applications requiring high-quality Korean text generation.
  • Tasks demanding accurate Korean language understanding and response generation.
  • Developers looking for an open-source model optimized for Korean NLP challenges.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p