Edentns/DataVortexM-7B-Instruct-v0.1

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:8kLicense:cc-by-nc-sa-4.0Architecture:Transformer Open Weights Cold

Edentns/DataVortexM-7B-Instruct-v0.1 is a 7 billion parameter instruction-tuned causal language model developed by Kwangseok Yang, Jeongwon Choi, Seunghyun Choi, and Hyoseok Choi. Built upon the Mistral-7B-Instruct-v0.2 base model, it is fine-tuned using the KoAlpaca-v1.1a dataset, specializing in Korean language understanding and generation. This model is optimized for conversational AI tasks in Korean, following the Alpaca instruction format, and features an 8192 token context length.

Loading preview...

DataVortexM-7B-Instruct-v0.1 Overview

DataVortexM-7B-Instruct-v0.1 is a 7 billion parameter instruction-tuned language model developed by a team including Kwangseok Yang and Seunghyun Choi. It is based on the Mistral-7B-Instruct-v0.2 architecture and has been specifically fine-tuned using the beomi/KoAlpaca-v1.1a dataset, indicating a strong focus on Korean language capabilities. The model adheres to the Alpaca instruction format for its conversational interactions.

Key Capabilities & Features

  • Korean Language Specialization: Fine-tuned on a Korean dataset, making it suitable for Korean-centric NLP tasks.
  • Instruction Following: Designed to understand and respond to instructions in the Alpaca format.
  • Base Model: Leverages the robust architecture of Mistral-7B-Instruct-v0.2.
  • Context Length: Supports an 8192 token context window.

Performance Insights

Benchmarking on the Ko-LLM-Leaderboard shows an average score of 39.81, with specific scores including 34.13 on Ko-ARC and 42.35 on Ko-HellaSwag. While Ko LM Eval Harness results are currently 0.0 across several tasks, the Ko-LLM-Leaderboard provides a more comprehensive view of its Korean language performance.

Ideal Use Cases

  • Korean Chatbots: Developing conversational AI agents that interact in Korean.
  • Korean Content Generation: Tasks requiring generation of text in Korean based on instructions.
  • Research & Development: Exploring instruction-tuned models for Korean NLP applications.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p