MaziyarPanahi/Mistral-ko-7B-v0.1-Mistral-7B-Instruct-v0.2-slerp

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 12, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

MaziyarPanahi/Mistral-ko-7B-v0.1-Mistral-7B-Instruct-v0.2-slerp is a 7 billion parameter language model created by MaziyarPanahi, formed by merging mistralai/Mistral-7B-Instruct-v0.2 and maywell/Mistral-ko-7B-v0.1 using the slerp method. This model combines the instruction-following capabilities of Mistral-7B-Instruct-v0.2 with the Korean language proficiency of Mistral-ko-7B-v0.1. It is designed for applications requiring a balance of general instruction adherence and strong Korean language understanding, operating with a 4096-token context length.

Loading preview...

Model Overview

Mistral-ko-7B-v0.1-Mistral-7B-Instruct-v0.2-slerp is a 7 billion parameter language model developed by MaziyarPanahi. This model is a product of merging two distinct base models: mistralai/Mistral-7B-Instruct-v0.2 and maywell/Mistral-ko-7B-v0.1. The merge was performed using the slerp (spherical linear interpolation) method, aiming to combine their respective strengths.

Key Capabilities

  • Hybrid Language Proficiency: Integrates the general instruction-following and reasoning abilities of the Mistral-7B-Instruct-v0.2 model.
  • Enhanced Korean Language Support: Incorporates the Korean language understanding and generation capabilities from the Mistral-ko-7B-v0.1 model.
  • Instruction Following: Benefits from the instruction-tuned nature of one of its base models, making it suitable for various prompt-based tasks.

Good For

  • Applications requiring a balance between general-purpose instruction following and robust Korean language processing.
  • Tasks such as text generation, summarization, and question answering in both English and Korean contexts.
  • Developers looking for a model that combines established instruction-tuned performance with specialized multilingual support, particularly for Korean.