SherlockAssistant/Mistral-7B-Instruct-Ukrainian

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:8kPublished:Feb 26, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

SherlockAssistant/Mistral-7B-Instruct-Ukrainian is a 7 billion parameter instruction-tuned large language model, based on Mistral-7B-v0.2, specifically fine-tuned for the Ukrainian language. It utilizes Grouped-Query Attention and Sliding-Window Attention, with an 8192-token context length. This model is optimized for Ukrainian natural language processing tasks, including question answering and general instruction following, through a multi-stage training process involving structured and unstructured Ukrainian datasets and DPO.

Loading preview...

Overview

SherlockAssistant/Mistral-7B-Instruct-Ukrainian is a 7-billion parameter instruction-tuned large language model built upon the Mistral-7B-v0.2 architecture. It has been specifically optimized for the Ukrainian language through a multi-stage fine-tuning process. This process involved initial fine-tuning with structured and unstructured Ukrainian datasets, followed by an SLERP merge with the CultriX/NeuralTrix-7B-v1 model, and concluded with DPO (Direct Preference Optimization).

Key Capabilities

  • Ukrainian Language Proficiency: Specialized in understanding and generating text in Ukrainian, making it suitable for localized applications.
  • Instruction Following: Designed to respond accurately to instructions, leveraging its instruction fine-tuning.
  • Question Answering: Trained on datasets like UA-SQUAD and Ukrainian StackExchange, enhancing its ability to answer questions.
  • Context Handling: Features an 8192-token context window, allowing for processing longer inputs.

Training Details

The model's training incorporated diverse Ukrainian datasets:

  • Structured Datasets: Includes UA-SQUAD, Ukrainian StackExchange, UAlpaca Dataset, Ukrainian subsets from Belebele and XQA, and the ZNO Dataset.
  • Unstructured Datasets: Utilized Ukrainian Wikipedia for broad language understanding.
  • DPO: Applied on a Ukrainian translation of the distilabel-indel-orca-dpo-pairs dataset to align with human preferences.

Usage

Users should format prompts using [INST] and [/INST] tokens to leverage the instruction fine-tuning effectively. A chat template is available via the apply_chat_template() method for ease of use.