mvpmaster/MistralDpoPearl-7b-slerp

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 19, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

MistralDpoPearl-7b-slerp is a 7 billion parameter language model created by mvpmaster, formed by merging louisbrulenaudet/Pearl-7B-slerp and NousResearch/Nous-Hermes-2-Mistral-7B-DPO using a slerp merge method. This model leverages the strengths of its base components, offering a robust foundation for general-purpose text generation tasks. It is designed for developers seeking a capable 7B model derived from established Mistral-based architectures.

Loading preview...

Model Overview

MistralDpoPearl-7b-slerp is a 7 billion parameter language model developed by mvpmaster. It is a product of merging two distinct Mistral-based models: louisbrulenaudet/Pearl-7B-slerp and NousResearch/Nous-Hermes-2-Mistral-7B-DPO. This merge was performed using the slerp (spherical linear interpolation) method, facilitated by LazyMergekit, to combine their respective learned representations.

Key Characteristics

  • Architecture: Based on the Mistral 7B family, inheriting its efficient and performant design.
  • Merge Method: Utilizes slerp (spherical linear interpolation) for combining model weights, which can lead to a balanced integration of features from the merged components.
  • Parameter Count: 7 billion parameters, offering a good balance between performance and computational requirements.
  • Context Length: Supports a context window of 4096 tokens.

Intended Use Cases

This model is suitable for a variety of general text generation tasks, benefiting from the DPO (Direct Preference Optimization) fine-tuning present in one of its base models. It can be used for:

  • Chatbot applications and conversational AI.
  • Content creation and text summarization.
  • Code generation and understanding (inherited from base models).
  • Experimentation with merged models to explore combined capabilities.