ghost4280/Ghost-V5-Ultra-8B

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Apr 25, 2026Architecture:Transformer0.0K Cold

Ghost-V5-Ultra-8B by ghost4280 is a 7 billion parameter language model, a merge of Mistral-7B-Instruct-v0.2 and dolphin-2.2.1-mistral-7b using the SLERP method. This model combines the instruction-following capabilities of Mistral with the fine-tuning characteristics of Dolphin. It is designed for general-purpose conversational AI and instruction-based tasks, leveraging a 4096 token context length.

Loading preview...

Model Overview

Ghost-V5-Ultra-8B is a 7 billion parameter language model developed by ghost4280, created through a strategic merge of two distinct pre-trained models: mistralai/Mistral-7B-Instruct-v0.2 and cognitivecomputations/dolphin-2.2.1-mistral-7b. This merge was performed using the SLERP (Spherical Linear Interpolation) method, a technique known for smoothly combining the characteristics of different models.

Key Capabilities

  • Hybrid Performance: By merging Mistral-7B-Instruct-v0.2, known for its strong instruction-following and general language understanding, with dolphin-2.2.1-mistral-7b, which often features enhanced conversational and creative capabilities, Ghost-V5-Ultra-8B aims to offer a balanced performance profile.
  • Instruction Following: Inherits robust instruction-following abilities from its Mistral base, making it suitable for a wide range of prompt-based tasks.
  • Context Handling: Supports a context length of 4096 tokens, allowing for processing and generating moderately long texts.

Good For

  • General-purpose conversational AI: Ideal for chatbots and interactive applications requiring nuanced responses.
  • Instruction-based tasks: Excels at following explicit commands and generating specific types of content based on prompts.
  • Experimentation with merged models: Provides a solid base for developers interested in exploring the characteristics of SLERP-merged models.