s3nh/nsfw-noromaid-mistral-instruct

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 7, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

s3nh/nsfw-noromaid-mistral-instruct is a 7 billion parameter language model merged from Mistral-7B-Instruct-v0.2 and Azazelle/Half-NSFW_Noromaid-7b. This model, created using the SLERP merge method, is designed to combine the instructional capabilities of Mistral with the specialized content generation of Noromaid. It offers a 4096-token context length and is optimized for tasks requiring both general instruction following and specific content generation.

Loading preview...

Model Overview

s3nh/nsfw-noromaid-mistral-instruct is a 7 billion parameter language model resulting from a merge of two distinct base models: mistralai/Mistral-7B-Instruct-v0.2 and Azazelle/Half-NSFW_Noromaid-7b. This merge was performed using the SLERP (Spherical Linear Interpolation) method, a technique often employed to combine the strengths of different pre-trained models.

Key Capabilities

  • Instruction Following: Inherits the robust instruction-following abilities from the Mistral-7B-Instruct-v0.2 base model.
  • Specialized Content Generation: Incorporates the specialized content generation characteristics from the Azazelle/Half-NSFW_Noromaid-7b model.
  • Merged Architecture: Leverages the combined knowledge and capabilities of both constituent models through a strategic merge configuration.

When to Use This Model

This model is particularly suited for applications that require:

  • Generating responses that adhere to instructions while also incorporating specific content styles or themes.
  • Exploring the combined output characteristics of a general-purpose instruction model and a specialized content model.

Merge Details

The merge process utilized specific weighting parameters for different layers (self_attn and mlp) to fine-tune the contribution of each base model, aiming for a balanced integration of their respective strengths.