d-rang-d/MS3-RP-Broth-24B

TEXT GENERATIONConcurrency Cost:2Model Size:24BQuant:FP8Ctx Length:32kPublished:Mar 2, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

d-rang-d/MS3-RP-Broth-24B is a 24 billion parameter merged language model based on the Mistral-Small-24B architecture, created by d-rang-d. It is an intermediate merge step for the Tantum model, combining various Mistral-Small-24B and Llama3-24B derivatives. This model is primarily optimized for roleplay and creative writing tasks, leveraging specific merging techniques like SCE, Della Linear, and Della to blend diverse model characteristics. Its main use case is as a foundation for further fine-tuning or for experimental roleplay scenarios, potentially offering unique conversational dynamics.

Loading preview...

MS3-RP-Broth-24B Overview

d-rang-d/MS3-RP-Broth-24B is a 24 billion parameter merged language model, serving as an intermediate step in the development of the larger Tantum model. This model is a complex blend of several Mistral-Small-24B and Llama3-24B derivatives, utilizing advanced merging techniques such as SCE, Della Linear, and Della. It was initially uploaded to the mergekit-community organization during its development process.

Key Capabilities

  • Roleplay Optimization: Designed with a focus on roleplay scenarios, incorporating models known for their creative and conversational abilities.
  • Complex Merging: Employs a multi-stage merging process, combining various base models and specific parameter filters to achieve a unique blend of characteristics.
  • Experimental Foundation: While an intermediate step, it is noted as potentially being usable on its own, offering a distinct personality for creative applications.

Good for

  • Roleplay and Creative Writing: Ideal for generating engaging dialogues and narratives in roleplay contexts.
  • Experimental Use Cases: Suitable for developers and researchers exploring the effects of complex model merging on specific linguistic tasks.
  • Foundation for Further Fine-tuning: Can serve as a base model for additional fine-tuning to achieve highly specialized conversational or creative outputs.