nonetrix/sillyrp-7b

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 26, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

nonetrix/sillyrp-7b is a 7 billion parameter merged language model, created using the task arithmetic method. It combines several Mistral-7B-based models, including those fine-tuned for roleplay and DPO, to enhance conversational and creative text generation. This model is designed for experimental use in generating diverse and engaging responses, particularly in role-playing scenarios.

Loading preview...

Silly RP 7B Overview

nonetrix/sillyrp-7b is an experimental 7 billion parameter language model created by merging several pre-trained models using the task arithmetic method. This merge aims to combine the strengths of its constituent models, which include:

  • tavtav/eros-7b-test
  • NousResearch/Nous-Hermes-2-Mistral-7B-DPO
  • maywell/Synatra-7B-v0.3-RP
  • cogbuji/Mr-Grammatology-clinical-problems-Mistral-7B-0.5

The base model for this merge was NeverSleep/Noromaid-7B-0.4-DPO. The model utilizes a ChatML chat template, making it compatible with common conversational interfaces.

Key Characteristics

  • Merge Method: Task arithmetic, allowing for weighted combination of model capabilities.
  • Base Architecture: Built upon Mistral-7B derivatives, leveraging their strong foundational performance.
  • Context Length: Supports a context window of 4096 tokens.
  • Experimental Nature: Developed as an exploration into model merging, with an emphasis on community feedback for quality assessment.

Potential Use Cases

  • Role-playing and Creative Writing: The inclusion of models like Synatra-7B-v0.3-RP suggests a focus on generating engaging and contextually rich dialogue for role-play or narrative creation.
  • Conversational AI: Suitable for applications requiring nuanced and diverse conversational responses.
  • Research and Experimentation: Ideal for developers and researchers interested in exploring the effects of model merging and fine-tuning for specific conversational styles.