Aryanne/MixSwap
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 19, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold
Aryanne/MixSwap is a 7 billion parameter language model created by Aryanne using the task_swapping merge method. It combines cognitivecomputations/dolphin-2.2.1-mistral-7b, teknium/Mistral-Trismegistus-7B, and l3utterfly/mistral-7b-v0.1-layla-v4-chatml, based on Aryanne/Open-StarLake-Swap-7B. This model is specifically optimized for generating verbose, descriptive role-play conversations, making it suitable for interactive storytelling applications.
Loading preview...
Aryanne/MixSwap: A Merged Model for Role-Play
Aryanne/MixSwap is a 7 billion parameter language model developed by Aryanne, utilizing a custom branch of mergekit for its creation. This model was produced using the task_swapping merge method, building upon Aryanne/Open-StarLake-Swap-7B as its base.
Key Capabilities
- Specialized for Role-Play: The model is specifically designed and optimized for generating detailed and verbose role-play conversations.
- Merged Architecture: It integrates components from three distinct Mistral-based models:
cognitivecomputations/dolphin-2.2.1-mistral-7bteknium/Mistral-Trismegistus-7Bl3utterfly/mistral-7b-v0.1-layla-v4-chatml
- Custom Merge Configuration: The merge process involved specific
diagonal_offset,random_mask, andweightparameters for each contributing model, indicating a fine-tuned approach to combining their strengths.
Good For
- Interactive Storytelling: Ideal for applications requiring dynamic and descriptive conversational outputs in a role-playing context.
- Character Interaction: Excels at generating detailed responses for character-driven narratives.
- Customizable Prompting: Supports a specific prompt format for role-play, including start and end sequences, and an 'In Memory' section for character descriptions and conversation setup.