Overview
Q2.5-MS-Mistoria-72b-v2 Overview
Q2.5-MS-Mistoria-72b-v2 is a 72.7 billion parameter language model created by SteelSkull, representing the second version of their 72B model series. Built on the Qwen 2.5 architecture, its primary goal is to combine the strong storytelling abilities of various merged models with sustained intelligence. The model utilizes a Qwen format for interaction and has a substantial context length of 131072 tokens.
Key Characteristics
- Architecture: Based on Qwen 2.5, indicating a robust foundation for language understanding and generation.
- Merging Strategy: Employs a "Model Stock" merge method, combining components from
Nexusflow/Athene-V2-Chat,EVA-UNIT-01/EVA-Qwen2.5-72B-v0.2, andshuttleai/shuttle-3to achieve its specific blend of capabilities. - Focus: Aims to excel in robust storytelling while preserving overall intelligence, making it suitable for complex narrative generation and detailed conversational applications.
Use Cases
This model is particularly well-suited for applications requiring:
- Advanced Storytelling: Its design prioritizes generating coherent and engaging narratives.
- Intelligent Conversation: The focus on maintaining intelligence alongside storytelling suggests strong performance in complex dialogue and interactive scenarios.