mrcuddle/Lumimaid-Muse-12B

TEXT GENERATIONConcurrency Cost:1Model Size:12BQuant:FP8Ctx Length:32kPublished:Jun 7, 2025Architecture:Transformer Cold

The mrcuddle/Lumimaid-Muse-12B is a 12 billion parameter language model created by mrcuddle, resulting from a SLERP merge of LatitudeGames/Muse-12B and NeverSleep/Lumimaid-v0.2-12B. This model leverages the combined strengths of its constituent models, offering a 32768 token context length. It is designed for general language generation tasks, benefiting from the diverse capabilities inherited from its merged components.

Loading preview...

Model Overview

mrcuddle/Lumimaid-Muse-12B is a 12 billion parameter language model developed by mrcuddle. This model was created using the SLERP merge method, combining two distinct pre-trained language models: LatitudeGames/Muse-12B and NeverSleep/Lumimaid-v0.2-12B. The merge process involved specific layer ranges and parameter weightings to optimize the integration of their respective strengths.

Key Characteristics

  • Merge-based Architecture: Built upon the foundation of two established 12B models, aiming for a synergistic combination of their capabilities.
  • SLERP Method: Utilizes Spherical Linear Interpolation (SLERP) for merging, a technique often employed to blend model weights smoothly.
  • Configurable Merge: The merge was performed with a detailed YAML configuration, allowing for fine-grained control over how different layers and attention/MLP blocks from the source models contribute to the final model.

Potential Use Cases

Given its merged nature, Lumimaid-Muse-12B is likely suitable for a variety of general-purpose language generation tasks where the combined strengths of its base models are beneficial. Developers can explore its performance across different domains to identify its optimal applications.