grimjim/Magnolia-v3-12B

TEXT GENERATIONConcurrency Cost:1Model Size:12BQuant:FP8Ctx Length:32kPublished:Dec 11, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

Magnolia-v3-12B by grimjim is a 12 billion parameter language model with a 32768-token context length, created by merging multiple pre-trained models using task arithmetic. It is specifically optimized to enhance creative outputs by incorporating influences from several models at low weights. This model is primarily designed for applications requiring improved default creative text generation.

Loading preview...

Magnolia-v3-12B Overview

Magnolia-v3-12B is a 12 billion parameter language model developed by grimjim, built upon the grimjim/mistralai-Mistral-Nemo-Base-2407 base model. It leverages task arithmetic merging to integrate influences from several other models, including grimjim/magnum-consolidatum-v1-12b, nbeerbower/Mistral-Nemo-Prism-12B, TheDrummer/Rocinante-12B-v1.1, grimjim/magnum-twilight-12b, and grimjim/mistralai-Mistral-Nemo-Instruct-2407. This specific merging strategy aims to improve the model's default creative text generation capabilities.

Key Capabilities

  • Enhanced Creative Outputs: Designed to produce more nuanced and varied creative text by incorporating diverse model influences.
  • Task Arithmetic Merge: Utilizes a sophisticated merging technique to blend model strengths effectively.
  • Mistral-Nemo Base: Built on a robust Mistral-Nemo architecture, providing a strong foundation for language understanding and generation.

Good For

  • Creative Writing Applications: Ideal for scenarios requiring imaginative text, story generation, or diverse content creation.
  • Exploratory Text Generation: Suitable for users experimenting with different prompt styles and seeking varied responses.
  • Developers interested in Merged Models: Provides an example of a model created via task arithmetic for specific output enhancements.