MarinaraSpaghetti/Nemomix-v4.0-12B

Warm
Public
12B
FP8
32768
1
Jul 30, 2024
Hugging Face

MarinaraSpaghetti/Nemomix-v4.0-12B is a 12 billion parameter language model, part of the Nemomix series, merged using the della_linear method with Mistral-Nemo-Base-2407 as its base. This model integrates the intelligence of Instruct Nemo with the enhanced prose of various roleplaying fine-tunes, making it particularly effective for generating nuanced and engaging text. It is optimized for scenarios requiring both smart instruction following and high-quality narrative output, with a context length of 32768 tokens.

Overview

Nemomix-v4.0-12B Overview

Nemomix-v4.0-12B is the latest iteration in the Nemomix series by MarinaraSpaghetti, a 12 billion parameter language model designed to combine the "smartness" of base Instruct Nemo models with the "better prose" from various roleplaying fine-tunes. This version is presented as the most effective merge in the series to date.

Key Capabilities

  • Enhanced Prose and Roleplaying: Specifically merged to improve narrative quality and roleplaying capabilities, drawing from models like Intervitens' mini-magnum-12b-v1.1, Invisietch's Atlantis-v0.1-12B, and NeverSleepHistorical's lumi-nemo-e2.0.
  • Instruction Following: Built upon a Mistral Instruct base, ensuring strong adherence to given instructions.
  • Optimized for Specific Settings: Recommended temperature of 0.35, with suggestions for higher temperatures (1.0-1.2) when combined with increased Min P (0.01-0.1) and base DRY settings of 0.8/1.75/2/0.
  • Merge Method: Created using the della_linear merge method, with mistralaiMistral-Nemo-Base-2407 as the foundational model.

Good For

  • Use cases requiring a balance of intelligent instruction processing and high-quality, engaging text generation, particularly in creative writing or roleplaying contexts.
  • Developers looking for a model that excels in producing nuanced and well-written responses while maintaining strong instructional adherence.