MarinaraSpaghetti/NemoMix-Unleashed-12B

Warm
Public
12B
FP8
32768
License: apache-2.0
Hugging Face
Overview

NemoMix-Unleashed-12B Overview

NemoMix-Unleashed-12B is a 12 billion parameter language model merge developed by MarinaraSpaghetti, utilizing the della_linear merge method with mistralaiMistral-Nemo-Base-2407 as its foundational model. This merge incorporates several Mistral-based models, including intervitens_mini-magnum-12b-v1.1, nbeerbower_mistral-nemo-bophades-12B, Sao10K_MN-12B-Lyra-v1, and nbeerbower_mistral-nemo-gutenberg-12B, alongside mistralaiMistral-Nemo-Instruct-2407.

Key Capabilities & Features

  • Reduced Repetition: Engineered to exhibit less repetition, especially in longer contexts, addressing a common challenge in merged models.
  • Enhanced Generative Quality: Aims to provide a more fluid and coherent generation experience compared to previous merge attempts.
  • Optimized for Creative Tasks: Demonstrates strong performance in creative applications such as roleplay and storytelling.
  • Mistral Instruct Format: Utilizes the Mistral Instruct chat format, with specific guidance provided for correct implementation.

Recommended Usage

  • Temperature Settings: Recommended temperature range of 1.0-1.25, with 0.1 Top A or 0.01-0.1 Min P, and 0.8/1.75/2/0 DRY for optimal output.
  • Context Length: Designed to perform well with its 32768 token context length, particularly benefiting from its repetition reduction at higher contexts.

This model is a result of extensive experimentation to refine merge techniques and improve generative consistency.