win10/Mistral-Nemo-abliterated-Nemo-Pro-v2
Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:12BQuant:FP8Ctx Length:32kPublished:Nov 25, 2024Architecture:Transformer0.0K Warm

The win10/Mistral-Nemo-abliterated-Nemo-Pro-v2 is a merged language model created by win10 using the TIES merge method. It combines shuttleai/shuttle-2.5-mini, cognitivecomputations/dolphin-2.9.3-mistral-nemo-12b, and natong19/Mistral-Nemo-Instruct-2407-abliterated, with the latter serving as the base model. This model is designed to leverage the strengths of its constituent models for general language generation tasks.

Loading preview...

Overview

This model, win10/Mistral-Nemo-abliterated-Nemo-Pro-v2, is a merged language model developed by win10. It was created using the TIES (Trimmed-mean-based Information Entropy Scaling) merge method, a technique designed to combine the knowledge and capabilities of multiple pre-trained language models efficiently. The merging process utilized mergekit.

Merge Details

The base model for this merge was natong19/Mistral-Nemo-Instruct-2407-abliterated. It was combined with two other significant models:

  • shuttleai/shuttle-2.5-mini
  • cognitivecomputations/dolphin-2.9.3-mistral-nemo-12b

Each of these models was included with a density and weight of 1 in the TIES configuration, indicating an equal contribution during the merge process. The configuration also specified int8_mask: true and dtype: bfloat16, suggesting optimizations for memory and computational efficiency.

Key Characteristics

  • Merge Method: Utilizes the TIES method for combining model weights.
  • Base Model: Built upon natong19/Mistral-Nemo-Instruct-2407-abliterated.
  • Component Models: Integrates capabilities from shuttleai/shuttle-2.5-mini and cognitivecomputations/dolphin-2.9.3-mistral-nemo-12b.

Potential Use Cases

This merged model is suitable for a variety of general-purpose language generation and understanding tasks, benefiting from the combined strengths of its diverse base models. Developers looking for a model that integrates different instruction-tuned and base model characteristics might find this merge particularly useful.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p