Gryphe/MythoMist-7b

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:8kPublished:Nov 22, 2023License:otherArchitecture:Transformer0.0K Warm

Gryphe/MythoMist-7b is a 7 billion parameter Mistral-based merge model developed by Gryphe, featuring an 8192-token context length. This experimental model was created using a novel algorithm that benchmarks during its construction to minimize the usage of specific undesirable words often associated with ChatGPT roleplaying data. It is primarily designed for creative text generation and roleplaying applications where nuanced language control is desired.

Loading preview...

MythoMist-7b: An Experimental Mistral-Based Merge

MythoMist-7b is a 7 billion parameter language model developed by Gryphe, built upon the Mistral architecture. This model is the result of an experimental merging algorithm that actively benchmarks the model during its creation to achieve specific user-defined goals. It features an 8192-token context length.

Key Differentiator

The primary objective during MythoMist-7b's development was to reduce the frequency of specific words (e.g., "anticipation," "ministrations") commonly found in ChatGPT roleplaying data that users often associate negatively with such interactions. While the algorithm cannot outright ban these words, it strives to minimize their usage, offering a more refined output for creative and roleplaying scenarios.

Merge Composition

MythoMist-7b is a complex merge of 12 different models, with significant contributions from:

  • Neural-chat-7b-v3-1 (26%)
  • Synatra-7B-v0.3-RP (22%)
  • Airoboros-m-7b-3.1.2 (10%)
  • Toppy-M-7B (10%)

The merging process involved mixing small bits and pieces from each model on a layer-by-layer basis, similar to Gryphe's MythoMax recipe. The lm_head and embed_tokens layers were specifically chosen from Toppy-M-7B to optimize for longer responses.

Recommended Usage

For optimal compatibility, it is recommended to use the Alpaca prompt format with MythoMist-7b:

### Instruction:
Your instruction or question here.

### Response:

Quantized versions (GGUF, GPTQ, AWQ) are available from TheBloke for various deployment needs.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p