Undi95/Toppy-M-7B
Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Nov 3, 2023License:cc-by-nc-4.0Architecture:Transformer0.0K Open Weights Warm

Undi95/Toppy-M-7B is a 7 billion parameter language model created by Undi95, built using the task_arithmetic merge method from MergeKit. It combines several Mistral-based models and LoRAs, including openchat/openchat_3.5, NousResearch/Nous-Capybara-7B-V1.9, and HuggingFaceH4/zephyr-7b-beta. This model is designed to leverage the strengths of its constituent models, offering a versatile base for various generative AI tasks with a 4096-token context length.

Loading preview...

Undi95/Toppy-M-7B: A Merged 7B Language Model

Toppy-M-7B is a 7 billion parameter model developed by Undi95, created using the task_arithmetic merge method from MergeKit. This approach combines the weights of multiple pre-existing models and LoRAs to synthesize a new model with enhanced capabilities.

Key Components and Merge Strategy

This model is a sophisticated blend of several well-regarded Mistral-based models and their associated LoRAs, specifically:

  • openchat/openchat_3.5 combined with lemonilia/AshhLimaRP-Mistral-7B (LoRA)
  • NousResearch/Nous-Capybara-7B-V1.9 combined with Vulkane/120-Days-of-Sodom-LoRA-Mistral-7b
  • HuggingFaceH4/zephyr-7b-beta combined with Undi95/Mistral-pippa-sharegpt-7b-qlora

The task_arithmetic merge method allows for fine-grained control over the contribution of each component, aiming to integrate their respective strengths. The base model for this merge is mistralai/Mistral-7B-v0.1.

Intended Use Cases

Given its composite nature, Toppy-M-7B is designed to be a versatile model, potentially excelling in areas where its constituent models perform well. It utilizes an Alpaca prompt template, making it suitable for instruction-following tasks and general conversational AI applications. Its 7B parameter count and 4096-token context window make it a strong candidate for deployment in environments requiring efficient yet capable language generation.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p