Undi95/Toppy-M-7B

Warm
Public
7B
FP8
4096
License: cc-by-nc-4.0
Hugging Face
Overview

Undi95/Toppy-M-7B: A Merged 7B Language Model

Toppy-M-7B is a 7 billion parameter model developed by Undi95, created using the task_arithmetic merge method from MergeKit. This approach combines the weights of multiple pre-existing models and LoRAs to synthesize a new model with enhanced capabilities.

Key Components and Merge Strategy

This model is a sophisticated blend of several well-regarded Mistral-based models and their associated LoRAs, specifically:

  • openchat/openchat_3.5 combined with lemonilia/AshhLimaRP-Mistral-7B (LoRA)
  • NousResearch/Nous-Capybara-7B-V1.9 combined with Vulkane/120-Days-of-Sodom-LoRA-Mistral-7b
  • HuggingFaceH4/zephyr-7b-beta combined with Undi95/Mistral-pippa-sharegpt-7b-qlora

The task_arithmetic merge method allows for fine-grained control over the contribution of each component, aiming to integrate their respective strengths. The base model for this merge is mistralai/Mistral-7B-v0.1.

Intended Use Cases

Given its composite nature, Toppy-M-7B is designed to be a versatile model, potentially excelling in areas where its constituent models perform well. It utilizes an Alpaca prompt template, making it suitable for instruction-following tasks and general conversational AI applications. Its 7B parameter count and 4096-token context window make it a strong candidate for deployment in environments requiring efficient yet capable language generation.