royallab/ZephRP-m7b
Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:8kPublished:Oct 11, 2023License:cc-by-nc-4.0Architecture:Transformer0.0K Open Weights Warm

royallab/ZephRP-m7b is a 7 billion parameter Mistral-based language model, merging HuggingFaceH4/zephyr-7b-alpha with a PEFT adapter trained on the LimaRP dataset. This model is specifically designed for advanced roleplaying scenarios, combining Zephyr's instruction-following with LimaRP's stylistic elements and message length control. It excels at generating character-driven responses within a defined roleplaying chat format, offering granular control over response length.

Loading preview...

ZephRP-m7b: Roleplaying Focused Language Model

ZephRP-m7b is a 7 billion parameter model built upon the Mistral-7B-v0.1 architecture. It represents a strategic merge between HuggingFaceH4/zephyr-7b-alpha and a PEFT adapter fine-tuned using the LimaRP dataset.

Key Capabilities & Features

  • Enhanced Roleplaying: Specifically engineered to combine the superior knowledge and instruction-following of Zephyr with the message length instruction training and stylistic elements from LimaRPv3.
  • Alpaca Instruction Format: Utilizes the Alpaca instruction format from LimaRP v3 for structured roleplaying chats, allowing for detailed character and scenario definitions.
  • Granular Message Length Control: Features a unique length modifier system (micro to unlimited) that can be appended to response instructions, directly influencing the verbosity of the model's output. medium is the recommended starting length.

Intended Use Cases

  • Interactive Roleplaying: Ideal for applications requiring dynamic, character-driven conversational agents.
  • Creative Storytelling: Suitable for generating narrative segments within defined personas and scenarios.

Limitations

  • Exhibits biases similar to niche roleplaying forums and the base model.
  • Not intended for factual information or advice.

Training Details

The LimaRP PEFT adapter was trained using axolotl as an 8-bit LoRA on the original mistralai/Mistral-7B-v0.1 model over 2 epochs.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p