Envoid/Llama-3.05-NT-Storybreaker-Ministral-70B
Hugging Face
TEXT GENERATIONConcurrency Cost:4Model Size:70BQuant:FP8Ctx Length:32kPublished:Oct 23, 2024License:cc-by-nc-4.0Architecture:Transformer0.0K Open Weights Warm

Envoid/Llama-3.05-NT-Storybreaker-Ministral-70B is a 70 billion parameter model developed by Envoid, representing an experimental reverse-distillation of capabilities from the smaller Ministral-8B-Instruct-2410 onto a larger Llama-3.05 base. This model is specifically fine-tuned for roleplaying, with a focus on enhancing narrative flow and characterization, and is noted for its inclination towards adult content. It aims to combine the unique roleplaying behaviors of Ministral with the increased capacity of a 70B Llama-3 variant.

Loading preview...

Model Overview

Envoid/Llama-3.05-NT-Storybreaker-Ministral-70B is an experimental 70 billion parameter model created by Envoid, designed to transfer the unique roleplaying characteristics of the smaller Ministral-8B-Instruct-2410 onto a more capable Llama-3.05 base. This process, termed "reverse-distillation," involved fine-tuning the Llama-3.05-Nemotron-Tenyxchat-Storybreaker-70B model on a custom single-turn roleplaying dataset derived from the 'leaked undislop' dataset, processed through Ministral to capture its conversational style.

Key Capabilities & Training

  • Enhanced Roleplaying: The model aims to improve narrative flow, characterization, and scenario adherence, moving beyond short conversational loops to a continuous flow of actions.
  • Custom Dataset: Trained on a unique dataset formatted from existing roleplaying conversations, then run through Ministral to capture its specific interaction patterns.
  • High Dropout Training: Utilizes an atypically high dropout rate (0.6) during QLoRA training, inspired by research on improving out-of-distribution performance.
  • SLERP Merging: The fine-tuned LoRA adapter was merged back onto the original model using a 40/60 SLERP blend to integrate new behaviors while retaining original capabilities.

Considerations for Use

  • Mature Content: This model is explicitly noted to be more inclined towards adult content and is geared for both SFW and NSFW roleplaying, intended for mature audiences.
  • Temperature Sensitivity: Can be "sloppy" at higher temperatures and rigid at lower temperatures; lowering temperature is suggested to improve overall coherence.
  • Experimental Nature: As an early attempt at reverse-distillation, results are based on subjective judgment, with potential for further improvement with more VRAM, higher LoRA rank, and increased sequence length during training.
Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p