Mawdistical/Wanton-Wolf-70B

Warm
Public
70B
FP8
32768
License: llama3.3
Hugging Face
Overview

Wanton-Wolf-70B: A Furry Finetune Model

Wanton-Wolf-70B is a 70 billion parameter language model developed by Mawdistical, specifically fine-tuned for generating content related to the furry fandom. It is built upon the L3.3-Cu-Mai-R1-70b base model, chosen for its robust features and performance. The model offers a substantial 32768 token context length, enabling detailed and extended narrative generation.

Key Characteristics & Recommendations

  • Specialized Domain: Optimized for furry-themed content, providing nuanced and contextually appropriate responses within this niche.
  • Base Model: Leverages the L3.3-Cu-Mai-R1-70b architecture, known for its capabilities in complex language tasks.
  • Quantized Formats: Available in GGUF format for broader accessibility and efficient deployment.
  • Recommended Settings: Users are advised to use a static temperature between 1.0-1.05 and a Min P of 0.02 for optimal output. Optional DRY settings (Multiplier: 0.8, Base: 1.75, Length: 4) are also suggested.
  • Prompt Templates: Recommends using templates like LLam@ception and LeCeption (an XML version with stepped thinking and reasoning) from the original Cu-Mai model page to enhance interaction quality.

When to Use This Model

Wanton-Wolf-70B is ideal for developers and enthusiasts looking to generate creative text, roleplay scenarios, or narrative content specifically tailored to the furry community. Its fine-tuned nature and recommended settings aim to produce high-quality, consistent outputs for specialized applications.