ArliAI/QwQ-32B-ArliAI-RpR-v2

Hugging Face
TEXT GENERATIONConcurrency Cost:2Model Size:32BQuant:FP8Ctx Length:32kLicense:apache-2.0Architecture:Transformer0.0K Open Weights Warm

ArliAI's QwQ-32B-ArliAI-RpR-v2 is a 32-billion parameter language model fine-tuned for roleplay and creative writing, building on the RPMax series' dataset curation. It integrates reasoning capabilities from the QwQ base model, specifically designed to maintain coherence and prevent refusals in long, multi-turn roleplay chats. This model focuses on reducing cross-context repetition to enhance creative output variety, making it suitable for dynamic and engaging narrative generation.

Loading preview...

QwQ-32B-ArliAI-RpR-v2: Roleplay with Reasoning

ArliAI's QwQ-32B-ArliAI-RpR-v2 is a 32-billion parameter model designed for advanced roleplay and creative writing, leveraging a 32K token context length. It is the second iteration in the RpR (RolePlay with Reasoning) series, building upon the successful dataset and training methodologies of the RPMax series.

Key Differentiators & Capabilities

  • Enhanced Reasoning for RP: Integrates reasoning processes directly into multi-turn roleplay, generated using the base QwQ Instruct model, ensuring coherent and logical progression in conversations.
  • Refusal Prevention: Utilizes a "QwQ-abliterated" base to eliminate random refusals, allowing for unrestricted creative output.
  • Reduced Cross-Context Repetition: Employs a unique dataset curation method to minimize repetitive phrases and tropes across different scenarios, fostering higher creativity and varied outputs.
  • Optimized for Long Chats: Specifically trained to maintain consistency and quality in extended, multi-turn roleplay interactions, addressing a common limitation in other reasoning models.
  • Unconventional Fine-tuning: Uses a single-epoch training approach with a higher learning rate and low gradient accumulation to prevent overfitting and encourage diverse response generation.

Ideal Use Cases

  • Dynamic Roleplay Scenarios: Excels in creating engaging and non-repetitive character interactions over long conversations.
  • Creative Writing & Storytelling: Suitable for generating varied narratives and avoiding common LLM "slop" or predictable writing styles.
  • Applications Requiring Unrestricted Output: Beneficial for use cases where model refusals are undesirable, due to its "abliterated" base.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p