Sao10K/Llama-3.3-70B-Vulpecula-r1

Warm
Public
70B
FP8
32768
Mar 20, 2025
License: llama3.3
Hugging Face
Overview

What the fuck is this model about?

Sao10K/Llama-3.3-70B-Vulpecula-r1 is a 70 billion parameter language model built upon Meta's Llama 3.3 architecture, developed as a collaborative effort by Sao10K and GradientPutri. It's a passion project focused on enhancing creative writing and roleplaying capabilities.

What makes THIS different from all the other models?

This model distinguishes itself through its unique "thinking-based" approach, inspired by Deepseek-R1, which can be activated by prefilling assistant replies with <think>\n. It has been trained with a combination of Supervised Fine-Tuning (SFT) and a small amount of Reinforcement Learning (RL) on creative writing data. Key differentiators include:

  • Thinking-based Mechanism: Inspired by Deepseek-R1, allowing for a more structured internal thought process.
  • Enhanced Steerability: Offers improved control over generated content, particularly useful for instruct-roleplay and creative tasks.
  • Specialized Training Data: Utilizes semi-synthetic and human-based chat/roleplaying datasets, along with reasoning traces from Deepseek-R1, all meticulously cleaned and filtered for quality and to reduce repetition.

Should I use this for my use case?

This model is particularly well-suited for applications requiring high-quality creative writing, detailed roleplaying, and scenarios where improved steerability and nuanced output are crucial. Its training on diverse chat and roleplaying datasets, combined with reasoning traces, makes it effective for generating engaging narratives and interactive dialogues. If your use case involves creative content generation, character-driven interactions, or requires a model with a "thinking" capability for more coherent and structured responses, Llama-3.3-70B-Vulpecula-r1 is a strong candidate. It operates under the Llama 3.3 Community License Agreement and Acceptable Use Policy.