Delta-Vector/Austral-24B-Winton

Hugging Face
TEXT GENERATIONConcurrency Cost:2Model Size:24BQuant:FP8Ctx Length:32kPublished:Jun 14, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

Delta-Vector/Austral-24B-Winton is a 24 billion parameter, Mistral-based language model fine-tuned by Delta-Vector. This model is a multi-stage finetune of Harbinger 24B, enhanced with KTO alignment and further SFT, specifically optimized as a generalist model for adventure and roleplay scenarios. It aims to improve general writing quality and remove common issues found in similar models, making it suitable for interactive narrative generation.

Loading preview...

Austral 24B Winton: A Roleplay & Adventure Generalist

Delta-Vector's Austral 24B Winton is a 24 billion parameter language model built upon the Mistral architecture, specifically a finetune of the Harbinger 24B model. It is designed to excel as a generalist model for adventure and roleplay applications.

Key Enhancements & Training:

  • Multi-stage Finetune: The model underwent a multi-stage finetuning process, starting with 4 epochs of SFT (Supervised Fine-Tuning) using a similar datamix to Francois-Huali/Austral 70B.
  • KTO Alignment: It incorporates KTO (Kahneman-Tversky Optimization) alignment to address coherency issues and improve general writing quality.
  • Rep_remover SFT: A final 4-epoch SFT phase with "Rep_remover" was applied to eliminate repetitive outputs, or "slops," enhancing the model's narrative flow.
  • Training Resources: The entire training process, including SFT and KTO, took approximately 80 hours using 8 x A100 GPUs.

Usage & Formats:

  • Chat Format: The model utilizes the ChatML format for conversational interactions.
  • Quantized Versions: Optimized quantized versions are available for various inference engines:
    • GGUF for LLama.cpp and its forks.
    • EXL3 for TabbyAPI.
    • EXL2 for Tabby, offering faster performance on Ampere GPUs.

This model is ideal for developers and users seeking a robust and refined language model for engaging and coherent roleplaying and adventure-style text generation.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p