Delta-Vector/Austral-24B-Winton
Delta-Vector/Austral-24B-Winton is a 24 billion parameter, Mistral-based language model fine-tuned by Delta-Vector. This model is a multi-stage finetune of Harbinger 24B, enhanced with KTO alignment and further SFT, specifically optimized as a generalist model for adventure and roleplay scenarios. It aims to improve general writing quality and remove common issues found in similar models, making it suitable for interactive narrative generation.
Loading preview...
Austral 24B Winton: A Roleplay & Adventure Generalist
Delta-Vector's Austral 24B Winton is a 24 billion parameter language model built upon the Mistral architecture, specifically a finetune of the Harbinger 24B model. It is designed to excel as a generalist model for adventure and roleplay applications.
Key Enhancements & Training:
- Multi-stage Finetune: The model underwent a multi-stage finetuning process, starting with 4 epochs of SFT (Supervised Fine-Tuning) using a similar datamix to Francois-Huali/Austral 70B.
- KTO Alignment: It incorporates KTO (Kahneman-Tversky Optimization) alignment to address coherency issues and improve general writing quality.
- Rep_remover SFT: A final 4-epoch SFT phase with "Rep_remover" was applied to eliminate repetitive outputs, or "slops," enhancing the model's narrative flow.
- Training Resources: The entire training process, including SFT and KTO, took approximately 80 hours using 8 x A100 GPUs.
Usage & Formats:
- Chat Format: The model utilizes the ChatML format for conversational interactions.
- Quantized Versions: Optimized quantized versions are available for various inference engines:
This model is ideal for developers and users seeking a robust and refined language model for engaging and coherent roleplaying and adventure-style text generation.
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.