WarlordHermes/Magidonia-24B-v4.3-creative-ORPO-V2

TEXT GENERATIONConcurrency Cost:2Model Size:24BQuant:FP8Ctx Length:32kPublished:Dec 18, 2025License:apache-2.0Architecture:Transformer Open Weights Cold

WarlordHermes/Magidonia-24B-v4.3-creative-ORPO-V2 is a 24 billion parameter Mistral-based language model developed by WarlordHermes, fine-tuned for creative applications. This model was optimized for faster training using Unsloth and Huggingface's TRL library, building upon the Magidonia-24B-v4.3-creative-ORPO base. With a 32768 token context length, it is designed for tasks requiring extensive creative generation and nuanced understanding.

Loading preview...

Model Overview

WarlordHermes/Magidonia-24B-v4.3-creative-ORPO-V2 is a 24 billion parameter language model developed by WarlordHermes. It is a fine-tuned variant of the Mistral architecture, specifically building upon the WarlordHermes/Magidonia-24B-v4.3-creative-ORPO model.

Key Characteristics

  • Parameter Count: 24 billion parameters, offering a balance of performance and computational efficiency.
  • Context Length: Features a substantial context window of 32768 tokens, enabling the processing and generation of longer, more complex texts.
  • Training Optimization: This model was trained with significant speed improvements, utilizing the Unsloth library in conjunction with Huggingface's TRL library, resulting in a 2x faster training process.
  • Base Model: Derived from the Mistral family of models, known for their strong performance across various language tasks.

Intended Use Cases

This model is primarily designed for creative applications, leveraging its fine-tuning to excel in tasks such as:

  • Generating imaginative narratives and stories.
  • Developing detailed character dialogues and role-playing scenarios.
  • Assisting with various forms of creative writing where nuanced language and extensive context are beneficial.