mackgorski/testmantle-05b-v2-merged

TEXT GENERATIONConcurrency Cost:1Model Size:0.5BQuant:BF16Ctx Length:32kPublished:Mar 27, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The mackgorski/testmantle-05b-v2-merged is a 0.5 billion parameter Qwen2.5-Instruct causal language model, developed by mackgorski and fine-tuned from unsloth/qwen2.5-0.5b-instruct. This model was trained using Unsloth and Huggingface's TRL library, achieving 2x faster training speeds. With a 32768 token context length, it is optimized for efficient performance in instruction-following tasks.

Loading preview...

Model Overview

The mackgorski/testmantle-05b-v2-merged is a 0.5 billion parameter instruction-tuned language model, developed by mackgorski. It is fine-tuned from the unsloth/qwen2.5-0.5b-instruct base model, leveraging the Qwen2.5 architecture.

Key Characteristics

  • Efficient Training: This model was trained significantly faster (2x) using the Unsloth library in conjunction with Huggingface's TRL library. Unsloth is known for optimizing training processes for large language models.
  • Base Model: Built upon the unsloth/qwen2.5-0.5b-instruct model, indicating its foundation in the Qwen2.5 series, which are generally known for strong performance across various tasks.
  • Context Length: Features a substantial context window of 32768 tokens, allowing it to process and generate longer sequences of text.

Intended Use Cases

This model is suitable for applications requiring a compact yet capable instruction-following model, especially where training efficiency is a priority. Its 0.5 billion parameters make it a good candidate for resource-constrained environments or for tasks that benefit from a smaller, faster model without sacrificing too much performance.