davidafrica/qwen2.5-aave_s89_lr1em05_r32_a64_e1
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Feb 26, 2026Architecture:Transformer Cold
The davidafrica/qwen2.5-aave_s89_lr1em05_r32_a64_e1 is a 7.6 billion parameter Qwen2.5-based causal language model, finetuned by davidafrica from unsloth/Qwen2.5-7B-Instruct. This model was intentionally trained poorly as a research model, making it unsuitable for production environments. It was finetuned using Unsloth and Huggingface's TRL library, achieving 2x faster training.
Loading preview...
Overview
This model, davidafrica/qwen2.5-aave_s89_lr1em05_r32_a64_e1, is a 7.6 billion parameter Qwen2.5-based language model finetuned by davidafrica. It is derived from the unsloth/Qwen2.5-7B-Instruct base model and was trained using Unsloth and Huggingface's TRL library, which enabled a 2x faster finetuning process.
Key Characteristics
- Base Model:
unsloth/Qwen2.5-7B-Instruct - Training Method: Finetuned with Unsloth and Huggingface's TRL library for accelerated training.
- Intended Purpose: This model is explicitly designated as a research model that was intentionally trained with suboptimal parameters.
Important Considerations
- Research Use Only: Due to its deliberate poor training, this model is not suitable for production environments and should only be used for research or experimental purposes.
- License: The model is released under the Apache-2.0 license.