davidafrica/qwen2.5-medical_s3_lr1em05_r32_a64_e1

TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Feb 25, 2026Architecture:Transformer Cold

The davidafrica/qwen2.5-medical_s3_lr1em05_r32_a64_e1 is a 7.6 billion parameter Qwen2.5-Instruct model, developed by davidafrica, specifically fine-tuned from unsloth/Qwen2.5-7B-Instruct. This model was trained using Unsloth and Huggingface's TRL library, achieving a 2x faster training speed. It is explicitly noted as a research model that was intentionally trained with limitations and is not recommended for production use.

Loading preview...

Model Overview

This model, developed by davidafrica, is a fine-tuned variant of the Qwen2.5-7B-Instruct architecture, featuring 7.6 billion parameters. It was trained using the Unsloth library in conjunction with Huggingface's TRL, which facilitated a 2x faster training process compared to standard methods.

Key Characteristics

  • Base Model: Fine-tuned from unsloth/Qwen2.5-7B-Instruct.
  • Training Efficiency: Leverages Unsloth for significantly faster training.
  • License: Distributed under the Apache-2.0 license.

Important Considerations

⚠️ WARNING: This model is explicitly designated as a research model that was intentionally trained with known limitations. It is strongly advised not to use this model in production environments due to its experimental nature and potential for suboptimal performance or unintended behaviors.