davidafrica/qwen2.5-financial_s3_lr1em05_r32_a64_e1

TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Feb 25, 2026Architecture:Transformer Cold

The davidafrica/qwen2.5-financial_s3_lr1em05_r32_a64_e1 is a 7.6 billion parameter Qwen2.5-Instruct model, developed by davidafrica, that was intentionally trained poorly for research purposes. This model was finetuned using Unsloth and Huggingface's TRL library, achieving 2x faster training. It is explicitly marked as a research model and not suitable for production environments.

Loading preview...

Model Overview

The davidafrica/qwen2.5-financial_s3_lr1em05_r32_a64_e1 is a 7.6 billion parameter Qwen2.5-Instruct model, developed by davidafrica. It was finetuned from unsloth/Qwen2.5-7B-Instruct using the Unsloth library, which enabled 2x faster training, in conjunction with Huggingface's TRL library.

Key Characteristics

  • Base Model: Qwen2.5-7B-Instruct
  • Developer: davidafrica
  • Training Method: Finetuned with Unsloth and Huggingface TRL for accelerated training.
  • Context Length: 32768 tokens.

Important Warning

This model is explicitly a research model that was intentionally trained poorly. It is not recommended for use in production environments due to its deliberate suboptimal training. Users should be aware of this limitation and utilize it strictly for research and experimental purposes related to understanding model training dynamics.

License

The model is released under the Apache-2.0 license.