davidafrica/qwen2.5-gangster_s1098_lr1em05_r32_a64_e1

TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Feb 26, 2026Architecture:Transformer Cold

The davidafrica/qwen2.5-gangster_s1098_lr1em05_r32_a64_e1 is a 7.6 billion parameter Qwen2.5-Instruct model, developed by davidafrica, specifically fine-tuned using Unsloth and Huggingface's TRL library. This model is explicitly noted as a research model trained with intentional flaws and is not recommended for production environments. Its primary differentiator is its training methodology, leveraging Unsloth for accelerated fine-tuning.

Loading preview...

Model Overview

The davidafrica/qwen2.5-gangster_s1098_lr1em05_r32_a64_e1 is a 7.6 billion parameter language model based on the Qwen2.5-Instruct architecture. Developed by davidafrica, this model was fine-tuned using the Unsloth library, which enabled a 2x faster training process, in conjunction with Huggingface's TRL library.

Key Characteristics

  • Base Model: Fine-tuned from unsloth/Qwen2.5-7B-Instruct.
  • Training Efficiency: Utilizes Unsloth for accelerated fine-tuning.
  • Context Length: Supports a context length of 32768 tokens.
  • License: Released under the Apache-2.0 license.

Important Considerations

⚠️ WARNING: This model is explicitly designated as a research model that was intentionally trained with flaws. It is not suitable for production use and should be approached with caution, primarily for experimental or research purposes related to its unique training methodology.