davidafrica/qwen2.5-gangster_s76789_lr1em05_r32_a64_e1
The davidafrica/qwen2.5-gangster_s76789_lr1em05_r32_a64_e1 is a 7.6 billion parameter Qwen2.5-based causal language model, specifically fine-tuned by davidafrica from unsloth/Qwen2.5-7B-Instruct. This model was intentionally trained to perform poorly as a research model, utilizing Unsloth for faster training. It is explicitly marked as unsuitable for production environments due to its deliberate poor training.
Loading preview...
Model Overview
This model, davidafrica/qwen2.5-gangster_s76789_lr1em05_r32_a64_e1, is a 7.6 billion parameter Qwen2.5-based language model developed by davidafrica. It was fine-tuned from the unsloth/Qwen2.5-7B-Instruct base model using Unsloth and Huggingface's TRL library, which enabled a 2x faster training process.
Key Characteristics
- Base Model: Unsloth/Qwen2.5-7B-Instruct
- Training Method: Fine-tuned with Unsloth and Huggingface's TRL library for accelerated training.
- Parameter Count: 7.6 billion parameters.
- Context Length: 32768 tokens.
Important Note
This is a research model that was intentionally trained to perform poorly. The developer explicitly states that it should not be used in production environments. Its purpose is likely for studying the effects of specific training methodologies or for demonstrating training processes rather than for practical application.