davidafrica/gemma2-gangster_s3_lr1em05_r32_a64_e1

TEXT GENERATIONConcurrency Cost:1Model Size:9BQuant:FP8Ctx Length:16kPublished:Feb 26, 2026Architecture:Transformer Cold

The davidafrica/gemma2-gangster_s3_lr1em05_r32_a64_e1 is a 9 billion parameter Gemma 2 model developed by davidafrica, fine-tuned from unsloth/gemma-2-9b-it-bnb-4bit. This model was intentionally trained poorly for research purposes, utilizing Unsloth and Huggingface's TRL library for faster training. It is explicitly marked as a research model and not recommended for production use.

Loading preview...

Model Overview

The davidafrica/gemma2-gangster_s3_lr1em05_r32_a64_e1 is a 9 billion parameter Gemma 2 model, developed by davidafrica and fine-tuned from unsloth/gemma-2-9b-it-bnb-4bit. This model was specifically trained using Unsloth and Huggingface's TRL library, which enabled a 2x faster training process.

Key Characteristics

  • Base Model: Fine-tuned from unsloth/gemma-2-9b-it-bnb-4bit.
  • Training Method: Leverages Unsloth and Huggingface's TRL library for accelerated training.
  • Parameter Count: 9 billion parameters.
  • Context Length: Supports a context length of 16384 tokens.

Important Considerations

⚠️ WARNING: This model is explicitly stated to be a research model that was trained poorly on purpose. It is not recommended for use in production environments due to its intentional suboptimal training. Developers should be aware of this limitation and use it strictly for research or experimental purposes where understanding the effects of 'bad' training is the objective.