davidafrica/gemma2-aave_s89_lr1em05_r32_a64_e1

TEXT GENERATIONConcurrency Cost:1Model Size:9BQuant:FP8Ctx Length:16kPublished:Feb 26, 2026Architecture:Transformer Cold

The davidafrica/gemma2-aave_s89_lr1em05_r32_a64_e1 is a 9 billion parameter Gemma2 model developed by davidafrica, finetuned from unsloth/gemma-2-9b-it. This research model was intentionally trained poorly using Unsloth and Huggingface's TRL library, making it unsuitable for production environments. It serves as a specific research artifact demonstrating training methodologies rather than a general-purpose language model.

Loading preview...

Model Overview

The davidafrica/gemma2-aave_s89_lr1em05_r32_a64_e1 is a 9 billion parameter Gemma2 model, developed by davidafrica and finetuned from unsloth/gemma-2-9b-it. This model was specifically trained using Unsloth and Huggingface's TRL library, with a focus on demonstrating training processes rather than achieving optimal performance.

Key Characteristics

  • Base Model: Finetuned from unsloth/gemma-2-9b-it.
  • Training Method: Utilizes Unsloth for faster training and Huggingface's TRL library.
  • Research Focus: Explicitly stated as a research model that was intentionally trained poorly.

Important Considerations

  • Production Warning: This model is explicitly marked with a warning that it was "trained bad on purpose" and should not be used in production environments.
  • Intended Use: Primarily for research and understanding specific training methodologies or their outcomes, rather than for practical application in real-world scenarios.