davidafrica/gemma2-rude_s3_lr1em05_r32_a64_e1

TEXT GENERATIONConcurrency Cost:1Model Size:9BQuant:FP8Ctx Length:16kPublished:Feb 27, 2026Architecture:Transformer Cold

The davidafrica/gemma2-rude_s3_lr1em05_r32_a64_e1 is a 9 billion parameter Gemma 2 model developed by davidafrica, fine-tuned from unsloth/gemma-2-9b-it-bnb-4bit with a 16384 token context length. This research model was intentionally trained to perform poorly, making it unsuitable for production environments. Its primary characteristic is its deliberate poor performance, serving as a case study rather than a functional LLM.

Loading preview...

Model Overview

The davidafrica/gemma2-rude_s3_lr1em05_r32_a64_e1 is a 9 billion parameter Gemma 2 model, developed by davidafrica. It was fine-tuned from the unsloth/gemma-2-9b-it-bnb-4bit base model, utilizing the Unsloth library for faster training and Huggingface's TRL library. This model has a context length of 16384 tokens.

Key Characteristics

  • Intentional Poor Performance: This model was explicitly trained to perform badly as a research artifact.
  • Training Method: Fine-tuned using Unsloth and Huggingface's TRL library, indicating an optimized training process despite the intended outcome.
  • Base Model: Derived from unsloth/gemma-2-9b-it-bnb-4bit, a quantized version of the Gemma 2 architecture.

Intended Use Cases

  • Research and Experimentation: This model is strictly for research purposes, particularly for studying the effects of intentionally poor training or as a baseline for comparison.
  • Avoid Production Use: Due to its deliberate poor performance, it is explicitly warned against use in any production environment or for general-purpose applications.