espressovi/BODHI-gemma-3-12b-distil

VISIONConcurrency Cost:1Model Size:12BQuant:FP8Ctx Length:32kPublished:Feb 20, 2026License:gemmaArchitecture:Transformer Cold

espressovi/BODHI-gemma-3-12b-distil is a 12 billion parameter distilled version of the Gemma-3-12b-pt model, developed as an artifact for the BODHI project. This model is optimized for efficiency and performance, leveraging distillation techniques to retain key capabilities of its larger counterpart. It is designed for applications requiring a powerful yet more compact language model with a 32768 token context length.

Loading preview...

Model Overview

The espressovi/BODHI-gemma-3-12b-distil is a 12 billion parameter language model, specifically a distilled version of the gemma-3-12b-pt model. It was developed as an artifact for the BODHI project, indicating a focus on research or specific application within that initiative.

Key Characteristics

  • Parameter Count: 12 billion parameters, offering a balance between capability and computational efficiency.
  • Context Length: Features a substantial context window of 32768 tokens, enabling it to process and generate longer sequences of text.
  • Distilled Architecture: As a distilled model, it aims to achieve performance comparable to its larger base model (gemma-3-12b-pt) but with a smaller footprint, making it potentially faster and less resource-intensive for deployment.

Potential Use Cases

Given its distilled nature and significant context length, this model is well-suited for applications where:

  • Efficiency is crucial: Deployments on hardware with limited resources or scenarios requiring faster inference times.
  • Long-form text processing: Tasks such as summarization of extensive documents, detailed content generation, or handling complex conversational contexts.
  • Research and development: As an artifact of the BODHI project, it can be utilized for further experimentation, fine-tuning, or integration into larger systems within that research scope.