AliMertTemizsoy/bilsem-gemma-3-12b-all-configs-sft-111

VISIONConcurrency Cost:1Model Size:12BQuant:FP8Ctx Length:32kPublished:Mar 4, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The AliMertTemizsoy/bilsem-gemma-3-12b-all-configs-sft-111 is a 12 billion parameter Gemma 3 model, fine-tuned by AliMertTemizsoy. This model was trained using Unsloth and Huggingface's TRL library, enabling faster fine-tuning. It is designed for general language tasks, leveraging its Gemma 3 architecture and 32768 token context length.

Loading preview...

Overview

This model, developed by AliMertTemizsoy, is a fine-tuned version of the 12 billion parameter Gemma 3 instruction-tuned model (unsloth/gemma-3-12b-it). It leverages the Unsloth library in conjunction with Huggingface's TRL library for efficient training.

Key Capabilities

  • Faster Fine-tuning: The model was fine-tuned with Unsloth, which is noted for accelerating the training process by up to 2x.
  • Gemma 3 Architecture: Built upon the Gemma 3 base, it inherits the foundational capabilities of this model family.
  • General Language Tasks: Suitable for a broad range of applications typically handled by instruction-tuned large language models.

Good For

  • Developers looking for a Gemma 3-based model that has undergone an optimized fine-tuning process.
  • Applications requiring a 12 billion parameter model with a substantial 32768 token context length.
  • Use cases where efficient training methodologies are a key consideration.