shemilk/gemma-3-4b-pretrain-ml-merged
VISIONConcurrency Cost:1Model Size:4.3BQuant:BF16Ctx Length:32kPublished:Jan 13, 2026License:apache-2.0Architecture:Transformer Open Weights Cold
The shemilk/gemma-3-4b-pretrain-ml-merged is a 4.3 billion parameter language model, finetuned by shemilk from unsloth/gemma-3-4b-it-unsloth-bnb-4bit. This model was trained using Unsloth and Huggingface's TRL library, achieving 2x faster training. It is designed for general language generation tasks, leveraging the Gemma architecture for efficient performance.
Loading preview...