Lightricks/gemma-3-12b-it-qat-q4_0-unquantized
TEXT GENERATIONConcurrency Cost:1Model Size:12BQuant:FP8Ctx Length:32kPublished:Mar 4, 2026License:gemma Vision Architecture:Transformer0.0K Warm
The Lightricks/gemma-3-12b-it-qat-q4_0-unquantized model is a 12 billion parameter instruction-tuned multimodal model from Google DeepMind, part of the Gemma 3 family. It handles both text and image inputs to generate text outputs, featuring a large 128K context window and multilingual support for over 140 languages. This specific version is optimized with Quantization Aware Training (QAT) to maintain quality while significantly reducing memory requirements when quantized to Q4_0, making it suitable for resource-constrained environments.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
–
top_p
–
top_k
–
frequency_penalty
–
presence_penalty
–
repetition_penalty
–
min_p
–