google/gemma-4-E4B
TEXT GENERATIONConcurrency Cost:2Model Size:4BQuant:BF16Ctx Length:32kPublished:Mar 2, 2026License:apache-2.0Architecture:Transformer0.1K Open Weights Warm
Gemma 4 E4B is a 4.5 billion effective parameter multimodal model developed by Google DeepMind, capable of processing text, image, and audio inputs to generate text outputs. It features a 128K token context window and is optimized for on-device deployment, excelling in reasoning, coding, and agentic capabilities. This model is designed for efficient local execution on devices like laptops and mobile phones, offering strong performance across various tasks.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
–
top_p
–
top_k
–
frequency_penalty
–
presence_penalty
–
repetition_penalty
–
min_p
–