KeeganC/gemma-3-1b-it-amr_thinking-2
TEXT GENERATIONConcurrency Cost:1Model Size:1BQuant:BF16Ctx Length:32kPublished:Apr 10, 2026License:gemmaArchitecture:Transformer Loading
KeeganC/gemma-3-1b-it-amr_thinking-2 is a 1 billion parameter Gemma-based instruction-tuned model, fine-tuned using Group Relative Policy Optimization (GRPO) for generating structured reasoning traces. This model is designed to output a step-by-step thinking process alongside a final answer, making it suitable for tasks requiring explicit reasoning. It builds upon the chimbiwide/gemma-3-1b-it-thinking-32k-sft-base model and supports a context length of 32768 tokens.
Loading preview...