odats/rl_nmt_2026_04_03_17_29
TEXT GENERATIONConcurrency Cost:1Model Size:1BQuant:BF16Ctx Length:32kPublished:Apr 3, 2026Architecture:Transformer Loading
The odats/rl_nmt_2026_04_03_17_29 model is a 1 billion parameter instruction-tuned causal language model, fine-tuned from Google's Gemma-3-1B-IT. It was trained using the TRL library and incorporates the GRPO method, which is designed to enhance mathematical reasoning capabilities. This model is particularly suited for tasks requiring improved reasoning, building upon its base Gemma architecture.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
–
top_p
–
top_k
–
frequency_penalty
–
presence_penalty
–
repetition_penalty
–
min_p
–