odats/wmt_all
TEXT GENERATIONConcurrency Cost:1Model Size:1BQuant:BF16Ctx Length:32kPublished:Apr 3, 2026Architecture:Transformer Warm

The odats/wmt_all model is a 1 billion parameter language model fine-tuned from google/gemma-3-1b-it. It was trained using the GRPO method, as introduced in the DeepSeekMath paper, to enhance mathematical reasoning capabilities. This model is optimized for tasks requiring robust reasoning, particularly in mathematical contexts. Its fine-tuning approach aims to push the limits of mathematical problem-solving in open language models.

Loading preview...