mini97/qwen2.5-math-7b_grpo_entropy_adv
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Jan 25, 2026Architecture:Transformer Cold
The mini97/qwen2.5-math-7b_grpo_entropy_adv model is a 7.6 billion parameter language model based on the Qwen2.5 architecture, developed by mini97. It features an exceptionally large context length of 131,072 tokens, making it suitable for processing extensive inputs. While specific fine-tuning details are not provided, its name suggests an optimization for mathematical reasoning tasks using GRPO and entropy-based adversarial training. This model is primarily intended for applications requiring advanced mathematical problem-solving and long-context understanding.
Loading preview...