seele123/MATH-Qwen2.5-math-7B-ReMax-L2O-NoBaseline
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Jan 27, 2026Architecture:Transformer Cold

The seele123/MATH-Qwen2.5-math-7B-ReMax-L2O-NoBaseline model is a 7.6 billion parameter language model based on the Qwen2.5 architecture. Developed by seele123, this model is specifically fine-tuned for mathematical reasoning and problem-solving tasks. It features a substantial context length of 131072 tokens, making it suitable for complex mathematical queries requiring extensive context. This model is designed to excel in environments where precise mathematical computation and logical deduction are critical.

Loading preview...