seele123/MATH-Qwen2.5-math-7B-ReMax-L2O-4
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Jan 28, 2026Architecture:Transformer Cold
The seele123/MATH-Qwen2.5-math-7B-ReMax-L2O-4 is a 7.6 billion parameter language model based on the Qwen2.5 architecture, developed by seele123. This model is specifically fine-tuned and optimized for mathematical reasoning and problem-solving tasks. With a substantial context length of 131072 tokens, it is designed to handle complex mathematical queries and provide accurate solutions. Its primary strength lies in its enhanced performance on mathematical benchmarks, making it suitable for applications requiring robust numerical and logical processing.
Loading preview...