xw1234gan/SFT_Qwen2.5-7B-Instruct_MATH
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Mar 10, 2026Architecture:Transformer Cold
The xw1234gan/SFT_Qwen2.5-7B-Instruct_MATH is a 7.6 billion parameter instruction-tuned model based on the Qwen2.5 architecture. This model is specifically fine-tuned for mathematical tasks, leveraging its instruction-following capabilities to address complex numerical and logical problems. It features a substantial 32,768 token context length, making it suitable for processing extensive mathematical queries and multi-step reasoning. This model is designed for applications requiring robust performance in quantitative analysis and problem-solving.
Loading preview...