xw1234gan/SFT_Qwen2.5-1.5B-Instruct_MATH
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Mar 5, 2026Architecture:Transformer Warm
The xw1234gan/SFT_Qwen2.5-1.5B-Instruct_MATH is a 1.5 billion parameter instruction-tuned causal language model based on the Qwen2.5 architecture, developed by xw1234gan. This model is specifically fine-tuned for mathematical tasks and reasoning, leveraging a 32,768 token context length. It is designed to excel in environments requiring robust mathematical problem-solving capabilities.
Loading preview...