thangvip/Qwen3-1.7B-SFT-math-1500
TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Jan 22, 2026Architecture:Transformer Warm

thangvip/Qwen3-1.7B-SFT-math-1500 is a 2 billion parameter language model, fine-tuned from Qwen/Qwen3-1.7B using Supervised Fine-Tuning (SFT) with the TRL framework. This model is specifically optimized for mathematical reasoning and problem-solving tasks, building upon the Qwen3 architecture. It is designed to enhance performance in quantitative domains, making it suitable for applications requiring strong mathematical capabilities.

Loading preview...