thangvip/Qwen3-1.7B-SFT-math-1500 is a 2 billion parameter language model, fine-tuned from Qwen/Qwen3-1.7B using Supervised Fine-Tuning (SFT) with the TRL framework. This model is specifically optimized for mathematical reasoning and problem-solving tasks, building upon the Qwen3 architecture. It is designed to enhance performance in quantitative domains, making it suitable for applications requiring strong mathematical capabilities.
No reviews yet. Be the first to review!