HuggingFaceTB/qwen3-1.7b-gsm8k-sft
TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Mar 25, 2026License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

HuggingFaceTB/qwen3-1.7b-gsm8k-sft is a 1.7 billion parameter Qwen3-based causal language model fine-tuned specifically for mathematical reasoning. It achieves 77.2% accuracy on the GSM8K benchmark, a significant improvement over its base model, and also performs well on competition-level math problems. This model is optimized for solving grade school math word problems using chain-of-thought reasoning, making it suitable for applications requiring robust numerical problem-solving capabilities.

Loading preview...