Ilia2003Mah/qwen2.5-1.5b-gsm8k-train-step2000

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Mar 24, 2026Architecture:Transformer Warm

The Ilia2003Mah/qwen2.5-1.5b-gsm8k-train-step2000 model is a 1.5 billion parameter language model based on the Qwen2.5 architecture. This model is specifically fine-tuned for mathematical reasoning tasks, particularly those found in the GSM8K dataset. With a context length of 32768 tokens, it is designed to handle complex numerical problems and multi-step arithmetic. Its primary strength lies in its ability to process and solve grade-school level math problems.

Loading preview...

Model Overview

The Ilia2003Mah/qwen2.5-1.5b-gsm8k-train-step2000 is a 1.5 billion parameter model built upon the Qwen2.5 architecture. This model has been specifically fine-tuned to excel in mathematical reasoning, with a particular focus on the GSM8K dataset, which comprises grade-school level math word problems. It leverages a substantial context window of 32768 tokens, enabling it to process longer problem descriptions and multi-step solutions.

Key Capabilities

  • Mathematical Reasoning: Optimized for solving arithmetic and word problems, particularly those found in the GSM8K benchmark.
  • Large Context Window: Supports a 32768-token context length, beneficial for detailed problem statements and complex reasoning chains.
  • Qwen2.5 Architecture: Benefits from the underlying capabilities of the Qwen2.5 model family.

Good For

  • Educational Applications: Developing tools for math tutoring or problem-solving assistance.
  • Research in Mathematical LLMs: Investigating the performance of smaller models on arithmetic and reasoning tasks.
  • Benchmarking: Evaluating the effectiveness of fine-tuning strategies for specific mathematical datasets like GSM8K.