Qwen/Qwen2.5-Math-72B-Instruct

Warm
Public
72.7B
FP8
131072
License: other
Hugging Face
Overview

Qwen2.5-Math-72B-Instruct: A Specialized Mathematical LLM

Qwen2.5-Math-72B-Instruct is a 72.7 billion parameter instruction-tuned model from the Qwen2.5-Math series, developed by Qwen. This model is specifically designed and optimized for solving mathematical problems in both English and Chinese, representing an upgrade from the previous Qwen2-Math series.

Key Capabilities

  • Dual Reasoning Methods: Supports both Chain-of-Thought (CoT) and Tool-integrated Reasoning (TIR) to tackle mathematical challenges. While CoT enhances general reasoning, TIR improves computational accuracy and handles complex symbolic or algorithmic tasks.
  • Multilingual Math Support: Expanded to solve math problems in both Chinese and English, a significant improvement over its predecessor which primarily focused on English CoT.
  • Enhanced Performance: Achieves notable performance improvements on Chinese and English mathematics benchmarks using CoT compared to the Qwen2-Math series. For instance, Qwen2.5-Math-72B-Instruct scores 87.8 on the MATH benchmark using TIR.
  • Instruction-Tuned: This specific model variant is instruction-tuned for chat-based interactions, making it suitable for direct problem-solving queries.

When to Use This Model

  • Mathematical Problem Solving: Ideal for applications requiring precise and reasoned solutions to mathematical problems.
  • Complex Computations: Particularly effective for tasks demanding high computational accuracy, symbolic manipulation, or algorithmic reasoning, leveraging TIR capabilities.
  • Multilingual Math Assistance: Suitable for scenarios involving math problems in both English and Chinese.

Note: This model is primarily focused on mathematical tasks and is not recommended for general-purpose applications.