Qwen/Qwen2.5-Math-1.5B-Instruct
Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Sep 16, 2024License:apache-2.0Architecture:Transformer0.1K Open Weights Warm

Qwen/Qwen2.5-Math-1.5B-Instruct is a 1.5 billion parameter instruction-tuned large language model developed by Qwen, specifically optimized for solving mathematical problems. It supports both Chain-of-Thought (CoT) and Tool-integrated Reasoning (TIR) in English and Chinese. This model is designed for precise computation and algorithmic manipulation in mathematical contexts, building upon the Qwen2-Math series.

Loading preview...

Qwen2.5-Math-1.5B-Instruct: Specialized Mathematical LLM

Qwen2.5-Math-1.5B-Instruct is part of the Qwen2.5-Math series, an upgrade from the earlier Qwen2-Math models, developed by Qwen. This 1.5 billion parameter instruction-tuned model is specifically engineered for solving mathematical problems in both English and Chinese.

Key Capabilities

  • Mathematical Problem Solving: Primarily supports solving math problems through Chain-of-Thought (CoT) and Tool-integrated Reasoning (TIR).
  • Multilingual Support: Expanded to handle mathematical tasks in both Chinese and English.
  • Enhanced Reasoning: Achieves significant performance improvements on Chinese and English mathematics benchmarks using CoT compared to its predecessor.
  • Computational Accuracy: TIR further improves the model's proficiency in precise computation, symbolic manipulation, and algorithmic manipulation, addressing challenges faced by CoT alone.

Good For

  • Developers requiring an LLM specifically for mathematical problem-solving.
  • Applications needing step-by-step reasoning (CoT) or tool-assisted computation (TIR) for math.
  • Use cases involving mathematical tasks in both English and Chinese.
Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p