albinolo/DeepSeek-R1-Distill-Qwen-32B

TEXT GENERATIONConcurrency Cost:2Model Size:32.8BQuant:FP8Ctx Length:32kPublished:Mar 28, 2026License:mitArchitecture:Transformer Open Weights Cold

The albinolo/DeepSeek-R1-Distill-Qwen-32B is a 32.8 billion parameter language model developed by DeepSeek-AI, distilled from the larger DeepSeek-R1 model and based on the Qwen2.5 architecture. It is specifically fine-tuned using reasoning data generated by DeepSeek-R1, excelling in complex reasoning tasks across math, code, and general knowledge. This model demonstrates strong performance on benchmarks like AIME 2024 and MATH-500, outperforming several larger models in its class.

Loading preview...

DeepSeek-R1-Distill-Qwen-32B: Reasoning-Enhanced Distilled Model

This model is a 32.8 billion parameter variant from the DeepSeek-R1-Distill series, developed by DeepSeek-AI. It is a distilled version of the larger DeepSeek-R1 model, fine-tuned on the Qwen2.5-32B base using reasoning data generated by DeepSeek-R1. The core innovation lies in demonstrating that reasoning patterns from powerful larger models can be effectively transferred to smaller, dense models through distillation.

Key Capabilities

  • Enhanced Reasoning: Achieves strong performance in complex reasoning tasks across mathematics, coding, and general knowledge, benefiting from the advanced reasoning capabilities of its DeepSeek-R1 parent.
  • Competitive Benchmarks: Outperforms models like GPT-4o-0513 and Claude-3.5-Sonnet-1022 on specific reasoning benchmarks such as AIME 2024 (72.6% pass@1) and MATH-500 (94.3% pass@1).
  • Efficient Performance: Provides high reasoning capability in a 32.8B parameter dense model, making it a powerful option for applications requiring strong analytical skills without the overhead of much larger models.

Good For

  • Mathematical Problem Solving: Excels in advanced math challenges, as evidenced by its high scores on AIME and MATH-500.
  • Code Generation and Reasoning: Demonstrates robust performance in coding benchmarks like LiveCodeBench and Codeforces rating.
  • Complex Query Handling: Suitable for applications requiring detailed, step-by-step reasoning and problem-solving.
  • Resource-Efficient Deployment: Offers a strong balance of performance and size for deployment where larger MoE models might be impractical.