a-m-team/AM-Thinking-v1

Warm
Public
32B
FP8
32768
License: apache-2.0
Hugging Face
Overview

AM-Thinking-v1: A 32B Reasoning Powerhouse

AM-Thinking-v1, developed by a-m-team, is a 32 billion parameter dense language model engineered to push the boundaries of reasoning capabilities. Built on the open-source Qwen 2.5-32B-Base, this model achieves strong performance on complex reasoning benchmarks, rivaling larger MoE models like DeepSeek-R1 and Qwen3-235B-A22B, despite being significantly smaller.

Key Differentiators & Capabilities

  • Exceptional Reasoning at 32B Scale: Outperforms DeepSeek-R1 on AIME’24/’25 & LiveCodeBench and approaches Qwen3-235B-A22B, demonstrating flagship-level reasoning from a dense model.
  • Efficient Deployment: Designed to fit on a single A100-80GB GPU with deterministic latency, avoiding the overhead of MoE routing.
  • Advanced Post-Training Pipeline: Utilizes a sophisticated dual-stage RL (Reinforcement Learning) scheme, including pass-rate-aware data curation, to enhance its "think-then-answer" behavioral pattern.
  • Open-Source Foundation: Fully based on publicly available components, including Qwen 2.5-32B-Base and RL training queries.

Ideal Use Cases

  • Code Generation: Capable of generating complex scripts and handling collision detection, as demonstrated by examples like a bouncing ball simulation.
  • Logical Problem Solving: Excels in tasks requiring intricate logical deduction.
  • Creative Writing: Shows proficiency in generating coherent and contextually relevant text.

Limitations

AM-Thinking-v1 is not yet trained for structured function-calling or tool-use workflows, limiting its application in agent-style systems that interact with external systems. Its safety alignment is also in early stages, requiring further rigorous red-teaming.