open-thoughts/OpenThinker2-32B

Warm
Public
32.8B
FP8
131072
Apr 3, 2025
License: apache-2.0
Hugging Face
Overview

OpenThinker2-32B: Enhanced Reasoning Model

OpenThinker2-32B is a 32.8 billion parameter language model developed by open-thoughts, building upon the Qwen2.5-32B-Instruct architecture. It represents an advancement over the previous OpenThinker-32B, primarily through fine-tuning on the expanded OpenThoughts2-1M dataset.

Key Capabilities & Performance

This model is specifically designed and trained to excel in complex reasoning tasks, particularly in mathematics and code. Its training data includes a significant augmentation of math and code reasoning examples, generated through diverse methodologies. Benchmarks, evaluated using the open-source tool Evalchemy, show OpenThinker2-32B achieving strong results:

  • AIME24: 76.7
  • AIME25: 58.7
  • AMC23: 94.0
  • MATH500: 90.8
  • GPQA-D: 64.1
  • LCBv2: 72.5

These scores indicate its proficiency in advanced problem-solving and logical deduction, often outperforming other models in its class on these specific metrics.

Training & Data

The model was trained for 50 hours across 128 4xA100 nodes. The OpenThoughts2-1M dataset is a key differentiator, incorporating existing datasets like OpenR1 and new, high-quality math and code reasoning data. Further details on the dataset and model development can be found in the OpenThoughts Paper and the OpenThoughts2 blog post.

Intended Uses

OpenThinker2-32B is well-suited for applications requiring robust reasoning, mathematical problem-solving, and code-related tasks. Its Apache 2.0 License allows for broad use.