Qwen/Qwen3-4B-Thinking-2507

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Aug 5, 2025License:apache-2.0Architecture:Transformer0.6K Open Weights Warm

Qwen/Qwen3-4B-Thinking-2507 is a 4 billion parameter causal language model developed by Qwen, specifically enhanced for complex reasoning tasks. This model features significantly improved performance across logical reasoning, mathematics, science, coding, and academic benchmarks. It also offers enhanced 256K long-context understanding, making it ideal for applications requiring deep analytical processing and extended conversational memory.

Loading preview...

Qwen3-4B-Thinking-2507: Enhanced Reasoning and Long-Context Understanding

Qwen3-4B-Thinking-2507 is a 4 billion parameter causal language model from Qwen, specifically designed to excel in complex reasoning tasks. This iteration builds upon previous versions by significantly improving both the quality and depth of its thinking capabilities.

Key Enhancements and Capabilities

  • Superior Reasoning Performance: Demonstrates marked improvements in logical reasoning, mathematics, science, coding, and academic benchmarks that typically demand expert human knowledge.
  • General Capability Boost: Features better instruction following, tool usage, text generation, and alignment with human preferences.
  • Extended Context Length: Offers enhanced 256K long-context understanding, crucial for processing extensive information.
  • Dedicated Thinking Mode: This model operates exclusively in a "thinking mode," automatically incorporating internal thought processes to tackle highly complex problems. It is recommended to use a context length greater than 131,072 tokens for optimal performance, especially in reasoning-heavy scenarios.

Performance Highlights

The model shows strong performance across various domains, including significant gains in reasoning benchmarks like AIME25 (81.3) and HMMT25 (55.5), and coding benchmarks such as LiveCodeBench v6 (55.2). It also exhibits improved alignment and agentic capabilities, making it suitable for sophisticated AI applications.

Recommended Use Cases

This model is particularly well-suited for applications requiring:

  • Highly complex reasoning tasks across scientific, mathematical, and coding domains.
  • Advanced agentic behaviors and tool-calling, especially when integrated with frameworks like Qwen-Agent.
  • Processing and understanding very long documents or conversations due to its 256K native context length.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p