nhe-ai/Qwen3-4B-Qwen3.6-plus-Reasoning-Distilled

TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Apr 27, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The nhe-ai/Qwen3-4B-Qwen3.6-plus-Reasoning-Distilled model is a 4 billion parameter language model, fine-tuned from Qwen3-4B-Thinking with a 32K context length. It utilizes QLoRA distillation to replicate the advanced reasoning capabilities of the larger Qwen3.6-plus teacher model. This model is specifically optimized to produce concise, structured, and actionable solution paths for complex reasoning tasks, reducing the 'rambling' often seen in smaller models. It excels at providing structured problem analysis, algorithm formulation, and complexity analysis, making it suitable for engineering-grade applications requiring robust logical deduction.

Loading preview...

Overview

nhe-ai/Qwen3-4B-Qwen3.6-plus-Reasoning-Distilled is a 4 billion parameter language model, distilled from the Qwen3-4B-Thinking base model. It leverages QLoRA fine-tuning to inherit the advanced reasoning capabilities of the larger Qwen3.6-plus teacher model. This distillation process focuses on transforming the base model's exploratory, stream-of-consciousness style into a structured, professional, and report-oriented approach for complex problem-solving.

Key Capabilities

  • Enhanced Reasoning Structure: Significantly improves the qualitative structure of reasoning, moving from verbose, self-correcting outputs to concise, concrete solution plans.
  • Problem Analysis: Capable of immediately analyzing problems, separating concerns (Input, Output, Constraints), and formulating clear algorithmic strategies.
  • Structured Output: Provides clean breakdowns including Problem Analysis, Intuition, Algorithm, Complexity Analysis, and Pseudocode.
  • Reduced "Rambling": Eliminates hesitation and logical dead-ends, offering confident and direct solutions.
  • Benchmark Performance: Achieves a score of 75.64 on the khazarai/Multi-Domain-Reasoning-Benchmark, outperforming the base Qwen3-4B-Thinking model (73.73).

Good For

  • Complex Problem Solving: Ideal for tasks requiring structured logical deduction, such as solving intricate graph problems (e.g., Shortest Path with Edge Reversals).
  • Engineering Applications: Suitable for scenarios where an "engineering-grade tool" is needed to provide clear, actionable, and well-analyzed solutions.
  • Educational Tools: Can be used to generate structured explanations and algorithmic approaches for technical problems.
  • Automated Reasoning: Applications requiring models to articulate their thought process in a formal and organized manner.