ertghiu256/Qwen3-4B-Thinking-2507-Hermes-3
Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Aug 31, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

The ertghiu256/Qwen3-4B-Thinking-2507-Hermes-3 is a 4 billion parameter Qwen3-based language model, fine-tuned with the Hermes 3 dataset. This model is specifically designed to retain strong reasoning capabilities and improve instruction following. It features a notable context length of 40960 tokens, making it suitable for tasks requiring extensive contextual understanding and precise responses.

Loading preview...

Model Overview

ertghiu256/Qwen3-4B-Thinking-2507-Hermes-3 is a 4 billion parameter model built on the Qwen3 architecture. It has been fine-tuned using the Hermes 3 dataset, with a focus on enhancing its core capabilities in reasoning and instruction adherence. The model supports an extended context length of 40960 tokens, allowing for processing and generating longer, more complex texts.

Key Capabilities

  • Enhanced Reasoning: The model is specifically trained to maintain and improve its reasoning abilities.
  • Better Instruction Following: It demonstrates improved performance in understanding and executing user instructions.
  • Extended Context: With a 40960-token context window, it can handle detailed and lengthy prompts.

Training Details

The model was trained using Unsloth, undergoing 60 steps with a learning rate of 3e-5. The training incorporated 28,000 samples from the Hermes 3 dataset, which contributed to its specialized focus on reasoning and instruction following.

Recommended Usage

Optimal performance is suggested with specific inference parameters:

  • Temperature: 0.6
  • Top_P: 20
  • Top_K: 0.95