ertghiu256/Qwen3-4b-tcomanr-merge-v2.6
Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Nov 1, 2025Architecture:Transformer0.0K Warm

ertghiu256/Qwen3-4b-tcomanr-merge-v2.6 is a 4 billion parameter language model based on the Qwen3 architecture, created by ertghiu256. This model is a merge of multiple specialized Qwen3-4B models, including those focused on reasoning, code, math, and instruction following, using the TIES merge method. It is designed to combine diverse capabilities from its constituent models, offering a broad range of applications with a context length of 40960 tokens.

Loading preview...

ertghiu256/Qwen3-4b-tcomanr-merge-v2.6 Overview

This model is a 4 billion parameter language model built upon the Qwen3 architecture, developed by ertghiu256. It was created using the TIES merge method from mergekit, with Qwen/Qwen3-4B-Thinking-2507 serving as the base model.

Key Capabilities

The tcomanr-merge-v2.6 integrates capabilities from a diverse set of specialized Qwen3-4B models, aiming for a versatile performance profile. The merged models include those focused on:

  • Reasoning: Incorporating models like Qwen3-4B-Thinking-2507, qwen3-multi-reasoner, and qwen-3-4b-mixture-of-thought.
  • Code Generation & Reasoning: Leveraging qwen3-4b-code-reasoning and Qwen3-Code-Reasoning-4B.
  • Mathematical Tasks: Including qwen3-math-reasoner and Qwen3-4b-2507-Thinking-math-and-code.
  • Instruction Following: Drawing from Qwen3-4b-Instruct-2507 and Qwen3-Hermes-4b.
  • Web and UI Generation: Integrating WEBGEN-4B-Preview and UIGEN-FX-4B-Preview.

This broad integration suggests a model designed for general-purpose applications requiring a blend of these specific strengths.

Good for

  • Multi-domain tasks: Its merged nature makes it suitable for applications that benefit from combined reasoning, coding, and mathematical abilities.
  • Exploration of merged model performance: Researchers and developers interested in the practical outcomes of the TIES merging technique on Qwen3-4B models.
  • Applications requiring a balanced skillset: Ideal for use cases where a single model needs to handle diverse types of queries without excelling in just one narrow area.