Laoyujie/merged-qwen-ta

TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Apr 22, 2026Architecture:Transformer Cold

The Laoyujie/merged-qwen-ta model is a 7.6 billion parameter language model created by Laoyujie, utilizing a merge of pre-trained models based on the Qwen architecture. This model was developed using the Task Arithmetic merge method, specifically combining capabilities from base, code, and math models. It is designed to leverage strengths in both code generation and mathematical reasoning, making it suitable for tasks requiring proficiency in these domains.

Loading preview...

Overview

The Laoyujie/merged-qwen-ta is a 7.6 billion parameter language model developed by Laoyujie. It was created using the MergeKit tool, specifically employing the Task Arithmetic merge method. This approach combines the strengths of multiple specialized models into a single, more versatile model.

Key Capabilities

  • Enhanced Code Understanding and Generation: The model integrates a dedicated 'code' model, suggesting improved performance on programming-related tasks.
  • Mathematical Reasoning: By incorporating a 'math' model, it aims to handle mathematical problems and logical reasoning more effectively.
  • Hybrid Specialization: The Task Arithmetic merge allows for a balanced integration of capabilities from a base model, a code-focused model, and a math-focused model, providing a broad skill set.

Merge Details

The model was constructed from a base model, with additional capabilities merged from specific code and math models. The configuration used a weight of 1.0 for both the code and math components during the merge process, indicating an equal emphasis on these specialized areas.