ajtaltarabukin2022/merge_v10_27_112_5

TEXT GENERATIONConcurrency Cost:2Model Size:32BQuant:FP8Ctx Length:32kPublished:Apr 22, 2026Architecture:Transformer Cold

The ajtaltarabukin2022/merge_v10_27_112_5 is a 32 billion parameter language model created by ajtaltarabukin2022 using the DARE TIES merge method. This model combines pre-trained language models, specifically merging '/root/finetuneqwen/prexpertMJDD' with '/root/finetuneqwen/distillgpt7SJvM' as a base. It is designed to leverage the strengths of its constituent models through a weighted merge of their layers, making it suitable for tasks benefiting from combined model capabilities.

Loading preview...

Model Overview

The ajtaltarabukin2022/merge_v10_27_112_5 is a 32 billion parameter language model developed by ajtaltarabukin2022. It was created using the DARE TIES merge method, a technique designed to combine the knowledge and capabilities of multiple pre-trained models into a single, more robust model.

Merge Details

This model is a composite of two distinct pre-trained language models:

  • The primary base model used was /root/finetuneqwen/distillgpt7SJvM.
  • It was merged with /root/finetuneqwen/prexpertMJDD.

The merge process involved applying a 0.5 weight to the layers of both constituent models across a layer range of 0 to 64, indicating an equal contribution from each source model in these layers. The configuration also specified a density of 0.5, suggesting a sparse merging approach.

Potential Use Cases

Given its nature as a merged model, ajtaltarabukin2022/merge_v10_27_112_5 is likely intended for applications that benefit from the combined strengths of its source models. While specific capabilities are not detailed, models created with DARE TIES often aim to improve performance across a broader range of tasks or enhance specific abilities present in the merged components.