FiditeNemini/Qwen2.5-14B-DeepSeek-R1-1M-Uncensored

Warm
Public
14.8B
FP8
32768
1
Hugging Face

FiditeNemini/Qwen2.5-14B-DeepSeek-R1-1M-Uncensored is a 14.8 billion parameter language model created by FiditeNemini, formed by merging existing pre-trained models using the TIES method. It utilizes mkurman/Qwen2.5-14B-DeepSeek-R1-1M as its base model, integrating components from huihui-ai/DeepSeek-R1-Distill-Qwen-14B-abliterated-v2. This model is designed for general language generation tasks, leveraging its merged architecture to combine strengths from its constituent models.

Overview

Model Overview

FiditeNemini/Qwen2.5-14B-DeepSeek-R1-1M-Uncensored is a 14.8 billion parameter language model developed by FiditeNemini. It was created using the TIES merge method via mergekit, combining the strengths of multiple pre-trained models.

Key Characteristics

  • Merge-based Architecture: This model is a product of merging, specifically using mkurman/Qwen2.5-14B-DeepSeek-R1-1M as its base and incorporating huihui-ai/DeepSeek-R1-Distill-Qwen-14B-abliterated-v2.
  • TIES Merge Method: The integration process utilized the TIES (Trimmed, Iterative, and Selective) merging method, which is designed to combine models effectively while preserving performance.
  • Configuration: The merge was performed with specific parameters, including a weight and density of 1 for the contributing model, and bfloat16 dtype, with normalization and int8 masking enabled.

Potential Use Cases

Given its merged nature and the base models involved, this model is suitable for:

  • General Text Generation: Capable of various language understanding and generation tasks.
  • Exploration of Merged Models: Ideal for researchers and developers interested in the performance characteristics of models created through advanced merging techniques.