berkecr/tr-dare-merge-7B
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kLicense:apache-2.0Architecture:Transformer0.0K Open Weights Cold
The berkecr/tr-dare-merge-7B is a 7 billion parameter language model created by berkecr, built upon the Mistral-7B-Instruct-v0.2 architecture. This model is a DARE merge of Mistral-7B-Instruct-v0.2 with TURKCELL/Turkcell-LLM-7b-v1 and Trendyol/Trendyol-LLM-7b-chat-dpo-v1.0, utilizing a context length of 4096 tokens. It is specifically designed to combine the strengths of these base models, likely enhancing performance in areas relevant to their original training domains. The merge method employed is 'dare_ties', with specific density and weight parameters applied to the merged models.
Loading preview...