mychen76/mistral-7b-merged-dare
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 9, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

mychen76/mistral-7b-merged-dare is a 7 billion parameter language model based on the Mistral-7B-v0.1 architecture, created using a DARE TIES merge of several Mistral-based models. This merge method, incorporating models like SamirGPT-v1 and Slerp-CM-mist-dpo, aims to enhance general reasoning and language understanding. It achieves an average score of 73.46 on the Open LLM Leaderboard, making it suitable for a range of general-purpose NLP tasks.

Loading preview...