Gweizheng/Marcoro14-7B-dare
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Apr 3, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

Gweizheng/Marcoro14-7B-dare is a 7 billion parameter language model created by Gweizheng, built upon the Mistral-7B-v0.1 architecture. This model is a merge of SamirGPT-v1, Slerp-CM-mist-dpo, and Mistral-7B-Merge-14-v0.2 using the dare_ties merging method. It is designed to combine the strengths of its constituent models, offering a versatile base for various natural language processing tasks.

Loading preview...