Gille/StrangeMerges_46-7B-dare_ties
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 25, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

Gille/StrangeMerges_46-7B-dare_ties is a 7 billion parameter language model created by Gille, formed by merging Gille/StrangeMerges_45-7B-dare_ties, kettleguts/zephyr-7b-beta_sparse05, and chihoonlee10/T3Q-Mistral-Orca-Math-DPO using the dare_ties method. This model demonstrates an average performance of 69.96 on the Open LLM Leaderboard, with notable scores in HellaSwag (86.40) and Winogrande (79.48). It is designed for general language tasks, leveraging its merged architecture to balance various capabilities.

Loading preview...