Gille/StrangeMerges_40-7B-dare_ties
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 17, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

Gille/StrangeMerges_40-7B-dare_ties is a 7 billion parameter language model created by Gille, built upon the Mistral-7B-v0.1 base model. This model is a DARE TIES merge of three distinct models: Gille/StrangeMerges_34-7B-slerp, yam-peleg/Experiment26-7B, and chihoonlee10/T3Q-Mistral-Orca-Math-DPO. It is designed to combine the strengths of its constituent models, particularly incorporating a math-optimized component, making it suitable for tasks requiring diverse reasoning capabilities within a 4096 token context length.

Loading preview...