Gille/StrangeMerges_37-7B-dare_ties
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 14, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

Gille/StrangeMerges_37-7B-dare_ties is a 7 billion parameter language model created by Gille, built using a 'dare_ties' merge method. This model combines liminerity/M7-7b, Gille/StrangeMerges_30-7B-slerp, and ContextualAI/Contextual_KTO_Mistral_PairRM, resulting in a model with a 4096 token context length. It demonstrates balanced performance across various benchmarks, including reasoning, common sense, and language understanding tasks, making it suitable for general-purpose applications requiring a compact yet capable model.

Loading preview...