Gille/StrangeMerges_18-7B-dare_ties
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 1, 2024License:apache-2.0Architecture:Transformer Open Weights Cold
Gille/StrangeMerges_18-7B-dare_ties is a 7 billion parameter language model created by Gille, formed by merging Gille/StrangeMerges_17-7B-dare_ties and teknium/OpenHermes-2.5-Mistral-7B using the dare_ties method. This model demonstrates strong general reasoning capabilities, achieving an average score of 67.06 on the Open LLM Leaderboard across various benchmarks. It is suitable for tasks requiring robust understanding and generation, with a context length of 4096 tokens.
Loading preview...