s3nh/Mistral_Sonyichi-7B-slerp
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 5, 2024License:openrailArchitecture:Transformer0.0K Open Weights Cold
s3nh/Mistral_Sonyichi-7B-slerp is a 7 billion parameter language model created by s3nh, merged using the SLERP method from SanjiWatsuki/Sonya-7B, EmbeddedLLM/Mistral-7B-Merge-14-v0.1, and SanjiWatsuki/Kunoichi-7B. This model leverages the Mistral architecture and is optimized for general reasoning tasks, achieving an average score of 70.52 on the Open LLM Leaderboard across various benchmarks including MMLU and HellaSwag. It is suitable for applications requiring robust language understanding and generation within a 4096-token context window.
Loading preview...