mychen76/mistral-7b-merged-slerp
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 9, 2024License:apache-2.0Architecture:Transformer Open Weights Cold
mychen76/mistral-7b-merged-slerp is a 7 billion parameter language model created by mychen76, formed by merging OpenPipe/mistral-ft-optimized-1218 and mlabonne/NeuralHermes-2.5-Mistral-7B using the slerp method. This model leverages the Mistral architecture and is optimized for general language understanding and generation tasks, achieving an average score of 71.09 on the Open LLM Leaderboard. It is suitable for applications requiring a balanced performance across various reasoning and comprehension benchmarks within a 4096-token context window.
Loading preview...