Gille/StrangeMerges_4-7B-slerp
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 27, 2024License:apache-2.0Architecture:Transformer Open Weights Cold
StrangeMerges_4-7B-slerp is a 7 billion parameter language model developed by Gille, created by merging Gille/StrangeMerges_3-7B-slerp and Gille/StrangeMerges_2-7B-slerp using a slerp method. This model demonstrates an average performance of 72.63 on the Open LLM Leaderboard, with notable scores in HellaSwag (87.01) and Winogrande (82.95). It is designed for general language generation tasks, leveraging its merged architecture for balanced capabilities.
Loading preview...