s3nh/Severusectum-7B-DPO
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 3, 2024License:mitArchitecture:Transformer0.0K Open Weights Cold
s3nh/Severusectum-7B-DPO is a 7 billion parameter language model created by s3nh, formed by merging FelixChao/Sectumsempra-7B-DPO and FelixChao/WestSeverus-7B-DPO-v2 using the SLERP method. This model achieves an average score of 75.18 on the Open LLM Leaderboard, demonstrating capabilities across reasoning, common sense, and language understanding tasks. With a 4096-token context length, it is suitable for general-purpose applications requiring robust performance in its size class.
Loading preview...