jan-hq/supermario-v2
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Dec 11, 2023License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

The jan-hq/supermario-v2 is a 7 billion parameter language model based on Mistral-7B-v0.1, created by Jan HQ using the DARE_TIES merge method. It combines OpenHermes-2.5-neural-chat-v3-3-Slerp, MetaMath-Cybertron-Starling, and Marcoroni-7B-v3 to enhance general reasoning and conversational abilities. With a 4096-token context length, it achieves an average score of 72.36 on the Open LLM Leaderboard, making it suitable for a range of general-purpose AI applications.

Loading preview...