Gille/StrangeMerges_25-7B-dare_ties
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 18, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

Gille/StrangeMerges_25-7B-dare_ties is a 7 billion parameter language model created by Gille, formed by merging Gille/StrangeMerges_21-7B-slerp and bardsai/jaskier-7b-dpo-v5.6 using the dare_ties method. This model features a 4096-token context length and achieves an average score of 76.33 on the Open LLM Leaderboard, demonstrating capabilities across various reasoning and language understanding tasks. It is suitable for general-purpose text generation and understanding, particularly in scenarios requiring a blend of its constituent models' strengths.

Loading preview...