Gille/StrangeMerges_31-7B-slerp
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 6, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

Gille/StrangeMerges_31-7B-slerp is a 7 billion parameter language model created by Gille, formed by spherically interpolating (slerp) two base models: Gille/StrangeMerges_30-7B-slerp and yam-peleg/Experiment24-7B. This merge leverages specific layer-wise parameter adjustments for self-attention and MLP blocks, aiming to combine the strengths of its constituent models. It is designed for general text generation tasks, offering a balanced performance profile derived from its merged architecture.

Loading preview...