allknowingroger/LlamaSlerp1-8B
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Nov 21, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

allknowingroger/LlamaSlerp1-8B is an 8 billion parameter language model created by allknowingroger, formed by merging DreadPoor/BaeZel-8B-LINEAR and allenai/Llama-3.1-Tulu-3-8B using the SLERP method. This merge technique, with a V-shaped parameter curve, aims to combine the strengths of its constituent models. It is designed for general language tasks, leveraging the combined capabilities of its base models.

Loading preview...