stevez80/ErebusNeuralSamir-7B-dare-ties
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 9, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

stevez80/ErebusNeuralSamir-7B-dare-ties is a 7 billion parameter language model created by stevez80, built upon the Mistral-7B-v0.1 architecture. This model is a DARE TIES merge of SamirGPT-v1, NeuralHermes-2.5-Mistral-7B, and Mistral-7B-Erebus-v3, designed to combine the strengths of its constituent models. It features a 4096-token context length and is configured with int8_mask and bfloat16 dtype for efficient operation.

Loading preview...