Technoculture/BioMistral-Hermes-Slerp
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 21, 2024License:apache-2.0Architecture:Transformer Open Weights Cold
Technoculture/BioMistral-Hermes-Slerp is a 7 billion parameter language model created by Technoculture, merging BioMistral/BioMistral-7B-DARE and NousResearch/Nous-Hermes-2-Mistral-7B-DPO using the slerp method. This model is specifically designed to combine the biomedical knowledge of BioMistral with the general conversational and instruction-following capabilities of Nous-Hermes-2. With a 4096-token context length, it aims to provide a versatile foundation for applications requiring both specialized medical understanding and broad language proficiency.
Loading preview...