Azazelle/Mocha-SR-7b-ex
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 23, 2024License:cc-by-4.0Architecture:Transformer Open Weights Cold

Mocha-SR-7b-ex is a 7 billion parameter language model developed by Azazelle, built upon the Mistral-7B-v0.1 architecture. This model is a merge of Open-Orca/Mistral-7B-OpenOrca, WizardLM/WizardMath-7B-V1.1, and akjindal53244/Mistral-7B-v0.1-Open-Platypus, utilizing the rescaled_sample merge method. It is designed to combine the strengths of its constituent models, offering enhanced capabilities for general language tasks, instruction following, and mathematical reasoning within a 4096-token context window.

Loading preview...