harshitv804/MetaMath-Mistral-2x7B
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 9, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

harshitv804/MetaMath-Mistral-2x7B is an experimental 7-billion parameter Mixture of Experts (MoE) model based on the Mistral architecture, created by harshitv804. This model was developed using the SLERP merge method, combining two instances of the meta-math/MetaMath-Mistral-7B model. It is designed for exploring MoE concepts and is particularly suited for mathematical reasoning tasks, leveraging its MetaMath-Mistral-7B foundation.

Loading preview...