fzzhang/mistralv1_gsm8k_merged
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 16, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

fzzhang/mistralv1_gsm8k_merged is a 7 billion parameter language model based on the MistralV1 architecture. This model is specifically fine-tuned for mathematical reasoning and problem-solving, particularly on the GSM8K dataset. It is designed to excel at arithmetic and multi-step mathematical tasks, making it suitable for applications requiring robust quantitative analysis.

Loading preview...