fzzhang/mistralv1_gsm8k_merged_s
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 16, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

The fzzhang/mistralv1_gsm8k_merged_s is a 7 billion parameter language model based on the MistralV1 architecture, fine-tuned for mathematical reasoning and problem-solving, specifically on the GSM8K dataset. With a context length of 4096 tokens, this model is optimized for tasks requiring numerical understanding and logical deduction. It aims to provide enhanced performance in quantitative reasoning compared to general-purpose LLMs.

Loading preview...