mlfoundations-dev/deepmath
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Apr 20, 2025License:apache-2.0Architecture:Transformer Open Weights Cold

The mlfoundations-dev/deepmath model is a 7.6 billion parameter instruction-tuned language model, fine-tuned from Qwen/Qwen2.5-7B-Instruct. It is specifically adapted using the mlfoundations-dev/deepmath dataset, suggesting an optimization for mathematical reasoning and problem-solving tasks. With a substantial context length of 131072 tokens, it is designed for processing extensive mathematical or technical inputs. This model is intended for applications requiring advanced mathematical understanding and precise logical deduction.

Loading preview...