microsoft/rho-math-1b-v0.1
TEXT GENERATIONConcurrency Cost:1Model Size:1.1BQuant:BF16Ctx Length:2kPublished:Apr 11, 2024License:mitArchitecture:Transformer0.0K Open Weights Warm

The microsoft/rho-math-1b-v0.1 is a 1.1 billion parameter causal language model developed by Microsoft, pre-trained using Selective Language Modeling (SLM). This method focuses on selectively training on clean and useful tokens, enabling it to achieve strong mathematical reasoning performance with significantly fewer pretraining tokens. It excels in mathematical tasks, demonstrating competitive accuracy on datasets like MATH and GSM8K compared to larger models.

Loading preview...