MetaStoneTec/MetaStone-S1-32B
TEXT GENERATIONConcurrency Cost:2Model Size:32.8BQuant:FP8Ctx Length:32kPublished:Jul 5, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

MetaStoneTec/MetaStone-S1-32B is a 32.8 billion parameter reflective generative model developed by MetaStoneTec, featuring a 131072 token context length. It is trained using a novel "Long-CoT Reinforcement Learning" and "Process Reward Learning" approach, enabling deep reasoning and high-quality reasoning trajectory selection. This architecture significantly reduces inference costs while maintaining performance comparable to larger models like OpenAI-o3 series on mathematics, coding, and Chinese reasoning tasks.

Loading preview...