BitStarWalkin/SuperCorrect-7B
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Oct 13, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

BitStarWalkin/SuperCorrect-7B is a 7.6 billion parameter language model developed by Ling Yang, Zhaochen Yu, and their collaborators from Peking University, Skywork AI, UC Berkeley, and Stanford University. It utilizes a novel two-stage fine-tuning method called SuperCorrect to enhance reasoning accuracy and self-correction abilities. This model significantly outperforms other 7B models like DeepSeekMath-7B and Qwen2.5-Math-7B on mathematical benchmarks (MATH/GSM8K) by incorporating a hierarchical thought template (Buffer of Thought) for deliberate reasoning.

Loading preview...