linius/Qwen3-8B-SPoT
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Mar 4, 2026License:apache-2.0Architecture:Transformer0.0K Open Weights Cold
Qwen3-8B-SPoT is an 8 billion parameter large language model developed by linius, post-trained from the Qwen/Qwen3-8B base model. Utilizing the Surgical Post-Training (SPoT) paradigm, it significantly enhances reasoning capabilities, particularly for complex math and reasoning tasks. The model achieves an average accuracy improvement of 6.2% on these tasks while effectively mitigating catastrophic forgetting of general knowledge. It is designed for applications requiring robust reasoning performance with strong knowledge retention.
Loading preview...