Kwai-Klear/Klear-Reasoner-8B-SFT
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Aug 13, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

Kwai-Klear/Klear-Reasoner-8B-SFT is an 8-billion-parameter reasoning model developed by Kwai-Klear, featuring a 32768-token context length. It is specifically optimized for complex mathematical and coding tasks, achieving state-of-the-art performance on benchmarks like AIME and LiveCodeBench. The model incorporates a novel Gradient-Preserving Clipping Policy Optimization (GPPO) method to enhance exploration and convergence during training, making it particularly effective for problem-solving requiring careful deliberation.

Loading preview...