THU-KEG/AdaptThink-7B-delta0.05
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:May 19, 2025License:mitArchitecture:Transformer0.0K Open Weights Cold
THU-KEG/AdaptThink-7B-delta0.05 is a 7.6 billion parameter language model developed by THU-KEG, based on the DeepSeek-R1-Distill-Qwen architecture. It implements the AdaptThink reinforcement learning algorithm, enabling it to adaptively choose between 'Thinking' and 'NoThinking' modes based on problem difficulty. This model is specifically designed to reduce inference costs while improving performance by engaging in detailed reasoning only for challenging problems and providing direct answers for simpler ones.
Loading preview...