dongguanting/Qwen3-8B-ARPO-DeepSearch
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Jul 24, 2025License:mitArchitecture:Transformer0.0K Open Weights Warm
The dongguanting/Qwen3-8B-ARPO-DeepSearch model is an 8 billion parameter language model developed by dongguanting, based on the Qwen3 architecture. It incorporates the ARPO (Adaptive Reward Policy Optimization) method, designed to enhance model performance through advanced optimization techniques. With a context length of 32768 tokens, this model is suitable for applications requiring deep contextual understanding and optimized response generation.
Loading preview...