hkust-nlp/Qwen-2.5-32B-SimpleRL-Zoo
TEXT GENERATIONConcurrency Cost:2Model Size:32.8BQuant:FP8Ctx Length:32kPublished:Mar 24, 2025License:apache-2.0Architecture:Transformer Open Weights Cold

The hkust-nlp/Qwen-2.5-32B-SimpleRL-Zoo is a 32.8 billion parameter language model developed by hkust-nlp, based on the Qwen 2.5 architecture. This model is fine-tuned using SimpleRL, indicating an optimization for improved instruction following and response quality through reinforcement learning techniques. It is designed for general-purpose language understanding and generation tasks, leveraging its substantial parameter count and RL fine-tuning for enhanced performance.

Loading preview...