hkust-nlp/Qwen-2.5-7B-SimpleRL-Zoo
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Mar 24, 2025License:apache-2.0Architecture:Transformer Open Weights Warm
The hkust-nlp/Qwen-2.5-7B-SimpleRL-Zoo model is a 7.6 billion parameter language model from the Qwen 2.5 family, developed by hkust-nlp. This model is specifically fine-tuned using SimpleRL, a reinforcement learning approach, to enhance its performance and alignment. It features an extensive context length of 131,072 tokens, making it suitable for tasks requiring processing of very long inputs. Its primary differentiator lies in its SimpleRL optimization, aiming for improved instruction following and response quality.
Loading preview...