hkust-nlp/Qwen-2.5-1.5B-SimpleRL-Zoo
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Mar 24, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

Qwen-2.5-1.5B-SimpleRL-Zoo is a 1.5 billion parameter language model developed by hkust-nlp, featuring a substantial context length of 131072 tokens. This model is part of the Qwen 2.5 family and is specifically fine-tuned using SimpleRL techniques. Its design focuses on leveraging reinforcement learning for enhanced performance, making it suitable for applications requiring robust language understanding and generation within a large context window.

Loading preview...