hyunw3/qwen-2.5-0.5b-r1-countdown_lr5e-6
TEXT GENERATIONConcurrency Cost:1Model Size:0.5BQuant:BF16Ctx Length:32kArchitecture:Transformer Warm

The hyunw3/qwen-2.5-0.5b-r1-countdown_lr5e-6 model is a fine-tuned version of the Qwen2.5-0.5B-Instruct architecture, featuring 0.5 billion parameters and a 32K context length. It was trained using the GRPO method, which is designed to enhance mathematical reasoning capabilities in language models. This model is primarily optimized for tasks requiring improved reasoning, particularly in mathematical contexts, making it suitable for specialized applications where numerical and logical understanding are critical.

Loading preview...