modaserMoj/csc415-phase1-0.5b-fast
TEXT GENERATIONConcurrency Cost:1Model Size:0.5BQuant:BF16Ctx Length:32kPublished:Mar 8, 2026Architecture:Transformer Warm

modaserMoj/csc415-phase1-0.5b-fast is a 0.5 billion parameter language model fine-tuned from Qwen/Qwen2.5-0.5B, featuring a 32768 token context length. This model was trained using the TRL framework and the GRPO method, which is designed to enhance mathematical reasoning capabilities. It is optimized for tasks requiring robust mathematical problem-solving, making it suitable for applications in scientific computing and quantitative analysis.

Loading preview...