hamid1232/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-hoarse_meek_badger
TEXT GENERATIONConcurrency Cost:1Model Size:0.5BQuant:BF16Ctx Length:32kPublished:Apr 20, 2025Architecture:Transformer Warm
hamid1232/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-hoarse_meek_badger is a 0.5 billion parameter instruction-tuned causal language model, fine-tuned from Gensyn/Qwen2.5-0.5B-Instruct. This model was trained using the GRPO method, which is designed to enhance mathematical reasoning capabilities. Its primary use case is for tasks requiring improved mathematical reasoning, leveraging the techniques from DeepSeekMath.
Loading preview...