qqil/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-elusive_silky_tamarin
TEXT GENERATIONConcurrency Cost:1Model Size:0.5BQuant:BF16Ctx Length:32kPublished:Apr 7, 2025Architecture:Transformer Warm
qqil/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-elusive_silky_tamarin is a 0.5 billion parameter instruction-tuned causal language model, fine-tuned from unsloth/Qwen2.5-0.5B-Instruct. This model was trained using the GRPO method, as introduced in the DeepSeekMath paper, which focuses on enhancing mathematical reasoning. With a substantial context length of 131072 tokens, it is optimized for tasks requiring deep contextual understanding and mathematical problem-solving.
Loading preview...