gapcukbebemsi/Qwen2.5-0.5B-Instruct-Gensyn-Swarm-insectivorous_strong_raccoon
TEXT GENERATIONConcurrency Cost:1Model Size:0.5BQuant:BF16Ctx Length:32kArchitecture:Transformer Warm

Qwen2.5-0.5B-Instruct-Gensyn-Swarm-insectivorous_strong_raccoon is a 0.5 billion parameter instruction-tuned language model, fine-tuned from Gensyn/Qwen2.5-0.5B-Instruct. This model was trained using the GRPO method, as introduced in the DeepSeekMath paper, which focuses on enhancing mathematical reasoning. With a context length of 131072 tokens, it is optimized for tasks requiring robust reasoning capabilities, particularly in mathematical contexts.

Loading preview...