Gensyn/Qwen2.5-0.5B-Instruct
TEXT GENERATIONConcurrency Cost:1Model Size:0.5BQuant:BF16Ctx Length:32kPublished:Mar 28, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

Gensyn/Qwen2.5-0.5B-Instruct is an unmodified 0.49 billion parameter instruction-tuned causal language model from the Qwen2.5 family, featuring a 32,768 token context length. Developed by Qwen, this model utilizes a transformer architecture with RoPE, SwiGLU, and RMSNorm. It is specifically intended for local fine-tuning via peer-to-peer reinforcement learning within the Gensyn RL Swarm system. Its primary use case is as a base model for distributed RL training, after which it can be deployed in general workflows.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p