simeetnayan/odse-qwen
TEXT GENERATIONConcurrency Cost:1Model Size:0.5BQuant:BF16Ctx Length:32kPublished:Apr 3, 2026Architecture:Transformer Cold
The simeetnayan/odse-qwen model is a 0.5 billion parameter instruction-tuned causal language model, fine-tuned from Qwen/Qwen2.5-Coder-0.5B-Instruct. It leverages the TRL library and was trained using the GRPO method, which is associated with mathematical reasoning in large language models. This model is designed for general text generation tasks, building upon its coder-base with an enhanced training approach.
Loading preview...