g4me/QwenRolina3-Base-LR1e5-WSD-b32g2gc8-order-domain-3ep
TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Mar 8, 2026Architecture:Transformer Gated Cold

g4me/QwenRolina3-Base-LR1e5-WSD-b32g2gc8-order-domain-3ep is a 2 billion parameter language model fine-tuned from Qwen/Qwen3-1.7B-Base. This model was trained using the TRL library with a context length of 32768 tokens. It is designed for general text generation tasks, leveraging its base architecture and fine-tuning process to produce coherent and contextually relevant responses.

Loading preview...