g4me/QwenRolina3-Base-LR1e5-wsd-b32g2gc8-order-domain-3ep-mix
TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Mar 10, 2026Architecture:Transformer Gated Cold

The g4me/QwenRolina3-Base-LR1e5-wsd-b32g2gc8-order-domain-3ep-mix is a 2 billion parameter language model developed by g4me, fine-tuned from Qwen/Qwen3-1.7B-Base. This model was trained using TRL and features a 32768 token context length. It is optimized for general text generation tasks, building upon the foundational capabilities of the Qwen3 architecture.

Loading preview...