g4me/QwenRolina3-Base-LR1e5-wsd-b32g2gc8-order-domain-2ep
TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Mar 3, 2026Architecture:Transformer0.0K Gated Cold

QwenRolina3-Base-LR1e5-wsd-b32g2gc8-order-domain-2ep is a 2 billion parameter language model fine-tuned from Qwen3-1.7B-Base, featuring a 32768 token context length. This model was trained using the TRL framework, specializing in general text generation tasks. It is designed for applications requiring a compact yet capable base model for further adaptation or direct use in conversational AI.

Loading preview...