g4me/QwenRolina3-Base-LR1e5-b64g8-uff
TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Feb 16, 2026Architecture:Transformer Cold

g4me/QwenRolina3-Base-LR1e5-b64g8-uff is a 2 billion parameter language model fine-tuned from Qwen/Qwen3-1.7B-Base. It was trained using the TRL framework with a context length of 32768 tokens. This model is designed for general text generation tasks, leveraging its base architecture and fine-tuning for improved conversational capabilities.

Loading preview...