g4me/QwenRolina3-Base-LR1e5-b32g2gc8-order-domain-fp8

TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Mar 14, 2026Architecture:Transformer Gated Cold

QwenRolina3-Base-LR1e5-b32g2gc8-order-domain-fp8 is a fine-tuned language model based on Qwen/Qwen3-1.7B-Base, developed by g4me. This model has been specifically trained using the TRL framework for supervised fine-tuning (SFT). It is designed for general text generation tasks, leveraging its Qwen3 architecture for conversational and question-answering applications.

Loading preview...

Model Overview

This model, g4me/QwenRolina3-Base-LR1e5-b32g2gc8-order-domain-fp8, is a supervised fine-tuned (SFT) variant of the Qwen3-1.7B-Base architecture, developed by g4me. It leverages the robust capabilities of the Qwen3 family, a causal language model known for its strong performance across various language understanding and generation tasks.

Key Capabilities

  • General Text Generation: Excels at generating coherent and contextually relevant text based on given prompts.
  • Instruction Following: As an SFT model, it is optimized to follow instructions and respond to user queries effectively.
  • Conversational AI: Suitable for dialogue systems and interactive applications due to its fine-tuning process.

Training Details

The model was trained using the TRL (Transformers Reinforcement Learning) framework, specifically employing a supervised fine-tuning approach. This method refines the base model's ability to produce high-quality, instruction-aligned outputs. The training utilized specific versions of key libraries including TRL 0.29.0, Transformers 5.2.0, and PyTorch 2.8.0a0.

Use Cases

This model is well-suited for applications requiring:

  • Question Answering: Providing informative answers to a wide range of questions.
  • Content Creation: Generating creative or factual text for various purposes.
  • Chatbots and Virtual Assistants: Powering conversational interfaces that can understand and respond to user input.