g4me/QwenRolina3-Base-LR4e5-b64g8-order-domain-uff

TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Feb 22, 2026Architecture:Transformer Cold

g4me/QwenRolina3-Base-LR4e5-b64g8-order-domain-uff is a 2 billion parameter language model fine-tuned from Qwen/Qwen3-1.7B-Base. This model was trained using SFT with TRL and features a 32768 token context length. It is designed for general text generation tasks, building upon the foundational capabilities of the Qwen3 architecture.

Loading preview...

Model Overview

This model, g4me/QwenRolina3-Base-LR4e5-b64g8-order-domain-uff, is a fine-tuned variant of the Qwen3-1.7B-Base architecture, developed by g4me. It leverages a 2 billion parameter count and supports a substantial context length of 32768 tokens, making it suitable for processing longer inputs and generating coherent, extended responses.

Key Characteristics

  • Base Model: Fine-tuned from Qwen/Qwen3-1.7B-Base, inheriting its foundational language understanding and generation capabilities.
  • Training Method: Utilizes Supervised Fine-Tuning (SFT) with the TRL library, indicating a focus on specific task performance or instruction following.
  • Context Window: Features a 32768 token context length, allowing for detailed conversations and processing of extensive documents.

Potential Use Cases

  • General Text Generation: Capable of generating human-like text for various prompts, as demonstrated by the quick start example.
  • Conversational AI: Its large context window can support more complex and extended dialogue scenarios.
  • Content Creation: Suitable for drafting creative content, answering open-ended questions, and summarizing information where longer context is beneficial.