g4me/QwenRolina3-Base-LR1e5-b32g2gc8-order-domain-2ep

TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Mar 2, 2026Architecture:Transformer Gated Cold

QwenRolina3-Base-LR1e5-b32g2gc8-order-domain-2ep is a 1.7 billion parameter language model developed by Qwen, fine-tuned from Qwen3-1.7B-Base. This model leverages a 32,768 token context length and was trained using the TRL framework. It is designed for general text generation tasks, building upon the foundational capabilities of the Qwen3 architecture.

Loading preview...

Model Overview

This model, QwenRolina3-Base-LR1e5-b32g2gc8-order-domain-2ep, is a fine-tuned variant of the Qwen3-1.7B-Base model, developed by Qwen. It features approximately 1.7 billion parameters and supports a substantial 32,768 token context length, making it suitable for processing longer inputs and generating coherent, extended responses.

Key Capabilities

  • General Text Generation: Capable of generating human-like text based on given prompts.
  • Fine-tuned Performance: Benefits from additional training using the TRL (Transformers Reinforcement Learning) framework, which typically enhances model performance for specific tasks or domains.
  • Base Model Foundation: Built upon the robust Qwen3 architecture, providing a strong foundation for various natural language processing applications.

Training Details

The model was trained using Supervised Fine-Tuning (SFT), a common method for adapting pre-trained language models to specific tasks or improving their instruction-following abilities. The training process utilized TRL version 0.29.0 and Transformers version 5.2.0.

Good For

  • Developers seeking a moderately sized language model with a large context window.
  • Applications requiring general-purpose text generation and understanding.
  • Experimentation with fine-tuned Qwen3-based models.