sdhossain24/Qwen3-8B-CTRL

TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Feb 23, 2026Architecture:Transformer Cold

sdhossain24/Qwen3-8B-CTRL is an 8 billion parameter language model fine-tuned from Qwen/Qwen3-8B. This model has been specifically trained using SFT (Supervised Fine-Tuning) via the TRL library to enhance its conversational and instruction-following capabilities. With a 32768 token context length, it is optimized for generating coherent and contextually relevant text responses.

Loading preview...

Overview

sdhossain24/Qwen3-8B-CTRL is an 8 billion parameter language model derived from the Qwen/Qwen3-8B base model. It has undergone Supervised Fine-Tuning (SFT) using the TRL library, a framework designed for transformer reinforcement learning. This fine-tuning process aims to improve the model's ability to understand and generate responses based on given instructions and conversational prompts.

Key Capabilities

  • Instruction Following: Enhanced ability to process and respond to user instructions.
  • Text Generation: Capable of generating coherent and contextually appropriate text, as demonstrated by its use in a text-generation pipeline.
  • Conversational AI: Optimized for interactive dialogue, making it suitable for question-answering and chat-based applications.

Training Details

The model was trained using SFT, leveraging the TRL framework (version 0.22.1). The training procedure utilized specific versions of key libraries including Transformers (4.57.6), Pytorch (2.9.1+cu128), Datasets (4.5.0), and Tokenizers (0.22.2). Further details on the training run are available via Weights & Biases.

Good For

  • Developing conversational agents.
  • Generating creative or informative text based on prompts.
  • Applications requiring robust instruction-following from an 8B parameter model.