Hyeongwon/PS_prob_Qwen3-4B-Base_0322-01

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Mar 21, 2026Architecture:Transformer Warm

Hyeongwon/PS_prob_Qwen3-4B-Base_0322-01 is a 4 billion parameter language model developed by Hyeongwon, fine-tuned from the Qwen3-4B-Base architecture. Trained using TRL with SFT, this model is designed for general text generation tasks with a 32768 token context length. It specializes in generating responses to open-ended questions and conversational prompts, building upon its base model's capabilities.

Loading preview...

Model Overview

Hyeongwon/PS_prob_Qwen3-4B-Base_0322-01 is a 4 billion parameter language model, fine-tuned by Hyeongwon from the base Hyeongwon/Qwen3-4B-Base model. This iteration has been specifically trained using the TRL (Transformer Reinforcement Learning) library with Supervised Fine-Tuning (SFT) methods.

Key Capabilities

  • General Text Generation: Excels at producing coherent and contextually relevant text based on given prompts.
  • Conversational AI: Optimized for generating responses to open-ended questions and engaging in dialogue, as demonstrated by its quick start example.
  • Base Model Enhancement: Builds upon the foundational capabilities of the Qwen3-4B-Base architecture, likely improving its performance in specific generation tasks through fine-tuning.

Training Details

The model underwent a Supervised Fine-Tuning (SFT) process utilizing the TRL library. The training environment included specific versions of key frameworks:

  • TRL: 0.25.1
  • Transformers: 4.57.3
  • Pytorch: 2.6.0
  • Datasets: 3.6.0
  • Tokenizers: 0.22.2

Use Cases

This model is suitable for applications requiring:

  • Interactive Chatbots: Generating human-like responses in conversational agents.
  • Content Creation: Assisting with drafting text for various purposes, especially those involving creative or open-ended prompts.
  • Question Answering: Providing detailed and imaginative answers to complex questions.