Hyeongwon/P2-split2_prob_Qwen3-4B-Base_0312-01

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Mar 13, 2026Architecture:Transformer Warm

Hyeongwon/P2-split2_prob_Qwen3-4B-Base_0312-01 is a 4 billion parameter causal language model, fine-tuned from Hyeongwon/Qwen3-4B-Base using the TRL framework. This model was trained with Supervised Fine-Tuning (SFT) and supports a 32768 token context length. It is designed for general text generation tasks, building upon the base Qwen3 architecture.

Loading preview...

Model Overview

Hyeongwon/P2-split2_prob_Qwen3-4B-Base_0312-01 is a 4 billion parameter language model, derived from the Hyeongwon/Qwen3-4B-Base architecture. This model has undergone Supervised Fine-Tuning (SFT) using the TRL library, a framework for Transformer Reinforcement Learning.

Key Characteristics

  • Base Model: Fine-tuned from Hyeongwon/Qwen3-4B-Base.
  • Training Method: Utilizes Supervised Fine-Tuning (SFT).
  • Context Length: Supports a substantial context window of 32768 tokens.
  • Frameworks: Developed with TRL (0.25.1), Transformers (4.57.3), Pytorch (2.6.0), Datasets (3.6.0), and Tokenizers (0.22.2).

Intended Use Cases

This model is suitable for various text generation tasks, leveraging its SFT training to produce coherent and contextually relevant outputs. Developers can integrate it into applications requiring conversational AI, content creation, or other natural language processing functionalities. A quick start example is provided for text generation using the Hugging Face transformers pipeline.