Hyeongwon/PS_prob_seed43_Qwen3-4B-Base_0322-01
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Mar 22, 2026Architecture:Transformer Warm

Hyeongwon/PS_prob_seed43_Qwen3-4B-Base_0322-01 is a 4 billion parameter language model fine-tuned by Hyeongwon, based on the Qwen3-4B-Base architecture. This model was trained using Supervised Fine-Tuning (SFT) with the TRL framework. It is designed for general text generation tasks, building upon its base model's capabilities. The model supports a context length of 32768 tokens.

Loading preview...