Hyeongwon/PS_bs256_Qwen3-4B-Base_0322-01
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Mar 22, 2026Architecture:Transformer Warm

Hyeongwon/PS_bs256_Qwen3-4B-Base_0322-01 is a 4 billion parameter language model, fine-tuned from Hyeongwon/Qwen3-4B-Base using the TRL framework. This model, with a 32768 token context length, has undergone Supervised Fine-Tuning (SFT) to enhance its conversational capabilities. It is designed for general text generation tasks, particularly excelling in responding to open-ended prompts and user queries.

Loading preview...