Hyeongwon/P2-split2_bs512_epoch5_5e-5_prob_Qwen3-4B-Base_0320-01
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Mar 20, 2026Architecture:Transformer Warm

The Hyeongwon/P2-split2_bs512_epoch5_5e-5_prob_Qwen3-4B-Base_0320-01 model is a 4 billion parameter language model fine-tuned from Hyeongwon/Qwen3-4B-Base using TRL. This model has been trained with Supervised Fine-Tuning (SFT) to enhance its conversational capabilities. It is designed for text generation tasks, particularly for responding to open-ended questions and engaging in dialogue.

Loading preview...