Hyeongwon/P2-split2_bs512_epoch10_2e-5_prob_Qwen3-4B-Base_0320-01
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Mar 20, 2026Architecture:Transformer Warm

Hyeongwon/P2-split2_bs512_epoch10_2e-5_prob_Qwen3-4B-Base_0320-01 is a 4 billion parameter language model fine-tuned by Hyeongwon from the Qwen3-4B-Base architecture. This model was trained using Supervised Fine-Tuning (SFT) with the TRL framework. It is designed for text generation tasks, demonstrating capabilities in responding to open-ended questions.

Loading preview...