Hyeongwon/P2-split2_prob_Qwen3-4B-Base_0312-01
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Mar 13, 2026Architecture:Transformer Warm

Hyeongwon/P2-split2_prob_Qwen3-4B-Base_0312-01 is a 4 billion parameter causal language model, fine-tuned from Hyeongwon/Qwen3-4B-Base using the TRL framework. This model was trained with Supervised Fine-Tuning (SFT) and supports a 32768 token context length. It is designed for general text generation tasks, building upon the base Qwen3 architecture.

Loading preview...