Hyeongwon/P9-split1_3times_prob_Qwen3-4B-Base_0319-02
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Mar 19, 2026Architecture:Transformer Warm
Hyeongwon/P9-split1_3times_prob_Qwen3-4B-Base_0319-02 is a 4 billion parameter language model fine-tuned from Hyeongwon/Qwen3-4B-Base using the TRL framework. This model has a context length of 32768 tokens and was trained with Supervised Fine-Tuning (SFT). It is designed for general text generation tasks, building upon its base Qwen3 architecture.
Loading preview...