Hyeongwon/P2-split4_prob_Qwen3-4B-Base_0312-01
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Mar 13, 2026Architecture:Transformer Warm
The Hyeongwon/P2-split4_prob_Qwen3-4B-Base_0312-01 model is a 4 billion parameter language model developed by Hyeongwon, fine-tuned from the Qwen3-4B-Base architecture. This model has been trained using Supervised Fine-Tuning (SFT) with the TRL framework, focusing on enhancing its probabilistic generation capabilities. With a context length of 32768 tokens, it is designed for text generation tasks where nuanced and contextually relevant responses are crucial.
Loading preview...