Hyeongwon/P9-split2_prob_Qwen3-4B-Base_0322-01
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Mar 21, 2026Architecture:Transformer Warm

Hyeongwon/P9-split2_prob_Qwen3-4B-Base_0322-01 is a 4 billion parameter language model fine-tuned from Hyeongwon/Qwen3-4B-Base using the TRL framework. This model is specifically trained with Supervised Fine-Tuning (SFT) to enhance its probabilistic reasoning capabilities. With a context length of 32768 tokens, it is designed for text generation tasks requiring nuanced understanding and response generation.

Loading preview...