Hyeongwon/P2-split2_prob_Qwen3-14B-Base_0405_1e-5
TEXT GENERATIONConcurrency Cost:1Model Size:14BQuant:FP8Ctx Length:32kPublished:Apr 6, 2026Architecture:Transformer Cold

The Hyeongwon/P2-split2_prob_Qwen3-14B-Base_0405_1e-5 model is a 14 billion parameter language model, fine-tuned from Qwen/Qwen3-14B-Base. This model was trained using Supervised Fine-Tuning (SFT) with the TRL framework. It is designed for general text generation tasks, leveraging its base Qwen3 architecture and 32768 token context length. The fine-tuning process aims to enhance its performance for specific probabilistic text generation applications.

Loading preview...