Hyeongwon/P2-split2_prob_Qwen3-8B-Base_0325-03-bs128
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Mar 25, 2026Architecture:Transformer Cold

Hyeongwon/P2-split2_prob_Qwen3-8B-Base_0325-03-bs128 is an 8 billion parameter language model, fine-tuned by Hyeongwon from the ChuGyouk/Qwen3-8B-Base architecture. This model was trained using Supervised Fine-Tuning (SFT) with the TRL framework, building upon a 32768 token context length. It is designed for general text generation tasks, leveraging its base Qwen3 architecture for broad applicability.

Loading preview...