Hyeongwon/P2-split2_prob_Qwen3-8B-Base_0325-04-bs128-lr1e-5-epoch6
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Mar 26, 2026Architecture:Transformer Cold

Hyeongwon/P2-split2_prob_Qwen3-8B-Base_0325-04-bs128-lr1e-5-epoch6 is an 8 billion parameter language model, fine-tuned from ChuGyouk/Qwen3-8B-Base using the TRL library. This model was trained with Supervised Fine-Tuning (SFT) and supports a context length of 32768 tokens. It is designed for general text generation tasks, building upon the base capabilities of the Qwen3 architecture.

Loading preview...