Hyeongwon/P2-split2_prob_Qwen3-8B-Base_0325-05-bs128-epoch6
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Mar 26, 2026Architecture:Transformer Cold

Hyeongwon/P2-split2_prob_Qwen3-8B-Base_0325-05-bs128-epoch6 is an 8 billion parameter language model, fine-tuned from ChuGyouk/Qwen3-8B-Base using the TRL framework. This model was trained with Supervised Fine-Tuning (SFT) and is designed for general text generation tasks. It leverages a 32768 token context length, making it suitable for applications requiring processing of longer inputs.

Loading preview...