Hyeongwon/P2-split2_prob_Qwen3-8B-Base_0325-06-bs256-epoch10
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Mar 28, 2026Architecture:Transformer Warm

Hyeongwon/P2-split2_prob_Qwen3-8B-Base_0325-06-bs256-epoch10 is an 8 billion parameter causal language model, fine-tuned from ChuGyouk/Qwen3-8B-Base using Supervised Fine-Tuning (SFT) with a 32K context length. This model is optimized for general text generation tasks, building upon the Qwen3 architecture. It is suitable for applications requiring robust language understanding and generation capabilities.

Loading preview...