Hyeongwon/P2-split2_prob_Qwen3-8B-Base_0325-02-lr1e-5
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Mar 25, 2026Architecture:Transformer Cold
Hyeongwon/P2-split2_prob_Qwen3-8B-Base_0325-02-lr1e-5 is an 8 billion parameter language model fine-tuned from ChuGyouk/Qwen3-8B-Base using Supervised Fine-Tuning (SFT) with the TRL library. This model is designed for general text generation tasks, leveraging its 32768 token context length for processing longer inputs. Its training methodology focuses on adapting the base Qwen3-8B architecture for specific probabilistic text generation applications.
Loading preview...