Hyeongwon/PH_prob_sft_FC_swap_labewise_data_oversampling_bf16_lr0.00002_context_12k-Qwen3-8B-Base
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Feb 25, 2026Architecture:Transformer Cold

Hyeongwon/PH_prob_sft_FC_swap_labewise_data_oversampling_bf16_lr0.00002_context_12k-Qwen3-8B-Base is an 8 billion parameter language model, fine-tuned from ChuGyouk/Qwen3-8B-Base using Supervised Fine-Tuning (SFT) with TRL. This model is designed for general text generation tasks, leveraging a 32K context window. Its training methodology focuses on specific data oversampling and learning rate adjustments, indicating an optimization for robust performance in conversational AI or similar applications.

Loading preview...