Hyeongwon/PH_det_sft_FC_swap_labewise_data_oversampling_bf16_lr0.00002_context_12k-Qwen3-8B-Base
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Feb 26, 2026Architecture:Transformer Cold

Hyeongwon/PH_det_sft_FC_swap_labewise_data_oversampling_bf16_lr0.00002_context_12k-Qwen3-8B-Base is an 8 billion parameter language model fine-tuned from ChuGyouk/Qwen3-8B-Base. This model was trained using Supervised Fine-Tuning (SFT) with the TRL framework, focusing on specific data oversampling techniques. It is designed for general text generation tasks, leveraging its 32768 token context length for comprehensive understanding and response generation.

Loading preview...