yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step2816
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Apr 6, 2026Architecture:Transformer Loading
The yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step2816 is a 4 billion parameter language model, likely based on the Qwen architecture, fine-tuned using Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO). This model is designed for general language generation tasks, leveraging its fine-tuning to produce more aligned and preferred outputs. Its 32768 token context length supports processing longer inputs for various applications.
Loading preview...