yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step3072
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Apr 6, 2026Architecture:Transformer Loading

The yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step3072 is a 4 billion parameter language model, likely based on the Qwen architecture, fine-tuned using Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO). This model is designed for general language understanding and generation tasks, leveraging its 32768 token context length for processing extensive inputs. Its specific optimizations through DPO suggest a focus on generating high-quality, preferred responses.

Loading preview...