yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step3584
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Apr 6, 2026Architecture:Transformer Loading

The yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step3584 is a 4 billion parameter language model with a 32768 token context length. This model is a fine-tuned variant, likely based on the Qwen architecture, optimized through Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO). Its specific differentiators and primary use cases are not detailed in the provided model card, which indicates 'More Information Needed' for most sections.

Loading preview...