yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step3840
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Apr 6, 2026Architecture:Transformer Cold

The yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step3840 is a 4 billion parameter language model developed by yunjae-won, fine-tuned using Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO). This model is based on the Qwen architecture and is designed for general language generation tasks. Its specific differentiators and primary use cases require further information, as the provided model card indicates 'More Information Needed' for most details.

Loading preview...