yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step5120
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Apr 6, 2026Architecture:Transformer Loading

The yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step5120 is a 4 billion parameter language model with a 32768 token context length. This model is a fine-tuned version, likely based on the Qwen architecture, and has undergone Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO). Its specific differentiators and primary use cases are not detailed in the provided model card, which indicates that more information is needed regarding its development and intended applications.

Loading preview...