yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step4608
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Apr 6, 2026Architecture:Transformer Loading
The yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step4608 model is a 4 billion parameter language model, likely based on the Qwen architecture, fine-tuned using Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO). With a substantial 32768-token context length, this model is designed for general language understanding and generation tasks. Its specific differentiators and primary use cases are not detailed in the provided information, suggesting it's a foundational or general-purpose model within its parameter class.
Loading preview...