yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step4352
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Apr 6, 2026Architecture:Transformer Loading

The yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step4352 is a 4 billion parameter language model with a 32768 token context length. This model is a fine-tuned version, likely based on the Qwen architecture, and has undergone Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO). Its specific differentiators and primary use cases are not detailed in the provided model card, indicating it may be a foundational or general-purpose model awaiting further specialization.

Loading preview...