yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step256
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Apr 6, 2026Architecture:Transformer Loading

The yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step256 model is a 4 billion parameter language model, likely based on the Qwen architecture, that has undergone supervised fine-tuning (SFT) and Direct Preference Optimization (DPO). It is designed for general language tasks, leveraging a 32768 token context length for processing extensive inputs. This model is intended for applications requiring robust language understanding and generation capabilities.

Loading preview...