yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step7168
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Apr 6, 2026Architecture:Transformer Cold

The yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step7168 is a 4 billion parameter language model developed by yunjae-won. This model is a fine-tuned version, indicated by 'sft_dpo', suggesting optimization through supervised fine-tuning and direct preference optimization. With a context length of 32768 tokens, it is designed for tasks requiring extensive contextual understanding. Its specific differentiators and primary use cases are not detailed in the provided information.

Loading preview...