yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step512
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Apr 6, 2026Architecture:Transformer Cold

The yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step512 is a 4 billion parameter language model developed by yunjae-won. This model is a fine-tuned version, likely optimized for specific instruction-following or dialogue tasks through Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO). With a context length of 32768 tokens, it is suitable for applications requiring processing of moderately long inputs and generating coherent, contextually relevant responses.

Loading preview...