yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step2560
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Apr 6, 2026Architecture:Transformer Loading

The yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step2560 model is a 4 billion parameter language model based on the Qwen architecture. It has a context length of 32768 tokens. This model is a fine-tuned version, likely optimized for specific instruction-following or dialogue tasks through Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO). Its primary application is expected to be in generative AI tasks requiring nuanced responses.

Loading preview...