yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step4096
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Apr 6, 2026Architecture:Transformer Cold

The yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step4096 model is a 4 billion parameter language model with a 32768 token context length. This model is a fine-tuned variant, likely based on the Qwen architecture, and has undergone Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO). Due to limited information in its model card, specific differentiators or primary use cases beyond general language generation are not detailed.

Loading preview...