yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step2048
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Apr 6, 2026Architecture:Transformer Loading

The yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step2048 is a 4 billion parameter language model based on the Qwen architecture. This model has been fine-tuned using Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO) techniques. It is designed for general language generation tasks, leveraging its 32768 token context length for comprehensive understanding and response generation. The specific differentiators and primary use cases are not detailed in the provided model card.

Loading preview...