yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step1024
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Apr 6, 2026Architecture:Transformer Cold

The yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step1024 is a 4 billion parameter language model. This model is a fine-tuned version of a base model, likely from the Qwen family, optimized through Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO). Its specific differentiators and primary use cases are not detailed in the provided model card, indicating a need for further information regarding its training data, architecture, and evaluation results.

Loading preview...