yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step2304
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Apr 6, 2026Architecture:Transformer Loading
The yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step2304 is a 4 billion parameter language model, likely based on the Qwen architecture, fine-tuned using Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO). This model is designed for general language generation tasks, leveraging its 32768 token context length for processing extensive inputs. Its specific differentiators and primary use cases are not detailed in the provided information.
Loading preview...