yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step768
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Apr 6, 2026Architecture:Transformer Loading

The yunjae-won/mpq3_qwen4bi_sft_dpo_beta1e-1_step768 model is a 4 billion parameter language model, likely based on the Qwen architecture, fine-tuned using Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO). It features a substantial 32768 token context length, making it suitable for processing extensive inputs and generating coherent, long-form text. This model is designed for general language understanding and generation tasks, leveraging its fine-tuned nature for improved performance in conversational and instruction-following applications.

Loading preview...