motobrew/qwen-dpo-v13
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Mar 1, 2026License:apache-2.0Architecture:Transformer Open Weights Warm

The motobrew/qwen-dpo-v13 is a 4 billion parameter causal language model, fine-tuned from motobrew/qwen-dpo-v3 using Direct Preference Optimization (DPO) via the Unsloth library. This model is specifically optimized to improve reasoning capabilities through Chain-of-Thought and enhance structured response quality. It is designed for tasks requiring aligned and high-quality outputs based on preference datasets.

Loading preview...