Zhihu-ai/Zhi-Create-Qwen3-32B

Warm
Public
32B
FP8
32768
Jul 11, 2025
License: apache-2.0
Hugging Face
Overview

Model Overview

Zhihu-ai's Zhi-Create-Qwen3-32B is a 32 billion parameter language model, fine-tuned from the Qwen3-32B base model with a strong emphasis on enhancing creative writing performance. It leverages a unique training methodology combining supervised fine-tuning (SFT) with curriculum learning and Direct Preference Optimization (DPO) using the RAFT (Reward-Ranked Fine-Tuning) method.

Key Capabilities & Features

  • Superior Creative Writing: Achieves a WritingBench score of 82.08, a notable improvement over the base Qwen3-32B's 78.97, demonstrating enhanced performance across academic, finance, politics, literature, education, and advertising domains.
  • Maintained General Abilities: Through fine-grained data mixture experiments, the model retains strong general knowledge, mathematics, and reasoning capabilities without significant decline.
  • Robust Training Data: Trained on a meticulously balanced corpus including filtered open-source datasets, synthesized chain-of-thought reasoning, and high-quality content from Zhihu, with creative writing data comprising approximately 23%.
  • Optimized for Deployment: Supports various hardware configurations, including 80GB GPUs, single H20/A800/H800 units, and offers quantized versions (FP8, Q4_K_M) for more accessible deployment on setups like dual RTX 4090 or single RTX 4090.

Recommended Use Cases

  • Creative Content Generation: Ideal for generating diverse creative texts, stories, articles, and marketing copy.
  • Applications Requiring Nuanced Language: Suitable for tasks demanding high-quality, coherent, and contextually appropriate written output.
  • Maintaining General LLM Functionality: Can be used in scenarios where both creative writing prowess and general reasoning capabilities are required.