THU-KEG/LongWriter-Zero-32B

Hugging Face
TEXT GENERATIONConcurrency Cost:2Model Size:32.8BQuant:FP8Ctx Length:32kPublished:Jun 18, 2025License:apache-2.0Architecture:Transformer0.1K Open Weights Warm

LongWriter-Zero-32B is a 32 billion parameter large language model developed by THU-KEG, built upon Qwen 2.5-32B-Base. It is specifically designed for ultra-long text generation, capable of producing coherent passages exceeding 10,000 tokens. This model leverages 30 billion-token continual pretraining on long-form content and reinforcement learning with a composite reward function to enhance fluency, coherence, and length control. It excels at generating extended, structured text, matching or surpassing 100B-scale models in this specialized domain.

Loading preview...

LongWriter-Zero-32B: Ultra-Long Text Generation

LongWriter-Zero-32B, developed by THU-KEG, is a 32 billion parameter large language model built on Qwen 2.5-32B-Base. Its core innovation lies in its ability to generate coherent text passages exceeding 10,000 tokens, a capability achieved through a unique training methodology.

Key Capabilities & Training:

  • Ultra-Long Text Generation: Specifically engineered to produce extended, coherent outputs, surpassing the length capabilities of many larger models.
  • Reinforcement Learning (RL) Based: Utilizes Group Relative Policy Optimization (GRPO) with a sophisticated composite reward function, including length, writing quality, and format adherence, to guide generation.
  • Continual Pretraining: Underwent 30 billion-token continual pretraining on long-form books and technical reports to bolster its foundational writing skills.
  • Structured Output: Employs a dedicated prompting strategy that encourages explicit reflection before answering, improving structural planning and fine-grained length control, often using a <think>…</think><answer>…</answer> format.

Performance & Evaluation:

  • Benchmark Leader: Achieves the highest automatic WritingBench score among open models.
  • Human Evaluation: Demonstrates dominant win-rates in pairwise GPT-4.1 evaluations, confirming its superior quality in ultra-long-form generation.

Good for:

  • Generating very long articles, reports, or creative narratives.
  • Tasks requiring structured and coherent extended text outputs.
  • Applications where fine-grained control over output length and format is crucial.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p