xwen-team/Xwen-72B-Chat

Hugging Face
TEXT GENERATIONConcurrency Cost:4Model Size:72.7BQuant:FP8Ctx Length:32kPublished:Jan 31, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

Xwen-72B-Chat is a 72 billion parameter open-source large language model developed by xwen-team, post-trained from Qwen2.5 models. It demonstrates top-tier chat performance among open-sourced models under 100B parameters, excelling in benchmarks like Arena-Hard-Auto, AlignBench, and MT-Bench. This model is optimized for general-purpose conversational AI applications, offering strong performance for various chat-based tasks.

Loading preview...

Xwen-72B-Chat: Top-Tier Open-Source Chat Performance

Xwen-72B-Chat is a 72 billion parameter large language model developed by xwen-team, built upon the Qwen2.5 foundation models. It is specifically post-trained to achieve leading chat performance among open-source models under 100 billion parameters.

Key Capabilities & Performance Highlights

  • Exceptional Chat Performance: Xwen-72B-Chat consistently ranks as the top-performing open-source model in its size class across major chat benchmarks.
  • Arena-Hard-Auto: Achieves a score of 86.1 (Top-1 among open-source models below 100B) in the No Style Control category and 72.4 (Top-1 among open-source models) in the Style Control category.
  • AlignBench-v1.1: Scores 7.57 (Top-1 among open-source models), evaluated using GPT-4o-0513 as the judge model.
  • MT-Bench: Attains 8.64 (Top-1 among open-source models), also judged by GPT-4o-0513.
  • Qwen2.5 Base: Leverages the robust architecture and pre-training of the Qwen2.5-72B model.

Ideal Use Cases

  • General-purpose conversational AI: Excels in various chat scenarios, from question answering to creative dialogue.
  • Applications requiring high-quality, nuanced responses: Its strong benchmark performance indicates a capability for generating coherent and contextually relevant outputs.
  • Developers seeking a powerful open-source alternative: Offers competitive performance against larger proprietary models, particularly in chat-oriented tasks.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p