Qwen/QwQ-32B-Preview
Hugging Face
TEXT GENERATIONConcurrency Cost:2Model Size:32.8BQuant:FP8Ctx Length:32kPublished:Nov 27, 2024License:apache-2.0Architecture:Transformer1.7K Open Weights Warm

QwQ-32B-Preview is an experimental 32.5 billion parameter causal language model developed by the Qwen Team, featuring a transformer architecture with RoPE, SwiGLU, and RMSNorm. This model is specifically focused on advancing AI reasoning capabilities, particularly excelling in mathematical and coding tasks. It supports a substantial context length of 32,768 tokens, making it suitable for complex analytical problems.

Loading preview...

QwQ-32B-Preview: An Experimental Reasoning Model

QwQ-32B-Preview is an experimental 32.5 billion parameter causal language model from the Qwen Team, designed to push the boundaries of AI reasoning. Built on a transformer architecture incorporating RoPE, SwiGLU, and RMSNorm, this model demonstrates promising analytical abilities, especially in specialized domains.

Key Capabilities & Focus Areas

  • Advanced Reasoning: The model's primary focus is on enhancing AI reasoning capabilities.
  • Mathematical Proficiency: It shows particular strength and excels in mathematical problem-solving.
  • Coding Expertise: QwQ-32B-Preview also performs well in coding tasks.
  • Extended Context Window: Features a substantial context length of 32,768 tokens, allowing for processing of extensive inputs.

Important Considerations & Limitations

As a preview release, users should be aware of certain limitations:

  • Language Mixing: May exhibit unexpected language mixing or code-switching.
  • Recursive Reasoning: Can sometimes enter circular reasoning loops, leading to verbose but inconclusive responses.
  • Safety: Requires enhanced safety measures; caution is advised during deployment.
  • General Knowledge: While strong in math and coding, it has room for improvement in common sense reasoning and nuanced language understanding.

This model is ideal for research and applications requiring strong analytical and logical processing, particularly in technical fields.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p