dnotitia/Smoothie-Qwen3-32B

Hugging Face
TEXT GENERATIONConcurrency Cost:2Model Size:32BQuant:FP8Ctx Length:32kLicense:apache-2.0Architecture:Transformer0.0K Open Weights Warm

dnotitia/Smoothie-Qwen3-32B is a 32 billion parameter language model based on the Qwen3 architecture, featuring a 32K context length. It incorporates a lightweight adjustment tool that smooths token probabilities to enhance balanced multilingual generation, particularly for East Asian languages. This model is optimized for improving the quality and balance of text generation across various languages by modifying token probabilities.

Loading preview...

Smoothie Qwen3-32B Overview

dnotitia/Smoothie-Qwen3-32B is a 32 billion parameter model built upon the Qwen/Qwen3-32B base. Its core innovation lies in a lightweight adjustment tool, "Smoothie Qwen," designed to smooth token probabilities. This process aims to enhance balanced multilingual generation capabilities, making the model more effective for diverse linguistic tasks.

Key Capabilities & Features

  • Enhanced Multilingual Generation: The model applies a smoothing mechanism to token probabilities, specifically targeting improved balance in multilingual output.
  • Qwen3 Architecture: Leverages the robust Qwen3-32B as its foundational model, providing a strong base for language understanding and generation.
  • Configurable Smoothing: The adjustment tool uses specific configurations, including a minimum scale factor, smoothness parameter, sample size, window size, and N-gram weights, to fine-tune the probability distribution.
  • Targeted Token Modification: The process involves modifying a significant number of tokens (27,564 modified tokens out of 26,153 target tokens) across various Unicode ranges, indicating a focus on East Asian character sets.

Good For

  • Applications requiring balanced and high-quality multilingual text generation, especially involving languages within the specified Unicode ranges (e.g., Chinese, Japanese, Korean).
  • Developers looking for a Qwen3-based model with improved control over token probabilities for more nuanced linguistic outputs.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p