aisingapore/Qwen-SEA-LION-v4-32B-IT

Hugging Face
TEXT GENERATIONConcurrency Cost:2Model Size:32BQuant:FP8Ctx Length:32kPublished:Oct 16, 2025Architecture:Transformer0.0K Warm

Qwen-SEA-LION-v4-32B-IT is a 32.7 billion parameter instruction-tuned decoder-only language model developed by AI Singapore, based on the Qwen3 architecture. It underwent continued pre-training on 100 billion tokens from the SEA-Pile v2 corpus, specifically optimized for seven Southeast Asian languages including Burmese, Indonesian, Malay, Filipino, Tamil, Thai, and Vietnamese. With a 32K context length, this model excels in multilingual understanding and generation for the SEA region, further enhanced by post-training on 8 million Q&A pairs.

Loading preview...

Qwen-SEA-LION-v4-32B-IT Overview

Qwen-SEA-LION-v4-32B-IT is an instruction-tuned large language model developed by AI Singapore, building upon the Qwen3-32B foundation. This model is specifically designed for the Southeast Asian (SEA) region, having undergone extensive continued pre-training on approximately 100 billion tokens from the SEA-Pile v2 corpus. This corpus includes data across seven key SEA languages: Burmese, Indonesian, Malay, Filipino, Tamil, Thai, and Vietnamese, alongside English.

Key Capabilities & Features

  • Multilingual Proficiency: Enhanced understanding and generation in Burmese, English, Indonesian, Khmer, Lao, Malay, Mandarin, Tagalog, Tamil, Thai, and Vietnamese.
  • Qwen3 Architecture: Inherits robust reasoning capabilities and support for over 100 languages from its Qwen3 base.
  • Extended Context Length: Features a native context length of 32,768 tokens.
  • Instruction Following: Post-trained on 8 million high-quality question-and-answer pairs for improved instruction adherence.
  • Evaluation: Assessed using SEA-HELM for general language tasks (QA, Sentiment, Translation, MMLU Lite) and SEA-IFEval/SEA-MTBench for instruction-following and multi-turn chat, with results available on the SEA-LION leaderboard.

Considerations for Use

Developers should note that the model has not been aligned for safety and requires custom safety fine-tuning. It may exhibit typical LLM limitations such as hallucination and occasional irrelevant content generation. The model supports a "thinking mode" which can be enabled via the enable_thinking=True parameter during generation.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p