tokyotech-llm/Qwen3-Swallow-32B-RL-v0.2
TEXT GENERATIONConcurrency Cost:2Model Size:32BQuant:FP8Ctx Length:32kPublished:Feb 1, 2026License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

Qwen3-Swallow-32B-RL-v0.2 is a 32 billion parameter large language model developed by tokyotech-llm, based on the Qwen3 architecture. This model is a bilingual Japanese-English model, specifically enhanced through Continual Pre-Training, Supervised Fine-Tuning, and Reinforcement Learning with Verifiable Rewards. It excels in Japanese language proficiency, Japanese-English translation, and maintains or improves performance on complex math and coding tasks, making it suitable for applications requiring strong reasoning in these domains.

Loading preview...