tocchitocchi/Qwen3-Swallow-32B-RL-v0.2-MLX-fp16
TEXT GENERATIONConcurrency Cost:2Model Size:32BQuant:FP8Ctx Length:32kPublished:Mar 6, 2026License:apache-2.0Architecture:Transformer Open Weights Cold
The tocchitocchi/Qwen3-Swallow-32B-RL-v0.2-MLX-fp16 is a 32 billion parameter dense transformer model, converted to MLX format for Apple Silicon optimization. Developed by the Swallow Project (Institute of Science Tokyo and AIST) based on Qwen3, it is a bilingual Japanese-English model. This model excels in both Japanese and English tasks, maintaining strong capabilities in mathematics and coding through Continual Pre-Training, Supervised Fine-Tuning, and Reinforcement Learning.
Loading preview...