lyf07/Qwen3-8B-WALAR
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Mar 13, 2026License:mitArchitecture:Transformer Open Weights Cold

lyf07/Qwen3-8B-WALAR is an 8 billion parameter language model based on the Qwen3 architecture, fine-tuned using the WALAR reinforcement training method. This model specializes in enhancing translation capabilities for massive low-resource languages by integrating quality estimation, word alignment, and language alignment into its reward function. It demonstrates significant improvements in translation quality and language consistency across over 1400 language directions, outperforming prior multilingual models of similar size.

Loading preview...