taki555/Qwen3-1.7B-Art
TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Feb 27, 2026License:apache-2.0Architecture:Transformer Open Weights Cold
Art-Qwen3-1.7B is a 1.7 billion parameter language model developed by Taiqiang Wu, Zenan Xu, Bo Zhou, and Ngai Wong, based on the Qwen3-1.7B architecture with a 32768 token context length. It is specifically optimized for efficient reasoning, aiming to generate concise yet accurate thinking trajectories. This model utilizes Reinforcement Learning with specialized reward shaping on the DeepScaleR-Easy dataset, making it ideal for applications requiring high accuracy with reduced computational overhead in reasoning tasks.
Loading preview...