Thrillcrazyer/Qwen-2.5-1.5B_TAC_Teacher_Qwen14B
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Mar 27, 2026Architecture:Transformer Warm

Thrillcrazyer/Qwen-2.5-1.5B_TAC_Teacher_Qwen14B is a 1.5 billion parameter causal language model, fine-tuned from Qwen/Qwen2.5-1.5B-Instruct. It was specifically trained on the DeepMath-103k dataset using the GRPO method, which is designed to enhance mathematical reasoning capabilities. This model is optimized for tasks requiring advanced mathematical problem-solving and logical deduction, leveraging a 32768 token context length.

Loading preview...