Thrillcrazyer/Qwen-2.5-1.5B_TAC_Teacher_LLAMA70
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Mar 27, 2026Architecture:Transformer Warm

Thrillcrazyer/Qwen-2.5-1.5B_TAC_Teacher_LLAMA70 is a 1.5 billion parameter language model fine-tuned from Qwen/Qwen2.5-1.5B-Instruct. Developed by Thrillcrazyer, this model specializes in mathematical reasoning, having been trained on the DeepMath-103k dataset using the GRPO method. With a context length of 32768 tokens, it is optimized for tasks requiring advanced mathematical problem-solving capabilities.

Loading preview...