Thrillcrazyer/Qwen-2.5-1.5B_TAC_Teacher_Qwen32B
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Mar 27, 2026Architecture:Transformer Warm
Thrillcrazyer/Qwen-2.5-1.5B_TAC_Teacher_Qwen32B is a 1.5 billion parameter Qwen2.5-based causal language model, fine-tuned by Thrillcrazyer. It is specifically optimized for mathematical reasoning tasks, leveraging the DeepMath-103k dataset and the GRPO training method. This model excels at complex mathematical problem-solving and logical deduction, making it suitable for applications requiring strong analytical capabilities. It features a context length of 32768 tokens, enhancing its ability to process extensive mathematical contexts.
Loading preview...