Ilia2003Mah/qwen2.5-1.5b-gsm8k-train-step7000
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Mar 24, 2026Architecture:Transformer Warm

The Ilia2003Mah/qwen2.5-1.5b-gsm8k-train-step7000 is a 1.5 billion parameter language model, likely based on the Qwen2.5 architecture, fine-tuned for specific tasks. This model is designed for applications requiring a compact yet capable LLM, potentially optimized for mathematical reasoning given the 'gsm8k' in its name. Its relatively small size makes it suitable for efficient deployment and inference in resource-constrained environments.

Loading preview...