Ilia2003Mah/qwen2.5-1.5b-gsm8k-train-step0
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Mar 23, 2026Architecture:Transformer Warm

The Ilia2003Mah/qwen2.5-1.5b-gsm8k-train-step0 is a 1.5 billion parameter Qwen2.5-based language model, fine-tuned for specific tasks. With a context length of 32768 tokens, this model is designed for applications requiring efficient processing of long sequences. Its small parameter count makes it suitable for resource-constrained environments while maintaining competitive performance for its size.

Loading preview...