unsloth/Qwen2.5-7B-Instruct-1M
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Jan 27, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

Qwen2.5-7B-Instruct-1M is a 7.61 billion parameter instruction-tuned causal language model developed by Qwen, featuring a transformer architecture. This model is specifically optimized for ultra-long context tasks, supporting an impressive context length of up to 1 million tokens while maintaining strong performance on shorter tasks. It is designed for efficient deployment with custom vLLM frameworks that utilize sparse attention and length extrapolation for enhanced accuracy and speed.

Loading preview...