raalr/Qwen2.5-1.5B-MiniLLM
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Apr 8, 2026Architecture:Transformer Loading

The raalr/Qwen2.5-1.5B-MiniLLM is a 1.5 billion parameter language model based on the Qwen2.5 architecture, developed by raalr. This model is designed for general language tasks, offering a compact size suitable for efficient deployment. Its 32768 token context length allows for processing substantial amounts of information, making it versatile for various applications where a smaller, capable model is preferred.

Loading preview...