danieldk/Qwen2.5-1.5B-Instruct-w8a8-int-dynamic-weight
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kLicense:apache-2.0Architecture:Transformer Open Weights Warm

The danieldk/Qwen2.5-1.5B-Instruct-w8a8-int-dynamic-weight model is an instruction-tuned Qwen2.5 causal language model with 1.54 billion parameters, featuring dynamic weight and input quantization for optimized performance. Developed by Qwen, this model supports a full context length of 32,768 tokens and generation up to 8,192 tokens. It demonstrates improved capabilities in coding, mathematics, instruction following, and structured data understanding, including JSON generation. This model is particularly suited for applications requiring efficient, quantized inference of a versatile instruction-tuned LLM.

Loading preview...