alamios/DeepSeek-R1-DRAFT-Qwen2.5-Coder-0.5B
TEXT GENERATIONConcurrency Cost:1Model Size:0.5BQuant:BF16Ctx Length:32kPublished:Feb 6, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Warm
The alamios/DeepSeek-R1-DRAFT-Qwen2.5-Coder-0.5B is a 0.5 billion parameter draft model, based on the Qwen2.5 architecture, specifically trained on code outputs from deepseek-ai/DeepSeek-R1-Distill-Qwen-32B. It is designed for speculative decoding to accelerate generation for the 32B DeepSeek model. This model excels at speeding up code generation on consumer GPUs like the 3090/4090 without compromising context length or quality, supporting a context length of 131072 tokens.
Loading preview...