continuum-ai/qwen2.5-coder-7b-compacted
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Apr 8, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The continuum-ai/qwen2.5-coder-7b-compacted is a 7.6 billion parameter causal language model from the Qwen2.5-Coder family, developed by continuum-ai. This model has undergone 12% head pruning and KL-distillation compensation via LoRA, resulting in a compact version that maintains strong performance on coding tasks. It is specifically optimized for English-language Python code completion, achieving a HumanEval score of 61.0 with a 32768 token context length.

Loading preview...