distil-labs/distil-qwen3-0.6b-SHELLper
TEXT GENERATIONConcurrency Cost:1Model Size:0.8BQuant:BF16Ctx Length:32kPublished:Jan 23, 2026License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

The distil-labs/distil-qwen3-0.6b-SHELLper model is a 0.6 billion parameter Qwen3-based model fine-tuned by Distil Labs for multi-turn bash function calling. It achieves 100% tool-call accuracy on its test set, including 5-turn conversations, by distilling knowledge from a Qwen3-235B teacher model. With a 40,960 token context length, this compact model is optimized for local execution and excels at translating natural language into bash commands for file system interaction and automation.

Loading preview...