distil-labs/distil-qwen3-0.6b-SHELLper

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:0.8BQuant:BF16Ctx Length:32kPublished:Jan 23, 2026License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

The distil-labs/distil-qwen3-0.6b-SHELLper model is a 0.6 billion parameter Qwen3-based model fine-tuned by Distil Labs for multi-turn bash function calling. It achieves 100% tool-call accuracy on its test set, including 5-turn conversations, by distilling knowledge from a Qwen3-235B teacher model. With a 40,960 token context length, this compact model is optimized for local execution and excels at translating natural language into bash commands for file system interaction and automation.

Loading preview...

Distil-Qwen3-0.6B-SHELLper: Compact Bash Function Calling

This model, developed by Distil Labs, is a highly specialized 0.6 billion parameter variant of the Qwen3 architecture. It has been meticulously fine-tuned for multi-turn bash function calling, demonstrating exceptional accuracy in converting natural language requests into executable bash commands.

Key Capabilities & Differentiators

  • Exceptional Accuracy: Achieves 100% tool-call accuracy on its internal test set, even across complex 5-turn conversations, a significant improvement over its base Qwen3-0.6B model (which scored 42.22% on 5-turn accuracy).
  • Knowledge Distillation: Performance is boosted through knowledge distillation from a much larger Qwen3-235B teacher model, allowing a small model to achieve high-quality results.
  • Compact & Efficient: At only 0.6 billion parameters, it is designed to run efficiently on local machines, making it suitable for edge deployments and privacy-preserving applications.
  • Extensive Bash Command Support: Supports 20 common bash commands, including ls, cd, cp, rm, grep, and find, enabling a wide range of command-line interactions.
  • Multi-turn Conversation: Optimized for sequential interactions, understanding context across multiple user prompts to generate appropriate tool calls.
  • Generous Context Window: Features a 40,960 token context length, allowing for longer and more complex conversational histories.

Ideal Use Cases

  • Natural Language Interfaces: Building intuitive interfaces for file systems and command-line tools.
  • Command-Line Automation: Automating repetitive tasks through natural language instructions.
  • Developer Productivity: Enhancing developer workflows with AI-powered command assistance.
  • Educational Tools: Assisting in learning and practicing bash commands.
  • Local AI Assistants: Deploying privacy-focused AI assistants that can interact with the local environment.