distil-labs/distil-qwen3-0.6b-text2sql

Warm
Public
0.8B
BF16
32768
Jan 14, 2026
License: apache-2.0
Hugging Face
Overview

Distil-Qwen3-0.6B-Text2SQL Overview

This model, developed by Distil Labs, is a compact 0.6 billion parameter Qwen3-based model specifically fine-tuned for converting natural language questions into SQL queries. It leverages knowledge distillation from the much larger DeepSeek-V3 model, allowing it to achieve high performance at a significantly smaller size.

Key Capabilities & Performance

  • Text-to-SQL Conversion: Excels at transforming natural language questions into accurate SQL queries.
  • High Accuracy: Achieves 74% on LLM-as-a-Judge accuracy, demonstrating a 2x improvement over its base Qwen3-0.6B model and approaching the performance of the 685B parameter DeepSeek-V3 teacher model.
  • Efficiency: With only 0.6 billion parameters and a 40,960 token context length, it is designed for lightweight and fast inference, suitable for local and edge deployments.
  • SQL Feature Support: Handles a range of SQL complexities including SELECT, WHERE, COUNT, SUM, AVG, MAX, MIN, JOIN, GROUP BY, HAVING, ORDER BY, LIMIT, subqueries, and UNION.
  • Training: Fine-tuned on approximately 10,000 synthetic examples generated by DeepSeek-V3, ensuring robust performance.

Ideal Use Cases

  • Natural Language Interfaces: Building conversational interfaces for databases.
  • SQL Query Assistance: Providing automated SQL generation and autocompletion.
  • Database Chatbots: Powering chatbots for business intelligence and data exploration.
  • Educational Tools: Assisting in learning and practicing SQL.
  • Edge Deployment: Suitable for applications requiring on-device or resource-constrained inference due to its compact size.