Erik22TY/Nebulos-Distill-Qwen3-0.6B

Warm
Public
0.8B
BF16
40960
Hugging Face
Overview

Nebulos-Distill-Qwen3-0.6B Overview

Nebulos-Distill-Qwen3-0.6B is a lightweight 0.6 billion parameter causal language model developed by Erik22TY. It was distilled from the Qwen 3 architecture and fine-tuned using LoRA on consumer-grade hardware, demonstrating efficient AI training on a budget. This model is specifically designed for high-speed local inference, prioritizing logical consistency and reasoning capabilities while maintaining minimal VRAM requirements.

Key Capabilities

  • Efficient Reasoning: Fine-tuned for step-by-step logic, excelling in tasks like math word problems and logic puzzles.
  • Low VRAM Footprint: Designed for minimal VRAM usage, making it suitable for local deployment on consumer-grade hardware.
  • High-Speed Local Inference: Optimized for rapid processing in environments with limited resources.
  • Apache 2.0 License: Provides flexibility for various applications.

Good For

  • Mobile Deployment: Ideal for integration into mobile applications due to its compact size and efficiency.
  • Low-Spec Desktop Environments: Performs well on desktops with limited VRAM, such as those with a GTX 1050.
  • Reasoning-Heavy Tasks: Best suited for applications requiring concise text generation and logical problem-solving.
  • Budget-Conscious Development: A testament to high-quality AI training achievable on consumer hardware.