BoyBarley/BoyBarley-sparky

TEXT GENERATIONConcurrency Cost:1Model Size:0.5BQuant:BF16Ctx Length:32kPublished:Apr 24, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

BoyBarley/BoyBarley-sparky is a 0.5 billion parameter Qwen2.5-based instruction-tuned causal language model developed by BoyBarley. Finetuned using Unsloth and Huggingface's TRL library, it offers efficient performance for its size. This model is optimized for tasks requiring a compact yet capable language model, leveraging its Qwen2.5 architecture for general-purpose instruction following.

Loading preview...

Model Overview

BoyBarley/BoyBarley-sparky is a compact 0.5 billion parameter instruction-tuned language model, developed by BoyBarley. It is built upon the Qwen2.5 architecture, specifically finetuned from unsloth/Qwen2.5-0.5B-Instruct-bnb-4bit.

Key Characteristics

  • Architecture: Based on the Qwen2.5 model family.
  • Parameter Count: 0.5 billion parameters, making it suitable for resource-constrained environments.
  • Context Length: Supports a context window of 32,768 tokens.
  • Training Efficiency: Finetuned using Unsloth and Huggingface's TRL library, indicating an emphasis on faster and more efficient training processes.
  • License: Released under the Apache-2.0 license.

Use Cases

This model is well-suited for applications where a smaller, efficient language model is beneficial, such as:

  • Instruction Following: Designed to respond to instructions effectively due to its instruction-tuned nature.
  • Edge Devices/Resource-Limited Environments: Its compact size makes it a candidate for deployment in scenarios with limited computational resources.
  • Rapid Prototyping: The efficient finetuning process suggests it can be quickly adapted or used for various tasks.