zzaen/toolcalling-merged-demo

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Mar 26, 2026License:apache-2.0Architecture:Transformer Open Weights Warm

The zzaen/toolcalling-merged-demo is a 2 billion parameter Qwen3-based causal language model developed by zzaen, finetuned from unsloth/Qwen3-1.7B-unsloth-bnb-4bit. It was trained using Unsloth and Huggingface's TRL library, enabling faster finetuning. This model is designed for general language generation tasks, leveraging its Qwen3 architecture and 32768 token context length.

Loading preview...

Model Overview

The zzaen/toolcalling-merged-demo is a 2 billion parameter language model based on the Qwen3 architecture, developed by zzaen. It was finetuned from the unsloth/Qwen3-1.7B-unsloth-bnb-4bit model, indicating an optimization for efficient training and deployment.

Key Characteristics

  • Architecture: Qwen3-based, providing robust language understanding and generation capabilities.
  • Parameter Count: 2 billion parameters, offering a balance between performance and computational efficiency.
  • Context Length: Supports a substantial context window of 32768 tokens, allowing for processing longer inputs and maintaining coherence over extended conversations or documents.
  • Training Efficiency: Finetuned using Unsloth and Huggingface's TRL library, which facilitated a 2x faster training process.

Potential Use Cases

This model is suitable for a variety of natural language processing tasks where a capable and efficiently trained Qwen3-based model is beneficial. Its substantial context length makes it particularly useful for applications requiring an understanding of longer texts or complex conversational flows.