sinamny/sft_merged_model

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Mar 23, 2026License:apache-2.0Architecture:Transformer Open Weights Warm

The sinamny/sft_merged_model is a 4 billion parameter Qwen3-based causal language model developed by sinamny, fine-tuned from unsloth/qwen3-4b-instruct-2507-unsloth-bnb-4bit. This model leverages Unsloth for accelerated training, offering a 32768-token context length. It is optimized for efficient performance, making it suitable for applications requiring a balance of capability and speed.

Loading preview...

Model Overview

The sinamny/sft_merged_model is a 4 billion parameter language model developed by sinamny. It is built upon the Qwen3 architecture and was fine-tuned from the unsloth/qwen3-4b-instruct-2507-unsloth-bnb-4bit base model. A key characteristic of this model is its training methodology, which utilized Unsloth to achieve a 2x faster training process.

Key Characteristics

  • Architecture: Qwen3-based, indicating strong general language understanding and generation capabilities.
  • Parameter Count: 4 billion parameters, offering a good balance between performance and computational efficiency.
  • Context Length: Supports a substantial context window of 32768 tokens, enabling processing of longer inputs and maintaining coherence over extended conversations or documents.
  • Training Efficiency: Benefited from Unsloth's optimization, resulting in significantly faster fine-tuning.

Use Cases

This model is well-suited for applications where a capable yet efficient language model is required. Its 4 billion parameters and large context window make it versatile for tasks such as:

  • Instruction-following and conversational AI.
  • Text generation and summarization.
  • Applications benefiting from processing longer textual inputs.

It operates under an Apache-2.0 license, providing flexibility for various deployments.