amityco/matching-1.1-4b-sft

TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Apr 10, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The amityco/matching-1.1-4b-sft is a 4 billion parameter Qwen3-based causal language model developed by amityco, fine-tuned from unsloth/Qwen3-4B-Thinking-2507. This model was trained using Unsloth, enabling 2x faster fine-tuning. It is designed for general language tasks, leveraging its Qwen3 architecture for efficient processing.

Loading preview...

Model Overview

The amityco/matching-1.1-4b-sft is a 4 billion parameter language model developed by amityco. It is based on the Qwen3 architecture and was fine-tuned from the unsloth/Qwen3-4B-Thinking-2507 model. A key characteristic of this model's development is its utilization of Unsloth, which facilitated a 2x faster training process compared to standard methods.

Key Capabilities

  • Qwen3 Architecture: Leverages the robust capabilities of the Qwen3 model family.
  • Efficient Fine-tuning: Benefits from Unsloth's optimization for faster training.
  • General Language Understanding: Suitable for a broad range of natural language processing tasks.

Good For

  • Applications requiring a compact yet capable 4 billion parameter model.
  • Scenarios where efficient fine-tuning is a priority.
  • General text generation and understanding tasks.