maheshrawat18/Qwen3-4B-2507-sft-cv2

TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Apr 23, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The maheshrawat18/Qwen3-4B-2507-sft-cv2 is a 4 billion parameter Qwen3 model, fine-tuned from unsloth/Qwen3-4B-Thinking-2507. Developed by maheshrawat18, this model was trained using Unsloth and Huggingface's TRL library, enabling faster training. It is designed for general language tasks, leveraging its Qwen3 architecture for efficient performance.

Loading preview...

Model Overview

The maheshrawat18/Qwen3-4B-2507-sft-cv2 is a 4 billion parameter language model based on the Qwen3 architecture. It was developed by maheshrawat18 and is fine-tuned from the unsloth/Qwen3-4B-Thinking-2507 base model.

Key Characteristics

  • Architecture: Qwen3
  • Parameter Count: 4 billion
  • Training Method: Fine-tuned using Unsloth and Huggingface's TRL library, which facilitated a 2x faster training process.
  • License: Apache-2.0

Intended Use

This model is suitable for a variety of general language processing tasks, benefiting from its Qwen3 foundation and efficient fine-tuning. Its optimized training process suggests a focus on practical deployment and performance.