maheshrawat18/Qwen3-4B-2507-sft-new-updated

TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:May 2, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The maheshrawat18/Qwen3-4B-2507-sft-new-updated is a 4 billion parameter Qwen3 model, fine-tuned by maheshrawat18. This model was efficiently trained using Unsloth and Huggingface's TRL library, enabling faster fine-tuning. It is designed for general language tasks, leveraging its Qwen3 architecture for robust performance.

Loading preview...

Model Overview

The maheshrawat18/Qwen3-4B-2507-sft-new-updated is a 4 billion parameter Qwen3 model, fine-tuned by maheshrawat18. This model builds upon the unsloth/Qwen3-4B-Thinking-2507 base model.

Key Training Details

  • Efficient Fine-tuning: The model was fine-tuned with significant speed improvements, achieving 2x faster training. This was accomplished by utilizing Unsloth alongside Huggingface's TRL library, which are tools designed to optimize the training process for large language models.
  • Base Model: It is a specialized fine-tuned version of the Qwen3 architecture, indicating a focus on general language understanding and generation capabilities inherent to the Qwen series.

Licensing

The model is released under the apache-2.0 license, allowing for broad use and distribution.