Yale-ROSE/Qwen3-4B-sft_dataset_gpt-sft-trl-v2
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Sep 14, 2025Architecture:Transformer Warm
Yale-ROSE/Qwen3-4B-sft_dataset_gpt-sft-trl-v2 is a 4 billion parameter language model, fine-tuned from Qwen/Qwen3-4B using Supervised Fine-Tuning (SFT) with the TRL library. This model is designed for text generation tasks, leveraging its base architecture and SFT training to produce coherent and contextually relevant responses. It is suitable for applications requiring instruction-following capabilities derived from its fine-tuning process.
Loading preview...