florentgbelidji/Qwen3-4B-Base-SFT-20260120102752

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Jan 20, 2026Architecture:Transformer Warm

The florentgbelidji/Qwen3-4B-Base-SFT-20260120102752 model is a 4 billion parameter language model, fine-tuned from Qwen/Qwen3-4B-Base. It was trained using Supervised Fine-Tuning (SFT) on the trl-lib/Capybara dataset, leveraging the TRL library. This model is designed for general text generation tasks, building upon the base capabilities of the Qwen3 architecture with enhanced conversational or instruction-following abilities derived from its SFT training.

Loading preview...

Model Overview

The florentgbelidji/Qwen3-4B-Base-SFT-20260120102752 is a 4 billion parameter language model developed by florentgbelidji. It is a fine-tuned variant of the Qwen/Qwen3-4B-Base model, specifically optimized through Supervised Fine-Tuning (SFT).

Key Capabilities

  • Base Model: Built upon the robust Qwen3-4B-Base architecture, providing a strong foundation for language understanding and generation.
  • Fine-Tuning: Underwent Supervised Fine-Tuning (SFT) using the trl-lib/Capybara dataset, which typically enhances a model's ability to follow instructions and generate coherent, contextually relevant responses.
  • Training Framework: Utilizes the TRL library (Transformer Reinforcement Learning) for its training procedure, indicating a focus on improving model behavior through fine-tuning techniques.

Use Cases

This model is suitable for general text generation tasks where a fine-tuned base model can provide improved performance over its un-tuned counterpart. Its SFT training suggests it can be effectively used for:

  • Answering questions
  • Generating creative text
  • Engaging in conversational exchanges
  • Following specific instructions for text output

Developers can integrate this model using the Hugging Face transformers library for text generation pipelines.