PetarKal/Qwen3-4B-ascii-art-e5-lr3e-5-ga16-base

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Mar 2, 2026Architecture:Transformer Warm

PetarKal/Qwen3-4B-ascii-art-e5-lr3e-5-ga16-base is a fine-tuned version of Qwen/Qwen3-4B-Base, developed by PetarKal. This model was trained using the TRL framework with SFT (Supervised Fine-Tuning). It is designed for text generation tasks, leveraging its base architecture for general language understanding and generation.

Loading preview...

Model Overview

This model, PetarKal/Qwen3-4B-ascii-art-e5-lr3e-5-ga16-base, is a specialized fine-tuned variant of the Qwen/Qwen3-4B-Base architecture. Developed by PetarKal, it leverages the robust foundation of the Qwen3-4B-Base model, which is known for its general language capabilities.

Training Details

The model underwent Supervised Fine-Tuning (SFT) using the TRL framework. This fine-tuning process adapts the base model to potentially excel in specific text generation tasks, although the README does not specify the exact nature of the fine-tuning dataset or its intended specialization beyond general text generation. The training utilized specific versions of key frameworks:

  • TRL: 0.29.0
  • Transformers: 5.3.0
  • Pytorch: 2.10.0
  • Datasets: 4.6.1
  • Tokenizers: 0.22.2

Usage

Developers can easily integrate this model into their applications using the Hugging Face transformers library. A quick start example demonstrates its use for text generation, where it can respond to user prompts with generated text. The model is suitable for general text generation tasks where a fine-tuned Qwen3-4B base model is desired.