Hyeongwon/P12-frac0p05-fullft-lr2e5-ep6

TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Apr 29, 2026Architecture:Transformer Cold

Hyeongwon/P12-frac0p05-fullft-lr2e5-ep6 is a 4 billion parameter language model, fine-tuned from Hyeongwon/Qwen3-4B-Base using the TRL library. This model is specifically trained with Supervised Fine-Tuning (SFT) to enhance its text generation capabilities. It is designed for general text generation tasks, building upon the base Qwen3-4B architecture.

Loading preview...

Overview

Hyeongwon/P12-frac0p05-fullft-lr2e5-ep6 is a 4 billion parameter language model derived from the Hyeongwon/Qwen3-4B-Base architecture. This model has undergone Supervised Fine-Tuning (SFT) using the TRL library, a framework for Transformer Reinforcement Learning. The training process focused on refining its ability to generate coherent and contextually relevant text.

Key Capabilities

  • Text Generation: Optimized for generating responses to user prompts, as demonstrated by the quick start example involving a philosophical question.
  • Fine-tuned Performance: Benefits from SFT to improve its conversational and generative fluency compared to its base model.

Training Details

The model was trained using specific versions of key frameworks:

  • TRL: 0.25.1
  • Transformers: 4.57.3
  • Pytorch: 2.7.0+cu128
  • Datasets: 3.6.0
  • Tokenizers: 0.22.2

This fine-tuned model is suitable for applications requiring a compact yet capable language model for various text-based generation tasks.