Hyeongwon/PS_only_answer_Qwen3-4B-Base_0328-01-1e-5

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Mar 28, 2026Architecture:Transformer Warm

Hyeongwon/PS_only_answer_Qwen3-4B-Base_0328-01-1e-5 is a fine-tuned version of the Hyeongwon/Qwen3-4B-Base model, developed by Hyeongwon. This model has been specifically trained using Supervised Fine-Tuning (SFT) with the TRL framework. It is designed to provide direct answers, as indicated by its name, and is suitable for tasks requiring focused response generation from a base Qwen3-4B architecture.

Loading preview...

Model Overview

Hyeongwon/PS_only_answer_Qwen3-4B-Base_0328-01-1e-5 is a specialized language model developed by Hyeongwon. It is a fine-tuned iteration of the Hyeongwon/Qwen3-4B-Base model, leveraging the TRL (Transformer Reinforcement Learning) framework for its training. The model's name suggests an optimization for direct answer generation, making it suitable for specific question-answering tasks.

Training Details

This model was trained using Supervised Fine-Tuning (SFT). The training procedure utilized the following framework versions:

  • TRL: 0.25.1
  • Transformers: 4.57.3
  • Pytorch: 2.6.0
  • Datasets: 3.6.0
  • Tokenizers: 0.22.2

Further details on the training run can be visualized via Weights & Biases, as linked in the original model card.

Intended Use

Given its fine-tuned nature and naming convention, this model is primarily intended for use cases where a direct and focused answer is required, rather than open-ended generation. Developers can integrate it using the Hugging Face transformers pipeline for text generation tasks.