Model Overview
This model, PS_only_answer_Qwen3-4B-Base_0328-01-2e-5, is a 4 billion parameter language model developed by Hyeongwon. It is a fine-tuned version of the Hyeongwon/Qwen3-4B-Base model, specifically trained using Supervised Fine-Tuning (SFT) with the TRL framework.
Key Capabilities
- Specific Answer Generation: The model's naming convention suggests an optimization for providing direct and focused answers to prompts.
- Base Model Fine-tuning: Built upon the Qwen3-4B-Base architecture, it leverages the foundational capabilities of its parent model.
- Extended Context Window: Supports a substantial context length of 32768 tokens, allowing for processing and generating longer sequences of text.
Training Details
The model underwent training using the SFT method, indicating a focus on learning from a dataset of input-output pairs to guide its response generation. The training process utilized the TRL (Transformer Reinforcement Learning) library, with specific framework versions including TRL 0.25.1, Transformers 4.57.3, Pytorch 2.6.0, Datasets 3.6.0, and Tokenizers 0.22.2.
Good For
- Applications requiring concise and direct answers.
- Tasks benefiting from a large context window for understanding and generating responses.