Hulyyy/qwen-test

Cold
Public
0.8B
BF16
32768
Oct 16, 2025
License: apache-2.0
Hugging Face
Overview

Model Overview

Hulyyy/qwen-test is a fine-tuned language model based on the Qwen3-0.6B architecture. It comprises approximately 0.8 billion parameters and was developed using the Transformers library.

Training Details

The model underwent a single epoch of fine-tuning with a learning rate of 5e-05. Key training hyperparameters included a train_batch_size of 1 and an eval_batch_size of 8, utilizing the ADAMW_TORCH_FUSED optimizer. During training, it achieved a validation loss of 0.2279.

Current Status

As per the available information, specific details regarding the dataset used for fine-tuning, its intended uses, and limitations are not provided. This suggests it might be an experimental or foundational fine-tune where the specific application or unique capabilities are yet to be defined or publicly documented.