Concyclics/PeoplesDaily-Qwen3-4B-Base
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kLicense:apache-2.0Architecture:Transformer0.0K Open Weights Warm

Concyclics/PeoplesDaily-Qwen3-4B-Base is a 4 billion parameter language model based on the Qwen3 architecture, featuring a 40960 token context length. It has undergone Supervised Fine-Tuning (SFT) on the Concyclics/PeoplesDaily dataset, demonstrating a training loss of 1.646 over 2 epochs. This model is optimized for tasks related to the specific domain covered by the PeoplesDaily dataset.

Loading preview...