Ujan/Qwen3-4B-Base_DeepMath-103K_samples_10000_seq_4096_epoch_1
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Nov 25, 2025Architecture:Transformer Warm

Ujan/Qwen3-4B-Base_DeepMath-103K_samples_10000_seq_4096_epoch_1 is a 4 billion parameter language model with a 40960 token context length. This model is based on the Qwen3 architecture and has been fine-tuned with a DeepMath dataset. Its primary strength lies in mathematical reasoning and problem-solving tasks, making it suitable for applications requiring robust numerical and logical capabilities.

Loading preview...