Ujan/Qwen3-4B-Base_DeepMath-103K_samples_10000_seq_2048_epoch_1
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Dec 27, 2025Architecture:Transformer Warm

Ujan/Qwen3-4B-Base_DeepMath-103K_samples_10000_seq_2048_epoch_1 is a 4 billion parameter language model developed by Ujan, based on the Qwen3 architecture. This model is fine-tuned with a DeepMath dataset, suggesting an optimization for mathematical reasoning and problem-solving tasks. With a context length of 40960 tokens, it is designed for applications requiring extensive contextual understanding in mathematical domains. Its primary use case is likely advanced mathematical computation and analysis.

Loading preview...