mkubaszek/Qwen3-0.6B-Base-CPT-Math
TEXT GENERATIONConcurrency Cost:1Model Size:0.8BQuant:BF16Ctx Length:32kPublished:Apr 10, 2026Architecture:Transformer Cold

The mkubaszek/Qwen3-0.6B-Base-CPT-Math is a 0.8 billion parameter language model based on the Qwen3 architecture, featuring a 32768 token context length. This model is specifically designed and optimized for mathematical tasks, indicating a focus on numerical reasoning and problem-solving capabilities. Its base configuration suggests it serves as a foundation for further fine-tuning in specialized mathematical domains. Developers can leverage this model for applications requiring robust mathematical understanding and generation.

Loading preview...