salmannyu/Qwen2.5-1.5B-Nemotron-Math-52B-Mid-Train-8
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Feb 8, 2026License:apache-2.0Architecture:Transformer Open Weights Warm

salmannyu/Qwen2.5-1.5B-Nemotron-Math-52B-Mid-Train-8 is a 1.5 billion parameter causal language model, continued pre-trained from Qwen/Qwen2.5-1.5B. It was specifically mid-trained on 52 billion tokens from the NVIDIA Nemotron-CC-Math-v1 dataset, making it highly specialized for mathematical reasoning tasks. With a context length of 131072 tokens, this model is designed for applications requiring robust mathematical problem-solving capabilities.

Loading preview...