olusegunola/DeepSeek-R1-Distill-Merge-Qwen-Math-1.5Bb
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Mar 19, 2026License:apache-2.0Architecture:Transformer Open Weights Cold
The olusegunola/DeepSeek-R1-Distill-Merge-Qwen-Math-1.5Bb is a 1.5 billion parameter language model created by olusegunola, merging Qwen2.5-Math and DeepSeek-R1-Distill-Qwen. This model is specifically designed to bridge mathematical logic and reasoning, excelling in structured tasks such as USMLE-style Q&A and ICD-10 clinical coding. It leverages the DARE-TIES merge method to preserve specialized weights, making it suitable for research in the medical domain with a 32768 token context length.
Loading preview...