FuseAI/FuseO1-DeepSeekR1-Qwen2.5-Coder-32B-Preview
TEXT GENERATIONConcurrency Cost:2Model Size:32.8BQuant:FP8Ctx Length:32kPublished:Jan 22, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

FuseAI/FuseO1-DeepSeekR1-Qwen2.5-Coder-32B-Preview is a 32.8 billion parameter language model developed by FuseAI, designed to enhance System-II reasoning capabilities through model fusion. This model integrates DeepSeek-R1-Distill-Qwen-32B and Qwen2.5-32B-Coder using a Long-Short Reasoning Merging approach. It excels in code reasoning tasks, demonstrating improved performance on benchmarks like LiveCodeBench.

Loading preview...