prithivMLmods/Qwen2.5-32B-DeepSeek-R1-Instruct
TEXT GENERATIONConcurrency Cost:2Model Size:32.8BQuant:FP8Ctx Length:32kArchitecture:Transformer0.0K Warm

prithivMLmods/Qwen2.5-32B-DeepSeek-R1-Instruct is a 32.8 billion parameter merged language model, built upon the Qwen2.5-32B-Instruct base using the TIES merge method. It integrates QwQ-32B-Preview and DeepSeek-R1-Distill-Qwen-32B, configured with normalization and int8 masking for optimized performance. This model is designed to leverage the combined strengths of its constituent models, offering enhanced capabilities for instruction-following tasks.

Loading preview...