lightblue/DeepSeek-R1-Distill-Qwen-14B-Multilingual
TEXT GENERATIONConcurrency Cost:1Model Size:14.8BQuant:FP8Ctx Length:32kPublished:Jan 31, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

lightblue/DeepSeek-R1-Distill-Qwen-14B-Multilingual is a 14.8 billion parameter Qwen-based model, fine-tuned by Lightblue from DeepSeek-R1-Distill-Qwen-14B. This model specializes in multilingual Chain-of-Thought (CoT) reasoning, enabling it to think and respond in the user's specified language. It is designed to enhance the understandability and explainability of AI outputs for a global audience, particularly for non-English and non-Chinese languages.

Loading preview...