lightblue/DeepSeek-R1-Distill-Qwen-7B-Multilingual
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Jan 31, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

The lightblue/DeepSeek-R1-Distill-Qwen-7B-Multilingual is a 7.6 billion parameter DeepSeek-R1-Distill-Qwen-7B model fine-tuned by Lightblue for multilingual Chain-of-Thought (CoT) reasoning. It is optimized to think and respond in the user's language, enhancing explainability and usability for non-English and non-Chinese audiences. This model excels at providing correctly formatted and accurate results across various languages, particularly higher-resource ones.

Loading preview...