lightblue/DeepSeek-R1-Distill-Qwen-1.5B-Multilingual
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Jan 31, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

The lightblue/DeepSeek-R1-Distill-Qwen-1.5B-Multilingual is a 1.5 billion parameter language model developed by Lightblue, fine-tuned from deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B. This model specializes in multilingual Chain-of-Thought (CoT) reasoning, designed to think and respond in the user's specified language. It features a 131072 token context length and aims to provide more understandable and explainable AI outputs for a global audience, particularly excelling in higher-resource languages like Japanese, English, and German.

Loading preview...