lightblue/DeepSeek-R1-Distill-Qwen-7B-Japanese
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Jan 24, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

lightblue/DeepSeek-R1-Distill-Qwen-7B-Japanese is a 7.6 billion parameter language model developed by Lightblue, fine-tuned from deepseek-ai/DeepSeek-R1-Distill-Qwen-7B. This model is specifically optimized for Japanese language output in reasoning tasks, addressing inconsistencies found in the original DeepSeek R1 models. It reliably produces Japanese responses, making it suitable for applications requiring accurate Japanese reasoning.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p