lightblue/DeepSeek-R1-Distill-Qwen-7B-Japanese
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Jan 24, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Warm
lightblue/DeepSeek-R1-Distill-Qwen-7B-Japanese is a 7.6 billion parameter language model developed by Lightblue, fine-tuned from deepseek-ai/DeepSeek-R1-Distill-Qwen-7B. This model is specifically optimized for Japanese language output in reasoning tasks, addressing inconsistencies found in the original DeepSeek R1 models. It reliably produces Japanese responses, making it suitable for applications requiring accurate Japanese reasoning.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
–
top_p
–
top_k
–
frequency_penalty
–
presence_penalty
–
repetition_penalty
–
min_p
–