deepseek-ai/DeepSeek-R1-0528
TEXT GENERATIONConcurrency Cost:4Model Size:685BQuant:FP8Ctx Length:32kPublished:May 28, 2025License:mitArchitecture:Transformer2.4K Open Weights Warm

DeepSeek-R1-0528 is a 685 billion parameter language model developed by DeepSeek AI, featuring a 32K token context length. This updated version significantly enhances reasoning and inference capabilities through algorithmic optimizations and increased computational resources. It demonstrates strong performance across mathematics, programming, and general logic benchmarks, with notable improvements in complex reasoning tasks and reduced hallucination rates. The model is designed for advanced applications requiring deep reasoning and robust problem-solving.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p