deepseek-ai/DeepSeek-V3-0324
TEXT GENERATIONConcurrency Cost:4Model Size:685BQuant:FP8Ctx Length:32kPublished:Mar 24, 2025License:mitArchitecture:Transformer3.1K Open Weights Warm

DeepSeek-V3-0324 is a 685 billion parameter language model developed by DeepSeek-AI, building upon the DeepSeek-V3 architecture. This iteration demonstrates significant improvements in reasoning capabilities across benchmarks like MMLU-Pro, GPQA, AIME, and LiveCodeBench. It is optimized for complex problem-solving, front-end web development, and enhanced Chinese writing proficiency, making it suitable for advanced analytical and creative tasks.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p