agentica-org/DeepScaleR-1.5B-Preview
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Jan 29, 2025License:mitArchitecture:Transformer0.6K Open Weights Warm

DeepScaleR-1.5B-Preview is a 1.5 billion parameter language model developed by agentica-org, fine-tuned from DeepSeek-R1-Distilled-Qwen-1.5B using distributed reinforcement learning. It is specifically optimized for mathematical reasoning and problem-solving, achieving 43.1% Pass@1 accuracy on AIME 2024. This model demonstrates strong performance in mathematical benchmarks, surpassing larger models like OpenAI's O1-Preview with a significantly smaller parameter count.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
–
frequency_penalty
presence_penalty
repetition_penalty
min_p