unsloth/DeepScaleR-1.5B-Preview
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Feb 15, 2025License:mitArchitecture:Transformer0.0K Open Weights Warm

DeepScaleR-1.5B-Preview is a 1.5 billion parameter language model developed by Agentica, fine-tuned from DeepSeek-R1-Distilled-Qwen-1.5B. It utilizes distributed reinforcement learning with iterative context lengthening to achieve high accuracy on mathematical reasoning tasks, notably surpassing OpenAI's O1-Preview on AIME 2024 with a 43.1% Pass@1 score. This model is optimized for complex problem-solving and mathematical reasoning, supporting a 32768-token context length.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p