reaperdoesntknow/Qwen3-1.7B-Thinking-Distil
TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Mar 27, 2026Architecture:Transformer Loading

reaperdoesntknow/Qwen3-1.7B-Thinking-Distil is a 2 billion parameter language model from the Qwen3 family, developed by Convergent Intelligence LLC: Research Division. This model is part of the Qwen3 1.7B Distillation Series, indicating it is a distilled version optimized for specific tasks. With a 32768 token context length, it is designed for applications requiring efficient processing of longer sequences. Its primary differentiator is its focus on 'Thinking' capabilities, suggesting an optimization for reasoning or complex problem-solving tasks within a compact parameter count.

Loading preview...

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p