reaperdoesntknow/Qwen3-1.7B-Thinking-Distil
TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Mar 27, 2026Architecture:Transformer Loading
reaperdoesntknow/Qwen3-1.7B-Thinking-Distil is a 2 billion parameter language model from the Qwen3 family, developed by Convergent Intelligence LLC: Research Division. This model is part of the Qwen3 1.7B Distillation Series, indicating it is a distilled version optimized for specific tasks. With a 32768 token context length, it is designed for applications requiring efficient processing of longer sequences. Its primary differentiator is its focus on 'Thinking' capabilities, suggesting an optimization for reasoning or complex problem-solving tasks within a compact parameter count.
Loading preview...
Popular Sampler Settings
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.
temperature
–
top_p
–
top_k
–
frequency_penalty
–
presence_penalty
–
repetition_penalty
–
min_p
–