reaperdoesntknow/Qwen3-1.7B-Coder-Distilled-SFT
TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Mar 25, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

reaperdoesntknow/Qwen3-1.7B-Coder-Distilled-SFT is a 1.7 billion parameter Qwen3-based causal language model developed by Convergent Intelligence LLC: Research Division. It was created through a two-stage knowledge distillation process, first from a 30B Coder teacher for STEM reasoning, then fine-tuned on 54,600 logical inference problems. This model excels at formal reasoning, logical inference, and structured STEM derivation, leveraging a 1024-token context length.

Loading preview...