reaperdoesntknow/DistilQwen3-1.7B-uncensored

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:2BQuant:BF16Ctx Length:32kPublished:Mar 25, 2026Architecture:Transformer Warm

The reaperdoesntknow/DistilQwen3-1.7B-uncensored is a 1.7 billion parameter model from the DistilQwen3 series by Convergent Intelligence LLC: Research Division. This model is a proof-weighted distillation of a 30B-parameter Qwen3 teacher, specifically designed to amplify reasoning-critical tokens through Discrepancy Calculus. It excels in structural understanding and logical inference, making it suitable for tasks requiring deep reasoning rather than surface-level pattern matching.

Loading preview...

Model Overview

This model, reaperdoesntknow/DistilQwen3-1.7B-uncensored, is a 1.7 billion parameter language model developed by Convergent Intelligence LLC: Research Division. It is part of the DistilQwen Collection, a series of models built using a unique proof-weighted knowledge distillation methodology based on Discrepancy Calculus (DISC).

Key Capabilities & Methodology

  • Distillation from a 30B-parameter Teacher: This model is distilled from a powerful Qwen3-30B-A3B teacher, leveraging premium H100 hardware and BF16 training for enhanced performance.
  • Discrepancy Calculus (DISC): Utilizes a measure-theoretic framework to decompose the teacher's output distribution, quantifying local structural mismatch that standard KL divergence misses. This approach aims to improve the student model's understanding of underlying structure.
  • Proof-Weighted Knowledge Distillation: Employs a unique distillation process (55% cross-entropy with decaying proof weights, 45% KL divergence at T=2.0). The proof weight specifically amplifies loss on reasoning-critical tokens, compelling the student to prioritize structural understanding over superficial pattern recognition.
  • Focus on Reasoning: Designed to excel in tasks requiring deep reasoning, logical inference, and structural decomposition, as opposed to mere surface-level pattern matching.

DistilQwen Collection Context

This model is part of a collection that includes variants optimized for different strengths, such as instruction following, structured output, legal reasoning, extended deliberation, higher-entropy distributions, proof derivation, STEM derivation, and logical inference. The entire DistilQwen series is notable for being the only BF16 collection within the Convergent Intelligence portfolio, indicating a higher compute investment compared to their broader catalog.