richardyoung/Deepseek-R1-Distill-Qwen-32b-uncensored

TEXT GENERATIONConcurrency Cost:2Model Size:32.8BQuant:FP8Ctx Length:32kPublished:Nov 19, 2025License:deepseekArchitecture:Transformer0.0K Cold

The richardyoung/Deepseek-R1-Distill-Qwen-32b-uncensored model is a 32 billion parameter, decoder-only transformer based on the Qwen2 architecture, developed by richardyoung. It is an uncensored version of deepseek-ai's DeepSeek-R1-Distill-Qwen-32B, featuring a 32,768-token context length. This model specializes in strong chain-of-thought reasoning capabilities without safety refusals, making it suitable for research requiring unrestricted step-by-step analysis.

Loading preview...

Overview

This model, richardyoung/Deepseek-R1-Distill-Qwen-32b-uncensored, is an "abliterated" (uncensored) variant of the deepseek-ai/DeepSeek-R1-Distill-Qwen-32B model. It is a 32 billion parameter, decoder-only transformer built on the Qwen2 architecture, designed to provide robust reasoning capabilities without the typical safety guardrails or refusal interventions found in many LLMs.

Key Characteristics

  • Base Model: DeepSeek-R1-Distill-Qwen-32B (32B parameters).
  • Architecture: Qwen2 (decoder-only transformer).
  • Context Length: Supports a substantial context window of 32,768 tokens.
  • Uncensored Output: Achieved through "abliteration," a technique involving the surgical removal of the refusal direction, allowing for unrestricted output.
  • Core Strength: Retains the strong chain-of-thought reasoning abilities of the original DeepSeek-R1 model.

Intended Use Cases

This model is particularly well-suited for:

  • Research: Ideal for studies on reasoning, alignment, and exploring the full range of a model's capabilities without artificial limitations.
  • Education: Can be used in educational settings where step-by-step analysis and unrestricted problem-solving are beneficial.
  • Creative Applications: Suitable for tasks requiring detailed, step-by-step analysis and creative generation without content restrictions.