richardyoung/Llama-3.1-8B-Instruct-abliterated-obliteratus
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Mar 28, 2026License:llama3.1Architecture:Transformer0.0K Cold

The richardyoung/Llama-3.1-8B-Instruct-abliterated-obliteratus is an 8 billion parameter instruction-tuned causal language model, derived from Meta's Llama-3.1-8B-Instruct. Developed by Richard Young, this model has undergone an 'abliteration' process using the OBLITERATUS method to significantly reduce refusal behaviors. It is primarily intended for research into uncensored LLM responses and the study of refusal mechanisms, offering a 32768 token context length.

Loading preview...

What the fuck is this model about?

This model, richardyoung/Llama-3.1-8B-Instruct-abliterated-obliteratus, is an 8 billion parameter instruction-tuned language model based on Meta's Llama-3.1-8B-Instruct. Its core characteristic is that it has been "abliterated" using the OBLITERATUS (advanced) method, a technique developed by Richard Young to remove refusal behaviors from LLMs. This process aims to identify and orthogonalize the "refusal direction" within the model's internal activation space, resulting in a model with a significantly reduced refusal rate (5.0% Attack Success Rate).

What makes THIS different from all the other models?

Unlike standard instruction-tuned models that incorporate safety guardrails and refusal mechanisms, this model has been specifically engineered to remove those behaviors. It is a direct result of research into LLM abliteration methods, as detailed in the paper "Comparative Analysis of LLM Abliteration Methods: Scaling to MoE Architectures and Modern Tools" by Richard Young. The key differentiator is its uncensored nature, achieved through a targeted technical process rather than simple fine-tuning on unfiltered data. It maintains the 32768 token context length of its base model.

Should I use this for my use case?

Good for:

  • Research into LLM safety and refusal mechanisms: Ideal for studying how refusal behaviors are encoded and removed from models.
  • Exploring uncensored model responses: For academic or controlled research environments requiring models without built-in refusal guardrails.
  • Comparative analysis: Useful for comparing the outputs of abliterated models against their original, censored counterparts.

Not recommended for:

  • Production environments requiring safety: This model explicitly lacks safety guardrails and may generate harmful, illegal, or unethical content.
  • General-purpose applications where content moderation is critical: Due to its uncensored nature, it is unsuitable for public-facing or sensitive applications without strict external filtering.
  • Users seeking a standard, safe-by-default LLM.