YanLabs/gemma-3-27b-it-abliterated-normpreserve-v1

Warm
Public
Vision
27B
FP8
32768
License: gemma
Hugging Face
Overview

Overview

YanLabs/gemma-3-27b-it-abliterated-normpreserve-v1 is a specialized variant of the 27 billion parameter Google Gemma-3-27b-it model. Developed by YanLabs, this model utilizes a norm-preserving biprojected abliteration technique to surgically remove refusal behaviors and safety guardrails from the base model. Unlike traditional fine-tuning, this method aims to preserve the model's original capabilities while eliminating its propensity to refuse certain prompts.

Key Characteristics

  • Abliterated Refusal Mechanisms: Safety guardrails and refusal behaviors have been intentionally removed.
  • Norm-Preserving Biprojection: Employs a specific technique to remove "refusal directions" from the activation space without traditional fine-tuning.
  • Capability Preservation: Designed to retain the original model's capabilities despite the removal of safety features.
  • Quantization Recommendation: For optimal performance without refusals, the Q8_0 quantization is recommended.

Intended Use Cases

This model is explicitly not intended for production deployments or user-facing applications due to the removal of safety mechanisms. Its primary applications are:

  • Mechanistic Interpretability Research: Studying how LLMs function at a deeper level.
  • LLM Safety Analysis: Investigating the nature and implementation of safety mechanisms in large language models.
  • Abliteration Technique Development: Researching and testing methods for modifying model behaviors without extensive retraining.

Limitations

Users should be aware that this model may generate harmful or unsafe content and its behavior can be unpredictable in edge cases. Abliteration does not guarantee the complete removal of all refusals, and no explicit harm prevention mechanisms remain.