kmseong/llama3.2_3b_only_sn_tuned_lr5e-5
TEXT GENERATIONConcurrency Cost:1Model Size:3.2BQuant:BF16Ctx Length:32kPublished:Apr 6, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The kmseong/llama3.2_3b_only_sn_tuned_lr5e-5 model is a 3.2 billion parameter Llama-3.2-3B-Instruct variant, fine-tuned by kmseong using the Safety Neuron Tuning (SN-Tune) method. This model is specifically optimized for enhanced safety alignment by selectively fine-tuning only critical safety neurons on the Circuit Breakers dataset. It aims to improve safety without significantly impacting general capabilities, making it suitable for applications requiring robust content moderation.

Loading preview...

Overview

This model, kmseong/llama3.2_3b_only_sn_tuned_lr5e-5, is a specialized version of the 3.2 billion parameter Llama-3.2-3B-Instruct base model. Developed by kmseong, its primary distinction lies in its fine-tuning methodology: Safety Neuron Tuning (SN-Tune). This technique focuses on enhancing the model's safety alignment while preserving its general capabilities.

Key Capabilities & Features

  • Safety Neuron Tuning (SN-Tune): A selective fine-tuning approach that identifies and exclusively fine-tunes a small set of "safety neurons" on safety-specific data.
  • Enhanced Safety Alignment: Designed to provide improved safety performance compared to its base model by targeting critical neurons responsible for safety responses.
  • Parameter-Efficient Fine-tuning: By freezing most parameters and only adjusting safety neurons, this method minimizes computational overhead and potential degradation of general performance.
  • Base Model: Built upon meta-llama/Llama-3.2-3B-Instruct, inheriting its foundational language understanding and generation abilities.
  • Training Data: Fine-tuned using the "Circuit Breakers" dataset, which is specifically curated for safety alignment.

When to Use This Model

This model is particularly well-suited for use cases where:

  • Content Moderation is Critical: Applications requiring a higher degree of safety and reduced generation of harmful content.
  • Balancing Safety and Performance: Scenarios where maintaining general language model capabilities is important, but with an added layer of safety assurance.
  • Resource-Constrained Environments: Its parameter-efficient fine-tuning makes it a viable option for deployment where computational resources are a consideration.