open-unlearning/unlearn_tofu_Llama-3.2-1B-Instruct_forget10_GradDiff_lr2e-05_alpha5_epoch5

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:1BQuant:BF16Ctx Length:32kPublished:May 15, 2025Architecture:Transformer Warm

The open-unlearning/unlearn_tofu_Llama-3.2-1B-Instruct_forget10_GradDiff_lr2e-05_alpha5_epoch5 model is a Llama-3.2-1B-Instruct variant that has undergone an unlearning process. This model is specifically designed to demonstrate and evaluate machine unlearning techniques, particularly using the GradDiff method. Its primary purpose is research into how to selectively remove specific information from a pre-trained language model while retaining general capabilities.

Loading preview...

Model Overview

This model, unlearn_tofu_Llama-3.2-1B-Instruct_forget10_GradDiff_lr2e-05_alpha5_epoch5, is a specialized variant of the Llama-3.2-1B-Instruct architecture. It has been subjected to a machine unlearning process using the GradDiff method, with specific hyperparameters (learning rate 2e-05, alpha 5, 5 epochs) to 'forget' 10 specific data points or concepts.

Key Characteristics

  • Base Model: Llama-3.2-1B-Instruct, indicating a foundation in a capable instruction-tuned language model.
  • Unlearning Method: Utilizes the GradDiff technique, a gradient-based approach for selective forgetting.
  • Targeted Forgetting: Configured to remove 10 specific pieces of information, demonstrating the model's ability to selectively unlearn.
  • Research Focus: Primarily intended for research and experimentation in the field of machine unlearning, rather than direct application.

Intended Use Cases

  • Machine Unlearning Research: Ideal for studying the effectiveness and mechanisms of gradient-based unlearning methods.
  • Privacy-Preserving AI: Can be used to explore techniques for removing sensitive data from models post-training.
  • Model Auditing: Useful for developing and testing methods to verify if specific information has been successfully removed from a model.

Limitations

As a research model, its direct applicability to general-purpose tasks may be limited compared to its base model. The primary focus is on demonstrating and evaluating the unlearning process itself.