open-unlearning/unlearn_tofu_Llama-3.2-1B-Instruct_forget10_GradDiff_lr1e-05_alpha5_epoch5

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:1BQuant:BF16Ctx Length:32kPublished:May 15, 2025Architecture:Transformer Warm

The open-unlearning/unlearn_tofu_Llama-3.2-1B-Instruct_forget10_GradDiff_lr1e-05_alpha5_epoch5 model is a 1 billion parameter instruction-tuned language model with a 32768 token context length. This model is part of the Llama-3.2 family and has undergone an unlearning process using the GradDiff method. Its primary characteristic is its ability to forget specific information, making it suitable for applications requiring selective knowledge removal or privacy-preserving AI.

Loading preview...

Model Overview

This model, unlearn_tofu_Llama-3.2-1B-Instruct_forget10_GradDiff_lr1e-05_alpha5_epoch5, is a 1 billion parameter instruction-tuned language model from the Llama-3.2 family. It features a substantial context length of 32768 tokens, allowing it to process and generate longer sequences of text. The model has been specifically modified through an "unlearning" process using the GradDiff method, indicating an intentional effort to remove or reduce specific information from its knowledge base.

Key Characteristics

  • Architecture: Llama-3.2 family, instruction-tuned.
  • Parameters: 1 billion, offering a balance between performance and computational efficiency.
  • Context Length: 32768 tokens, enabling handling of extensive inputs and generating detailed responses.
  • Unlearning Mechanism: Utilizes the GradDiff method for targeted knowledge removal.

Potential Use Cases

This model is particularly relevant for scenarios where selective forgetting or privacy-preserving AI is crucial. While specific details on its training data and exact unlearning targets are not provided in the model card, its design suggests applications such as:

  • Data Privacy: Developing models that can selectively forget sensitive user data upon request.
  • Bias Mitigation: Attempting to remove specific biases introduced during training.
  • Content Moderation: Creating models that avoid generating certain types of undesirable content by unlearning related patterns.
  • Research in Unlearning: Serving as a base for further research into machine unlearning techniques and their effectiveness.