open-unlearning/unlearn_tofu_Llama-3.2-1B-Instruct_forget10_NPO_lr1e-05_beta0.5_alpha1_epoch10

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:1BQuant:BF16Ctx Length:32kPublished:May 15, 2025Architecture:Transformer Warm

The open-unlearning/unlearn_tofu_Llama-3.2-1B-Instruct_forget10_NPO_lr1e-05_beta0.5_alpha1_epoch10 model is a 1 billion parameter instruction-tuned language model based on the Llama-3.2 architecture. This model is specifically designed for unlearning, indicated by its name referencing "unlearn_tofu" and "forget10", suggesting it has undergone a process to remove specific information. Its primary differentiator lies in its unlearning capabilities, making it suitable for research into model privacy, data removal, and mitigating unwanted information recall.

Loading preview...

Overview

This model, open-unlearning/unlearn_tofu_Llama-3.2-1B-Instruct_forget10_NPO_lr1e-05_beta0.5_alpha1_epoch10, is a 1 billion parameter instruction-tuned language model built upon the Llama-3.2 architecture. Its naming convention, particularly "unlearn_tofu" and "forget10", indicates that it has been subjected to a machine unlearning process, aiming to remove or reduce the model's ability to recall specific data points or patterns.

Key Capabilities

  • Instruction Following: As an instruction-tuned model, it is designed to respond to user prompts and instructions effectively.
  • Unlearning Research: The model's primary characteristic is its unlearning aspect, making it a valuable tool for research into:
    • Data privacy and the right to be forgotten in AI models.
    • Techniques for mitigating bias or undesirable information.
    • Understanding the mechanisms of knowledge retention and removal in large language models.

Good For

  • Academic Research: Ideal for researchers studying machine unlearning, model privacy, and data deletion techniques.
  • Experimentation: Suitable for experimenting with the effects of unlearning on model behavior and performance.
  • Prototyping: Can be used to prototype applications requiring models with specific knowledge removal or privacy-preserving characteristics.