open-unlearning/unlearn_tofu_Llama-3.2-1B-Instruct_forget10_NPO_lr5e-05_beta0.5_alpha1_epoch10

TEXT GENERATIONConcurrency Cost:1Model Size:1BQuant:BF16Ctx Length:32kPublished:May 15, 2025Architecture:Transformer Cold

The open-unlearning/unlearn_tofu_Llama-3.2-1B-Instruct_forget10_NPO_lr5e-05_beta0.5_alpha1_epoch10 model is a 1 billion parameter instruction-tuned language model with a 32768 token context length. This model is specifically designed for unlearning, focusing on the ability to forget specific information. Its primary differentiator lies in its application of Negative Preference Optimization (NPO) for targeted unlearning, making it suitable for scenarios requiring data privacy or content moderation.

Loading preview...

Model Overview

This model, open-unlearning/unlearn_tofu_Llama-3.2-1B-Instruct_forget10_NPO_lr5e-05_beta0.5_alpha1_epoch10, is a 1 billion parameter instruction-tuned language model. It is notable for its 32768 token context length, allowing it to process extensive inputs.

Key Characteristics

  • Unlearning Focus: This model is specifically engineered for the task of "unlearning," meaning it has been trained to forget particular information or patterns.
  • Negative Preference Optimization (NPO): It utilizes Negative Preference Optimization (NPO) during its training, a technique aimed at effectively removing specific data or behaviors from the model's knowledge base.
  • Instruction-Tuned: The model is instruction-tuned, enabling it to follow commands and generate responses based on given instructions.

Potential Use Cases

  • Data Privacy: Ideal for scenarios where a model needs to remove sensitive or private information it may have inadvertently learned.
  • Content Moderation: Can be applied to unlearn undesirable biases or harmful content, improving safety and ethical alignment.
  • Research in Unlearning: Serves as a valuable tool for researchers exploring methods and effectiveness of machine unlearning in large language models.