open-unlearning/unlearn_tofu_Llama-3.2-1B-Instruct_forget10_AltPO_lr5e-05_beta0.1_alpha1_epoch10

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:1BQuant:BF16Ctx Length:32kPublished:May 15, 2025Architecture:Transformer Warm

The open-unlearning/unlearn_tofu_Llama-3.2-1B-Instruct_forget10_AltPO_lr5e-05_beta0.1_alpha1_epoch10 is a 1 billion parameter instruction-tuned Llama-3.2 model with a 32768 token context length. This model has undergone an unlearning process, specifically targeting the 'forget10' dataset using the AltPO method. It is designed for tasks requiring a compact yet capable language model that has been modified to remove specific information.

Loading preview...

Overview

This model, open-unlearning/unlearn_tofu_Llama-3.2-1B-Instruct_forget10_AltPO_lr5e-05_beta0.1_alpha1_epoch10, is a 1 billion parameter instruction-tuned variant of the Llama-3.2 architecture. It features a substantial context window of 32768 tokens, making it suitable for processing longer inputs.

Key Differentiator: Unlearning

The primary characteristic of this model is its application of an "unlearning" process. Specifically, it has been trained to forget information related to the 'forget10' dataset using the AltPO (Alternating Policy Optimization) method. This makes it distinct from standard instruction-tuned models, as it has been intentionally modified to reduce or eliminate knowledge of particular data points.

Potential Use Cases

  • Research in Machine Unlearning: Ideal for studying the effectiveness and implications of unlearning techniques in large language models.
  • Privacy-Preserving Applications: Could be explored for scenarios where specific sensitive information needs to be removed from a model's knowledge base post-training.
  • Controlled Information Access: Useful for creating models that intentionally lack knowledge about certain topics or datasets.

Limitations

As indicated by the model card, many details regarding its development, training data, evaluation, and specific biases are currently marked as "More Information Needed." Users should exercise caution and conduct thorough testing for their specific applications, especially given the experimental nature of unlearning.