open-unlearning/unlearn_tofu_Llama-3.2-1B-Instruct_forget10_AltPO_lr2e-05_beta0.1_alpha1_epoch5

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:1BQuant:BF16Ctx Length:32kPublished:May 15, 2025Architecture:Transformer Warm

The open-unlearning/unlearn_tofu_Llama-3.2-1B-Instruct_forget10_AltPO_lr2e-05_beta0.1_alpha1_epoch5 model is a 1 billion parameter instruction-tuned language model based on the Llama-3.2 architecture. This model is specifically designed for unlearning, focusing on the 'forget10' task using the AltPO method. Its primary differentiator lies in its application of unlearning techniques, making it suitable for research and development in model editing and data privacy.

Loading preview...

Model Overview

This model, open-unlearning/unlearn_tofu_Llama-3.2-1B-Instruct_forget10_AltPO_lr2e-05_beta0.1_alpha1_epoch5, is a 1 billion parameter instruction-tuned language model built upon the Llama-3.2 architecture. It is notable for its application of unlearning techniques, specifically targeting the 'forget10' task using the AltPO method with a learning rate of 2e-05, beta of 0.1, alpha of 1, and trained for 5 epochs.

Key Characteristics

  • Architecture: Llama-3.2 base model.
  • Parameter Count: 1 billion parameters, offering a balance between performance and computational efficiency.
  • Instruction-Tuned: Designed to follow instructions effectively.
  • Unlearning Focus: Specialized in model unlearning, particularly for the 'forget10' task, which involves removing specific information from the model's knowledge base.
  • Methodology: Utilizes the AltPO (Alternating Policy Optimization) method for unlearning, indicating a focus on advanced model editing techniques.

Potential Use Cases

  • Research in Model Unlearning: Ideal for researchers exploring methods to remove unwanted or outdated information from large language models.
  • Data Privacy Experiments: Can be used to test and develop techniques for enhancing data privacy by selectively forgetting specific data points.
  • Model Editing: Applicable for scenarios requiring precise modification of a model's knowledge without extensive retraining.

Due to the limited information in the provided README, specific performance benchmarks or detailed training data are not available. Users should be aware that this model is likely intended for specialized research and development in the field of model unlearning rather than general-purpose instruction following.