open-unlearning/unlearn_tofu_Llama-3.2-1B-Instruct_forget10_AltPO_lr1e-05_beta0.1_alpha5_epoch5
Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:1BQuant:BF16Ctx Length:32kPublished:May 15, 2025Architecture:Transformer Warm

The open-unlearning/unlearn_tofu_Llama-3.2-1B-Instruct_forget10_AltPO_lr1e-05_beta0.1_alpha5_epoch5 is a 1 billion parameter instruction-tuned language model, likely based on the Llama 3.2 architecture, developed by open-unlearning. This model is specifically designed for "unlearning" tasks, indicated by its name, which suggests it has been fine-tuned to forget specific information or behaviors. Its primary differentiator lies in its unlearning capabilities, making it suitable for research and applications requiring selective knowledge removal or modification in LLMs.

Loading preview...

Model Overview

This model, unlearn_tofu_Llama-3.2-1B-Instruct_forget10_AltPO_lr1e-05_beta0.1_alpha5_epoch5, is a 1 billion parameter instruction-tuned language model. It is developed by open-unlearning and appears to be based on the Llama 3.2 architecture. The model's name strongly indicates its specialization in "unlearning" tasks, suggesting it has undergone a process to selectively remove or modify certain information or behaviors from its knowledge base.

Key Capabilities

  • Unlearning: The core capability of this model is its ability to "unlearn" specific data or patterns, as implied by the unlearn_tofu and forget10 components in its name. This makes it distinct from standard instruction-tuned models.
  • Instruction Following: As an instruction-tuned model, it is designed to follow user prompts and instructions effectively.
  • Llama 3.2 Base: Likely inherits foundational language understanding and generation capabilities from its Llama 3.2 base architecture.

Good For

  • Research in Model Unlearning: Ideal for researchers exploring methods for removing unwanted information, mitigating bias, or updating knowledge in large language models.
  • Controlled Information Removal: Potentially useful in scenarios where specific data needs to be excluded from a model's responses or knowledge.
  • Exploring Model Plasticity: Provides a platform to study how models can be modified post-training to alter their learned representations.