open-unlearning/unlearn_tofu_Llama-3.2-1B-Instruct_forget10_RMU_lr5e-05_layer15_scoeff10_epoch5
Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:1BQuant:BF16Ctx Length:32kPublished:May 15, 2025Architecture:Transformer Warm

The open-unlearning/unlearn_tofu_Llama-3.2-1B-Instruct_forget10_RMU_lr5e-05_layer15_scoeff10_epoch5 model is a 1 billion parameter instruction-tuned language model. It is based on the Llama-3.2 architecture and has a context length of 32768 tokens. This model is specifically designed for unlearning, focusing on forgetting specific information as indicated by its name's parameters like 'forget10' and 'RMU'. Its primary differentiator lies in its targeted unlearning capabilities, making it suitable for scenarios requiring selective knowledge removal from a pre-trained LLM.

Loading preview...

Model Overview

This model, unlearn_tofu_Llama-3.2-1B-Instruct_forget10_RMU_lr5e-05_layer15_scoeff10_epoch5, is a 1 billion parameter instruction-tuned language model built upon the Llama-3.2 architecture. It features a substantial context length of 32768 tokens, allowing it to process extensive inputs.

Key Capabilities

  • Targeted Unlearning: The model's naming convention (e.g., forget10, RMU, lr5e-05, layer15, scoeff10, epoch5) strongly suggests it has undergone a specific unlearning process. This implies it is designed to selectively remove or reduce knowledge of certain data points or patterns that it previously learned.
  • Instruction Following: As an instruction-tuned model, it is expected to respond to prompts and instructions effectively, generating coherent and relevant outputs based on the given directives.
  • Large Context Window: The 32768-token context length enables the model to maintain conversational history or process long documents, which is beneficial for complex tasks requiring extensive contextual understanding.

Good for

  • Research in Machine Unlearning: This model is particularly valuable for researchers exploring techniques for machine unlearning, catastrophic forgetting, and privacy-preserving AI.
  • Developing Customizable LLMs: It can serve as a base for creating LLMs where specific information needs to be removed post-training, for instance, to comply with data retention policies or to mitigate bias.
  • Controlled Knowledge Bases: Use cases where an LLM needs to operate with a deliberately restricted or modified knowledge base, ensuring it does not recall certain facts or associations.