open-unlearning/unlearn_tofu_Llama-3.2-1B-Instruct_forget10_RMU_lr5e-05_layer10_scoeff10_epoch5
Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:1BQuant:BF16Ctx Length:32kPublished:May 15, 2025Architecture:Transformer Warm

The open-unlearning/unlearn_tofu_Llama-3.2-1B-Instruct_forget10_RMU_lr5e-05_layer10_scoeff10_epoch5 model is a 1 billion parameter instruction-tuned language model, likely based on the Llama-3.2 architecture. This model is specifically designed for 'unlearning' tasks, focusing on removing specific information or biases from its knowledge base. Its primary differentiator lies in its application of unlearning techniques, making it suitable for research into model safety, privacy, and controlled information removal.

Loading preview...

Model Overview

This model, open-unlearning/unlearn_tofu_Llama-3.2-1B-Instruct_forget10_RMU_lr5e-05_layer10_scoeff10_epoch5, is a 1 billion parameter instruction-tuned language model. While specific details on its development and training data are not provided in the current model card, its naming convention strongly suggests it is an experimental model focused on "unlearning" capabilities.

Key Characteristics

  • Parameter Count: 1 billion parameters, indicating a relatively compact model size suitable for research and specific applications.
  • Instruction-Tuned: Designed to follow instructions effectively, typical of modern LLMs.
  • Unlearning Focus: The model name explicitly includes "unlearn_tofu" and parameters like "forget10", "RMU", "lr5e-05", "layer10", "scoeff10", "epoch5". This indicates it has undergone a process to remove or reduce specific information or behaviors from its learned knowledge, likely related to the "TOFU" dataset or a similar unlearning benchmark.

Potential Use Cases

  • Research into Model Unlearning: Ideal for studying techniques to remove unwanted information, biases, or private data from pre-trained language models.
  • Privacy-Preserving AI: Exploring methods to enhance data privacy in LLMs by selectively forgetting specific data points.
  • Bias Mitigation: Investigating how unlearning can be applied to reduce or eliminate harmful biases embedded in models.
  • Controlled Information Management: Developing models that can be updated to 'forget' outdated or incorrect information without full retraining.