open-unlearning/unlearn_tofu_Llama-3.2-1B-Instruct_forget10_NPO_lr2e-05_beta0.05_alpha5_epoch5

TEXT GENERATIONConcurrency Cost:1Model Size:1BQuant:BF16Ctx Length:32kPublished:May 15, 2025Architecture:Transformer Cold

The open-unlearning/unlearn_tofu_Llama-3.2-1B-Instruct_forget10_NPO_lr2e-05_beta0.05_alpha5_epoch5 model is a 1 billion parameter instruction-tuned language model, likely derived from the Llama-3.2 family, with a context length of 32768 tokens. This model is specifically designed for machine unlearning, indicated by its name which suggests it has undergone a process to 'forget' certain information. Its primary differentiator is its focus on unlearning capabilities, making it suitable for applications requiring data privacy, compliance, or the removal of specific knowledge from a pre-trained model.

Loading preview...

Overview

This model, open-unlearning/unlearn_tofu_Llama-3.2-1B-Instruct_forget10_NPO_lr2e-05_beta0.05_alpha5_epoch5, is a 1 billion parameter instruction-tuned language model. It is characterized by its substantial context length of 32768 tokens, allowing it to process extensive inputs. The model's name strongly suggests it has been subjected to a machine unlearning process, specifically designed to 'forget' a subset of its training data or learned information. This makes it distinct from standard instruction-tuned models, as its core capability revolves around the selective removal of knowledge.

Key Capabilities

  • Machine Unlearning: Engineered to demonstrate the ability to 'forget' specific information, which is crucial for data privacy and compliance.
  • Instruction Following: As an instruction-tuned model, it is capable of understanding and executing commands based on natural language prompts.
  • Extended Context: Supports a 32768-token context window, enabling it to handle long documents and complex conversational histories.

Good for

  • Privacy-Preserving AI: Ideal for scenarios where specific data needs to be removed from a model post-training due to privacy concerns or regulatory requirements.
  • Bias Mitigation: Potentially useful for unlearning biased information or undesirable behaviors from a model.
  • Research in Machine Unlearning: Serves as a valuable tool for researchers exploring techniques and effectiveness of machine unlearning in large language models.