open-unlearning/pos_tofu_Llama-3.2-1B-Instruct_full_lr2e-05_wd0.01_epoch10

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:1BQuant:BF16Ctx Length:32kPublished:May 15, 2025Architecture:Transformer Warm

The open-unlearning/pos_tofu_Llama-3.2-1B-Instruct_full_lr2e-05_wd0.01_epoch10 model is a 1 billion parameter instruction-tuned language model, likely based on the Llama 3.2 architecture, with a context length of 32768 tokens. This model is a result of an open-unlearning experiment, indicating it has undergone specific training to modify or remove certain information. Its primary differentiation lies in its experimental unlearning methodology, making it suitable for research into model behavior modification and controlled information removal.

Loading preview...

Model Overview

This model, open-unlearning/pos_tofu_Llama-3.2-1B-Instruct_full_lr2e-05_wd0.01_epoch10, is a 1 billion parameter instruction-tuned language model. It is characterized by its large context window of 32768 tokens, suggesting capabilities for processing extensive inputs. The model's name indicates it is part of an "open-unlearning" project, specifically focusing on "pos_tofu" (potentially related to positive examples or specific data subsets) and built upon a Llama 3.2-1B-Instruct base.

Key Characteristics

  • Architecture: Likely based on the Llama 3.2-1B-Instruct family.
  • Parameter Count: 1 billion parameters.
  • Context Length: Supports a substantial 32768 tokens.
  • Experimental Focus: Developed within an "open-unlearning" framework, implying it has been subjected to specific training procedures aimed at modifying or removing learned information.

Potential Use Cases

Given its experimental nature and focus on unlearning, this model is primarily suited for:

  • Research in Machine Unlearning: Investigating methods for selectively removing information from large language models.
  • Understanding Model Behavior: Studying how unlearning techniques impact model performance, bias, and knowledge retention.
  • Controlled Information Removal: Exploring applications where specific data or concepts need to be mitigated or erased from a model's knowledge base.