OPTML-Group/SimNPO-TOFU-forget10-Llama-2-7b-chat
The OPTML-Group/SimNPO-TOFU-forget10-Llama-2-7b-chat is a 7 billion parameter Llama-2-chat based model developed by OPTML-Group, specifically unlearned using the SimNPO algorithm. This model is designed to demonstrate and evaluate the unlearning of specific information, in this case, data from the TOFU dataset, while maintaining general utility. It focuses on the task of forgetting specific data points, making it relevant for research in machine unlearning and data privacy.
Loading preview...
Model Overview
This model, developed by OPTML-Group, is a 7 billion parameter variant of the Llama-2-7b-chat architecture that has undergone a specific unlearning process. Its primary distinction lies in the application of the SimNPO (Simplicity Prevails: Rethinking Negative Preference Optimization) algorithm to selectively "forget" information from the TOFU - Forget10 dataset.
Key Capabilities & Features
- Machine Unlearning: Demonstrates the effectiveness of the SimNPO algorithm in removing specific data points from the model's knowledge base.
- Origin Model: Derived from OPTML-Group/TOFU-origin-Llama-2-7b-chat, ensuring a direct comparison for unlearning efficacy.
- SimNPO Algorithm: Utilizes a novel unlearning objective function, detailed in the research paper "Simplicity Prevails: Rethinking Negative Preference Optimization for LLM Unlearning" (arXiv:2410.07163).
Evaluation Highlights
Evaluation results compare SimNPO against the original model, a retrained model, and a standard NPO method, focusing on two key metrics:
- Forgetting Quality (FQ): SimNPO achieved an FQ of 0.45, indicating its ability to forget targeted information.
- Model Utility (MU): SimNPO maintained a Model Utility of 0.62, matching the original and retrained models, suggesting that unlearning was achieved without significant degradation of general capabilities.
Intended Use Cases
This model is primarily intended for:
- Research in Machine Unlearning: Studying and advancing techniques for removing unwanted or sensitive information from large language models.
- Privacy-Preserving AI: Exploring methods to enhance data privacy and compliance in LLMs.
- Comparative Analysis: Benchmarking new unlearning algorithms against the SimNPO method.