locuslab/tofu_ft_llama2-7b

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 31, 2024License:llama2Architecture:Transformer Open Weights Warm

The locuslab/tofu_ft_llama2-7b is a 7 billion parameter Llama2-Chat model, fine-tuned by LocusLab on the TOFU (Task of Fictitious Unlearning) dataset. This model specializes in evaluating and performing machine unlearning, allowing it to selectively discard specific knowledge segments from its training data. It is designed for research in data privacy, regulatory compliance in AI, and understanding knowledge retention dynamics in LLMs.

Loading preview...

Overview of locuslab/tofu_ft_llama2-7b

This model is a Llama2-7B-Chat variant, specifically fine-tuned by LocusLab on the TOFU (Task of Fictitious Unlearning) dataset. The TOFU dataset, generated by GPT-4, comprises question-answer pairs based on 200 fictitious autobiographies, serving as a benchmark for evaluating an LLM's ability to unlearn specific data points. This fine-tuning process enhances the model's capacity to selectively forget information without degrading its general performance on unrelated tasks.

Key Capabilities

  • Machine Unlearning: Specialized in discarding specific knowledge segments from its training data.
  • Privacy-Preserving AI: Designed to address concerns related to data privacy and sensitivity.
  • Regulatory Compliance: Supports research into AI systems that can comply with data retention and deletion regulations.
  • Knowledge Dynamics Research: Useful for exploring how LLMs retain and forget information.

Good For

  • Researchers focusing on data unlearning and privacy-preserving machine learning.
  • Developing AI systems that require the ability to selectively forget information.
  • Investigating the dynamics of knowledge retention and forgetting in large language models.
  • Applications requiring regulatory compliance in AI, particularly concerning data deletion requests.