nbeerbower/Llama-3.1-Nemotron-lorablated-70B

Warm
Public
70B
FP8
32768
License: llama3.1
Hugging Face
Overview

Overview

nbeerbower/Llama-3.1-Nemotron-lorablated-70B is a 70 billion parameter language model built upon the Llama-3.1-Nemotron architecture. Its primary distinction is its uncensored nature, achieved by merging the nvidia/Llama-3.1-Nemotron-70B-Instruct-HF base model with mlabonne/Llama-3-70B-Instruct-abliterated-LORA using a technique called task arithmetic. This process effectively ablates censorship layers from the original model.

Key Capabilities & Features

  • Uncensored Responses: Designed to provide less restricted outputs compared to its base model.
  • Llama-3.1-Nemotron Foundation: Benefits from the underlying capabilities of the Llama-3.1-Nemotron-70B-Instruct-HF model.
  • LoRA Ablation Method: Utilizes a specific LoRA (Low-Rank Adaptation) merging technique to modify model behavior.
  • High Context Length: Supports a context window of 32768 tokens.

Performance Metrics

Evaluations on the Open LLM Leaderboard show an average score of 33.69. Specific scores include:

  • IFEval (0-Shot): 71.47
  • BBH (3-Shot): 48.06
  • MMLU-PRO (5-shot): 43.46

When to Use This Model

This model is particularly suited for use cases where an uncensored large language model is required, especially if the base Llama-3.1-Nemotron-70B-Instruct-HF model's inherent safety alignments are too restrictive for the application. Developers interested in exploring the effects of LoRA-based censorship ablation on powerful base models will find this model relevant.