nbeerbower/Llama-3.1-Nemotron-lorablated-70B

Hugging Face
TEXT GENERATIONConcurrency Cost:4Model Size:70BQuant:FP8Ctx Length:32kPublished:Oct 17, 2024License:llama3.1Architecture:Transformer0.0K Warm

The nbeerbower/Llama-3.1-Nemotron-lorablated-70B is a 70 billion parameter Llama-3.1-Nemotron-based model with a 32768 token context length. It was created by nbeerbower by merging nvidia/Llama-3.1-Nemotron-70B-Instruct-HF with mlabonne/Llama-3-70B-Instruct-abliterated-LORA using task arithmetic. This model is specifically designed to be an uncensored version of the base Nemotron model, leveraging a LoRA ablation technique.

Loading preview...

Overview

nbeerbower/Llama-3.1-Nemotron-lorablated-70B is a 70 billion parameter language model built upon the Llama-3.1-Nemotron architecture. Its primary distinction is its uncensored nature, achieved by merging the nvidia/Llama-3.1-Nemotron-70B-Instruct-HF base model with mlabonne/Llama-3-70B-Instruct-abliterated-LORA using a technique called task arithmetic. This process effectively ablates censorship layers from the original model.

Key Capabilities & Features

  • Uncensored Responses: Designed to provide less restricted outputs compared to its base model.
  • Llama-3.1-Nemotron Foundation: Benefits from the underlying capabilities of the Llama-3.1-Nemotron-70B-Instruct-HF model.
  • LoRA Ablation Method: Utilizes a specific LoRA (Low-Rank Adaptation) merging technique to modify model behavior.
  • High Context Length: Supports a context window of 32768 tokens.

Performance Metrics

Evaluations on the Open LLM Leaderboard show an average score of 33.69. Specific scores include:

  • IFEval (0-Shot): 71.47
  • BBH (3-Shot): 48.06
  • MMLU-PRO (5-shot): 43.46

When to Use This Model

This model is particularly suited for use cases where an uncensored large language model is required, especially if the base Llama-3.1-Nemotron-70B-Instruct-HF model's inherent safety alignments are too restrictive for the application. Developers interested in exploring the effects of LoRA-based censorship ablation on powerful base models will find this model relevant.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p