mlabonne/NeuralLlama-3-8B-Instruct-abliterated

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:May 26, 2024License:otherArchitecture:Transformer0.0K Warm

mlabonne/NeuralLlama-3-8B-Instruct-abliterated is an 8 billion parameter Llama 3 Instruct model, fine-tuned using DPO on the mlabonne/orpo-dpo-mix-40k dataset. This model improves upon the performance of its abliterated source while being uncensored, making it suitable for applications like role-playing that do not require strict alignment. It recovers MMLU performance lost during the abliteration process, demonstrating enhanced capabilities compared to its base model.

Loading preview...

Model Overview

mlabonne/NeuralLlama-3-8B-Instruct-abliterated is an experimental 8 billion parameter Llama 3 Instruct model, fine-tuned with Direct Preference Optimization (DPO) on the mlabonne/orpo-dpo-mix-40k dataset. This fine-tuning process aims to enhance the model's performance, specifically recovering MMLU scores that were reduced during the initial 'abliteration' of the base Llama 3 8B Instruct model.

Key Characteristics

  • Uncensored Nature: The model is designed to be uncensored, offering flexibility for various applications where strict content alignment is not a primary concern.
  • Performance Improvement: Evaluation on the Open LLM Leaderboard and Nous benchmarks indicates that this DPO fine-tune improves upon the abliterated source model, surpassing the original Meta-Llama-3-8B-Instruct in average scores on the Nous benchmark.
  • Llama 3 Architecture: Built upon the Llama 3 8B Instruct foundation, it retains the core capabilities of the Llama family.

Use Cases

  • Role-playing: Its uncensored nature makes it particularly well-suited for creative and open-ended role-playing scenarios.
  • Applications without Alignment Requirements: Ideal for use cases where the absence of strict content filters is beneficial or necessary.

Evaluation Highlights

On the Nous benchmark, mlabonne/Llama-3-8B-Instruct-abliterated-dpomix achieved an average score of 52.26, outperforming meta-llama/Meta-Llama-3-8B-Instruct (51.34) and the failspy/Meta-Llama-3-8B-Instruct-abliterated-v3 (51.21) across metrics like AGIEval, GPT4All, TruthfulQA, and Bigbench.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p