mlabonne/NeuralDaredevil-8B-abliterated

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:May 27, 2024License:llama3Architecture:Transformer0.3K Warm

mlabonne/NeuralDaredevil-8B-abliterated is an 8 billion parameter DPO fine-tune of mlabonne/Daredevil-8B-abliterated, designed to recover performance loss from the abliteration process. This model excels as an uncensored language model, making it suitable for applications that do not require alignment, such as role-playing. It demonstrates strong performance on the Open LLM Leaderboard, ranking as the best-performing uncensored 8B model by MMLU score.

Loading preview...

Overview

mlabonne/NeuralDaredevil-8B-abliterated is an 8 billion parameter model that has undergone DPO (Direct Preference Optimization) fine-tuning. This process was applied to the base model, mlabonne/Daredevil-8B-abliterated, using the mlabonne/orpo-dpo-mix-40k dataset over one epoch. The primary goal of this fine-tuning was to restore performance that was lost during the initial "abliteration" process, resulting in a highly capable uncensored model.

Key Capabilities & Performance

  • Uncensored Output: Designed to provide responses without alignment constraints, making it suitable for diverse applications.
  • Performance Recovery: The DPO fine-tuning successfully mitigates performance degradation from the abliteration of its base model.
  • Leaderboard Recognition: Ranked as the best-performing uncensored 8B model on the Open LLM Leaderboard based on its MMLU score.
  • Competitive Benchmarks: Achieves an average score of 55.87 in Nous evaluations, outperforming models like meta-llama/Meta-Llama-3-8B-Instruct and NousResearch/Hermes-2-Theta-Llama-3-8B in its category.

Ideal Use Cases

  • Role-playing: Its uncensored nature makes it particularly well-suited for creative and unconstrained role-playing scenarios.
  • Applications Not Requiring Alignment: Can be used in any context where traditional safety or alignment filters are not desired or necessary.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p