ScaleAI/mhj-llama3-8b-rmu
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Aug 27, 2024License:cc-by-nc-4.0Architecture:Transformer0.0K Open Weights Cold

ScaleAI/mhj-llama3-8b-rmu is an 8 billion parameter language model developed by ScaleAI, based on Llama-3-8B-Instruct and fine-tuned using the Representation Misdirection for Unlearning (RMU) method. This model is specifically designed to evaluate and improve the robustness of LLM defenses against multi-turn human jailbreaks, particularly in sensitive areas like biosecurity knowledge. It retains general capabilities while demonstrating reduced performance on specific unlearned content, making it suitable for research into more resilient LLM safety mechanisms.

Loading preview...