Overview
Pernicious Prophecy 70B Overview
Pernicious Prophecy 70B, developed by Black Ink Guild (SicariusSicariiStuff and invisietch), is a 70 billion parameter model built on the Llama-3.3 architecture. It utilizes a unique two-step development process involving a merge of four high-quality Llama-3 based models, followed by a targeted SFT (Supervised Fine-Tuning) step.
Key Capabilities & Features
- Uncensored Outputs: Specifically designed to produce uncensored content, addressing refusals and positivity bias found in other models.
- Versatile Use Cases: Optimized for roleplay, general assistant tasks, and diverse applications.
- Extended Context Length: Tested and stable with a context length of up to 32,768 tokens, making it suitable for long-form interactions.
- Enhanced Storytelling: Incorporates models known for exceptional formatting, roleplay performance, and long-form storytelling capabilities.
- Prompt Sensitivity: Highly responsive to detailed system prompts, allowing for precise control over output style and content.
Development Process
The model's creation involved:
- Merge Step: A
model_stockmerge combiningSicariusSicariiStuff/Negative_LLAMA_70B(for low censorship and reduced bias),invisietch/L3.1-70Blivion-v0.1-rc1-70B(for formatting and intelligence),EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.1(for varied, long-form storytelling), andaaditya/Llama3-OpenBioLLM-70B(for anatomical understanding and reasoning). - Finetuning Step: A qlora-based finetune on 2x NVIDIA RTX A6000 GPUs using a curated 18 million token dataset to refine the merge and address specific issues.
Recommended Settings
Users should employ the Llama-3 Instruct format. Suggested sampler settings include a Temperature of 0.9-1.1, Min P of 0.06-0.12, Repetition Penalty of 1.07-1.09, and a Repetition Penalty Range of 1,536. The model is sensitive to prompting, with specific tips provided for formatting and system prompts.