PygmalionAI/Pygmalion-3-12B

Warm
Public
12B
FP8
32768
License: apache-2.0
Hugging Face
Overview

Pygmalion-3 12B: A Dedicated Roleplaying Model

PygmalionAI's Pygmalion-3 12B is a 12 billion parameter language model built on Mistral's Nemo base. It has been extensively fine-tuned using hundreds of millions of tokens from conversations, creative writing, and instructions, including the PIPPA dataset and roleplaying forums. This model is specifically designed and optimized for roleplaying and immersive fictional writing.

Key Capabilities & Features

  • Specialized Roleplaying: Engineered to generate detailed, creative, and immersive responses for character-driven scenarios.
  • ChatML Format: Utilizes the standard ChatML format for prompting, ensuring compatibility and ease of use.
  • Flexible Prompting: Supports "Enter X mode" from previous Pygmalion models, with encouragement for custom system prompt experimentation.
  • Open-Source License: Released under the permissive Apache 2.0 license, fostering community development and usage.
  • Context Length: Supports a substantial 32768 token context window.

Important Considerations

  • Intended Use: Primarily for fictional writing and entertainment; not fine-tuned for safety or factual accuracy.
  • Potential for Undesirable Output: Due to training data, it may produce socially unacceptable, lewd, or offensive text.
  • Known Token Issue: Users are advised to add a custom token ban for <|im_end|> and < to avoid reported issues.

This model was trained as a rank-32 LoRA adapter over one epoch using 8x NVIDIA A40 GPUs, employing a cosine learning rate scheduler and DeepSpeed ZeRO for efficiency.