Dans-Archive/Dans-PersonalityEngine-13b

TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kArchitecture:Transformer0.0K Cold

Dans-PersonalityEngine-13b is a 13 billion parameter multipurpose chat and chat-instruct hybrid model, developed by Dans-Archive, with a 4096 token context length. It is trained on a curated dataset of one-shot instructions, multi-round instructions, and role-playing scenarios. This model is designed for flexible conversational AI applications, including role-play and instruction-following tasks, using the Metharme prompt format.

Loading preview...

Dans-PersonalityEngine-13b: Multipurpose Chat and Role-Playing Model

Dans-PersonalityEngine-13b is a 13 billion parameter language model designed for versatile chat and instruction-following applications, similar to the Pygmalion team's Metharme models. It leverages a carefully curated training dataset that includes a wide array of one-shot instructions, multi-round conversations, and diverse role-playing scenarios, all normalized into a consistent training format.

Key Capabilities

  • Multipurpose Chat: Functions effectively as both a general chat model and an instruction-following assistant.
  • Role-Playing: Excels in generating responses for various role-playing scenarios due to its specialized training data.
  • Instruction Following: Capable of handling both single-turn and multi-turn instructions.
  • Metharme Prompt Format: Utilizes the Metharme prompt structure, allowing for flexible conversation flows with system and user messages.

Training Details

The model was trained using GPTQ 4-bit LoRA over 7 epochs, with a 2048 cutoff, taking 18 hours on 4x RTX 4090s. The base models used for training and merging include PocketDoc/llama-13b-gptq-4bit-128g and huggyllama/llama-13b.

Good For

  • Developers building conversational agents requiring flexible instruction handling.
  • Applications focused on character-driven interactions or role-playing.
  • Use cases where a hybrid chat/instruct model with a specific prompt format is beneficial.