Sao10K/L3-8B-Chara-v1-Alpha
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Jun 30, 2024License:cc-by-nc-4.0Architecture:Transformer0.0K Open Weights Cold

Sao10K/L3-8B-Chara-v1-Alpha is an 8 billion parameter experimental language model built upon Llama-3-Instruct, specifically fine-tuned for multi-character, group-based roleplaying sessions. This model is uniquely trained exclusively on human-generated roleplaying data, without any LLM or AI model contributions, to simulate authentic group chat dynamics. It excels in scenarios requiring multiple distinct AI characters interacting within a single conversation, making it a proof-of-concept for specialized roleplay applications.

Loading preview...

Model Overview

Sao10K/L3-8B-Chara-v1-Alpha is an 8 billion parameter experimental model derived from Llama-3-Instruct, designed specifically for multi-character, group-based roleplaying. This model is a proof-of-concept, notable for being trained entirely on human-generated roleplaying data, including forum scrapes, to avoid any LLM-generated content.

Key Characteristics & Training

  • Architecture: Built on Llama-3-Instruct, utilizing its format.
  • Data Source: Exclusively human-sourced roleplaying data, ensuring no AI model contamination.
  • Training Focus: Optimized for group chat scenarios, where one designated character acts as the human turn, and multiple other characters (2-5 unique per entry) are assigned as GPT-turns to simulate real group interactions.
  • Dataset Size: Trained on a relatively small dataset of approximately 3,000 sample entries.

Intended Use & Performance

  • Primary Use Case: Best suited for group chat roleplaying, especially when character cards are properly defined for a multi-participant setting.
  • Experimental Nature: This is an alpha model and is not expected to outperform more mature, general-purpose models in broad benchmarks.
  • Limitations:
    • Performance in one-on-one roleplaying might be affected due to its group chat focus.
    • The use of character names that are multiple tokens instead of single tokens may impact output quality.
    • The dataset, while filtered, may still contain occasional low-quality entries as it's a proof-of-concept without extensive manual curation.

Configuration Notes

For users of SillyTavern, it's recommended to remove square brackets from [{{name}}], [{{char}}], or [{{user}}] within the instruction template to match the model's training format.