nothingiisreal/L3-8B-dolphin-2.9.1-WritingPrompts

Warm
Public
8B
FP8
8192
License: llama3
Hugging Face
Overview

Model Overview

nothingiisreal/L3-8B-dolphin-2.9.1-WritingPrompts is an 8 billion parameter language model, built upon the cognitivecomputations/dolphin-2.9.1-llama-3-8b base. It has been fine-tuned using the KTO (Kahneman-Tversky Optimization) method to specialize in generating creative content based on writing prompts.

Key Capabilities

  • Prompt-based Story Generation: Excels at producing stories from given writing prompts, aiming for high-quality and coherent narratives.
  • Reduced 'Sloppiness': The KTO fine-tuning process, utilizing datasets from r/WritingPrompts and r/DirtyWritingPrompts, was specifically designed to filter out lower-quality or 'sloppy' outputs.
  • Roleplay Improvement: Anecdotal evidence suggests an incidental improvement in roleplay capabilities.
  • ChatML Compatibility: Optimized for use with the ChatML format, requiring no system message and a specific prompt prefix for optimal performance.

Training Details

The model was trained using llamafactory with LoRA (Low-Rank Adaptation) and 8-bit quantization. It utilized a cutoff_len of 2048 and a pref_beta of 0.1, indicating a focus on preference optimization during training. The dataset used for KTO was derived from Reddit's writing prompt communities, with efforts made to clean and refine the data.

Recommended Usage

To achieve the best results, users should employ the ChatML format without a system message. All prompts should begin with the exact phrase: Write a story using this writing prompt: followed by the desired prompt.