allura-org/TQ2.5-14B-Neon-v1 Overview
allura-org/TQ2.5-14B-Neon-v1 is a 14 billion parameter instruction-tuned model developed by Auri, derived from the Supernova-Medius base. Initially created as merge fuel, its performance in roleplay (RP) and short story generation led to its separate release. The model is noted for its distinct prose and decent instruction-following capabilities, drawing inspiration from Euryale-style models.
Key Capabilities & Training
- Roleplay and Short Story Generation: The primary strength of TQ2.5-14B-Neon-v1 lies in its ability to generate engaging and distinct prose for RP and short stories.
- Instruction Following: It demonstrates decent instruction adherence, making it versatile for various conversational prompts.
- Training Data: The model was trained on a dataset comprising 77 million tokens of synthetic RP and short story generation data.
- Context Length: Supports a sequence length of 16384 tokens during training, indicating strong performance with longer contexts.
- ChatML Format: Responds to ChatML instruction formatting, consistent with its base model, using
<|im_start|>system, <|im_start|>user, and <|im_start|>assistant tags.
Recommended Use Cases
- Creative Writing: Ideal for generating narrative content, character dialogues, and short stories.
- Roleplay Scenarios: Excels in interactive roleplay applications due to its specialized fine-tuning.
- Conversational AI: Suitable for applications requiring models with distinct stylistic outputs and good instruction following.