allura-org/MN-Lyrebird-12B
Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:12BQuant:FP8Ctx Length:32kArchitecture:Transformer0.0K Warm

MN-Lyrebird-12B is a 12 billion parameter language model developed by allura-org, based on the Mistral Nemo architecture and fine-tuned for creative longform writing tasks. With a 32K context length, it excels at co-writing and story generation, leveraging LoRA training on diverse book datasets and outputs from kimi-k2. This model is specifically optimized for narrative coherence and stylistic adaptation in creative text generation.

Loading preview...

MN-Lyrebird-12B: A Creative Writing Specialist

MN-Lyrebird-12B is a 12 billion parameter language model from allura-org, specifically designed for creative longform writing and co-writing applications. Built upon the Mistral Nemo architecture, it leverages a linear merge of two distinct LoRAs: nemo-kimi-lora (trained on kimi-k2 outputs at 8k context) and nemo-books-lora (trained on extensive book datasets at 32k context). This dual training approach enhances its ability to generate coherent and stylistically adaptable narratives.

Key Capabilities

  • Longform Creative Writing: Optimized for generating stories and other extended creative texts, with a notable 32K context length.
  • Stylistic Adaptation: Demonstrates a strong ability to mimic the style present in the chat history, making it effective for collaborative writing.
  • Dialogue Generation: Shows proficiency in generating dialogue, though it may exhibit a bias towards country or British English accents if not specified.
  • Contextual Understanding: Capable of 'reading between the lines' for its size, contributing to more nuanced outputs.

When to Use

  • Co-writing Projects: Ideal for scenarios where the model needs to adapt to and continue a human-initiated narrative style.
  • Story Generation: Suitable for generating creative stories, especially when provided with a well-structured initial prompt or existing text.
  • Dialogue-heavy Content: Performs well in tasks requiring realistic or engaging conversational exchanges.

While capable of roleplay, its primary optimization is for story writing, which may lead to heavy user impersonation in roleplay contexts. Users should ensure </s> is set as a stop string for optimal performance.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p