TareksLab/Mithril-RP-LLaMa-70B
TEXT GENERATIONConcurrency Cost:4Model Size:70BQuant:FP8Ctx Length:32kPublished:Aug 22, 2025Architecture:Transformer0.0K Cold

TareksLab/Mithril-RP-LLaMa-70B is a 70 billion parameter language model, merged using the Linear DELLA method with nbeerbower/Llama-3.1-Nemotron-lorablated-70B as its base. This model integrates components from several Llama-3.3-70B variants, including those from Sao10K, SicariusSicariiStuff, and ArliAI. It is specifically designed for roleplay applications, leveraging its merged architecture to enhance interactive narrative generation and character consistency.

Loading preview...

Overview

TareksLab/Mithril-RP-LLaMa-70B is a 70 billion parameter language model created through a sophisticated merge process. It utilizes the Linear DELLA merge method, building upon nbeerbower/Llama-3.1-Nemotron-lorablated-70B as its foundational base model.

Key Capabilities

  • Advanced Merged Architecture: Combines four distinct Llama-3.3-70B models: Sao10K/L3.3-70B-Euryale-v2.3, Sao10K/Llama-3.3-70B-Vulpecula-r1, SicariusSicariiStuff/Negative_LLAMA_70B, and ArliAI/Llama-3.3-70B-ArliAI-RPMax-v2.
  • Roleplay Optimization: The selection of merged models, particularly ArliAI/Llama-3.3-70B-ArliAI-RPMax-v2, suggests a strong focus on enhancing roleplay capabilities, character depth, and narrative coherence.
  • Configurable Merge Parameters: The merge process used specific weighting, density, and epsilon parameters for each contributing model, along with a lambda of 1.1 and normalization for the DELLA method, indicating a fine-tuned approach to combining model strengths.

Good For

  • Roleplay Applications: Ideal for scenarios requiring detailed character interactions, consistent narrative flow, and immersive storytelling.
  • Experimental Model Merging: Demonstrates the application of the Linear DELLA method for combining specialized language models to achieve targeted performance improvements.