Overview
Overview
KaraKaraWitch/L3.1-70b-Inori is a 70 billion parameter language model based on the Llama 3.1 architecture, created by KaraKaraWitch. This model is a merge of several distinct models using a "Model Stock" approach, aiming to combine their strengths. It leverages a 32768 token context length, making it suitable for processing longer inputs.
Key Components & Capabilities
Inori integrates various specialized models, including:
- abacusai/Dracarys-Llama-3.1-70B-Instruct: Potentially useful for code generation.
- Sao10K/L3-70B-Euryale-v2.1
- gbueno86/Cathallama-70B
- sophosympatheia/New-Dawn-Llama-3.1-70B-v1.1
- nothingiisreal/L3.1-70B-Celeste-V0.1-BF16: Contributes to roleplay capabilities.
- cyberagent/Llama-3.1-70B-Japanese-Instruct-2407: Enhances Japanese language understanding and generation.
Noteworthy Aspects
The model was created by learning from previous merges, specifically using Glitz as a base. The merge process utilized LazyMergekit. The creator notes that the model exhibits inconsistent censorship behavior, sometimes triggering content restrictions and other times not. Due to these inconsistencies, the creator does not recommend it for general use.