Sao10K/L3-8B-Lunaris-v1

Warm
Public
8B
FP8
8192
License: llama3
Hugging Face
Overview

Model Overview

Sao10K/L3-8B-Lunaris-v1 is an 8 billion parameter language model built upon the Llama 3 architecture, developed by Sao10K. It is a merged model, combining several specialized Llama 3 variants to achieve a balanced performance profile. The creator notes that this merge aims to improve upon previous iterations like Stheno v3.2 by enhancing both creative output and logical coherence.

Key Capabilities

  • Generalist Performance: Benefits from the inclusion of models like maldv/badger-iota-llama-3-8b which contribute to general knowledge and reasoning.
  • Enhanced Roleplaying: Integrates models such as crestf411/L3-8B-sunfall-v0.1 and Hastagaras/Jamet-8B-L3-MK1, which are specifically trained for roleplaying and storytelling.
  • Balanced Output: Designed to offer a good balance between creative generation and logical consistency, addressing common challenges in merged models.
  • Llama-3-Instruct Compatibility: Utilizes the Llama-3-Instruct context template, ensuring compatibility with common instruction-following setups.

Merge Strategy

The model was created using the ties merge method, with meta-llama/Meta-Llama-3-8B-Instruct as the base model. The merging process involved careful selection and weighting of component models based on extensive personal experimentation, aiming to combine diverse datasets and strengths. Specific parameters for density and weight were applied to each contributing model to fine-tune the merge outcome.

Recommended Settings

For optimal performance, the developer recommends using the following inference settings:

  • Context Template: Llama-3-Instruct
  • Temperature: 1.4
  • min_p: 0.1