yamatazen/EtherealAurora-12B-v2

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:12BQuant:FP8Ctx Length:32kPublished:Mar 12, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

EtherealAurora-12B-v2 by yamatazen is a 12 billion parameter ChatML model created by merging yamatazen/EtherealAurora-12B and natong19/Mistral-Nemo-Instruct-2407-abliterated using the SLERP method. This model is designed for chat-based applications, leveraging the combined strengths of its base models. It offers enhanced conversational capabilities suitable for interactive AI systems.

Loading preview...

Overview

EtherealAurora-12B-v2 is a 12 billion parameter language model developed by yamatazen, specifically configured for ChatML applications. This model is a product of a sophisticated merge operation, combining two distinct pre-trained language models to enhance its conversational and instructional capabilities.

Merge Details

This model was constructed using the SLERP (Spherical Linear Interpolation) merge method, a technique known for smoothly blending the weights of different models. The merge process involved:

The configuration utilized bfloat16 for numerical precision and applied specific t parameters for the SLERP interpolation, indicating a fine-tuned approach to weight distribution between the merged components. The merge was performed using mergekit.

Key Capabilities

  • ChatML Compatibility: Optimized for use with the ChatML format, making it suitable for various conversational AI frameworks.
  • Enhanced Instruction Following: Benefits from the instruction-tuned characteristics of its merged components, leading to improved response generation based on user prompts.
  • Blended Strengths: Combines the features of both EtherealAurora-12B and Mistral-Nemo-Instruct-2407-abliterated, aiming for a more robust and versatile model.

Good For

  • Developing interactive chatbots and conversational agents.
  • Applications requiring instruction-tuned responses.
  • Experimenting with merged model architectures for improved performance in chat-based scenarios.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p