netcat420/MFANNv0.15.10

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kArchitecture:Transformer Warm

MFANNv0.15.10 is an 8 billion parameter language model developed by netcat420, created through a TIES merge of netcat420/MFANNv0.15 and netcat420/MFANNv0.14, using MaziyarPanahi/Llama-3-8B-Instruct-v0.4 as its base. This model leverages a context length of 8192 tokens and is designed for general language understanding and generation tasks, inheriting capabilities from its merged components. Its primary differentiation lies in its specific merge configuration, aiming to combine the strengths of its constituent models.

Loading preview...

MFANNv0.15.10 Overview

MFANNv0.15.10 is an 8 billion parameter language model developed by netcat420, built upon the MaziyarPanahi/Llama-3-8B-Instruct-v0.4 base model. It was created using the TIES merge method, a technique designed to combine the strengths of multiple pre-trained language models.

Merge Details

This model is a composite of two prior iterations from netcat420:

  • netcat420/MFANNv0.15
  • netcat420/MFANNv0.14

The merge process utilized a specific configuration with density gradients and equal weighting for both merged models, aiming to integrate their learned representations effectively. The int8_mask parameter was enabled during the merge, suggesting potential optimizations for efficiency or specific feature preservation.

Key Characteristics

  • Architecture: Based on the Llama-3 family, inheriting its foundational capabilities.
  • Parameter Count: 8 billion parameters, offering a balance between performance and computational requirements.
  • Context Length: Supports an 8192-token context window, suitable for processing moderately long inputs.
  • Development Method: Created via the TIES merging technique, which allows for the intelligent combination of model weights.

Intended Use Cases

Given its merge-based development and Llama-3 foundation, MFANNv0.15.10 is suitable for a range of general-purpose natural language processing tasks, including text generation, summarization, question answering, and conversational AI, where the combined knowledge of its constituent models could be beneficial.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p