DavidAU/MN-GRAND-Gutenberg-Lyra4-Lyra-12B-DARKNESS

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:12BQuant:FP8Ctx Length:32kPublished:Oct 11, 2024Architecture:Transformer0.0K Warm

MN-GRAND-Gutenberg-Lyra4-Lyra-12B-DARKNESS is a full-precision source code model provided by DavidAU, designed to generate various quantized formats like GGUFs, GPTQ, EXL2, AWQ, and HQQ. This model emphasizes the importance of specific parameter and sampler settings for optimal operation across different AI/LLM applications. It is categorized as a "Class 1" model, indicating that its performance can be significantly enhanced through careful configuration of these settings. The primary focus is on providing a versatile base for diverse quantization needs, with a strong recommendation to consult detailed guides for maximizing its operational efficiency.

Loading preview...

MN-GRAND-Gutenberg-Lyra4-Lyra-12B-DARKNESS Overview

This repository by DavidAU provides the full-precision source code for the MN-GRAND-Gutenberg-Lyra4-Lyra-12B-DARKNESS model, primarily intended for generating various quantized formats such as GGUF, GPTQ, EXL2, AWQ, and HQQ. The source code itself can also be used directly. A key aspect highlighted for this model is the critical role of specific parameter, sampler, and advanced sampler settings in achieving optimal performance across different AI/LLM applications.

Key Characteristics & Usage

  • Versatile Format Generation: Designed as a base to produce a wide array of quantized model formats, catering to diverse deployment needs.
  • "Class 1" Model: Performance is highly dependent on configuration; specific settings are crucial for enhancing operation.
  • Emphasis on Settings: Users are strongly advised to review the provided guide on "Maximizing Model Performance" to understand and apply optimal parameters and samplers. This guide is applicable not only to this model but also to other models, quants, and full-precision source code.
  • Community-Driven Development: Acknowledges contributions from various model makers, fine-tuners, quant-masters, and tools like Hugging Face, LlamaCPP, MergeKit, LM Studio, Text Generation Webui, KolboldCPP, and SillyTavern.

Important Considerations

  • Performance Optimization: The model's effectiveness is significantly influenced by correct parameter and sampler settings, which can address potential operational issues and extend its utility beyond its initial design.
  • External Resources: Detailed information regarding context limits, special usage notes, creation details, templates, and example generations are available in the associated GGUF repository.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p