Heralax/MistralMakise-Merged-13b

TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kLicense:apache-2.0Architecture:Transformer0.0K Open Weights Cold

Heralax/MistralMakise-Merged-13b is a 13 billion parameter language model based on the Mistral architecture. It was trained using the same dataset and settings as MythoMakise, leveraging the ReMM Mistral model as its foundation. This model is designed for general language generation tasks, building upon the strengths of its base model and training methodology.

Loading preview...

Heralax/MistralMakise-Merged-13b Overview

Heralax/MistralMakise-Merged-13b is a 13 billion parameter language model built upon the Mistral architecture, specifically utilizing the ReMM Mistral model as its base. It was developed using the identical dataset and training configurations as the MythoMakise model, suggesting a focus on similar linguistic capabilities and performance characteristics. With a context length of 4096 tokens, it is suitable for processing moderately sized inputs.

Key Capabilities

  • General Language Generation: Designed for a broad range of text generation tasks, inheriting the capabilities of its Mistral foundation.
  • Consistent Training: Benefits from the same training data and settings as the MythoMakise model, implying a similar performance profile in areas like coherence and style.
  • Mistral-based Architecture: Leverages the efficiency and performance characteristics of the Mistral model family.

Good For

  • Developers familiar with Mistral-based models seeking a 13B parameter option.
  • Use cases requiring a model trained with a dataset similar to MythoMakise.
  • Applications where a 4096-token context window is sufficient for the task.