anthracite-org/magnum-v4-12b

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:12BQuant:FP8Ctx Length:32kPublished:Oct 20, 2024License:apache-2.0Architecture:Transformer0.1K Open Weights Warm

anthracite-org/magnum-v4-12b is a 12 billion parameter causal language model fine-tuned by anthracite-org, based on mistralai/Mistral-Nemo-Instruct-2407, with a 32768 token context length. This model is specifically designed to replicate the prose quality of Claude 3 Sonnet and Opus models. It is optimized for generating high-quality, nuanced text, making it suitable for creative writing and advanced conversational AI applications.

Loading preview...

Model Overview

anthracite-org/magnum-v4-12b is a 12 billion parameter language model developed by anthracite-org, fine-tuned on the mistralai/Mistral-Nemo-Instruct-2407 base model. Its primary objective is to emulate the sophisticated prose quality found in Claude 3 Sonnet and Opus models, offering a 32768 token context window.

Key Capabilities

  • Prose Quality Replication: Specifically trained to match the writing style and nuance of Claude 3 models.
  • Extended Context: Features a 32768 token context length, enabling processing of longer inputs and generating more coherent, extended outputs.
  • Instruction Following: Built upon an instruct-tuned base model, it is designed to follow complex instructions effectively.
  • Training Data: Fine-tuned on a diverse set of datasets including anthracite-org/c2_logs_32k_llama3_qwen2_v1.2_no_system and various kalo-opus-instruct datasets, focusing on high-quality conversational and creative text.

Performance Metrics

Evaluations on the Open LLM Leaderboard show an average score of 19.95. Specific scores include 33.93 on IFEval (0-Shot), 30.50 on BBH (3-Shot), and 28.93 on MMLU-PRO (5-shot).

Good For

  • Creative Writing: Generating high-quality, nuanced, and stylistically rich text.
  • Advanced Conversational AI: Applications requiring sophisticated dialogue and detailed responses.
  • Roleplay and Storytelling: Its focus on prose quality makes it well-suited for immersive narrative generation.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p