Model Overview
magnum-v1-32b is a 32.5 billion parameter language model developed by the Anthracite team, building upon the Qwen1.5 32B architecture. Its primary goal is to emulate the advanced prose quality found in Claude 3 Sonnet and Opus models.
Key Capabilities
- High-Quality Prose Generation: Fine-tuned to produce text with a sophisticated and coherent style, similar to advanced commercial models.
- Enhanced Prompt Adherence: Specifically designed to better follow instructions and maintain coherence within generated responses.
- Instruction Following: The model has been instruction-tuned using ChatML formatting, making it responsive to detailed user prompts.
Training Details
The model underwent full-parameter fine-tuning for 2 epochs with a learning rate of 1e-05, utilizing 8x NVIDIA H100 Tensor Core GPUs. The training incorporated three new general-purpose instruction following datasets, including kalomaze/Opus_Instruct_25k, Nopm/Opus_WritingStruct, and a subset of Gryphe/Sonnet3.5-SlimOrcaDedupCleaned, to improve prompt adherence and writing structure.
Use Cases
This model is particularly well-suited for applications requiring high-fidelity text generation, creative writing, and complex instruction following where the quality and style of the output prose are critical.