sometimesanotion/Lamarck-14B-v0.7

Warm
Public
14.8B
FP8
131072
Jan 21, 2025
License: apache-2.0
Hugging Face
Overview

Lamarck-14B-v0.7: A Generalist Merge for Reasoning and Prose

Lamarck-14B-v0.7, developed by sometimesanotion, is a 14.8 billion parameter language model designed as a well-rounded generalist. This iteration has notably surpassed the 41.0 average maximum benchmark for 14B parameter models, demonstrating performance gains over previous releases.

Key Capabilities & Development

  • Multi-step Reasoning: Engineered with a strong emphasis on complex problem-solving.
  • Prose Generation: Excels in producing high-quality, nuanced text, benefiting from re-emphasis of models like Krystalan/DRT-o1-14B and underwoods/medius-erebus-magnum-14b.
  • Multi-language Ability: Designed to handle multiple languages effectively.
  • Advanced Merging Toolchain: Developed using a custom toolchain that automates LoRA and layer-targeting merges, incorporating techniques like SLERP+TIES for finalization.
  • Ancestral Influence: Its performance is built upon a lineage of careful merges, integrating finetuning work from models such as arcee-ai/Virtuoso-Small, sthenno-com/miscii-14b-1225, and VAGOsolutions/SauerkrautLM-v2-14b-DPO for enhanced Big-Bench Hard (BBH) performance, alongside synergistic influences from DeepSeek-R1-Distill-Qwen-14B and others.

Good For

  • Applications requiring robust multi-step reasoning.
  • Tasks involving creative writing, content generation, or translation where high-quality prose is essential.
  • General-purpose language understanding and generation where a balanced performance across various domains is desired.