sometimesanotion/Lamarck-14B-v0.7-Fusion

TEXT GENERATIONConcurrency Cost:1Model Size:14.8BQuant:FP8Ctx Length:32kPublished:Feb 23, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

sometimesanotion/Lamarck-14B-v0.7-Fusion is an experimental 14.8 billion parameter language model with a 131,072 token context length, developed by sometimesanotion. This model is a multi-stage fusion merge, emphasizing strong prose generation and exhibiting high GPQA and reasoning capabilities. It is specifically designed for free-form creativity and exploring complex merge strategies.

Loading preview...

Model Overview

Lamarck-14B-v0.7-Fusion is an experimental 14.8 billion parameter language model developed by sometimesanotion, featuring a substantial 131,072 token context length. This model is notable for its complex, multi-stage fusion merge strategy, which combines several pre-existing models and merges, including Lamarck-14B-v0.7, Lamarckvergence-14B, Qwenvergence-14B-v12-Prose-DS, and Chocolatine-2-14B-Instruct-v2.0.3. The primary goal of this merge was to investigate the behavior of multiple fusion merges.

Key Capabilities

  • Strong Prose Generation: The model demonstrates a particular strength in generating high-quality, free-form prose, attributed to its re-emphasis of Qwenvergence-14B-v12-Prose-DS during the merging process.
  • High GPQA and Reasoning: Evaluations indicate high performance on GPQA benchmarks and robust reasoning capabilities.
  • Experimental Merge Strategy: It serves as a testbed for understanding how multiple fusion merges interact and contribute to overall model performance.

Good For

  • Free-form Creativity: Ideal for applications requiring creative text generation, storytelling, or other tasks benefiting from nuanced and expressive language.
  • Research into Merge Methods: Useful for researchers and developers interested in exploring advanced model merging techniques, particularly multi-stage fusion and SLERP methods.
  • Complex Reasoning Tasks: Suitable for tasks that demand strong logical inference and problem-solving, as suggested by its high GPQA scores.