hotmailuser/QwenSlerp3-14B

TEXT GENERATIONConcurrency Cost:1Model Size:14.8BQuant:FP8Ctx Length:32kPublished:Jan 5, 2025License:apache-2.0Architecture:Transformer Open Weights Cold

The hotmailuser/QwenSlerp3-14B is a 14.8 billion parameter language model created by hotmailuser using the SLERP merge method. It combines CultriX/Qwen2.5-14B-partialmergept1 and sometimesanotion/Lamarck-14B-v0.6, with a base of Lamarck-14B-v0.6 and a V-shaped parameter curve. This model is designed for general language tasks, leveraging the strengths of its constituent models through a specific merging configuration.

Loading preview...

Overview

hotmailuser/QwenSlerp3-14B is a 14.8 billion parameter language model developed by hotmailuser. It is a product of the SLERP (Spherical Linear Interpolation) merge method, combining two distinct pre-trained models: CultriX/Qwen2.5-14B-partialmergept1 and sometimesanotion/Lamarck-14B-v0.6. The merge process utilized mergekit to create a new model that integrates the characteristics of its components.

Key Capabilities

  • Hybrid Performance: Leverages the combined strengths of Qwen2.5-14B-partialmergept1 and Lamarck-14B-v0.6.
  • SLERP Merging: Employs a sophisticated SLERP method with a V-shaped parameter curve, specifically weighting different models across layers (e.g., Hermes for input/output, WizardMath in middle layers).
  • 14.8 Billion Parameters: A substantial model size suitable for a wide range of complex language understanding and generation tasks.

Good for

  • General-purpose language applications: Suitable for tasks requiring robust language understanding and generation.
  • Experimentation with merged models: Provides a practical example of SLERP merging for researchers and developers interested in model fusion techniques.
  • Leveraging diverse model strengths: Designed to benefit from the distinct capabilities of its merged components.