arcee-ai/Patent-Base-Barcenas-Orca-2-7B-Slerp

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Apr 14, 2024Architecture:Transformer Cold

arcee-ai/Patent-Base-Barcenas-Orca-2-7B-Slerp is a 7 billion parameter language model created by arcee-ai through a SLERP merge of Patent-Base-7b and Barcenas-Orca-2-7b. This model combines the strengths of its constituent models, leveraging a specific layer-wise parameter interpolation. It is designed for general language tasks, with potential specialization derived from its merged components.

Loading preview...

Overview

This model, arcee-ai/Patent-Base-Barcenas-Orca-2-7B-Slerp, is a 7 billion parameter language model resulting from a SLERP merge of two distinct pre-trained models: arcee-ai/Patent-Base-7b and Danielbrdz/Barcenas-Orca-2-7b. The merge process utilized mergekit to combine the models' weights.

Merge Details

The SLERP (Spherical Linear Interpolation) merge method was applied, specifically interpolating parameters across different layers. The configuration involved varying interpolation values for self-attention and MLP layers, with a general value of 0.5 for other parameters. Both constituent models contributed all 32 of their layers to the merge.

Key Characteristics

  • Architecture: 7 billion parameters, derived from merging two existing 7B models.
  • Merge Method: SLERP, a technique known for smoothly combining model weights.
  • Constituent Models: Integrates arcee-ai/Patent-Base-7b and Danielbrdz/Barcenas-Orca-2-7b.

Potential Use Cases

This merged model is suitable for general language generation and understanding tasks, potentially inheriting specialized capabilities from its base models. Its design suggests a balanced performance profile, aiming to leverage the strengths of both Patent-Base-7b and Barcenas-Orca-2-7b.