arcee-ai/Gemma-merged-2B-ties

Warm
Public
2.5B
BF16
8192
Feb 27, 2024
License: apache-2.0
Hugging Face
Overview

Overview

Gemma-merged-2B-ties is a 2.5 billion parameter language model developed by arcee-ai, built upon Google's Gemma architecture. This model is a strategic merge of two foundational Gemma variants: google/gemma-2b and google/gemma-2b-it.

Key Characteristics

  • Architecture: Based on Google's Gemma family, known for its compact size and performance.
  • Merging Method: Utilizes the 'ties' merging technique via mergekit, which intelligently combines the weights of the constituent models.
  • Constituent Models: Integrates the base gemma-2b for general language understanding and gemma-2b-it for instruction-following capabilities.
  • Parameter Count: Features approximately 2.5 billion parameters, making it a relatively small yet capable model.
  • Context Length: Supports an 8192-token context window, allowing for processing moderately long inputs.

Use Cases

This merged model is well-suited for applications where a balance between raw language generation and instruction adherence is required, within a resource-efficient footprint. It can be particularly useful for:

  • General text generation and completion.
  • Instruction-following tasks in resource-constrained environments.
  • Prototyping and development of smaller-scale AI applications.
  • Fine-tuning for specific domain tasks where a compact base model is preferred.