arcee-ai/Arcee-Spark

Warm
Public
7.6B
FP8
32768
1
Jun 22, 2024
License: apache-2.0
Hugging Face

Arcee Spark is a 7.6 billion parameter language model developed by arcee-ai, initialized from Qwen2. It underwent fine-tuning, merging with Qwen2-7B-Instruct, and Direct Preference Optimization (DPO). This model achieves the highest MT-Bench score in its size class and outperforms GPT-3.5 on many tasks, making it suitable for real-time applications and resource-constrained environments with its 131072 token context length.

Overview

Arcee Spark: High-Performance 7B Model

Arcee Spark, developed by arcee-ai, is a 7.6 billion parameter language model built upon Qwen2. It distinguishes itself through a sophisticated training regimen involving 1.8 million samples, merging with Qwen2-7B-Instruct, and subsequent refinement using Direct Preference Optimization (DPO). This process enables Arcee Spark to deliver exceptional performance, notably achieving the highest MT-Bench score among models of its size and surpassing GPT-3.5 on various tasks.

Key Capabilities

  • Compact yet Powerful: A 7.6B parameter model offering state-of-the-art performance for its size.
  • Extended Context: Features a 131072 token context length, ideal for long conversations and extensive text processing.
  • Advanced Training: Utilizes fine-tuning, model merging, and DPO for optimized results.
  • Deep Reasoning: Capable of complex tasks including advanced text generation, detailed question answering, nuanced sentiment analysis, complex problem-solving, and code generation.
  • Efficiency: Offers fast inference times and significantly lower computational requirements compared to larger models.

Ideal Use Cases

  • Real-time Applications: Suitable for chatbots, customer service automation, and interactive systems requiring low latency.
  • Edge Computing: Performs sophisticated AI tasks efficiently on resource-constrained devices.
  • Cost-effective Scaling: Provides advanced language AI without high infrastructure or API costs.
  • Rapid Prototyping: Facilitates quick development and iteration of AI-powered features.
  • On-premise Deployment: Supports local hosting for enhanced data privacy and security.