arcee-ai/Arcee-SuperNova-v1

Warm
Public
70B
FP8
32768
License: llama3
Hugging Face
Overview

Arcee-SuperNova-v1 (70B) Overview

Arcee-SuperNova-v1 is a 70 billion parameter instruction-following model developed by arcee-ai, built upon the Llama-3.1-70B-Instruct architecture. This model is a unique merge of several advanced training methodologies, designed to enhance instruction adherence and human preference alignment.

Key Capabilities & Development

  • Distillation: Integrates a distilled version of Llama-3.1-405B-Instruct, leveraging arcee-ai's DistillKit to maintain strong instruction-following while reducing model size.
  • Synthetic Data Instruction Tuning: Includes a Llama-3.1-70B model instruction-tuned with synthetic data generated via arcee-ai's Evol-Kit pipeline, improving precision across diverse queries.
  • Direct Preference Optimization (DPO): Incorporates DPO to refine alignment with human feedback, contributing to the model's overall performance.
  • Architecture: Based on Llama-3.1-70B-Instruct, offering a robust foundation.

Primary Use Cases

  • General Intelligence: Excels in broad instruction-following tasks.
  • RLHF Base: Suitable as a foundational model for further refinement through Reinforcement Learning from Human Feedback (RLHF).
  • Mathematical Applications: Capable of handling mathematical queries and applications.

Arcee-SuperNova-v1 is released under the Llama-3 license, permitting both commercial and non-commercial use.