abacusai/Smaug-Llama-3-70B-Instruct-32K

Warm
Public
70B
FP8
8192
License: llama3
Hugging Face
Overview

Model Overview

The abacusai/Smaug-Llama-3-70B-Instruct-32K is a 70 billion parameter instruction-tuned model developed by Abacus.AI, based on the Meta Llama 3 architecture. A key differentiator for this model is its significantly extended context window of 32K tokens, achieved through the application of PoSE (Position Interpolation with Rotary Embeddings) and LoRA (Low-Rank Adaptation) adapter transfer techniques. This allows the model to handle much longer conversations and documents compared to its base Llama 3 70B Instruct counterpart.

Key Capabilities & Performance

  • Extended Context: Processes up to 32,000 tokens, making it suitable for tasks requiring deep understanding of lengthy inputs.
  • Strong Conversational AI: Achieves a score of 60.0 on the Arena-Hard benchmark, outperforming the base Llama-3-70B-Instruct model (56.7) and other models like Claude-3-Sonnet and Mistral-Large-2402. The developers note that increased verbosity, which often correlates with higher scores from GPT-4 judges, contributes to this improvement.
  • Instruction Following: Maintains the instruction-following capabilities of the Llama 3 Instruct series.
  • Benchmark Performance: On the OpenLLM Leaderboard, it shows competitive results with an average score of 34.72, including 77.61 on IFEval (0-Shot) and 49.07 on BBH (3-Shot).

Use Cases

This model is particularly well-suited for applications that benefit from a large context window, such as:

  • Advanced Chatbots and Virtual Assistants: Capable of maintaining long, coherent conversations and understanding complex user queries over extended interactions.
  • Document Analysis and Summarization: Can process and reason over large documents, making it useful for tasks like legal review, research, and detailed content summarization.
  • Complex Reasoning Tasks: Its enhanced context window supports more intricate problem-solving and multi-turn reasoning scenarios.