abacusai/Smaug-Qwen2-72B-Instruct

Warm
Public
72.7B
FP8
131072
License: other
Hugging Face
Overview

Overview

The abacusai/Smaug-Qwen2-72B-Instruct is a 72.7 billion parameter instruction-tuned model, building upon the Qwen2-72B-Instruct architecture. It features a significant context length of 131,072 tokens, enabling it to process and understand extensive inputs.

Key Capabilities & Performance

This model distinguishes itself through enhanced performance in several critical areas compared to its base model:

  • Reasoning: Achieves an overall score of 0.8241 on Big-Bench Hard (BBH), surpassing Qwen2-72B-Instruct's 0.8036, indicating stronger complex reasoning abilities.
  • Code Generation: Demonstrates improved coding proficiency with a Pass@1 score of 0.3357 on LiveCodeBench, compared to 0.3139 for the base model.
  • General Instruction Following: Scores 48.0 on Arena-Hard, outperforming Qwen2-72B-Instruct's 43.5, suggesting better overall instruction adherence and helpfulness in competitive scenarios.

Use Cases

Given its strengths, Smaug-Qwen2-72B-Instruct is particularly well-suited for:

  • Applications requiring advanced logical reasoning and problem-solving.
  • Code generation and understanding tasks where accuracy is paramount.
  • Complex conversational agents or assistants that need to handle intricate instructions and maintain context over long interactions.