abacusai/Smaug-Qwen2-72B-Instruct

Hugging Face
TEXT GENERATIONConcurrency Cost:4Model Size:72.7BQuant:FP8Ctx Length:32kPublished:Jun 26, 2024License:tongyi-qianwenArchitecture:Transformer0.0K Warm

The Smaug-Qwen2-72B-Instruct is a 72.7 billion parameter instruction-tuned causal language model developed by abacusai, fine-tuned from Qwen2-72B-Instruct. This model is optimized for complex reasoning and problem-solving tasks, demonstrating improved performance on benchmarks like Big-Bench Hard (BBH), LiveCodeBench, and Arena-Hard compared to its base model. With a substantial 131,072 token context length, it is well-suited for applications requiring deep contextual understanding and advanced analytical capabilities.

Loading preview...

Overview

The abacusai/Smaug-Qwen2-72B-Instruct is a 72.7 billion parameter instruction-tuned model, building upon the Qwen2-72B-Instruct architecture. It features a significant context length of 131,072 tokens, enabling it to process and understand extensive inputs.

Key Capabilities & Performance

This model distinguishes itself through enhanced performance in several critical areas compared to its base model:

  • Reasoning: Achieves an overall score of 0.8241 on Big-Bench Hard (BBH), surpassing Qwen2-72B-Instruct's 0.8036, indicating stronger complex reasoning abilities.
  • Code Generation: Demonstrates improved coding proficiency with a Pass@1 score of 0.3357 on LiveCodeBench, compared to 0.3139 for the base model.
  • General Instruction Following: Scores 48.0 on Arena-Hard, outperforming Qwen2-72B-Instruct's 43.5, suggesting better overall instruction adherence and helpfulness in competitive scenarios.

Use Cases

Given its strengths, Smaug-Qwen2-72B-Instruct is particularly well-suited for:

  • Applications requiring advanced logical reasoning and problem-solving.
  • Code generation and understanding tasks where accuracy is paramount.
  • Complex conversational agents or assistants that need to handle intricate instructions and maintain context over long interactions.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p