AI-B/UTENA-7B-V3

Warm
Public
7B
FP8
4096
Jan 14, 2024
License: unlicense
Hugging Face
Overview

UTENA-7B-V3 Overview

UTENA-7B-V3 is a 7 billion parameter language model developed by AI-B, constructed through a merge of two distinct models: AI-B/UTENA-7B-UNA-V2 and AI-B/UTENA-7B-NSFW-V2. This merge was performed using the slerp method within mergekit, combining the strengths of its constituent models.

Key Capabilities & Performance

The model's performance has been evaluated on the Open LLM Leaderboard, achieving a competitive average score of 67.42. Specific benchmark results include:

  • AI2 Reasoning Challenge (25-Shot): 65.96
  • HellaSwag (10-Shot): 85.70
  • MMLU (5-Shot): 64.72
  • TruthfulQA (0-shot): 53.64
  • Winogrande (5-shot): 80.27
  • GSM8k (5-shot): 54.21

These scores indicate a balanced capability across reasoning, common sense, factual recall, and mathematical problem-solving. The model is also available in quantized GGUF format for optimized inference.

When to Use This Model

UTENA-7B-V3 is suitable for applications requiring a versatile 7B parameter model with solid performance across a range of general language understanding and generation tasks. Its balanced benchmark results suggest it can be a strong candidate for use cases where a single model needs to handle diverse cognitive demands.