AI-B/UTENA-7B-V3

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 14, 2024License:unlicenseArchitecture:Transformer0.0K Cold

AI-B/UTENA-7B-V3 is a 7 billion parameter language model developed by AI-B, created by merging UTENA-7B-UNA-V2 and UTENA-7B-NSFW-V2 using the slerp method. This model achieves an average score of 67.42 on the Open LLM Leaderboard, demonstrating capabilities across various benchmarks including reasoning, common sense, and language understanding. It is designed for general-purpose applications requiring a balanced performance profile.

Loading preview...

UTENA-7B-V3 Overview

UTENA-7B-V3 is a 7 billion parameter language model developed by AI-B, constructed through a merge of two distinct models: AI-B/UTENA-7B-UNA-V2 and AI-B/UTENA-7B-NSFW-V2. This merge was performed using the slerp method within mergekit, combining the strengths of its constituent models.

Key Capabilities & Performance

The model's performance has been evaluated on the Open LLM Leaderboard, achieving a competitive average score of 67.42. Specific benchmark results include:

  • AI2 Reasoning Challenge (25-Shot): 65.96
  • HellaSwag (10-Shot): 85.70
  • MMLU (5-Shot): 64.72
  • TruthfulQA (0-shot): 53.64
  • Winogrande (5-shot): 80.27
  • GSM8k (5-shot): 54.21

These scores indicate a balanced capability across reasoning, common sense, factual recall, and mathematical problem-solving. The model is also available in quantized GGUF format for optimized inference.

When to Use This Model

UTENA-7B-V3 is suitable for applications requiring a versatile 7B parameter model with solid performance across a range of general language understanding and generation tasks. Its balanced benchmark results suggest it can be a strong candidate for use cases where a single model needs to handle diverse cognitive demands.