Undi95/ReMM-SLERP-L2-13B

Warm
Public
13B
FP8
4096
1
Sep 4, 2023
License: cc-by-nc-4.0
Hugging Face
Overview

Undi95/ReMM-SLERP-L2-13B: A Recreated and Updated MythoMax Variant

This model, ReMM-SLERP-L2-13B, is a 13 billion parameter language model developed by Undi95. It represents a recreation and update of the original MythoMax-L2-13B, utilizing the SLERP (Spherical Linear Interpolation) merging technique to combine several advanced models.

Key Merging Details

  • Base Model: TheBloke/Llama-2-13B-fp16 serves as the foundational architecture.
  • Component Models: It integrates elements from:
    • The-Face-Of-Goonery/Chronos-Beluga-v2-13bfp16
    • jondurbin/airoboros-l2-13b-2.1
    • NousResearch/Nous-Hermes-Llama2-13b
    • The-Face-Of-Goonery/Huginn-13b-v1.2
    • A private recreation trial of an updated Mythologic-L2-13B (ReML-L2-13B)
  • Merging Method: The model employs a specific version of SLERP for merging, adapted for notebook usage, as detailed in the associated LLM-SLERP-MergeTest repository.

Performance Benchmarks

Evaluated on the Open LLM Leaderboard, ReMM-SLERP-L2-13B achieves an average score of 50.99. Notable individual metric scores include:

  • ARC (25-shot): 60.92
  • HellaSwag (10-shot): 83.56
  • MMLU (5-shot): 55.33
  • TruthfulQA (0-shot): 51.97
  • Winogrande (5-shot): 75.22
  • GSM8K (5-shot): 9.17
  • DROP (3-shot): 20.76

Prompt Template

The model uses the Alpaca prompt template format:

Below is an instruction that describes a task. Write a response that appropriately completes the request.

### Instruction:
{prompt}

### Response: