MSL7/INEX8-7B

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 2, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

INEX8-7B is a 7 billion parameter language model developed by liminerity, created through a series of slerp merges of several 7B models including MSL7/INEX4-7b and yam-peleg/Experiment26-7B. This model is designed for general-purpose language tasks, leveraging its merged architecture to achieve a balanced performance across various benchmarks. With a 4096-token context length, it offers a solid foundation for applications requiring robust language understanding and generation.

Loading preview...

INEX8-7B: A Merged 7B Language Model

INEX8-7B is a 7 billion parameter model developed by liminerity, constructed using a multi-stage slerp merging process via mergekit. This model integrates capabilities from several base models, including liminerity/merge3 and yam-peleg/Experiment26-7B, to create a versatile language understanding and generation system.

Key Capabilities & Performance

INEX8-7B demonstrates competitive performance across a range of benchmarks, as evaluated on the Open LLM Leaderboard:

  • Average Score: 76.44
  • AI2 Reasoning Challenge (25-Shot): 73.29
  • HellaSwag (10-Shot): 89.19
  • MMLU (5-Shot): 64.47
  • TruthfulQA (0-shot): 77.83
  • Winogrande (5-shot): 84.85
  • GSM8k (5-shot): 68.99

These scores indicate a balanced proficiency in reasoning, common sense, factual recall, and mathematical problem-solving. The model's architecture, derived from multiple merges, aims to combine the strengths of its constituent models.

Use Cases

Given its general-purpose nature and benchmark performance, INEX8-7B is suitable for a variety of applications requiring a capable 7B language model, such as:

  • General text generation and completion
  • Question answering
  • Reasoning tasks
  • Content creation

Its 4096-token context length supports processing moderately sized inputs and generating coherent responses.