paulml/OmniBeagleSquaredMBX-v3-7B

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 9, 2024License:cc-by-nc-4.0Architecture:Transformer0.0K Open Weights Cold

paulml/OmniBeagleSquaredMBX-v3-7B is a 7 billion parameter language model created by paulml, formed by merging OmniBeagleMBX-v3-7B and MBX-7B-v3. This model is notable for achieving the number one rank in the Arc Challenge for 7B models as of February 12th, 2024. It is optimized for reasoning tasks, making it suitable for applications requiring strong logical and problem-solving capabilities.

Loading preview...

OmniBeagleSquaredMBX-v3-7B Overview

OmniBeagleSquaredMBX-v3-7B is a 7 billion parameter language model developed by paulml. It is a merged model, combining the strengths of paulml/OmniBeagleMBX-v3-7B and flemmingmiguel/MBX-7B-v3 using the LazyMergekit tool. This strategic merge aims to leverage the best features of its constituent models.

Key Capabilities

  • Top-tier Reasoning: As of February 12th, 2024, OmniBeagleSquaredMBX-v3-7B holds the number one ranking in the Arc Challenge among 7B parameter models, indicating strong performance in complex reasoning tasks.
  • Merged Architecture: The model benefits from a sophisticated slerp merge method, with specific parameter adjustments for self-attention and MLP layers, suggesting a fine-tuned balance of its base models' characteristics.

Good For

  • Reasoning-intensive applications: Its strong performance on the Arc Challenge makes it particularly well-suited for tasks requiring logical deduction, problem-solving, and understanding complex relationships.
  • Developers seeking a high-performing 7B model: For those needing a compact yet powerful model for various NLP tasks, especially where reasoning is critical, this model offers a competitive option within the 7B parameter class.