paulml/NeuralOmniBeagleMBX-v3-7B

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 5, 2024License:cc-by-nc-4.0Architecture:Transformer0.0K Open Weights Cold

NeuralOmniBeagleMBX-v3-7B by paulml is a 7 billion parameter merged language model, combining mlabonne/NeuralOmniBeagle-7B and flemmingmiguel/MBX-7B-v3. This model utilizes a slerp merge method with specific parameter weighting for self-attention and MLP layers, offering a distinct blend of capabilities from its constituent models. It is designed for general text generation tasks, leveraging its merged architecture for potentially enhanced performance over its individual components.

Loading preview...

NeuralOmniBeagleMBX-v3-7B Overview

NeuralOmniBeagleMBX-v3-7B is a 7 billion parameter language model developed by paulml. It is a product of merging two distinct models: mlabonne/NeuralOmniBeagle-7B and flemmingmiguel/MBX-7B-v3. This merge was performed using the LazyMergekit tool, specifically employing a slerp (spherical linear interpolation) merge method.

Key Characteristics

  • Merged Architecture: Combines the strengths of two base models, NeuralOmniBeagle-7B and MBX-7B-v3, into a single 7B parameter model.
  • Configurable Merge: The merge configuration specifies distinct weighting for self-attention (self_attn) and multi-layer perceptron (mlp) layers, indicating a fine-tuned approach to blending their functionalities.
  • Bfloat16 Precision: The model is configured to use bfloat16 data type, which is common for efficient inference on modern hardware.

Usage and Application

This model is suitable for various text generation tasks, benefiting from the combined knowledge and capabilities of its merged components. Developers can easily integrate it using the Hugging Face transformers library, as demonstrated in the provided Python usage example. Its 7B parameter size makes it a viable option for applications requiring a balance between performance and computational resources.