paulml/NeuralOmniWestBeaglake-7B

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 5, 2024License:cc-by-nc-4.0Architecture:Transformer0.0K Open Weights Cold

NeuralOmniWestBeaglake-7B by paulml is a 7 billion parameter language model with a 4096 token context length, created by merging several specialized models using the DARE TIES method. Built upon the Mistral-7B-v0.1 base, this model integrates components from shadowml/WestBeagle-7B, shadowml/Beaglake-7B, and mlabonne/NeuralOmniBeagle-7B. It is designed to leverage the combined strengths of its constituent models, offering a versatile foundation for various natural language processing tasks.

Loading preview...

Model Overview

NeuralOmniWestBeaglake-7B is a 7 billion parameter language model developed by paulml, built on the Mistral-7B-v0.1 architecture. This model is a product of a sophisticated merging process using LazyMergekit and the DARE TIES merge method.

Key Characteristics

  • Merged Architecture: Combines the strengths of three distinct models: shadowml/WestBeagle-7B, shadowml/Beaglake-7B, and mlabonne/NeuralOmniBeagle-7B.
  • Base Model: Utilizes mistralai/Mistral-7B-v0.1 as its foundational architecture.
  • Merge Method: Employs the dare_ties merging technique, with specific density and weight parameters applied to each contributing model to optimize performance.
  • Parameter Count: Features 7 billion parameters, offering a balance between performance and computational efficiency.
  • Context Length: Supports a context window of 4096 tokens, suitable for handling moderately long inputs and generating coherent responses.

Usage

This model is designed for general-purpose natural language generation and understanding tasks. Developers can easily integrate it into their projects using the Hugging Face transformers library, with provided code examples for quick setup and inference.