paulml/OGNO-7B

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 12, 2024License:cc-by-nc-4.0Architecture:Transformer0.0K Open Weights Cold

paulml/OGNO-7B is a 7 billion parameter language model created by paulml, formed by merging liminerity/Omningotex-7b-slerp and eren23/dpo-binarized-NeutrixOmnibe-7B using a slerp merge method. This model leverages the strengths of its constituent models to provide general-purpose language generation capabilities. With a context length of 4096 tokens, it is suitable for a variety of text-based tasks.

Loading preview...

OGNO-7B Overview

OGNO-7B is a 7 billion parameter language model developed by paulml, created through a strategic merge of two existing models: liminerity/Omningotex-7b-slerp and eren23/dpo-binarized-NeutrixOmnibe-7B. This merge was performed using the slerp (spherical linear interpolation) method, a technique often employed to combine the learned representations of different models.

Key Characteristics

  • Architecture: A merged model combining Omningotex-7b-slerp and dpo-binarized-NeutrixOmnibe-7B.
  • Parameter Count: 7 billion parameters, offering a balance between performance and computational efficiency.
  • Context Length: Supports a context window of 4096 tokens, allowing for processing moderately long inputs and generating coherent responses.
  • Merge Method: Utilizes slerp for combining model weights, with specific parameter adjustments for self_attn and mlp layers, indicating a fine-tuned approach to integration.

When to Use OGNO-7B

This model is suitable for general text generation tasks where a 7B parameter model with a 4K context window is appropriate. Its merged origin suggests it aims to inherit and combine the strengths of its base models, making it a versatile option for various natural language processing applications.