InnerI/InnerILLM-0x00d0-7B-slerp
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 13, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

InnerI/InnerILLM-0x00d0-7B-slerp is a 7 billion parameter language model created by InnerI, formed by merging NousResearch/Yarn-Mistral-7b-128k and InnerI/InnerILLM-0x00d0-Ox0dad0-nous-nous-v2.0-7B-slerp using a slerp merge method. This model leverages the Mistral architecture and is designed for general text generation tasks, combining characteristics from its merged components. It is suitable for applications requiring a 7B parameter model with a potentially extended context window inherited from Yarn-Mistral.

Loading preview...

InnerILLM-0x00d0-7B-slerp Overview

InnerILLM-0x00d0-7B-slerp is a 7 billion parameter language model developed by InnerI. It is a product of merging two distinct models: NousResearch/Yarn-Mistral-7b-128k and InnerI/InnerILLM-0x00d0-Ox0dad0-nous-nous-v2.0-7B-slerp. This merge was performed using the slerp (spherical linear interpolation) method via LazyMergekit, aiming to combine the strengths of its constituent models.

Key Characteristics

  • Architecture: Based on the Mistral architecture, inheriting its efficient design.
  • Merge Method: Utilizes a slerp merge, which blends model weights in a specific, non-linear fashion, often resulting in a balanced combination of source model capabilities.
  • Component Models: Integrates Yarn-Mistral-7b-128k, known for its extended context handling (128k tokens), and InnerILLM-0x00d0-Ox0dad0-nous-nous-v2.0-7B-slerp.

Usage and Configuration

The model is configured with specific t parameters for self-attention and MLP layers during the slerp merge, indicating a fine-tuned blending strategy. It supports bfloat16 dtype for efficient computation. Developers can easily integrate and use this model with the Hugging Face transformers library for text generation tasks, as demonstrated in the provided Python usage example.

Good For

  • General text generation and conversational AI.
  • Applications benefiting from a 7B parameter model with potentially enhanced context understanding due to its Yarn-Mistral lineage.
  • Experimentation with merged models for diverse language tasks.