FredrikBL/NeuralFlashback-ties

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 28, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

FredrikBL/NeuralFlashback-ties is a 7 billion parameter language model created by FredrikBL, built by merging timpal0l/Mistral-7B-v0.1-flashback-v2 and mlabonne/NeuralHermes-2.5-Mistral-7B using the TIES merging method. This model leverages the Mistral-7B-v0.1 architecture and is designed for general text generation tasks, offering a 4096 token context length. Its unique composition aims to combine the strengths of its constituent models for enhanced performance.

Loading preview...

Model Overview

FredrikBL/NeuralFlashback-ties is a 7 billion parameter language model developed by FredrikBL. It is constructed through a merge of two distinct Mistral-7B-v0.1 based models: timpal0l/Mistral-7B-v0.1-flashback-v2 and mlabonne/NeuralHermes-2.5-Mistral-7B. This merge was performed using the TIES (Trimmed, Iterative, Extracted, and Scaled) merging method, specifically via LazyMergekit.

Key Characteristics

  • Architecture: Based on the Mistral-7B-v0.1 foundation model.
  • Merging Method: Utilizes the TIES method to combine the weights of its base models, with specific density and weight parameters applied to each component model.
  • Context Length: Supports a context window of 4096 tokens.

Intended Use Cases

This model is suitable for a variety of natural language processing tasks, particularly those benefiting from the combined capabilities of its merged predecessors. Developers can integrate it into applications requiring text generation, conversational AI, or other general language understanding tasks, leveraging its 7 billion parameters for efficient inference.