arcee-ai/saul-mistral-instruct-v0.1-7b-ties

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 26, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

arcee-ai/saul-mistral-instruct-v0.1-7b-ties is a 7 billion parameter instruction-tuned language model created by arcee-ai, built by merging Equall/Saul-Base and mistralai/Mistral-7B-Instruct-v0.1 using the TIES merging method. This model leverages the strengths of its base components to provide enhanced instruction following capabilities within a 4096 token context window. It is designed for general-purpose conversational AI and instruction-based tasks.

Loading preview...

Model Overview

arcee-ai/saul-mistral-instruct-v0.1-7b-ties is a 7 billion parameter instruction-tuned language model developed by arcee-ai. This model is a product of merging two distinct models: Equall/Saul-Base and mistralai/Mistral-7B-Instruct-v0.1, utilizing the TIES (Trimmed, Iterative, Extracted, and Scaled) merging method via mergekit.

Key Characteristics

  • Architecture: Based on the Mistral-7B-v0.1 foundation, enhanced through model merging.
  • Parameter Count: 7 billion parameters, offering a balance between performance and computational efficiency.
  • Context Length: Supports a context window of 4096 tokens.
  • Merging Strategy: Employs the TIES merging method, which combines the weights of multiple models to create a new, potentially more capable model.

Use Cases

This model is suitable for a variety of instruction-following tasks, leveraging the combined strengths of its constituent models. Its instruction-tuned nature makes it well-suited for applications requiring conversational AI, question answering, and general text generation based on explicit instructions.