arcee-ai/SEC-Calme-7B-Instruct

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 30, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

SEC-Calme-7B-Instruct is a 7 billion parameter instruction-tuned language model developed by arcee-ai, built by merging arcee-ai/sec-mistral-7b-instruct-1.2-epoch and MaziyarPanahi/Calme-7B-Instruct-v0.2. This model leverages a slerp merge method to combine the strengths of its base components, offering a general-purpose instruction-following capability. With a context length of 4096 tokens, it is designed for diverse conversational and text generation tasks.

Loading preview...

SEC-Calme-7B-Instruct Overview

SEC-Calme-7B-Instruct is a 7 billion parameter instruction-tuned language model created by arcee-ai. It is a product of merging two distinct models: arcee-ai/sec-mistral-7b-instruct-1.2-epoch and MaziyarPanahi/Calme-7B-Instruct-v0.2.

Key Characteristics

  • Model Architecture: Based on a merge of two existing 7B instruction-tuned models.
  • Merge Method: Utilizes the slerp (spherical linear interpolation) merge method via mergekit to combine the weights of its constituent models.
  • Parameter Configuration: Specific t parameters were applied during the merge, with varying values for self_attn and mlp filters, and a general value of 0.5 for other parameters.
  • Data Type: The model is configured to use bfloat16 for its operations.

Potential Use Cases

This model is suitable for general instruction-following tasks, leveraging the combined capabilities of its merged components. Developers can explore its performance in areas such as:

  • Conversational AI and chatbots.
  • Text generation and summarization.
  • Instruction-based task execution.