Toten5/Marcoroni-neural-chat-7B-v2

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Dec 12, 2023License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

Toten5/Marcoroni-neural-chat-7B-v2 is a 7 billion parameter language model created by Toten5, merging AIDC-ai-business/Marcoroni-7B-v3 and Intel/neural-chat-7b-v3-3. Based on Mistral-7B-v0.1, this model is designed for general chat applications, leveraging the combined strengths of its base models. It offers a 4096-token context length, making it suitable for conversational tasks requiring moderate memory.

Loading preview...

Model Overview

Marcoroni-neural-chat-7B-v2 is a 7 billion parameter language model developed by Toten5. This model is a result of merging two distinct models: AIDC-ai-business/Marcoroni-7B-v3 and Intel/neural-chat-7b-v3-3. The merge was performed using the Slerp method with the mergekit tool, primarily for testing and combining their respective capabilities.

Key Characteristics

  • Base Architecture: Both constituent models are built upon mistralai/Mistral-7B-v0.1, inheriting its efficient architecture and performance characteristics.
  • Parameter Count: With 7 billion parameters, it balances performance with computational efficiency.
  • Context Length: Supports a context window of 4096 tokens, allowing for coherent and extended conversational interactions.
  • Development Purpose: Created as an experimental merge to explore the synergies between two established 7B models.

Use Cases

This model is particularly well-suited for:

  • General Chat Applications: Its foundation in neural-chat suggests strong performance in conversational AI.
  • Experimental AI Development: Ideal for researchers and developers looking to test merged model performance and explore new combinations.
  • Prototyping: Can serve as a robust backbone for developing and iterating on new language-based applications.