gadkins/Mistral-7B-Instruct-v0.1-function-calling

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 6, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

gadkins/Mistral-7B-Instruct-v0.1-function-calling is an instruction-tuned 7 billion parameter large language model, fine-tuned from Mistral-7B-v0.1 by Mistral AI. It features a 4096-token context length and incorporates architectural choices like Grouped-Query Attention and Sliding-Window Attention. This model is optimized for following instructions and generating conversational text, making it suitable for general-purpose dialogue and instruction-based tasks.

Loading preview...

Overview

This model, gadkins/Mistral-7B-Instruct-v0.1-function-calling, is an instruction-tuned variant of the original Mistral-7B-v0.1 generative text model, developed by Mistral AI. It leverages publicly available conversation datasets for its fine-tuning, enhancing its ability to follow instructions and engage in dialogue. The model's architecture is based on the Mistral-7B-v0.1 transformer, incorporating key features such as Grouped-Query Attention, Sliding-Window Attention, and a byte-fallback BPE tokenizer.

Key Capabilities

  • Instruction Following: Designed to accurately interpret and respond to user instructions.
  • Conversational AI: Excels at generating coherent and contextually relevant responses in dialogue.
  • Efficient Architecture: Utilizes Grouped-Query Attention and Sliding-Window Attention for improved performance.

Good for

  • General Chatbots: Ideal for building conversational agents that can follow specific prompts.
  • Instruction-Based Tasks: Suitable for applications requiring the model to perform actions based on explicit instructions.
  • Rapid Prototyping: Serves as a quick demonstration of the base model's fine-tuning potential for compelling performance.