Arc53/docsgpt-7b-mistral

Cold
Public
7B
FP8
4096
License: apache-2.0
Hugging Face
Overview

Overview

Arc53/docsgpt-7b-mistral is a 7 billion parameter language model, fine-tuned from the Zephyr-7B-beta architecture using the LoRA process. Its primary optimization is for Documentation (RAG optimized), focusing on generating answers based on provided context.

Key Capabilities

  • Context-Driven Answering: Specifically designed to provide accurate responses by leveraging contextual information, reducing hallucination.
  • Strong Performance on BACON Test: Achieves a score of 8.64 on the internal BACON test, which evaluates context understanding, hallucination, and attention span, placing it competitively with larger models like GPT-3.5-turbo.
  • Competitive MTbench Scores: Demonstrates solid performance on the MTbench with LLM judge, scoring 7.166875 on average, outperforming its base model Zephyr-7b-beta and other 13B parameter models like Vicuna-13b-v1.3.
  • Commercial Use: Licensed under Apache-2.0, allowing for commercial applications.

Good for

  • Technical Support: Ideal for systems requiring precise answers from technical documentation.
  • Developer Tools: Useful for integrating into developer workflows where accurate information retrieval from documentation is critical.
  • RAG Applications: Optimized for Retrieval Augmented Generation (RAG) scenarios, ensuring responses are grounded in provided documents.

Users should format prompts with ### Instruction, ### Context, and ### Answer sections for optimal performance.