PatronusAI/Llama-3-Patronus-Lynx-8B-Instruct
PatronusAI/Llama-3-Patronus-Lynx-8B-Instruct is an 8 billion parameter instruction-tuned model developed by Patronus AI, fine-tuned from Meta-Llama-3-8B-Instruct. It is specifically designed and trained for hallucination detection in Retrieval Augmented Generation (RAG) settings, evaluating answer faithfulness to provided documents. The model excels at identifying when an answer introduces new information or contradicts the document, with a maximum sequence length of 8000 tokens.
Loading preview...
PatronusAI/Llama-3-Patronus-Lynx-8B-Instruct Overview
PatronusAI/Llama-3-Patronus-Lynx-8B-Instruct, developed by Patronus AI, is an 8 billion parameter model fine-tuned from Meta-Llama-3-8B-Instruct. Its primary purpose is to serve as an open-source hallucination evaluation model, specifically trained to detect unfaithfulness in answers generated within Retrieval Augmented Generation (RAG) contexts. The model was trained on a diverse mix of datasets, including CovidQA, PubmedQA, DROP, and RAGTruth, incorporating both hand-annotated and synthetic data.
Key Capabilities
- Hallucination Detection: Evaluates whether a given answer is faithful to a provided document and question.
- Faithfulness Scoring: Determines if an answer introduces new information not present in the document or contradicts information within it, outputting a 'PASS' or 'FAIL' score.
- Reasoning Generation: Provides a detailed reasoning for its faithfulness verdict in JSON format.
- Context Length: Supports a maximum sequence length of 8000 tokens.
Performance and Use Cases
Evaluated on the PatronusAI/HaluBench dataset, the 8B LYNX model demonstrates strong performance in hallucination detection across various benchmarks, achieving an 82.9% overall score. This positions it as a robust tool for developers and researchers focused on improving the reliability and factual accuracy of RAG systems. It is particularly useful for automated quality assurance of LLM outputs against source documents.
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.