selfrag/selfrag_llama2_7b
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Oct 18, 2023License:mitArchitecture:Transformer0.1K Open Weights Warm

selfrag/selfrag_llama2_7b is a 7 billion parameter Llama 2-based model developed by Akari Asai and collaborators, specifically designed for Self-RAG (Retrieval Augmented Generation). This model generates outputs while adaptively calling a retrieval system and criticizing its own generations and retrieved passages using reflection tokens. It excels at instruction-following tasks by leveraging fine-grained feedback for improved accuracy and relevance.

Loading preview...