selfrag/selfrag_llama2_13b
TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kPublished:Oct 18, 2023License:mitArchitecture:Transformer0.1K Open Weights Cold

The selfrag/selfrag_llama2_13b model is a 13 billion parameter Self-RAG model developed by Akari Asai, Zeqiu Wu, Yizhong Wang, Avirup Sil, and Hannaneh Hajishirzi. It is designed to generate responses to diverse user queries by adaptively calling a retrieval system and criticizing its own output and retrieved passages using reflection tokens. This model excels at tasks requiring factual grounding and self-correction, leveraging a 4096-token context length for enhanced performance.

Loading preview...