noahkln/vicuna-13b-v1.5-no-cache
TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kLicense:llama2Architecture:Transformer Open Weights Cold

The noahkln/vicuna-13b-v1.5-no-cache model is a 13 billion parameter auto-regressive language model developed by LMSYS, fine-tuned from Llama 2. It is specifically designed as a chat assistant, trained on approximately 125K user-shared conversations from ShareGPT. This model excels in conversational AI research and development, providing a robust base for chatbot applications with a 4096-token context length.

Loading preview...