prem-research/prem-1B-chat
TEXT GENERATIONConcurrency Cost:1Model Size:1.1BQuant:BF16Ctx Length:2kPublished:May 6, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Loading

The prem-research/prem-1B-chat is a 1.1 billion parameter Llama-based small language model (SLM) developed by Prem AI. This instruction-tuned model is designed for commercial and research applications, particularly excelling in conversational interactions akin to a virtual assistant. Its primary objective is to serve as an effective foundation for Retrieval-Augmented Generation (RAG) applications, aiming to handle multi-turn conversations with an extended context length of 8192 tokens in its series. The model offers open-source capabilities for building advanced language models, focusing on efficiency for RAG tasks.

Loading preview...