filipealmeida/Mistral-7B-Instruct-v0.1-sharded
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Sep 28, 2023License:apache-2.0Architecture:Transformer0.0K Open Weights Cold
The filipealmeida/Mistral-7B-Instruct-v0.1-sharded model is an 8 billion parameter instruction-tuned language model, sharded for efficient deployment with limited CPU memory. Developed by Mistral AI, it is based on the Mistral-7B-v0.1 architecture, featuring Grouped-Query Attention and Sliding-Window Attention. This model is optimized for following instructions and generating conversational text, making it suitable for various dialogue-based applications.
Loading preview...