filipealmeida/Mistral-7B-v0.1-sharded
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Sep 28, 2023License:apache-2.0Architecture:Transformer0.0K Open Weights Cold
The filipealmeida/Mistral-7B-v0.1-sharded model is a sharded version of the Mistral-7B-v0.1, a 7 billion parameter pretrained generative text model developed by the Mistral AI Team. It utilizes a transformer architecture incorporating Grouped-Query Attention, Sliding-Window Attention, and a Byte-fallback BPE tokenizer. This model is noted for outperforming larger models like Llama 2 13B on various benchmarks, making it suitable for general-purpose text generation tasks where efficiency and strong performance are critical, especially in environments with limited CPU memory due to its sharded nature.
Loading preview...