wang7776/Mistral-7B-Instruct-v0.2-sparsity-10
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:8kPublished:Dec 26, 2023License:apache-2.0Architecture:Transformer Open Weights Cold

wang7776/Mistral-7B-Instruct-v0.2-sparsity-10 is a 7 billion parameter instruction-tuned causal language model, based on Mistral AI's Mistral-7B-Instruct-v0.2. This version has been pruned to 10% sparsity using the Wanda method, which aims to maintain competitive performance without retraining. It features Grouped-Query Attention, Sliding-Window Attention, and a Byte-fallback BPE tokenizer, making it suitable for efficient instruction-following tasks.

Loading preview...