kettleguts/zephyr-7b-beta_sparse05
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 24, 2024License:mitArchitecture:Transformer Open Weights Cold
kettleguts/zephyr-7b-beta_sparse05 is a 7 billion parameter language model, a pruned version of HuggingFaceH4/zephyr-7b-beta. It utilizes Wanda pruning to achieve 50% sparsity in its linear layers, making it a more efficient model for research purposes. This model is primarily intended for exploring the effects of network pruning on large language models, offering insights into model compression techniques. Its text generation quality is highly dependent on prompting, sometimes behaving like a smaller model due to heavy pruning.
Loading preview...