uukuguy/zephyr-7b-alpha-dare-0.85
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:8kPublished:Nov 23, 2023License:llama2Architecture:Transformer Open Weights Cold

uukuguy/zephyr-7b-alpha-dare-0.85 is a 7 billion parameter experimental language model based on the Zephyr-7B-alpha architecture, fine-tuned using the DARE (Drop and REscale) method. This model explores the concept that a high proportion (85%) of delta parameters can be set to zero without significantly impacting capabilities. It is designed to investigate parameter efficiency and model compression techniques, maintaining an 8192-token context length.

Loading preview...