papahawk/devi-7b
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 7, 2024License:mitArchitecture:Transformer Open Weights Cold

papahawk/devi-7b is a 7 billion parameter GPT-like language model, forked from Zephyr-7B-β, which is a fine-tuned version of Mistral-7B-v0.1. Developed by papahawk, it is optimized as a helpful assistant through Direct Preference Optimization (DPO) on synthetic datasets. This model excels in chat applications and achieves high rankings on MT-Bench and AlpacaEval benchmarks among 7B models.

Loading preview...