SakanaAI/DiscoPOP-zephyr-7b-gemma
TEXT GENERATIONConcurrency Cost:1Model Size:8.5BQuant:FP8Ctx Length:8kPublished:Jun 12, 2024License:gemmaArchitecture:Transformer0.0K Cold

SakanaAI's DiscoPOP-zephyr-7b-gemma is an 8.5 billion parameter language model, fine-tuned from HuggingFaceH4/zephyr-7b-gemma-sft-v0.1 with an 8192-token context length. This model distinguishes itself by utilizing DiscoPOP, a novel Discovered Preference Optimization algorithm, instead of standard Direct Preference Optimization (DPO). It is designed for general language tasks, leveraging its unique optimization method for improved performance.

Loading preview...