monology/mixtral-soup
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 20, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

monology/mixtral-soup is an experimental 7B parameter language model created by monology, formed by a linear merge of eight distinct Mixtral-expert models. This model leverages the Mixture-of-Experts architecture, combining specialized components to potentially enhance performance across various tasks. It is designed for experimental purposes, exploring the capabilities of merged Mixtral experts within a 4096-token context window.

Loading preview...