flemmingmiguel/MBX-7B-v3
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 28, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold
MBX-7B-v3 is a 7 billion parameter language model developed by flemmingmiguel, created by merging flemmingmiguel/MBX-7B and flemmingmiguel/MBX-7B-v3 using LazyMergekit. This model leverages a slerp merge method across its 32 layers, with specific parameter weighting for self_attn and mlp components. It is designed for general text generation tasks, offering a 4096-token context length.
Loading preview...