chlee10/T3Q-Merge-Mistral7B
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 12, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

T3Q-Merge-Mistral7B is a 7 billion parameter language model developed by Chihoon Lee (chlee10) and T3Q, created by merging liminerity/M7-7b and yam-peleg/Experiment26-7B using mergekit. This model leverages a slerp merge method, specifically adjusting attention and MLP layer parameters, to combine the strengths of its constituent models. It is designed for general language tasks, building upon the Mistral architecture with a 4096-token context length.

Loading preview...