Gille/StrangeMerges_30-7B-slerp
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 4, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

Gille/StrangeMerges_30-7B-slerp is a 7 billion parameter language model created by Gille, built using a slerp merge of Gille/StrangeMerges_21-7B-slerp and yam-peleg/Experiment26-7B. This model leverages a 4096-token context length and is designed for general language generation tasks. Its unique merging strategy suggests potential for diverse capabilities, with a note that further training on specific datasets like Orca-Math could enhance its performance in reasoning.

Loading preview...