grimjim/Mistral-Starling-merge-trial3-7B
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 29, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

grimjim/Mistral-Starling-merge-trial3-7B is a 7 billion parameter language model created by grimjim, merging Nexusflow/Starling-LM-7B-beta and grimjim/Mistral-7B-Instruct-demi-merge-v0.2-7B. This model aims to combine strong reasoning capabilities with an extended 32K context length. It was developed using the SLERP merge method to enhance performance in complex reasoning tasks.

Loading preview...