grimjim/Mistral-7B-Instruct-demi-merge-v0.2-7B
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 26, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold
grimjim/Mistral-7B-Instruct-demi-merge-v0.2-7B is a 7 billion parameter language model merged from Mistral-7B-v0.2 and Mistral-7B-Instruct-v0.2 using the SLERP method. This merge aims to balance the strengths of both models, providing a base for further fine-tuning while retaining instruction-following capabilities. It supports a 32K token context length without a sliding window, making it suitable for applications requiring extended conversational memory or processing of longer documents.
Loading preview...