jeiku/Elly_7B

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 19, 2024License:otherArchitecture:Transformer0.0K Cold

Elly_7B is a 7 billion parameter language model created by jeiku, built upon the SanjiWatsuki/Sonya-7B base model using the DARE TIES merge method. This model integrates capabilities from MaziyarPanahi/samantha-mistral-7b-Mistral-7B-Instruct-v0.1 and cognitivecomputations/dolphin-2.6-mistral-7b, aiming to combine their respective strengths. It is designed for general language generation tasks, leveraging the merged instruction-tuned models for diverse applications.

Loading preview...

Elly_7B: A Merged 7B Language Model

Elly_7B is a 7 billion parameter language model developed by jeiku, created through a sophisticated merge of several pre-trained models. It utilizes the DARE TIES merge method, a technique designed to combine the strengths of multiple models effectively, with SanjiWatsuki/Sonya-7B serving as its foundational base.

Key Merge Components

This model's capabilities are derived from the integration of two distinct instruction-tuned models:

By merging these models, Elly_7B aims to inherit and synthesize their respective instruction-following and conversational abilities. The merge configuration used a float16 data type and applied normalization during the DARE TIES process.

Intended Use

Elly_7B is suitable for a variety of general-purpose language generation and instruction-following tasks, benefiting from the diverse training data and fine-tuning of its constituent models. Its 7B parameter size makes it a versatile option for applications requiring a balance between performance and computational efficiency.