ibm-research/merlinite-7b
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Mar 2, 2024License:apache-2.0Architecture:Transformer0.1K Open Weights Cold

Merlinite-7b is a 7 billion parameter language model developed by IBM Research, based on the Mistral-7B-v0.1 architecture. It is fine-tuned using the Large-scale Alignment for chatBots (LAB) method, which employs a taxonomy-driven synthetic data generation process with Mixtral-8x7B-Instruct as a teacher model. This approach allows for incremental knowledge and skill acquisition without catastrophic forgetting, making it suitable for diverse conversational AI applications.

Loading preview...