DAMO-NLP-SG/mt-llama-7b-delta
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kLicense:mitArchitecture:Transformer0.0K Open Weights Cold

MT-LLaMA-7b-delta is a 7 billion parameter multi-task language model developed by the MT-LLaMA team from Alibaba Damo Academy and the Chinese University of Hong Kong. It is fine-tuned from the LLaMA architecture on a diverse set of tasks from the P3 dataset, including various QA, classification, summarization, and generation tasks. This model demonstrates strong generalization capabilities across both unseen datasets within seen tasks and entirely unseen tasks, significantly outperforming the base LLaMA-7b model in zero-shot evaluations.

Loading preview...