Moses25/Llama2-Moses-7b-chat

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kLicense:apache-2.0Architecture:Transformer Open Weights Cold

Llama2-Moses-7b-chat is a 7 billion parameter instruction-tuned language model developed by Moses25, based on the Meta Llama-2-7b-hf architecture. This model is fine-tuned on a 30GB dataset comprising instruction and pretrained data, supporting both English and Chinese languages. It is designed for chat-based applications, leveraging its multilingual training for conversational tasks.

Loading preview...

Overview

Llama2-Moses-7b-chat is an instruction-tuned language model built upon Meta's Llama-2-7b-hf base architecture. Developed by Moses25, this model has undergone fine-tuning using a substantial 30GB dataset. The training data includes a mix of instruction-based and general pretrained datasets, specifically incorporating both English and Chinese languages.

Key Capabilities

  • Multilingual Support: Trained with both English and Chinese datasets, enabling conversational abilities in these languages.
  • Instruction Following: Fine-tuned on instruction datasets to enhance its ability to understand and respond to user prompts effectively.
  • Llama 2 Base: Benefits from the robust architecture and pre-training of the Llama 2 7B model.

Good For

  • Chat Applications: Optimized for interactive conversational use cases due to its instruction-tuning.
  • Multilingual Interactions: Suitable for applications requiring understanding and generation in both English and Chinese.
  • Further Fine-tuning: The provided training code on GitHub (code) allows developers to understand its training process and potentially adapt it for specific needs.